AIJune 17, 2025OpenBMB Releases MiniCPM4: Ultra-Efficient Language Models for Edge Devices with Sparse Attention and Fast Inference
AIJune 15, 2025Internal Coherence Maximization (ICM): A Label-Free, Unsupervised Training Framework for LLMs
AIJune 14, 2025OpenThoughts: A Scalable Supervised Fine-Tuning SFT Data Curation Pipeline for Reasoning Models
AIMay 27, 2025Qwen Researchers Proposes QwenLong-L1: A Reinforcement Learning Framework for Long-Context Reasoning in Large Language Models
AIApril 25, 2025Microsoft Research Introduces MMInference to Accelerate Pre-filling for Long-Context Vision-Language Models
AIApril 23, 2025Muon Optimizer Significantly Accelerates Grokking in Transformers: Microsoft Researchers Explore Optimizer Influence on Delayed Generalization