AIMarch 29, 2025NVIDIA AI Researchers Introduce FFN Fusion: A Novel Optimization Technique that Demonstrates How Sequential Computation in Large Language Models LLMs can be Effectively Parallelized
AIMarch 24, 2025Lyra: A Computationally Efficient Subquadratic Architecture for Biological Sequence Modeling
AIMarch 14, 2025Optimizing Test-Time Compute for LLMs: A Meta-Reinforcement Learning Approach with Cumulative Regret Minimization
AIFebruary 25, 2025Researchers from Moonshot AI Introduce Muon and Moonlight: Optimizing Large-Scale Language Models with Efficient Training Techniques
AIJanuary 20, 2025Google AI Proposes a Fundamental Framework for Inference-Time Scaling in Diffusion Models