GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models
Paper
• 2508.06471
• Published • 210
NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid
Mamba-Transformer Reasoning Model
Paper
• 2508.14444
• Published • 46
Gemini 2.5: Pushing the Frontier with Advanced Reasoning, Multimodality,
Long Context, and Next Generation Agentic Capabilities
Paper
• 2507.06261
• Published • 67
MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning
Attention
Paper
• 2506.13585
• Published • 274
Paper
• 2506.10910
• Published • 67
Paper
• 2505.09388
• Published • 339
MiMo: Unlocking the Reasoning Potential of Language Model -- From
Pretraining to Posttraining
Paper
• 2505.07608
• Published • 82
Phi-4-reasoning Technical Report
Paper
• 2504.21318
• Published • 54
Llama-Nemotron: Efficient Reasoning Models
Paper
• 2505.00949
• Published • 42
CLIMB: CLustering-based Iterative Data Mixture Bootstrapping for
Language Model Pre-training
Paper
• 2504.13161
• Published • 97
DeepSeek-R1 Thoughtology: Let's <think> about LLM Reasoning
Paper
• 2504.07128
• Published • 87
Rethinking Reflection in Pre-Training
Paper
• 2504.04022
• Published • 80
OLMoTrace: Tracing Language Model Outputs Back to Trillions of Training
Tokens
Paper
• 2504.07096
• Published • 77
Paper
• 2503.19786
• Published • 55
LIMO: Less is More for Reasoning
Paper
• 2502.03387
• Published • 62
Skywork Open Reasoner 1 Technical Report
Paper
• 2505.22312
• Published • 54
Every Sample Matters: Leveraging Mixture-of-Experts and High-Quality
Data for Efficient and Accurate Code LLM
Paper
• 2503.17793
• Published • 24
RedStone: Curating General, Code, Math, and QA Data for Large Language
Models
Paper
• 2412.03398
• Published • 2
Nemotron-CC-Math: A 133 Billion-Token-Scale High Quality Math
Pretraining Dataset
Paper
• 2508.15096
• Published • 7
RLBFF: Binary Flexible Feedback to bridge between Human Feedback &
Verifiable Rewards
Paper
• 2509.21319
• Published • 8
StarCoder 2 and The Stack v2: The Next Generation
Paper
• 2402.19173
• Published • 154
Nemotron-Cascade: Scaling Cascaded Reinforcement Learning for General-Purpose Reasoning Models
Paper
• 2512.13607
• Published • 37
Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning
Paper
• 2512.20848
• Published • 41
X-Coder: Advancing Competitive Programming with Fully Synthetic Tasks, Solutions, and Tests
Paper
• 2601.06953
• Published • 46