Recursive Language Models Meet Uncertainty: The Surprising Effectiveness of Self-Reflective Program Search for Long Context Paper • 2603.15653 • Published 15 days ago • 9
Nemotron-Cascade 2: Post-Training LLMs with Cascade RL and Multi-Domain On-Policy Distillation Paper • 2603.19220 • Published 2 days ago • 41
view article Article **LoRA Fine-Tuning BitNet b1.58 LLMs on Heterogeneous Edge GPUs via QVAC Fabric** 5 days ago • 11
Decoupling Reasoning and Confidence: Resurrecting Calibration in Reinforcement Learning from Verifiable Rewards Paper • 2603.09117 • Published 12 days ago • 9
view article Article **ColBERT-Zero: To Pre-train Or Not To Pre-train ColBERT models?** about 1 month ago • 19
GPT 5 Codex Collection Distilled models and datasets for GPT 5 Codex • 7 items • Updated Dec 20, 2025 • 5
🧮functiongemma ft mobile-actions Collection A collection of functiongemma-270m-it models fine-tuned on mobile actions dataset for Spanish, French and Italian • 3 items • Updated Jan 5 • 3
view article Article Aligning to What? Rethinking Agent Generalization in MiniMax M2 Oct 30, 2025 • 43
Z-Image: An Efficient Image Generation Foundation Model with Single-Stream Diffusion Transformer Paper • 2511.22699 • Published Nov 27, 2025 • 244
Nemotron RAG Collection Set of tools to build retrieval-augmented generation (RAG) systems, improve search and ranking accuracy, and extract structured data from complex docs • 9 items • Updated 1 day ago • 84
Olmo 3 Post-training Collection All artifacts for post-training Olmo 3. Datasets follow the model that resulted from training on them. • 32 items • Updated Dec 23, 2025 • 51
Black-Box On-Policy Distillation of Large Language Models Paper • 2511.10643 • Published Nov 13, 2025 • 52
view article Article How to train a new language model from scratch using Transformers and Tokenizers Feb 14, 2020 • 60