LongMamba: Enhancing Mamba's Long Context Capabilities via Training-Free Receptive Field Enlargement Paper • 2504.16053 • Published Apr 22, 2025
Minifinetuning: Low-Data Generation Domain Adaptation through Corrective Self-Distillation Paper • 2506.15702 • Published May 30, 2025
Scaling Up RL: Unlocking Diverse Reasoning in LLMs via Prolonged Training Paper • 2507.12507 • Published Jul 16, 2025 • 1
NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model Paper • 2508.14444 • Published Aug 20, 2025 • 40
FlexGS: Train Once, Deploy Everywhere with Many-in-One Flexible 3D Gaussian Splatting Paper • 2506.04174 • Published Jun 4, 2025
Universal Deep Research: Bring Your Own Model and Strategy Paper • 2509.00244 • Published Aug 29, 2025 • 14
BroRL: Scaling Reinforcement Learning via Broadened Exploration Paper • 2510.01180 • Published Oct 1, 2025 • 19
LaCache: Ladder-Shaped KV Caching for Efficient Long-Context Modeling of Large Language Models Paper • 2507.14204 • Published Jul 14, 2025
OmniVinci: Enhancing Architecture and Data for Omni-Modal Understanding LLM Paper • 2510.15870 • Published Oct 17, 2025 • 91
DLER: Doing Length pEnalty Right - Incentivizing More Intelligence per Token via Reinforcement Learning Paper • 2510.15110 • Published Oct 16, 2025 • 16
ProfBench: Multi-Domain Rubrics requiring Professional Knowledge to Answer and Judge Paper • 2510.18941 • Published Oct 21, 2025 • 8
Nemotron Elastic: Towards Efficient Many-in-One Reasoning LLMs Paper • 2511.16664 • Published Nov 20, 2025 • 27
Nemotron-Flash: Towards Latency-Optimal Hybrid Small Language Models Paper • 2511.18890 • Published Nov 24, 2025 • 34