Nemotron-Post-Training-v3 Collection Collection of datasets used in the post-training phase of Nemotron Nano v3. β’ 7 items β’ Updated 10 days ago β’ 54
Nemotron-Pre-Training-Datasets Collection Large scale pre-training datasets used in the Nemotron family of models. β’ 11 items β’ Updated 10 days ago β’ 85
NVIDIA Nemotron v3 Collection Open, Production-ready Enterprise Models β’ 6 items β’ Updated 2 days ago β’ 108
Bolmo: Byteifying the Next Generation of Language Models Paper β’ 2512.15586 β’ Published 16 days ago β’ 13
Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory Paper β’ 2504.19413 β’ Published Apr 28, 2025 β’ 36
SmolDocling: An ultra-compact vision-language model for end-to-end multi-modal document conversion Paper β’ 2503.11576 β’ Published Mar 14, 2025 β’ 124
TurboDiffusion: Accelerating Video Diffusion Models by 100-200 Times Paper β’ 2512.16093 β’ Published 15 days ago β’ 90
Emergent temporal abstractions in autoregressive models enable hierarchical reinforcement learning Paper β’ 2512.20605 β’ Published 10 days ago β’ 59
β Long-context post-training π§Ά β Collection Resources for post-training LLMs with long-context samples β’ 5 items β’ Updated Sep 14, 2025 β’ 6
VL-JEPA: Joint Embedding Predictive Architecture for Vision-language Paper β’ 2512.10942 β’ Published 22 days ago β’ 21
V-JEPA 2 Collection A frontier video understanding model developed by FAIR, Meta, which extends the pretraining objectives of https://ai.meta.com/blog/v-jepa-yann β’ 8 items β’ Updated Jun 13, 2025 β’ 178