Nemotron-Post-Training-v3 Collection Collection of datasets used in the post-training phase of Nemotron Nano v3. β’ 7 items β’ Updated 8 days ago β’ 53
Nemotron-Pre-Training-Datasets Collection Large scale pre-training datasets used in the Nemotron family of models. β’ 11 items β’ Updated 8 days ago β’ 84
NVIDIA Nemotron v3 Collection Open, Production-ready Enterprise Models β’ 6 items β’ Updated 1 day ago β’ 107
Bolmo: Byteifying the Next Generation of Language Models Paper β’ 2512.15586 β’ Published 15 days ago β’ 12
Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory Paper β’ 2504.19413 β’ Published Apr 28, 2025 β’ 36
SmolDocling: An ultra-compact vision-language model for end-to-end multi-modal document conversion Paper β’ 2503.11576 β’ Published Mar 14, 2025 β’ 123
TurboDiffusion: Accelerating Video Diffusion Models by 100-200 Times Paper β’ 2512.16093 β’ Published 14 days ago β’ 88
Emergent temporal abstractions in autoregressive models enable hierarchical reinforcement learning Paper β’ 2512.20605 β’ Published 9 days ago β’ 59
β Long-context post-training π§Ά β Collection Resources for post-training LLMs with long-context samples β’ 5 items β’ Updated Sep 14, 2025 β’ 6
VL-JEPA: Joint Embedding Predictive Architecture for Vision-language Paper β’ 2512.10942 β’ Published 21 days ago β’ 18
V-JEPA 2 Collection A frontier video understanding model developed by FAIR, Meta, which extends the pretraining objectives of https://ai.meta.com/blog/v-jepa-yann β’ 8 items β’ Updated Jun 13, 2025 β’ 178