๐Ÿฆ™Stalecollected in 3h

Nvidia Removes Rug-Pull from Nemotron License

PostLinkedIn
๐Ÿฆ™Read original on Reddit r/LocalLLaMA

๐Ÿ’กNemotron license now allows guardrail removal without terminationโ€”huge for fine-tunes.

โšก 30-Second TL;DR

What Changed

Removed guardrail bypass termination clause

Why It Matters

Operators gain freedom to modify models without license termination risks, boosting open-source LLM experimentation. This could accelerate community fine-tunes and uncensored variants. Wider adoption of Nemotron models likely increases.

What To Do Next

Download the BF16 Nemotron variant and verify the new license notice.

Who should care:Developers & AI Engineers

๐Ÿง  Deep Insight

Web-grounded analysis with 6 cited sources.

๐Ÿ”‘ Enhanced Key Takeaways

  • โ€ขNemotron 3 Super features a hybrid Mamba2-Transformer architecture with LatentMoE, activating only 12B of its 120B parameters during inference for 4x higher efficiency[3][4][6].
  • โ€ขModel trained on over 10 trillion tokens of synthetic data with post-training cutoff in February 2026, including 21 RL environments and full recipes published for reproducibility[3][4][6].
  • โ€ขAvailable on Hugging Face, build.nvidia.com, Perplexity, OpenRouter, and partners like Google Cloud Vertex AI, Oracle OCI, with self-hosting requiring 8x H100 GPUs minimum[3][4][5].

๐Ÿ› ๏ธ Technical Deep Dive

  • โ€ขArchitecture: Mamba2-Transformer Hybrid Latent Mixture of Experts (LatentMoE) with Multi-Token Prediction (MTP); 120B total parameters, 12B active at inference[3][4][6].
  • โ€ขOptimizations: Native NVFP4 pretraining for Blackwell GPUs (4x inference speedup vs FP8 on H100); hybrid backbone with Mamba for efficiency and Transformers for reasoning[4][6].
  • โ€ขTraining: Pre-training data cutoff June 2025, post-training February 2026; multi-environment RL with 1.2M rollouts across 21 configs using NeMo Gym/RL[4][6].
  • โ€ขHardware: Compatible with NVIDIA Ampere A100, Hopper H100-80GB, Blackwell; runtime NeMo 25.11.01 on Linux[4].
  • โ€ขPerformance: 85.6% on PinchBench for agentic tasks; configurable reasoning mode via chat template[4][6].

๐Ÿ”ฎ Future ImplicationsAI analysis grounded in cited sources

Nemotron 3 Super reduces inference costs by 5x for multi-agent systems
Its MoE activates only 12B of 120B parameters while handling 15x token volume of standard chats, enabling scalable enterprise agents[3][5].
Open weights accelerate custom agent development in code review and workflows
Full data, recipes, and permissive license allow fine-tuning by platforms like Perplexity, Palantir, and CodeRabbit without NVIDIA output ownership claims[3][5].

โณ Timeline

2025-12
Nemotron 3 Nano introduced as precursor to Super
2025-12-15
NVIDIA Nemotron Open Model License first modified
2026-03-11
Nemotron 3 Super 120B released on Hugging Face with open weights
2026-03-12
Model checkpoints published via NVIDIA NIM and partners
๐Ÿ“ฐ

Weekly AI Recap

Read this week's curated digest of top AI events โ†’

๐Ÿ‘‰Related Updates

AI-curated news aggregator. All content rights belong to original publishers.
Original source: Reddit r/LocalLLaMA โ†—

Nvidia Removes Rug-Pull from Nemotron License | Reddit r/LocalLLaMA | SetupAI | SetupAI