
Nvidia has launched the Nemotron 3 vary of open fashions, datasets and libraries, designed to help clear, environment friendly and specialised agentic AI methods throughout sectors.
The Nemotron 3 fashions are available three sizes: Nano, Tremendous and Extremely. They use a hybrid latent mixture-of-experts structure to assist builders create and run multi-agent AI methods at scale.

Entry deeper trade intelligence
Expertise unmatched readability with a single platform that mixes distinctive information, AI, and human experience.
The corporate positions Nemotron 3 for organisations shifting from single-model chatbots to multi-agent methods, the place builders face points corresponding to communication overhead, context drift, excessive inference prices and the necessity for transparency when automating advanced workflows.
Nvidia hyperlinks Nemotron 3 to its wider sovereign AI efforts, stating that organisations in areas from Europe to South Korea are adopting open, clear and environment friendly fashions that allow them to construct AI aligned with their very own information, rules and values.
Nvidia founder and CEO Jensen Huang commented: “Open innovation is the muse of AI progress.
“With Nemotron, we’re remodeling superior AI into an open platform that provides builders the transparency and effectivity they should construct agentic methods at scale.”
Open Nemotron 3 fashions give startups a technique to construct and iterate extra shortly on AI brokers and transfer from prototype to enterprise rollout.
The Nemotron 3 household consists of three MoE fashions. Nemotron 3 Nano is described as a small 30-billion-parameter mannequin that may activate as much as three billion parameters at a time for focused duties.
Nemotron 3 Tremendous has 100 billion parameters, with as much as 10 billion lively per token, and is geared toward multi-agent functions that depend on reasoning.
Nemotron 3 Extremely has 500 billion parameters, with as much as 50 billion lively per token, and is positioned as a reasoning engine for advanced AI functions. It’s optimised for makes use of corresponding to software program debugging, content material summarisation, AI assistant workflows and knowledge retrieval at low inference price.
The mannequin’s hybrid MoE structure improves effectivity and scaleability. It delivers as much as 4 instances increased token throughput than Nemotron 2 Nano and cuts reasoning-token technology by as much as 60%, decreasing inference prices.
Nemotron 3 Nano has a 1-million-token context window, which Nvidia says permits it to retain extra info and join information over lengthy, multistep duties.
Nemotron 3 Tremendous is suited to functions during which many brokers work collectively on advanced duties with low latency, whereas Nemotron 3 Extremely is geared toward AI workflows that want deep analysis and strategic planning.
Each Tremendous and Extremely use Nvidia’s 4-bit NVFP4 coaching format on the Nvidia Blackwell structure, which the corporate says cuts reminiscence necessities and accelerates coaching.
This format permits bigger fashions to be skilled on present infrastructure “with out compromising accuracy relative to higher-precision codecs.”
Nvidia states that the three-model lineup permits builders to pick out open fashions matched to particular workloads and to scale from dozens to a whole bunch of brokers, whereas gaining quicker, extra correct long-horizon reasoning for advanced workflows.
Alongside the fashions, Nvidia has launched coaching datasets and reinforcement studying libraries for constructing specialised AI brokers.
The corporate has offered three trillion tokens of Nemotron pretraining, post-training and reinforcement studying datasets, which embrace reasoning, coding and multistep workflow examples to help the creation of domain-specialised brokers.
The Nemotron Agentic Security Dataset provides real-world telemetry to assist groups consider and enhance the protection of advanced agent methods.
To help improvement, Nvidia has launched the NeMo Fitness center and NeMo RL open-source libraries, which provide coaching environments and post-training foundations for Nemotron fashions, together with NeMo Evaluator to test mannequin security and efficiency.
All these instruments and datasets at the moment are accessible on GitHub and Hugging Face.
Nemotron 3 is supported by LM Studio, llama.cpp, SGLang and vLLM. Prime Mind and Unsloth are integrating NeMo Fitness center’s coaching environments into their very own workflows, offering groups with quicker entry to reinforcement studying coaching.
Nemotron 3 Nano is at the moment accessible on Hugging Face and thru inference suppliers Baseten, DeepInfra, Fireworks, FriendliAI, OpenRouter and Collectively AI.
Nemotron can also be provided on enterprise AI and information infrastructure platforms Couchbase, DataRobot, H2O.ai, JFrog, Lambda and UiPath.
For patrons utilizing public cloud platforms, Nemotron 3 Nano will likely be accessible on AWS by way of Amazon Bedrock (serverless) and will likely be supported on CoreWeave, Crusoe, Google Cloud, Microsoft Foundry, Nebius, Nscale and Yotta. Nemotron 3 Nano can also be provided as an Nvidia NIM microservice for deployment on Nvidia-accelerated infrastructure.
Nvidia is anticipated to make Nemotron 3 Tremendous and Extremely accessible within the first half of 2026.
INVIDIA additionally individually introduced that it has acquired SchedMD, the developer of Slurm, an open-source workload administration system utilized in high-performance computing and AI. This acquisition helps the open-source software program ecosystem and advances AI-related work for researchers, builders and enterprises.
Nvidia plans to proceed growing and distributing Slurm as open-source, vendor-neutral software program, and to maintain it broadly accessible and supported throughout a spread of {hardware} and software program environments.

