We’re excited to partner with NVIDIA to bring @cf/nvidia/nemotron-3-120b-a12b to Workers AI. NVIDIA Nemotron 3 Super is a Mixture-of-Experts (MoE) model with a hybrid Mamba-transformer architecture, 120B total parameters, and 12B active parameters per forward pass.
The model is optimized for running many collaborating agents per application. It delivers high accuracy for reasoning, tool calling, and instruction following across complex multi-step tasks.
Key capabilities:
- Hybrid Mamba-transformer architecture delivers over 50% higher token generation throughput compared to leading open models, reducing latency for real-world applications
- Tool calling support for building AI agents that invoke tools across multiple conversation turns
- Multi-Token Prediction (MTP) accelerates long-form text generation by predicting several future tokens simultaneously in a single forward pass
- 32,000 token context window for retaining conversation history and plan states across multi-step agent workflows
Use Nemotron 3 Super through the Workers AI binding (env.AI.run()), the REST API, or the OpenAI-compatible endpoint.
For more information, refer to the Nemotron 3 Super model page.
Source: Cloudflare
Latest Posts
- Amazon SageMaker HyperPod now supports on-demand deep health checks

- Amazon SageMaker Unified Studio now supports multiple code spaces within projects for IAM domains

- Amazon IVS Low-Latency Streaming now supports server-side ad insertion

- Introducing GPU Health Monitoring and Auto Repair for Amazon ECS Managed Instances






