BazaarLinkBazaarLink
登入

Nemotron 3 Super 120b A12b

NVIDIA Nemotron 3 Super is a 120B-parameter open hybrid MoE model, activating just 12B parameters for maximum compute efficiency and accuracy in complex multi-agent applications. Built on a hybrid Mamba-Transformer Mixture-of-Experts architecture with multi-token prediction (MTP), it delivers over 50% higher token generation compared to leading open models. The model features a 1M token context window for long-term agent coherence, cross-document reasoning, and multi-step task planning. Latent MoE enables calling 4 experts for the inference cost of only one, improving intelligence and generalization. Multi-environment RL training across 10+ environments delivers leading accuracy on benchmarks including AIME 2025, TerminalBench, and SWE-Bench Verified. Fully open with weights, datasets, and recipes under the NVIDIA Open License, Nemotron 3 Super allows easy customization and secure deployment anywhere — from workstation to cloud.

定價資訊

輸入價格
$0.1000
/ 百萬 tokens
輸出價格
$0.5000
/ 百萬 tokens
上下文視窗
262K
tokens
供應商nvidia
模態text->text
發布日期2026年3月
Model IDnvidia/nemotron-3-super-120b-a12b

快速開始

只需更改 base_url 即可透過 BazaarLink API 使用 Nemotron 3 Super 120b A12b:

Python
from openai import OpenAI

client = OpenAI(
    base_url="https://bazaarlink.ai/api/v1",
    api_key="sk-bl-YOUR_API_KEY",
)

response = client.chat.completions.create(
    model="nvidia/nemotron-3-super-120b-a12b",
    messages=[{"role": "user", "content": "Hello!"}],
)
print(response.choices[0].message.content)
TypeScript
import OpenAI from "openai";

const client = new OpenAI({
  baseURL: "https://bazaarlink.ai/api/v1",
  apiKey: "sk-bl-YOUR_API_KEY",
});

const response = await client.chat.completions.create({
  model: "nvidia/nemotron-3-super-120b-a12b",
  messages: [{ role: "user", content: "Hello!" }],
});
console.log(response.choices[0].message.content);

為什麼透過 BazaarLink 使用 Nemotron 3 Super 120b A12b?

  • 美元計費(台幣報價)+ 統一發票 — 台灣團隊無需海外刷卡
  • OpenAI 相容 API — 無需改寫程式碼
  • 自動故障轉移 — 同模型多供應商備援
  • 中文客服支援 — 在地團隊即時協助
立即試用← 查看所有模型

Frequently Asked Questions

What is Nemotron 3 Super 120b A12b?

NVIDIA Nemotron 3 Super is a 120B-parameter open hybrid MoE model, activating just 12B parameters for maximum compute efficiency and accuracy in complex multi-agent applications. Built on a hybrid Mamba-Transformer Mixture-of-Experts architecture with multi-token prediction (MTP), it delivers over 50% higher token generation compared to leading open models. The model features a 1M token context window for long-term agent coherence, cross-document reasoning, and multi-step task planning. Latent MoE enables calling 4 experts for the inference cost of only one, improving intelligence and generalization. Multi-environment RL training across 10+ environments delivers leading accuracy on benchmarks including AIME 2025, TerminalBench, and SWE-Bench Verified. Fully open with weights, datasets, and recipes under the NVIDIA Open License, Nemotron 3 Super allows easy customization and secure deployment anywhere — from workstation to cloud.

How much does the Nemotron 3 Super 120b A12b API cost?

Nemotron 3 Super 120b A12b costs $0.1000 per 1K input tokens and $0.5000 per 1K output tokens when accessed through BazaarLink.

How do I use Nemotron 3 Super 120b A12b with the OpenAI SDK?

Set base_url to "https://bazaarlink.ai/api/v1" and use model ID "nvidia/nemotron-3-super-120b-a12b". All OpenAI SDK methods (chat.completions, embeddings, streaming) work without code changes.

What is the context window for Nemotron 3 Super 120b A12b?

Nemotron 3 Super 120b A12b supports a context window of 262,144 tokens.

Is Nemotron 3 Super 120b A12b available for free?

Nemotron 3 Super 120b A12b is a paid model. BazaarLink offers free trial credits on registration so you can test it without a credit card.

Support
Support
Hi! How can we help you?
Send a message and we'll get back to you soon.
Nemotron 3 Super 120b A12b API Pricing & Documentation — nvidia