Context Window
256,000
Tokens
Model Cost Profile
Developer: nvidia
Pricing updated Mar 11, 2026
Live Pricing
Input: $0.0000
Output: $0.0000
Pricing via OpenRouter API · Last synced Mar 11, 2026 · MMLU score via public benchmark data
The NVIDIA Nemotron 3 Nano 30B A3B model offers a substantial context window of 256,000 tokens, making it ideal for applications requiring extensive text comprehension and generation, such as document summarization and conversational AI. Given its free access, teams can leverage this model without incurring input or output costs, significantly lowering the barrier for experimentation and deployment in various projects. This model is particularly suitable for startups and research teams looking to integrate advanced AI capabilities without budget constraints, facilitating innovation in natural language processing tasks.
Context Window
256,000
Tokens
Input Price / 1M
$0.0000
Prompt tokens
Output Price / 1M
$0.0000
Completion tokens
Intelligence (MMLU)
79.4
Massive Multitask Language Understanding
Standardized evaluation scores for NVIDIA: Nemotron 3 Nano 30B A3B (free).
| Usage Type | Price / 1M Tokens |
|---|---|
| Input (Prompt) | $0.0000 |
| Output (Completion) | $0.0000 |
Price History
Current Input / 1M
$0.000000
Current Output / 1M
$0.000000
Estimate monthly spend for NVIDIA: Nemotron 3 Nano 30B A3B (free) based on your workload.
Estimated Monthly Cost
$0.00
25M input + 12M output tokens
Quick links for cost-down decisions before production rollout.
Common pricing and benchmark questions for NVIDIA: Nemotron 3 Nano 30B A3B (free).
NVIDIA: Nemotron 3 Nano 30B A3B (free) input pricing is $0.0000 per 1M tokens based on the latest synced provider data.
NVIDIA: Nemotron 3 Nano 30B A3B (free) output pricing is $0.0000 per 1M tokens based on the latest synced provider data.
NVIDIA: Nemotron 3 Nano 30B A3B (free) supports a context window of 256,000 tokens.
Use the comparison links on this page to open direct model-vs-model pricing and benchmark pages, then evaluate monthly spend projections for your workload.