Context Window
131,000
Tokens
Model Cost Profile
Developer: meta-llama
Pricing updated Mar 11, 2026
Live Pricing
Input: $4.00
Output: $4.00
Pricing via OpenRouter API ยท Last synced Mar 11, 2026 ยท MMLU score via public benchmark data
Meta: Llama 3.1 405B Instruct offers a substantial context window of 131,000 tokens, making it suitable for complex applications such as long-form content generation and detailed conversational agents. With an input and output pricing structure of $4.00 per million tokens, teams can effectively manage costs while leveraging advanced AI capabilities for diverse tasks. This model is ideal for organizations requiring extensive context handling and efficient budget allocation for high-volume API usage.
Context Window
131,000
Tokens
Input Price / 1M
$4.00
Prompt tokens
Output Price / 1M
$4.00
Completion tokens
Intelligence (MMLU)
73.2
Massive Multitask Language Understanding
Standardized evaluation scores for Meta: Llama 3.1 405B Instruct.
| Usage Type | Price / 1M Tokens |
|---|---|
| Input (Prompt) | $4.00 |
| Output (Completion) | $4.00 |
Price History
Current Input / 1M
$4.00
Current Output / 1M
$4.00
Estimate monthly spend for Meta: Llama 3.1 405B Instruct based on your workload.
Estimated Monthly Cost
$148
25M input + 12M output tokens
Quick links for cost-down decisions before production rollout.
Common pricing and benchmark questions for Meta: Llama 3.1 405B Instruct.
Meta: Llama 3.1 405B Instruct input pricing is $4.00 per 1M tokens based on the latest synced provider data.
Meta: Llama 3.1 405B Instruct output pricing is $4.00 per 1M tokens based on the latest synced provider data.
Meta: Llama 3.1 405B Instruct supports a context window of 131,000 tokens.
Use the comparison links on this page to open direct model-vs-model pricing and benchmark pages, then evaluate monthly spend projections for your workload.