Model Cost Profile

Meta: Llama 3.1 405B Instruct

Developer: meta-llama

Pricing updated Mar 11, 2026

Input rank: #316Output rank: #252

Live Pricing

Input: $4.00

Output: $4.00

Pricing via OpenRouter API ยท Last synced Mar 11, 2026 ยท MMLU score via public benchmark data

Meta: Llama 3.1 405B Instruct offers a substantial context window of 131,000 tokens, making it suitable for complex applications such as long-form content generation and detailed conversational agents. With an input and output pricing structure of $4.00 per million tokens, teams can effectively manage costs while leveraging advanced AI capabilities for diverse tasks. This model is ideal for organizations requiring extensive context handling and efficient budget allocation for high-volume API usage.

๐Ÿ”ง Tool Calling๐Ÿ“‹ Structured Output

Context Window

131,000

Tokens

Input Price / 1M

$4.00

Prompt tokens

Output Price / 1M

$4.00

Completion tokens

Intelligence (MMLU)

73.2

Massive Multitask Language Understanding

Benchmark Scores

Standardized evaluation scores for Meta: Llama 3.1 405B Instruct.

BenchmarkScoreRankSource
GPQA51.5#83 of 118artificial_analysis
MMLU73.2#74 of 121artificial_analysis

Price History

Meta: Llama 3.1 405B Instruct Pricing Trend

Input / 1M tokens0.0%Output / 1M tokens0.0%
Mar 7 โ€” Mar 11
$4.00$4.00$4.00Mar 7Mar 8Mar 9Mar 10Mar 11

Current Input / 1M

$4.00

Current Output / 1M

$4.00

Cheaper Alternatives to Compare

Quick links for cost-down decisions before production rollout.

FAQ

Common pricing and benchmark questions for Meta: Llama 3.1 405B Instruct.

How much does Meta: Llama 3.1 405B Instruct cost per 1M input tokens?

Meta: Llama 3.1 405B Instruct input pricing is $4.00 per 1M tokens based on the latest synced provider data.

How much does Meta: Llama 3.1 405B Instruct cost per 1M output tokens?

Meta: Llama 3.1 405B Instruct output pricing is $4.00 per 1M tokens based on the latest synced provider data.

What context window does Meta: Llama 3.1 405B Instruct support?

Meta: Llama 3.1 405B Instruct supports a context window of 131,000 tokens.

How can I compare Meta: Llama 3.1 405B Instruct with cheaper alternatives?

Use the comparison links on this page to open direct model-vs-model pricing and benchmark pages, then evaluate monthly spend projections for your workload.