Context Window
131,072
Tokens
Model Cost Profile
Developer: sao10k
Pricing updated Mar 11, 2026
Sao10K: Llama 3.3 Euryale 70B is designed for applications requiring extensive context handling, making it ideal for complex tasks such as document summarization and conversational AI with a context window of 131072 tokens. Teams utilizing this API model can expect a cost of $0.65 per million input tokens and $0.75 per million output tokens, which can significantly impact budget planning for large-scale projects. This pricing structure allows for flexible scaling, enabling organizations to optimize usage based on their specific data processing needs.
Context Window
131,072
Tokens
Input Price / 1M
$0.6500
Prompt tokens
Output Price / 1M
$0.7500
Completion tokens
Intelligence (MMLU)
Benchmark Pending
Massive Multitask Language Understanding
| Usage Type | Price / 1M Tokens |
|---|---|
| Input (Prompt) | $0.6500 |
| Output (Completion) | $0.7500 |
Price History
Current Input / 1M
$0.6500
Current Output / 1M
$0.7500
Estimate monthly spend for Sao10K: Llama 3.3 Euryale 70B based on your workload.
Estimated Monthly Cost
$25
25M input + 12M output tokens
Quick links for cost-down decisions before production rollout.
Common pricing and benchmark questions for Sao10K: Llama 3.3 Euryale 70B.
Sao10K: Llama 3.3 Euryale 70B input pricing is $0.6500 per 1M tokens based on the latest synced provider data.
Sao10K: Llama 3.3 Euryale 70B output pricing is $0.7500 per 1M tokens based on the latest synced provider data.
Sao10K: Llama 3.3 Euryale 70B supports a context window of 131,072 tokens.
Use the comparison links on this page to open direct model-vs-model pricing and benchmark pages, then evaluate monthly spend projections for your workload.