Context Window
131,072
Tokens
Model Cost Profile
Developer: z-ai
Pricing updated Mar 11, 2026
Live Pricing
Input: $0.3000
Output: $0.9000
Pricing via OpenRouter API ยท Last synced Mar 11, 2026 ยท MMLU score via public benchmark data
Z.ai: GLM 4.6V offers an extensive context window of 131072 tokens, making it suitable for applications requiring deep contextual understanding, such as legal document analysis and long-form content generation. Teams leveraging this API model can expect input costs of $0.30 per million tokens and output costs of $0.90 per million tokens, allowing for flexible budgeting based on usage patterns. This pricing structure is particularly advantageous for businesses that handle large volumes of text, enabling them to optimize costs while maximizing the model's capabilities.
Context Window
131,072
Tokens
Input Price / 1M
$0.3000
Prompt tokens
Output Price / 1M
$0.9000
Completion tokens
Intelligence (MMLU)
75.2
Massive Multitask Language Understanding
Standardized evaluation scores for Z.ai: GLM 4.6V.
| Usage Type | Price / 1M Tokens |
|---|---|
| Input (Prompt) | $0.3000 |
| Output (Completion) | $0.9000 |
Price History
Current Input / 1M
$0.3000
Current Output / 1M
$0.9000
Estimate monthly spend for Z.ai: GLM 4.6V based on your workload.
Estimated Monthly Cost
$18
25M input + 12M output tokens
Quick links for cost-down decisions before production rollout.
Common pricing and benchmark questions for Z.ai: GLM 4.6V.
Z.ai: GLM 4.6V input pricing is $0.3000 per 1M tokens based on the latest synced provider data.
Z.ai: GLM 4.6V output pricing is $0.9000 per 1M tokens based on the latest synced provider data.
Z.ai: GLM 4.6V supports a context window of 131,072 tokens.
Use the comparison links on this page to open direct model-vs-model pricing and benchmark pages, then evaluate monthly spend projections for your workload.