Context Window
202,752
Tokens
Model Cost Profile
Developer: z-ai
Pricing updated Mar 11, 2026
Z.ai: GLM 4.7 Flash is designed for applications requiring extensive context management, accommodating up to 202,752 tokens, making it suitable for complex document analysis and large-scale conversational AI. Teams utilizing this model can expect an input cost of $0.06 per million tokens and an output cost of $0.40 per million tokens, which allows for budget-conscious planning in high-volume scenarios. Its robust capabilities are ideal for industries such as finance, legal, and customer support, where detailed context and nuanced understanding are critical.
Context Window
202,752
Tokens
Input Price / 1M
$0.0600
Prompt tokens
Output Price / 1M
$0.4000
Completion tokens
Intelligence (MMLU)
Benchmark Pending
Massive Multitask Language Understanding
| Usage Type | Price / 1M Tokens |
|---|---|
| Input (Prompt) | $0.0600 |
| Output (Completion) | $0.4000 |
Price History
Current Input / 1M
$0.0600
Current Output / 1M
$0.4000
Estimate monthly spend for Z.ai: GLM 4.7 Flash based on your workload.
Estimated Monthly Cost
$6.30
25M input + 12M output tokens
Quick links for cost-down decisions before production rollout.
Common pricing and benchmark questions for Z.ai: GLM 4.7 Flash.
Z.ai: GLM 4.7 Flash input pricing is $0.0600 per 1M tokens based on the latest synced provider data.
Z.ai: GLM 4.7 Flash output pricing is $0.4000 per 1M tokens based on the latest synced provider data.
Z.ai: GLM 4.7 Flash supports a context window of 202,752 tokens.
Use the comparison links on this page to open direct model-vs-model pricing and benchmark pages, then evaluate monthly spend projections for your workload.