Model Cost Profile

Qwen: Qwen-Plus

Developer: qwen

Pricing updated Mar 11, 2026

Input rank: #195Output rank: #189

Live Pricing

Input: $0.4000

Output: $1.20

Pricing via OpenRouter API ยท Last synced Mar 11, 2026

Qwen-Plus, developed by qwen, offers an extensive context window of 1,000,000 tokens, making it suitable for applications requiring deep contextual understanding, such as long-form content generation and complex data analysis. Teams utilizing this API model can expect input costs of $0.40 per million tokens and output costs of $1.20 per million tokens, allowing for scalable budgeting based on usage patterns. This pricing structure enables organizations to effectively manage expenses while leveraging the model's capabilities for tasks like conversational AI and extensive document summarization.

๐Ÿ”ง Tool Calling๐Ÿ“‹ Structured Output

Context Window

1,000,000

Tokens

Input Price / 1M

$0.4000

Prompt tokens

Output Price / 1M

$1.20

Completion tokens

Intelligence (MMLU)

Benchmark Pending

Massive Multitask Language Understanding

Price History

Qwen: Qwen-Plus Pricing Trend

Input / 1M tokens0.0%Output / 1M tokens0.0%
Mar 7 โ€” Mar 11
$0.4000$0.8000$1.20Mar 7Mar 8Mar 9Mar 10Mar 11

Current Input / 1M

$0.4000

Current Output / 1M

$1.20

Cheaper Alternatives to Compare

Quick links for cost-down decisions before production rollout.

FAQ

Common pricing and benchmark questions for Qwen: Qwen-Plus.

How much does Qwen: Qwen-Plus cost per 1M input tokens?

Qwen: Qwen-Plus input pricing is $0.4000 per 1M tokens based on the latest synced provider data.

How much does Qwen: Qwen-Plus cost per 1M output tokens?

Qwen: Qwen-Plus output pricing is $1.20 per 1M tokens based on the latest synced provider data.

What context window does Qwen: Qwen-Plus support?

Qwen: Qwen-Plus supports a context window of 1,000,000 tokens.

How can I compare Qwen: Qwen-Plus with cheaper alternatives?

Use the comparison links on this page to open direct model-vs-model pricing and benchmark pages, then evaluate monthly spend projections for your workload.