Model Comparison

Author
Favicon for qwen
qwen
Context Length33K

Qwen-Max, based on Qwen2.5, provides the best inference performance among Qwen models, especially for complex multi-step tasks. It's a large-scale MoE model that has been pretrained on over 20 trillion tokens and further post-trained with curated Supervised Fine-Tuning (SFT) and Reinforcement Learning from Human Feedback (RLHF) methodologies. The parameter count is unknown.

Provider

Pricing

Input$1.60 / M tokens
Output$6.40 / M tokens
Images– –

Endpoint Features

Quantization– –
Max Tokens (input + output)33K
Max Output Tokens8K
Stream cancellation– –
Supports Tools– –
No Prompt Training
Reasoning– –