Qwen3 235B A22B Instruct 2507 FP8
235B MoE model with 22B activation featuring enhanced instruction following, reasoning, and 262K context for cost-efficient high-throughput inference.
About model
Enhanced Qwen3 model optimized for serverless inference with superior price-performance.
Model | AIME 2025 | GPQA Diamond | HLE | LiveCodeBench | MATH500 | SWE-bench verified |
|---|---|---|---|---|---|---|
Qwen3 235B A22B Instruct 2507 FP8 | 65.9% | Related open-source models | Competitor closed-source models | |||
90.5% | 34.2% | 78.7% | ||||
83.3% | 24.9% | 99.2% | 62.3% | |||
76.8% | 96.4% | 48.9% | ||||
49.2% | 2.7% | 32.3% | 89.3% | 31.0% |
API usage
Endpoint:
- TypeChatReasoning
- Main use casesChatSmall & FastMedium General PurposeFunction Calling
- FeaturesFunction CallingJSON Mode
- DeploymentServerless
- Parameters235B
- Context length262K
- Input price
$0.20 / 1M tokens
- Output price
$0.60 / 1M tokens
- Input modalitiesText
- Output modalitiesText
- ReleasedJuly 22, 2025
- Quantization levelFP8
- External link
- CategoryChat