Qwen: Qwen3 235B A22B

qwen/qwen3-235b-a22b

qwen/qwen3-235b-a22b-04-28
Created Apr 28, 202540,960 context
$0.13/M input tokens$0.60/M output tokens

Qwen3-235B-A22B is a 235B parameter mixture-of-experts (MoE) model developed by Qwen, activating 22B parameters per forward pass. It supports seamless switching between a "thinking" mode for complex reasoning, math, and code tasks, and a "non-thinking" mode for general conversational efficiency. The model demonstrates strong reasoning ability, multilingual support (100+ languages and dialects), advanced instruction-following, and agent tool-calling capabilities. It natively handles a 32K token context window and extends up to 131K tokens using YaRN-based scaling.

Recent activity on Qwen3 235B A22B

Tokens processed per day

Apr 28May 3May 8May 13May 18May 23May 28Jun 2Jun 7Jun 12Jun 17Jun 22Jun 270800M1.6B2.4B3.2B

More models from Qwen

    Qwen: Qwen3 235B A22B – Recent Activity | OpenRouter