SargalaySargalay

Command Palette

Search for a command to run...

Back to all models
qwen

Qwen: Qwen3 Coder 480B A35B

Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team. It is optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning over repositories. The model features 480 billion total parameters, with 35 billion active per forward pass (8 out of 160 experts). Pricing for the Alibaba endpoints varies by context length. Once a request is greater than 128k input tokens, the higher pricing is used.

qwen/qwen3-coder

Context Size

262.144K

Input Price

1,353 Ks/M

Output Price

6,150 Ks/M


Architecture

Text

Supported Parameters

frequency_penaltylogit_biasmax_tokensmin_ppresence_penaltyrepetition_penaltyresponse_formatseedstopstructured_outputstemperaturetool_choicetoolstop_ktop_p

Details

TokenizerQwen3
Provider Context262.144K tokens
ModeratedNo