Description
Qwen's most powerful code model, featuring 480B total parameters with 35B activated through Mixture of Experts (MoE) architecture.
Stats
9.9K Downloads
19 stars
Capabilities
Minimum system memory
Tags
Last updated
Updated on July 31byREADME
Qwen's most powerful code model, featuring 480B total parameters with 35B activated through Mixture of Experts (MoE) architecture.
Key Features:
Technical Specifications:
Note: This model operates in non-thinking mode only and does not generate <think></think> blocks.
Parameters
Custom configuration options included with this model
Sources
The underlying model files this model uses