1.3K Downloads
Qwen's most powerful code model, featuring 480B total parameters with 35B activated through Mixture of Experts (MoE) architecture.
Trained for Ttool use
Qwen's most powerful code model, featuring 480B total parameters with 35B activated through Mixture of Experts (MoE) architecture.
Key Features:
Technical Specifications:
Note: This model operates in non-thinking mode only and does not generate <think></think>
blocks.
The underlying model files this model uses
When you download this model, LM Studio picks the source that will best suit your machine (you can override this)
Custom configuration options included with this model