Description
Hybrid attention architecture, high-sparsity Mixture-of-Experts 80B model (active 3B). Currently supported for Mac only with MLX.
Stats
10.8K Downloads
10 stars
Capabilities
Minimum system memory
Tags
Last updated
Updated on September 15byREADME
The first model in the Qwen3-Next series featuring innovative hybrid attention architecture and high-efficiency Mixture-of-Experts design.
Delivers performance comparable to much larger models while maintaining exceptional efficiency:
Parameters
Custom configuration options included with this model
Sources
The underlying model files this model uses