8.4K Downloads
Description
Hybrid attention architecture, high-sparsity Mixture-of-Experts 80B model (active 3B). Currently supported for Mac only with MLX.
Use cases
Minimum system memory
Tags
Last update
Updated on September 15byREADME
The first model in the Qwen3-Next series featuring innovative hybrid attention architecture and high-efficiency Mixture-of-Experts design.
Delivers performance comparable to much larger models while maintaining exceptional efficiency:
Parameters
Custom configuration options included with this model
Sources
The underlying model files this model uses