LMmy choosing

Model Catalog

New & noteworthy local models you can run on your own machine.

minimax-m2
230B
MiniMax M2 is a 230B MoE (10B active) model built for coding and agentic workflows
12.8K
10
Updated 10 days ago
gpt-oss-safeguard
20B
120B
gpt-oss-safeguard-20b and gpt-oss-safeguard-120b are open safety models from OpenAI, building on gpt-oss. Trained to help classify text content based on customizable policies.
4.6K
12
2
Updated 18 days ago
Qwen3-VL
2B
4B
8B
30B
32B
Qwen's latest vision-language model. Includes comprehensive upgrades to visual perception, spatial reasoning, and image understanding.
223.5K
44
5
Updated 19 days ago
Granite 4.0
3B
3B
7B
32B
Granite 4.0 language models are lightweight, state-of-the-art open models that natively support multilingual capabilities, coding tasks, RAG, tool use, and JSON output.
41K
31
4
Updated 19 days ago
seed-oss
36B
Advanced reasoning model from ByteDance with flexible "thinking budget" control and ability to reflect on the length of its own reasoning
34.9K
17
Updated 19 days ago
Qwen3
4B
4B
30B
30B
235B
235B
The latest version of the Qwen3 model family, featuring 4B, 30B, and 235B dense and MoE models, both thinking and non-thinking variants.
285.3K
99
6
Updated 19 days ago
gpt-oss
20B
120B
OpenAI's first open source LLM. Comes in 2 sizes: 20B and 120B. Supports configurable reasoning effort (low, medium, high). Trained for tool use. Apache 2.0 licensed.
887.6K
198
2
Updated 19 days ago
Qwen3-Coder
30B
480B
State-of-the-art, Mixture-of-Experts local coding model with native support for 256K context length. Available in 30B (3B active) and 480B (35B active) sizes.
158.2K
71
2
Updated 19 days ago
Ernie-4.5
21B
Medium-size Mixture-of-Experts model from Baidu's new Ernie 4.5 line of foundation models.
11.9K
7
Updated 19 days ago
LFM2
350M
700M
1.2B
LFM2 is a new generation of hybrid models developed by Liquid AI, specifically designed for edge AI and on-device deployment. It sets a new standard in terms of quality, speed, and memory efficiency.
47.2K
35
3
Updated 19 days ago
devstral
23.6B
24B
Devstral is a coding model from Mistral AI. It excels at using tools to explore codebases, editing multiple files and power software engineering agents.
66.8K
30
2
Updated 19 days ago
gemma-3n
4.5B
6.9B
Gemma 3n is a generative AI model optimized for use in everyday devices, such as phones, laptops, and tablets.
135.8K
60
2
Updated 19 days ago
Mistral Small
24B
Mistrall Small is a 'knowledge-dense' 24B multi-modal (image input) local model that supports up to 128 token context length.
58.4K
17
Updated 19 days ago
Magistral
23.6B
24B
MistralAI's open-weight reasoning model. 24B dense transformer model supporting up to 128K token context window. The model is capable of long chains of reasoning traces before providing answers.
111.4K
41
2
Updated 19 days ago
mistral-nemo
12B
General purpose dense transformer designed for multilingual use cases. Built in collaboration between MistralAI and NVIDIA.
19.4K
2
Updated 19 days ago
qwen2.5-vl
3B
7B
32B
72B
Qwen2.5-VL is a performant vision-language model, capable of recognizing common objects and text. Supports context length of 128k tokens in a variety of human languages.
53.7K
13
4
Updated 19 days ago
gemma-3
270M
1B
4B
12B
27B
State-of-the-art image + text input models from Google, built from the same research and tech used to create the Gemini models
610.8K
82
5
Updated 19 days ago
phi-4-reasoning
3.8B
14.7B
14.7B
Phi-4-mini-reasoning is a lightweight open model built upon synthetic data with a focus on high-quality, reasoning dense data.
98.3K
22
3
Updated 19 days ago
phi-4
3B
14B
phi-4 is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&A datasets.
17.8K
5
2
Updated 19 days ago
Codestral
22B
Mistral AI's latest coding model, Codestral can handle both instructions and code completions with ease in over 80 programming languages.
26.4K
14
Updated 19 days ago
Mistral
7B
One of the most popular open-source LLMs, Mistral's 7B Instruct model's balance of speed, size, and performance makes it a great general-purpose daily driver.
62.3K
28
Updated 19 days ago
Qwen3 (1st Generation)
4B
8B
14B
30B
32B
235B
The first batch of Qwen3 models (Qwen3-2504), a collection of dense and MoE models ranging from 4B to 235B. These are general purpose models that score highly on benchmarks.
295.3K
37
6
Updated 19 days ago
deepseek-r1
7B
8B
8B
14B
32B
70B
Distilled version of the DeepSeek-R1-0528 model, created by continuing the post-training process on the Qwen3 8B Base model using Chain-of-Thought (CoT) from DeepSeek-R1-0528.
382.7K
96
6
Updated 19 days ago