LMmy choosing

Model Catalog

New & noteworthy local models you can run on your own machine.

Olmo 3
7B
7B
32B
Olmo 3 is a family of Open language models designed to enable the science of language models.
15.3K
11
3
Updated 9 days ago
olmOCR 2
7B
The olmOCR 2 model is a Vision Language Model (VLM) from Allen AI.
18.2K
5
Updated 10 days ago
minimax-m2
230B
MiniMax M2 is a 230B MoE (10B active) model built for coding and agentic workflows
19.7K
13
Updated 24 days ago
gpt-oss-safeguard
20B
120B
gpt-oss-safeguard-20b and gpt-oss-safeguard-120b are open safety models from OpenAI, building on gpt-oss. Trained to help classify text content based on customizable policies.
6.1K
17
2
Updated 1 month ago
Qwen3-VL
2B
4B
8B
30B
32B
Qwen's latest vision-language model. Includes comprehensive upgrades to visual perception, spatial reasoning, and image understanding.
328.7K
49
5
Updated 1 month ago
Granite 4.0
3B
3B
7B
32B
Granite 4.0 language models are lightweight, state-of-the-art open models that natively support multilingual capabilities, coding tasks, RAG, tool use, and JSON output.
46.9K
32
4
Updated 1 month ago
seed-oss
36B
Advanced reasoning model from ByteDance with flexible "thinking budget" control and ability to reflect on the length of its own reasoning
38.1K
19
Updated 1 month ago
Qwen3
4B
4B
30B
30B
235B
235B
The latest version of the Qwen3 model family, featuring 4B, 30B, and 235B dense and MoE models, both thinking and non-thinking variants.
309.4K
105
6
Updated 1 month ago
gpt-oss
20B
120B
OpenAI's first open source LLM. Comes in 2 sizes: 20B and 120B. Supports configurable reasoning effort (low, medium, high). Trained for tool use. Apache 2.0 licensed.
1M
218
2
Updated 1 month ago
Qwen3-Coder
30B
480B
State-of-the-art, Mixture-of-Experts local coding model with native support for 256K context length. Available in 30B (3B active) and 480B (35B active) sizes.
178.6K
77
2
Updated 1 month ago
Ernie-4.5
21B
Medium-size Mixture-of-Experts model from Baidu's new Ernie 4.5 line of foundation models.
13K
8
Updated 1 month ago
LFM2
350M
700M
1.2B
LFM2 is a new generation of hybrid models developed by Liquid AI, specifically designed for edge AI and on-device deployment. It sets a new standard in terms of quality, speed, and memory efficiency.
50.2K
36
3
Updated 1 month ago
devstral
23.6B
24B
Devstral is a coding model from Mistral AI. It excels at using tools to explore codebases, editing multiple files and power software engineering agents.
70.2K
30
2
Updated 1 month ago
gemma-3n
4.5B
6.9B
Gemma 3n is a generative AI model optimized for use in everyday devices, such as phones, laptops, and tablets.
146K
61
2
Updated 1 month ago
Mistral Small
24B
Mistrall Small is a 'knowledge-dense' 24B multi-modal (image input) local model that supports up to 128 token context length.
62K
17
Updated 1 month ago
Magistral
23.6B
24B
MistralAI's open-weight reasoning model. 24B dense transformer model supporting up to 128K token context window. The model is capable of long chains of reasoning traces before providing answers.
126.7K
44
2
Updated 1 month ago
mistral-nemo
12B
General purpose dense transformer designed for multilingual use cases. Built in collaboration between MistralAI and NVIDIA.
22.1K
2
Updated 1 month ago
qwen2.5-vl
3B
7B
32B
72B
Qwen2.5-VL is a performant vision-language model, capable of recognizing common objects and text. Supports context length of 128k tokens in a variety of human languages.
61.3K
17
4
Updated 1 month ago
gemma-3
270M
1B
4B
12B
27B
State-of-the-art image + text input models from Google, built from the same research and tech used to create the Gemini models
662K
84
5
Updated 1 month ago
phi-4-reasoning
3.8B
14.7B
14.7B
Phi-4-mini-reasoning is a lightweight open model built upon synthetic data with a focus on high-quality, reasoning dense data.
104.1K
23
3
Updated 1 month ago
phi-4
3B
14B
phi-4 is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&A datasets.
19.5K
7
2
Updated 1 month ago
Codestral
22B
Mistral AI's latest coding model, Codestral can handle both instructions and code completions with ease in over 80 programming languages.
29.3K
16
Updated 1 month ago
Mistral
7B
One of the most popular open-source LLMs, Mistral's 7B Instruct model's balance of speed, size, and performance makes it a great general-purpose daily driver.
68.4K
29
Updated 1 month ago
Qwen3 (1st Generation)
4B
8B
14B
30B
32B
235B
The first batch of Qwen3 models (Qwen3-2504), a collection of dense and MoE models ranging from 4B to 235B. These are general purpose models that score highly on benchmarks.
322.7K
37
6
Updated 1 month ago
deepseek-r1
7B
8B
8B
14B
32B
70B
Distilled version of the DeepSeek-R1-0528 model, created by continuing the post-training process on the Qwen3 8B Base model using Chain-of-Thought (CoT) from DeepSeek-R1-0528.
418.6K
107
6
Updated 1 month ago