LM Studio 0.3.16 Release Notes
LM Studio 0.3.16 is available now as a stable release. This version includes the public preview of community presets, automatic deletion of least recently used Runtime Extension Packs, and a way to use LLMs as text embedding models. It also introduces the lms chat
command to the CLI, allowing you to chat with a model in the terminal.
Ability to publish your presets to the LM Studio community hub and share them with others
Staff Picks polish: introduce new aggregate format that combines formats (e.g. GGUF and MLX) into a single entry
Custom Settings for models
Add "Offload KV Cache to GPU Memory" option to model load options and GPU settings
Add lms chat
command to the CLI to chat with a model in the terminal (Thanks @mayfer)
LM Studio REST API (/api/v0
): return model capabilities in GET /models
response
"capabilities": ["tool_use"]
Auto-deletion of least recently used Runtime Extension Packs
⌘/Ctrl + ,
)Show System Prompt button in chat top bar when sidebar is collapsed
Use Cmd / Ctrl + Shift + D
to create a duplicate of the current chat
Use Cmd / Ctrl + E
to open the System Prompt editor
Use Cmd / Ctrl + W
to close the System Prompt editor tab when it is open
App Settings (⌘/Ctrl + ,): sections now have their own tabs for easier navigation
Add button to access downloads panel in "User" UI mode
Added a dropdown in the model editor (⚙️ in My Models) to allow overriding the domain type of a model
Add "Reveal in Finder" context menu option on the chat sidebar body
[MLX] Register chat_template.jinja
as a source for chat templates
Bug Fixes:
Fix Qwen3 jinja template parsing error "OpenSquareBracket !== CloseStatement"
Fix nomic-bert-moe, nomic-embed-code not showing under embedding models
Fix flickering horizontal line at bottom of generating message
Fix inability to be create a new folder when no chat is selected
Fix prompt processing % UI not showing
Bugfix: [OpenAI-like REST API] Don't add empty function name string after the first packet of streamed function calls
Bugfix: Correctly filter out unsupported models when listing downloaded models
Bugfix: Don't assume Vulkan iGPU memory is Shared Memory
Fix architecture badge text alignment
Fix SmolVLM jinja render error "Expected iterable type in for loop: got StringValue"
Fixed GTE Qwen embedding models are not recognized as text embedding
Fixed a bug where if no engine extension pack for a specific format is installed, the indexing will fail
Fixed a bug where some strings were parsed as [object Object]
when using RAG
Fixed a bug showing duplicate models in Load Model in tray menu
[MLX] Fix installation issue that caused ModuleNotFoundError: No module named 'mlx_engine'