Documentation
Core
LM Studio REST API
OpenAI Compatible Endpoints
Anthropic Compatible Endpoints
Core
LM Studio REST API
OpenAI Compatible Endpoints
Anthropic Compatible Endpoints
GUI-less operation of LM Studio: run in the background, start on machine login, and load models on demand
LM Studio can be run as a background service without the GUI. There are two ways to do this:
llmster is the core of the LM Studio desktop app, packaged to be server-native, without reliance on the GUI. It can run on Linux boxes, cloud servers, GPU rigs, or your local machine without the GUI. See the LM Studio 0.4.0 release post for more details.

Linux / Mac
curl -fsSL https://lmstudio.ai/install.sh | bash
Windows
irm https://lmstudio.ai/install.ps1 | iex
lms daemon up
See the daemon CLI docs for full reference.
For setting up llmster as a startup task on Linux, see Linux Startup Task.
This works on Mac, Windows, and Linux machines with a graphical user interface. It's useful if you already have the desktop app installed and want it to run as a background service.
Head to app settings (Cmd / Ctrl + ,) and check the box to run the LLM server on login.

Enable the LLM server to start on machine login
When this setting is enabled, exiting the app will minimize it to the system tray, and the LLM server will continue to run in the background.
Your last server state will be saved and restored on app or service launch.
To achieve this programmatically:
lms server start
Applies to both options. Useful when using LM Studio as an LLM service with other frontends or applications.

Load models on demand
/v1/models will return all downloaded models, not only the ones loaded into memory/v1/models will return only the models loaded into memoryJIT loaded models will be auto-unloaded from memory by default after a set period of inactivity (learn more).
Chat with other LM Studio developers, discuss LLMs, hardware, and more on the LM Studio Discord server.
Please report bugs and issues in the lmstudio-bug-tracker GitHub repository.
This page's source is available on GitHub
On this page
Option 1: llmster (recommended)
Install llmster
Start llmster
Option 2: Desktop app in headless mode
Run the LLM service on machine login
Auto Server Start
Just-In-Time (JIT) model loading for REST endpoints
- When JIT loading is ON
- When JIT loading is OFF
- What about auto unloading?
Community