Documentation
Core
LM Studio REST API
OpenAI Compatible Endpoints
Anthropic Compatible Endpoints
Core
LM Studio REST API
OpenAI Compatible Endpoints
Anthropic Compatible Endpoints
Create responses with support for streaming, reasoning, prior response state, and optional Remote MCP tools.
POSTcurl http://localhost:1234/v1/responses \ -H "Content-Type: application/json" \ -d '{ "model": "openai/gpt-oss-20b", "input": "Provide a prime number less than 50", "reasoning": { "effort": "low" } }'
Use the id from a previous response as previous_response_id.
curl http://localhost:1234/v1/responses \ -H "Content-Type: application/json" \ -d '{ "model": "openai/gpt-oss-20b", "input": "Multiply it by 2", "previous_response_id": "resp_123" }'
curl http://localhost:1234/v1/responses \ -H "Content-Type: application/json" \ -d '{ "model": "openai/gpt-oss-20b", "input": "Hello", "stream": true }'
You will receive SSE events such as response.created, response.output_text.delta, and response.completed.
Enable Remote MCP in the app (Developer → Settings). Example payload using an MCP server tool:
curl http://localhost:1234/v1/responses \ -H "Content-Type: application/json" \ -d '{ "model": "ibm/granite-4-micro", "input": "What is the top trending model on hugging face?", "tools": [ { "type": "mcp", "server_label": "huggingface", "server_url": "https://huggingface.co/mcp", "allowed_tools": [ "model_search" ] } ] }'
This page's source is available on GitHub
On this page