NVIDIA DGX Spark
LM Studio is now supported on NVIDIA's DGX Spark! We're excited to be a launch partner for this new device. This also means that starting now, LM Studio ships for Linux on ARM (aarch64).
The Spark is a tiny but mighty Linux ARM box with 128GB of unified RAM, and 273 GB/s memory bandwidth. The form factor is quite small (similar to a Mac Mini), and it's styled like a miniature version the bigger DGX systems.
We brought up a new variant of LM Studio's llama.cpp engine with support for CUDA 13.
On the Spark, you can run the LM Studio GUI and use the Chat UI as you would on any other computer.
Additionally, you could utilize LM Studio's API server and make the Spark your own private LLM server and connect to it over the network.
First, grab the LM Studio Linux ARM64 AppImage from the download page and install it on your Spark.
Next, download a model like gpt-oss or Qwen3 Coder.
lms get openai/gpt-oss-20b
In the LM Studio GUI, head to the Developer tab and in the server settings, turn on Serve on Local Network. This will allow other devices on your network to connect to the Spark's LLM server.
Turn on Serve on Local Network in the Developer tab
In (3) above, you can see which IP address and port the server is running on. By default, it's port 1234.
P.S. you can also start the server from the command line:
lms server start # --port 1234
Assuming your network isn't locked down in such a way that prevents devices from talking to each other, you can connect to the Spark's LLM server from another device on the same subnet.
Give it a try by sending a request to list the available models:
curl http://<SPARK_IP_ADDRESS>:1234/v1/models
If you can reach the device, you're good to go!
Utilize LM Studio's SDKs (lmstudio-js, lmstudio-python) or OpenAI-compatible API to connect to the Spark's LLM server from your apps, just like you would a hosted LLM service.