Documentation
Getting Started
Predicting with LLMs
Agentic Flows
Text Embedding
Tokenization
Getting Started
Predicting with LLMs
Agentic Flows
Text Embedding
Tokenization
Get Load Config
Required Python SDK version: 1.2.0
LM Studio allows you to configure certain parameters when loading a model through the server UI or through the API.
You can retrieve the config with which a given model was loaded using the SDK.
In the below examples, the LLM reference can be replaced with an embedding model reference without requiring any other changes.
Context length is a special case that has its own method.
import lmstudio as lms
model = lms.llm()
print(model.get_load_config())