This is a plugin for Simon Willison's LLM command-line utility that lets you talk to models running on a local LMStudio server.
Make sure you have llm installed, then install this plugin from PyPI:
pip install llm-lmstudioOr, to install the latest development version directly from GitHub:
llm install -U git+https://github.com/agustif/llm-lmstudio.gitAlternatively, llm install llm-lmstudio will also find and install the plugin.
First, you need LMStudio running with a model loaded. The plugin talks to the LMStudio server API, which usually runs at http://localhost:1234.
Once the server is up, the plugin should automatically find the models you have loaded in LMStudio. You can check this using the llm models command:
llm models listYou should see your LMStudio models listed, prefixed with lmstudio/ (e.g., lmstudio/llava-v1.5-7b).
To run a prompt against a model:
# Replace 'your-model-id' with the actual ID shown in 'llm models list'
# e.g., llm -m lmstudio/llava-v1.5-7b "Describe this image"
llm -m lmstudio/your-model-id -o temperature 0.7 -o max_tokens 100 "Tell me a joke"To start an interactive chat session:
llm chat -m lmstudio/your-model-idYou can exit the chat by typing exit or quit.
The plugin supports vision-language models (VLMs).
- When using a VLM, you can attach images using the standard
llmattachment syntax:Or for a single prompt:llm chat -m lmstudio/your-vlm-id -a path/to/image.png "Describe this image"llm -m lmstudio/your-vlm-id -a path/to/image.png "What is in this picture?" - The plugin will encode the image and send it to the model.
- This feature's success depends on the specific VLM, its configuration in LM Studio, and LM Studio's API correctly handling image data.
- Models that support vision may have a
👁️(eye icon) in theirdisplay_suffixwhen inspected viallm inspect -m lmstudio/your-vlm-id, though this may not always render inllm models list.
Some models loaded in LMStudio can call tools. The plugin will surface those tool invocations and their results when interacting with such models.
Example:
$ llm --tool llm_version "What version of LLM is this?" --tdFor more information about tool calling support consult the llm documentation on tools.
If you have embedding models loaded in LMStudio (their names usually contain "embed"), the plugin will register them too. You can list them with:
llm embed-modelsYou should see models like lmstudio/text-embedding-nomic-embed-text-v1.5@f16.
To generate embeddings for text using one of these models:
llm embed -m lmstudio/your-embedding-model-id -c "This is the text to embed"The plugin connects to the LMStudio server API. By default, it tries http://localhost:1234.
You can configure the server URL(s) using the LMSTUDIO_API_BASE environment variable.
- For a single server:
export LMSTUDIO_API_BASE="http://your-server-address:port"
- For multiple servers (the plugin will try them in order):
export LMSTUDIO_API_BASE="http://server1:1234,http://server2:5678,https://server3:custom_port"
The variable accepts one or more http[s]://host:port values, separated by commas (spaces around commas are optional). The plugin automatically attempts to append /v1 or /api/v0 to the determined base URL(s) as needed when probing the server.
You can pass generation options supported by the LMStudio API (like temperature, max_tokens, top_p, stop) using the -o flag:
llm -m lmstudio/your-model-id -o temperature 0.7 -o max_tokens 100 "Tell me a joke"If a selected model is not currently loaded in LM Studio, the plugin will attempt to automatically load it by invoking lms load <model_id> (if lms CLI is installed and configured). You may see progress messages from LM Studio in your terminal during this process.
To set up this plugin for development:
- Clone the repository.
- Run tests
uv run pytest
Or do the classic complicated dance:
- Create a virtual environment:
python -m venv .venv - Activate it:
source .venv/bin/activate - Install dependencies, including dev dependencies: `pip install -e . --group dev
- Run tests:
pytest
The asynchronous tests in tests/test_llm_lmstudio_async.py (among others) use pytest-vcr to record and replay HTTP interactions with the LM Studio server. To record new cassettes:
- Ensure LM Studio is running with the target model(s) loaded (e.g.,
llava-v1.5-7b). - Temporarily set
record_mode='all'in the@pytest.mark.vcrdecorator for the relevant tests intests/test_llm_lmstudio_async.py. - Run
pytest -v -s tests/test_llm_lmstudio_async.py. - Cassettes will be saved in
tests/cassettes/. - Important: Change
record_modeback to'once'after recording. - Commit the new/updated cassettes.
- The reliability and capabilities of image support can vary significantly based on the specific VLM and its implementation within LM Studio.