Awesome
<p align="center"> <br/> <picture> <source media="(prefers-color-scheme: dark)" srcset="https://files.lmstudio.ai/lms-dark.png"> <source media="(prefers-color-scheme: light)" srcset="https://files.lmstudio.ai/lms-light.png"> <img alt="lmstudio cli logo" src="https://files.lmstudio.ai/lms-light.png" width="180"> </picture> <br/> <br/> </p> <p align="center"><bold><code>lms</code> - Command Line Tool for <a href="https://lmstudio.ai/">LM Studio</a></bold></p> <p align="center">Built with <bold><code><a href="https://github.com/lmstudio-ai/lmstudio.js">lmstudio.js</a></code></bold></p>Installation
lms
ships with LM Studio 0.2.22 and newer.
To set it up, run the built-in bootstrap
command like so:
-
Windows:
cmd /c %USERPROFILE%/.cache/lm-studio/bin/lms.exe bootstrap
-
Linux/macOS:
~/.cache/lm-studio/bin/lms bootstrap
To check if the bootstrapping was successful, run the following in a 👉 new terminal window 👈:
lms
Usage
You can use lms --help
to see a list of all available subcommands.
For details about each subcommand, run lms <subcommand> --help
.
Here are some frequently used commands:
lms status
- To check the status of LM Studio.lms server start
- To start the local API server.lms server stop
- To stop the local API server.lms ls
- To list all downloaded models.lms ls --detailed
- To list all downloaded models with detailed information.lms ls --json
- To list all downloaded models in machine-readable JSON format.
lms ps
- To list all loaded models available for inferencing.lms ps --json
- To list all loaded models available for inferencing in machine-readable JSON format.
lms load --gpu max
- To load a model with maximum GPU accelerationlms load <model path> --gpu max -y
- To load a model with maximum GPU acceleration without confirmation
lms unload <model identifier>
- To unload a modellms unload --all
- To unload all models
lms create
- To create a new project with LM Studio SDKlms log stream
- To stream logs from LM Studio