RooPrompts/docs/providers/lmstudio.md
2025-05-17 14:58:58 +05:30

40 lines
2.6 KiB
Markdown

---
sidebar_label: LM Studio
---
# Using LM Studio With Roo Code
Roo Code supports running models locally using LM Studio. LM Studio provides a user-friendly interface for downloading, configuring, and running local language models. It also includes a built-in local inference server that emulates the OpenAI API, making it easy to integrate with Roo Code.
**Website:** [https://lmstudio.ai/](https://lmstudio.ai/)
## Setting Up LM Studio
1. **Download and Install LM Studio:** Download LM Studio from the [LM Studio website](https://lmstudio.ai/).
2. **Download a Model:** Use the LM Studio interface to search for and download a model. Some recommended models include:
* CodeLlama models (e.g., `codellama:7b-code`, `codellama:13b-code`, `codellama:34b-code`)
* Mistral models (e.g., `mistralai/Mistral-7B-Instruct-v0.1`)
* DeepSeek Coder models (e.g., `deepseek-coder:6.7b-base`)
* Any other model that is supported by Roo, or for which you can set the context window.
Look for models in the GGUF format. LM Studio provides a search interface to find and download models.
3. **Start the Local Server:**
* Open LM Studio.
* Click the **"Local Server"** tab (the icon looks like `<->`).
* Select the model you downloaded.
* Click **"Start Server"**.
## Configuration in Roo Code
1. **Open Roo Code Settings:** Click the gear icon (<Codicon name="gear" />) in the Roo Code panel.
2. **Select Provider:** Choose "LM Studio" from the "API Provider" dropdown.
3. **Enter Model ID:** Enter the *file name* of the model you loaded in LM Studio (e.g., `codellama-7b.Q4_0.gguf`). You can find this in the LM Studio "Local Server" tab.
4. **(Optional) Base URL:** By default, Roo Code will connect to LM Studio at `http://localhost:1234`. If you've configured LM Studio to use a different address or port, enter the full URL here.
## Tips and Notes
* **Resource Requirements:** Running large language models locally can be resource-intensive. Make sure your computer meets the minimum requirements for the model you choose.
* **Model Selection:** LM Studio provides a wide range of models. Experiment to find the one that best suits your needs.
* **Local Server:** The LM Studio local server must be running for Roo Code to connect to it.
* **LM Studio Documentation:** Refer to the [LM Studio documentation](https://lmstudio.ai/docs) for more information.
* **Troubleshooting:** If you see a "Please check the LM Studio developer logs to debug what went wrong" error, you may need to adjust the context length settings in LM Studio.