LM Studio
How to Integrate LM Studio with Jan
LM Studio enables you to explore, download, and run local Large Language Models (LLMs). You can integrate Jan with LM Studio using two methods:
- Integrate the LM Studio server with Jan UI
- Migrate your downloaded model from LM Studio to Jan.
To integrate LM Studio with Jan follow the steps below:
In this guide, we're going to show you how to connect Jan to LM Studio using the second method. We'll use the Phi 2 - GGUF model from Hugging Face as our example.
Step 1: Server Setup
-
Access the
Local Inference Server
within LM Studio. -
Select your desired model.
-
Start the server after configuring the port and options.
-
Update the
openai.json
file in~/jan/engines
to include the LM Studio server's full URL.
{
"full_url": "http://localhost:<port>/v1/chat/completions"
}
Replace (port)
with your chosen port number. The default is 1234.
Step 2: Model Configuration
- Navigate to
~/jan/models
. - Create a folder named
lmstudio-(modelname)
, likelmstudio-phi-2
. - Inside, create a
model.json
file with these options:- Set
format
toapi
. - Specify
engine
asopenai
. - Set
state
toready
.
- Set
{
"sources": [
{
"filename": "phi-2-GGUF",
"url": "https://huggingface.co/TheBloke/phi-2-GGUF"
}
],
"id": "lmstudio-phi-2",
"object": "model",
"name": "LM Studio - Phi 2 - GGUF",
"version": "1.0",
"description": "TheBloke/phi-2-GGUF",
"format": "api",
"settings": {},
"parameters": {},
"metadata": {
"author": "Microsoft",
"tags": ["General", "Big Context Length"]
},
"engine": "openai"
}
For more details regarding the model.json
settings and parameters fields, please see here.
Step 3: Starting the Model
- Restart Jan and proceed to the Hub.
- Locate your model and click Use to activate it.
Migrating Models from LM Studio to Jan (version 0.4.6 and older)
Step 1: Model Migration
- In LM Studio, navigate to
My Models
and access your model folder. - Copy the model folder to
~/jan/models
. - Ensure the folder name matches the model name in the
.gguf
filename. Rename as necessary.
Step 2: Activating the Model
- Restart Jan and navigate to the Hub, where the model will be automatically detected.
- Click Use to use the model.
Direct Access to LM Studio Models from Jan (version 0.4.7+)
Starting from version 0.4.7, Jan enables direct import of LM Studio models using absolute file paths.
Step 1: Locating the Model Path
- Access
My Models
in LM Studio and locate your model folder. - Obtain the absolute path of your model.
Step 2: Model Configuration
- Go to
~/jan/models
. - Create a folder named
(modelname)
(e.g.,phi-2.Q4_K_S
). - Inside, craft a
model.json
file:- Set
id
to match the folder name. - Use the direct binary download link ending in
.gguf
as theurl
. You can now use the absolute filepath of the model file. - Set
engine
asnitro
.
- Set
{
"object": "model",
"version": 1,
"format": "gguf",
"sources": [
{
"filename": "phi-2.Q4_K_S.gguf",
"url": "<absolute-path-of-model-file>"
}
],
"id": "phi-2.Q4_K_S",
"name": "phi-2.Q4_K_S",
"created": 1708308111506,
"description": "phi-2.Q4_K_S - user self import model",
"settings": {
"ctx_len": 4096,
"embedding": false,
"prompt_template": "{system_message}\n### Instruction: {prompt}\n### Response:",
"llama_model_path": "phi-2.Q4_K_S.gguf"
},
"parameters": {
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 2048,
"stop": ["<endofstring>"],
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"size": 1615568736,
"author": "User",
"tags": []
},
"engine": "nitro"
}
For Windows users, ensure to include double backslashes in the URL property, such as C:\\Users\\username\\filename.gguf
.
Step 3: Starting the Model
- Restart Jan and proceed to the Hub.
- Locate your model and click Use to activate it.