This repository has been archived by the owner on Oct 14, 2024. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 7
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
docs: Updated the docs to reflect the new structure.
- Loading branch information
1 parent
9bad606
commit 17561c7
Showing
8 changed files
with
325 additions
and
237 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,10 @@ | ||
{ | ||
"manage-models": { | ||
"title": "Managing Models", | ||
"href": "/docs/models/manage-models" | ||
}, | ||
"model-parameters": { | ||
"title": "Model Parameters", | ||
"href": "/docs/models/model-parameters" | ||
} | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,145 @@ | ||
--- | ||
title: Managing Models | ||
description: Manage your interaction with AI locally. | ||
keywords: | ||
[ | ||
Jan, | ||
Customizable Intelligence, LLM, | ||
local AI, | ||
privacy focus, | ||
free and open source, | ||
private and offline, | ||
conversational AI, | ||
no-subscription fee, | ||
large language models, | ||
threads, | ||
chat history, | ||
thread history, | ||
] | ||
--- | ||
import { Callout, Steps } from 'nextra/components' | ||
|
||
# Overview | ||
This guide provides comprehensive instructions on adding, customizing, and deleting models within the Jan platform. | ||
|
||
## Add Models | ||
|
||
There are various ways to add models to Jan. | ||
|
||
Currently, Jan natively supports the following model formats: | ||
- GGUF (through a llama.cpp engine) | ||
- TensorRT (through a TRT-LLM engine) | ||
|
||
### Download from Jan Hub | ||
|
||
You can choose from a list of popular, recommended models directly from Jan app's Model Hub. | ||
These models are preconfigured with optimal runtime parameters. This is the easiest way to get started. | ||
|
||
1. Open the Jan app and navigate to the Hub. | ||
2. Select models, clicking the `v` dropdown for more information. Models with the `Recommended` label will likely run faster on your computer. | ||
3. After downloading a model, click **Use** to activate it. Ensure it's selected in the model dropdown for your thread. | ||
|
||
### Add a Model Manually | ||
You can also add a specific model that is not available within the **Hub** section by following the steps below: | ||
1. Open the Jan app. | ||
2. Click the **gear icon (⚙️)** on the bottom left of your screen. | ||
3. Under the **Settings screen**, click **Advanced Settings**. | ||
4. Open the Jan Data folder. | ||
5. Head to the `~/jan/models/`. | ||
6. Create a new model folder. | ||
7. Create a `model.json` file inside the folder. | ||
8. Insert the following `model.json` default code: | ||
```json | ||
{ | ||
"id": "<unique_identifier_of_the_model>", | ||
"object": "<type_of_object, e.g., model, tool>", | ||
"name": "<name_of_the_model>", | ||
"version": "<version_number>", | ||
"description": "<brief_description_of_the_model>", | ||
"format": "<format_of_the_model_api_or_other>", | ||
"settings": "<additional_settings_as_needed>", | ||
"parameters": { | ||
"max_tokens": "<maximum_number_of_tokens_the_model_can_generate>", | ||
"temperature": "<temperature_setting_for_randomness_in_generation>" | ||
}, | ||
"metadata": { | ||
"author": "<name_of_the_creator_or_organization>", | ||
"tags": ["<list_of_relevant_tags_describing_the_model>"] | ||
}, | ||
"engine": "<engine_or_platform_the_model_runs_on>", | ||
"source": "<url_or_source_of_the_model_information>" | ||
} | ||
``` | ||
<Callout> | ||
If you've set up your model's configuration in `nitro.json`, please note that `model.json` can **overwrite** the settings. | ||
</Callout> | ||
There are two important fields in `model.json` that you need to set: | ||
|
||
#### Settings | ||
|
||
This is the field where you can set your engine configurations. There are two important fields that you need to define for your local models: | ||
|
||
| Term | Description | | ||
| ----------------- | --------------------------------------------------------------------- | | ||
| `ctx_len` | Defined based on the model's context size. | | ||
| `prompt_template` | Defined based on the model's trained template (e.g., ChatML, Alpaca). | | ||
|
||
To set up the `prompt_template` based on your model, follow the steps below: | ||
1. Visit [Hugging Face](https://huggingface.co/), an open-source machine-learning platform. | ||
2. Find the current model that you're using (e.g., [Gemma 7b it](https://huggingface.co/google/gemma-7b-it)). | ||
3. Review the text and identify the template. | ||
|
||
#### Parameters | ||
|
||
`parameters` are the adjustable settings that affect how your model operates or processes the data. | ||
The fields in `parameters` are typically general and can be the same across models. An example is provided below: | ||
|
||
<Callout type='info'> | ||
To see the complete list of a model's parameters, please see [below](/docs/models#model-parameters). | ||
</Callout> | ||
|
||
```json | ||
"parameters":{ | ||
"temperature": 0.7, | ||
"top_p": 0.95, | ||
"stream": true, | ||
"max_tokens": 4096, | ||
"frequency_penalty": 0, | ||
"presence_penalty": 0 | ||
} | ||
``` | ||
|
||
### Import or Symlink Local Models | ||
|
||
You can also point to existing model binary files on your local filesystem. | ||
This is the easiest and most space-efficient way if you have already used other local AI applications. | ||
|
||
1. Navigate to the Hub. | ||
2. Click on `Import Model` at the top. | ||
3. Select the model or the folder containing multiple models. | ||
4. Optionally, check the box to symlink the model files instead of copying them over the Jan Data Folder. This saves disk space. | ||
|
||
<Callout type="warning"> | ||
Windows users should drag and drop the model file, as **Click to Upload** might not show the model files in Folder Preview. | ||
</Callout> | ||
|
||
### Download with Hugging Face URL | ||
You can download a model with a Hugging Face URL or the model name by following the steps below: | ||
1. Navigate to the [Hugging Face](https://huggingface.co/models). | ||
2. Select the model you want to use. | ||
3. Copy the Model name or the URL, for example: `MaziyarPanahi/Mixtral-8x22B-Instruct-v0.1-GGUF` or `https://huggingface.co/MaziyarPanahi/Mixtral-8x22B-Instruct-v0.1-GGUF`. | ||
<Callout type="warning"> | ||
Only `GGUF` models are supported for this feature. | ||
</Callout> | ||
4. Head back to Jan application. | ||
5. Click the **Hub** tab. | ||
6. Paste the **URL** or the **model name** you have copied into the search bar. | ||
7. A new window will display all the available model versions. | ||
8. Click the **Download** button to download the model. | ||
|
||
## Delete Models | ||
To delete a model: | ||
|
||
1. Go to **Settings**. | ||
2. Go to **My Models**. | ||
3. Select the three dots next and select `Delete model`. |
Oops, something went wrong.