docs: update based on suggestions
This commit is contained in:
parent
c8c9ec3064
commit
28832b3e18
@ -170,7 +170,7 @@ Navigate to the `~/jan/models` folder. Create a folder named `<modelname>`, for
|
|||||||
|
|
||||||
:::warning
|
:::warning
|
||||||
|
|
||||||
- Windows users may need to use double backslashes in the `url` property, for example: `C:\\Users\\username\\filename.gguf`.
|
- If you are using Windows, you need to use double backslashes in the url property, for example: `C:\\Users\\username\\filename.gguf`.
|
||||||
|
|
||||||
:::
|
:::
|
||||||
|
|
||||||
|
|||||||
@ -24,13 +24,14 @@ With [Ollama](https://ollama.com/), you can run large language models locally. I
|
|||||||
|
|
||||||
### 1. Start the Ollama Server
|
### 1. Start the Ollama Server
|
||||||
|
|
||||||
Firstly, you can select the model you want to use from the [Ollama library](https://ollama.com/library). Then, run your model by using the following command:
|
1. Select the model you want to use from the [Ollama library](https://ollama.com/library).
|
||||||
|
2. Run your model by using the following command:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
ollama run <model-name>
|
ollama run <model-name>
|
||||||
```
|
```
|
||||||
|
|
||||||
According to the [Ollama documentation on OpenAI compatibility](https://github.com/ollama/ollama/blob/main/docs/openai.md), you can use the `http://localhost:11434/v1/chat/completions` endpoint to interact with the Ollama server. Thus, modify the `openai.json` file in the `~/jan/engines` folder to include the full URL of the Ollama server.
|
3. According to the [Ollama documentation on OpenAI compatibility](https://github.com/ollama/ollama/blob/main/docs/openai.md), you can use the `http://localhost:11434/v1/chat/completions` endpoint to interact with the Ollama server. Thus, modify the `openai.json` file in the `~/jan/engines` folder to include the full URL of the Ollama server.
|
||||||
|
|
||||||
```json title="~/jan/engines/openai.json"
|
```json title="~/jan/engines/openai.json"
|
||||||
{
|
{
|
||||||
@ -40,13 +41,14 @@ According to the [Ollama documentation on OpenAI compatibility](https://github.c
|
|||||||
|
|
||||||
### 2. Modify a Model JSON
|
### 2. Modify a Model JSON
|
||||||
|
|
||||||
Navigate to the `~/jan/models` folder. Create a folder named `<ollam-modelname>`, for example, `lmstudio-phi-2` and create a `model.json` file inside the folder including the following configurations:
|
1. Navigate to the `~/jan/models` folder.
|
||||||
|
2. Create a folder named `<ollam-modelname>`, for example, `lmstudio-phi-2`.
|
||||||
|
3. Create a `model.json` file inside the folder including the following configurations:
|
||||||
|
|
||||||
- Ensure the filename must be `model.json`.
|
- Set the `id` property to the model name as Ollama model name.
|
||||||
- Ensure the `id` property is set to the model name as Ollama model name.
|
- Set the `format` property to `api`.
|
||||||
- Ensure the `format` property is set to `api`.
|
- Set the `engine` property to `openai`.
|
||||||
- Ensure the `engine` property is set to `openai`.
|
- Set the `state` property to `ready`.
|
||||||
- Ensure the `state` property is set to `ready`.
|
|
||||||
|
|
||||||
```json title="~/jan/models/llama2/model.json"
|
```json title="~/jan/models/llama2/model.json"
|
||||||
{
|
{
|
||||||
@ -77,7 +79,8 @@ Navigate to the `~/jan/models` folder. Create a folder named `<ollam-modelname>`
|
|||||||
|
|
||||||
### 3. Start the Model
|
### 3. Start the Model
|
||||||
|
|
||||||
Restart Jan and navigate to the Hub. Locate your model and click the Use button.
|
1. Restart Jan and navigate to the **Hub**.
|
||||||
|
2. Locate your model and click the **Use** button.
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user