added huggingface page and updated readme
23
README.md
@ -12,16 +12,14 @@
|
|||||||
</p>
|
</p>
|
||||||
|
|
||||||
<p align="center">
|
<p align="center">
|
||||||
<a href="https://jan.ai/docs/quickstart">Getting Started</a>
|
<a href="https://jan.ai/docs/quickstart">Getting Started</a>
|
||||||
- <a href="https://jan.ai/docs">Docs</a>
|
- <a href="https://jan.ai/docs">Docs</a>
|
||||||
- <a href="https://jan.ai/changelog">Changelog</a>
|
- <a href="https://jan.ai/changelog">Changelog</a>
|
||||||
- <a href="https://github.com/menloresearch/jan/issues">Bug reports</a>
|
- <a href="https://github.com/menloresearch/jan/issues">Bug reports</a>
|
||||||
- <a href="https://discord.gg/AsJ8krTT3N">Discord</a>
|
- <a href="https://discord.gg/AsJ8krTT3N">Discord</a>
|
||||||
</p>
|
</p>
|
||||||
|
|
||||||
Jan is a ChatGPT-alternative that runs 100% offline on your device. Our goal is to make it easy for a layperson to download and run LLMs and use AI with **full control** and **privacy**.
|
Jan is an AI assistant that runs 100% offline on your device. Download and run LLMs with **full control** and **privacy**.
|
||||||
|
|
||||||
**⚠️ Jan is in active development.**
|
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
@ -31,43 +29,32 @@ Because clicking a button is still the easiest way to get started:
|
|||||||
<tr>
|
<tr>
|
||||||
<td><b>Platform</b></td>
|
<td><b>Platform</b></td>
|
||||||
<td><b>Stable</b></td>
|
<td><b>Stable</b></td>
|
||||||
<td><b>Beta</b></td>
|
|
||||||
<td><b>Nightly</b></td>
|
<td><b>Nightly</b></td>
|
||||||
</tr>
|
</tr>
|
||||||
<tr>
|
<tr>
|
||||||
<td><b>Windows</b></td>
|
<td><b>Windows</b></td>
|
||||||
<td><a href='https://app.jan.ai/download/latest/win-x64'>jan.exe</a></td>
|
<td><a href='https://app.jan.ai/download/latest/win-x64'>jan.exe</a></td>
|
||||||
<td><a href='https://app.jan.ai/download/beta/win-x64'>jan.exe</a></td>
|
|
||||||
<td><a href='https://app.jan.ai/download/nightly/win-x64'>jan.exe</a></td>
|
<td><a href='https://app.jan.ai/download/nightly/win-x64'>jan.exe</a></td>
|
||||||
</tr>
|
</tr>
|
||||||
<tr>
|
<tr>
|
||||||
<td><b>macOS</b></td>
|
<td><b>macOS</b></td>
|
||||||
<td><a href='https://app.jan.ai/download/latest/mac-universal'>jan.dmg</a></td>
|
<td><a href='https://app.jan.ai/download/latest/mac-universal'>jan.dmg</a></td>
|
||||||
<td><a href='https://app.jan.ai/download/beta/mac-universal'>jan.dmg</a></td>
|
|
||||||
<td><a href='https://app.jan.ai/download/nightly/mac-universal'>jan.dmg</a></td>
|
<td><a href='https://app.jan.ai/download/nightly/mac-universal'>jan.dmg</a></td>
|
||||||
</tr>
|
</tr>
|
||||||
<tr>
|
<tr>
|
||||||
<td><b>Linux (deb)</b></td>
|
<td><b>Linux (deb)</b></td>
|
||||||
<td><a href='https://app.jan.ai/download/latest/linux-amd64-deb'>jan.deb</a></td>
|
<td><a href='https://app.jan.ai/download/latest/linux-amd64-deb'>jan.deb</a></td>
|
||||||
<td><a href='https://app.jan.ai/download/beta/linux-amd64-deb'>jan.deb</a></td>
|
|
||||||
<td><a href='https://app.jan.ai/download/nightly/linux-amd64-deb'>jan.deb</a></td>
|
<td><a href='https://app.jan.ai/download/nightly/linux-amd64-deb'>jan.deb</a></td>
|
||||||
</tr>
|
</tr>
|
||||||
<tr>
|
<tr>
|
||||||
<td><b>Linux (AppImage)</b></td>
|
<td><b>Linux (AppImage)</b></td>
|
||||||
<td><a href='https://app.jan.ai/download/latest/linux-amd64-appimage'>jan.AppImage</a></td>
|
<td><a href='https://app.jan.ai/download/latest/linux-amd64-appimage'>jan.AppImage</a></td>
|
||||||
<td><a href='https://app.jan.ai/download/beta/linux-amd64-appimage'>jan.AppImage</a></td>
|
|
||||||
<td><a href='https://app.jan.ai/download/nightly/linux-amd64-appimage'>jan.AppImage</a></td>
|
<td><a href='https://app.jan.ai/download/nightly/linux-amd64-appimage'>jan.AppImage</a></td>
|
||||||
</tr>
|
</tr>
|
||||||
</table>
|
</table>
|
||||||
|
|
||||||
Download from [jan.ai](https://jan.ai/) or [GitHub Releases](https://github.com/menloresearch/jan/releases).
|
Download from [jan.ai](https://jan.ai/) or [GitHub Releases](https://github.com/menloresearch/jan/releases).
|
||||||
|
|
||||||
## Demo
|
|
||||||
|
|
||||||
<video width="100%" controls>
|
|
||||||
<source src="./docs/public/assets/videos/enable-tool-call-for-models.mp4" type="video/mp4">
|
|
||||||
Your browser does not support the video tag.
|
|
||||||
</video>
|
|
||||||
|
|
||||||
## Features
|
## Features
|
||||||
|
|
||||||
|
|||||||
BIN
docs/src/pages/docs/_assets/hf_hub.png
Normal file
|
After Width: | Height: | Size: 203 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano.png
Normal file
|
After Width: | Height: | Size: 1.6 MiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano_2.png
Normal file
|
After Width: | Height: | Size: 171 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano_3.png
Normal file
|
After Width: | Height: | Size: 139 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano_4.png
Normal file
|
After Width: | Height: | Size: 152 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano_5.png
Normal file
|
After Width: | Height: | Size: 417 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano_6.png
Normal file
|
After Width: | Height: | Size: 405 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano_7.png
Normal file
|
After Width: | Height: | Size: 26 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano_8.png
Normal file
|
After Width: | Height: | Size: 661 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_nano_9.png
Normal file
|
After Width: | Height: | Size: 158 KiB |
BIN
docs/src/pages/docs/_assets/hf_jan_setup.png
Normal file
|
After Width: | Height: | Size: 642 KiB |
BIN
docs/src/pages/docs/_assets/hf_providers.png
Normal file
|
After Width: | Height: | Size: 1.5 MiB |
@ -26,5 +26,9 @@
|
|||||||
"openrouter": {
|
"openrouter": {
|
||||||
"title": "OpenRouter",
|
"title": "OpenRouter",
|
||||||
"href": "/docs/remote-models/openrouter"
|
"href": "/docs/remote-models/openrouter"
|
||||||
|
},
|
||||||
|
"huggingface": {
|
||||||
|
"title": "Hugging Face",
|
||||||
|
"href": "/docs/remote-models/huggingface"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
152
docs/src/pages/docs/remote-models/huggingface.mdx
Normal file
@ -0,0 +1,152 @@
|
|||||||
|
---
|
||||||
|
title: Hugging Face
|
||||||
|
description: Learn how to integrate Hugging Face models with Jan using the Router or Inference Endpoints.
|
||||||
|
keywords:
|
||||||
|
[
|
||||||
|
Hugging Face,
|
||||||
|
Jan,
|
||||||
|
Jan AI,
|
||||||
|
Hugging Face Router,
|
||||||
|
Hugging Face Inference Endpoints,
|
||||||
|
Hugging Face API,
|
||||||
|
Hugging Face Integration,
|
||||||
|
Hugging Face API Integration
|
||||||
|
]
|
||||||
|
---
|
||||||
|
|
||||||
|
import { Callout, Steps } from 'nextra/components'
|
||||||
|
import { Settings, Plus } from 'lucide-react'
|
||||||
|
|
||||||
|
# Hugging Face
|
||||||
|
|
||||||
|
Jan supports Hugging Face models through two methods: the new **HF Router** (recommended) and **Inference Endpoints**. Both methods require a Hugging Face token and **billing to be set up**.
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
## Option 1: HF Router (Recommended)
|
||||||
|
|
||||||
|
The HF Router provides access to models from multiple providers (Replicate, Together AI, SambaNova, Fireworks, Cohere, and more) through a single endpoint.
|
||||||
|
|
||||||
|
<Steps>
|
||||||
|
|
||||||
|
### Step 1: Get Your HF Token
|
||||||
|
|
||||||
|
Visit [Hugging Face Settings > Access Tokens](https://huggingface.co/settings/tokens) and create a token. Make sure you have billing set up on your account.
|
||||||
|
|
||||||
|
### Step 2: Configure Jan
|
||||||
|
|
||||||
|
1. Go to **Settings** > **Model Providers** > **HuggingFace**
|
||||||
|
2. Enter your HF token
|
||||||
|
3. Use this URL: `https://router.huggingface.co/v1`
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
You can find out more about the HF Router [here](https://huggingface.co/docs/inference-providers/index).
|
||||||
|
|
||||||
|
### Step 3: Start Using Models
|
||||||
|
|
||||||
|
Jan comes with three HF Router models pre-configured. Select one and start chatting immediately.
|
||||||
|
|
||||||
|
</Steps>
|
||||||
|
|
||||||
|
<Callout type='info'>
|
||||||
|
The HF Router automatically routes your requests to the best available provider for each model, giving you access to a wide variety of models without managing individual endpoints.
|
||||||
|
</Callout>
|
||||||
|
|
||||||
|
## Option 2: HF Inference Endpoints
|
||||||
|
|
||||||
|
For more control over specific models and deployment configurations, you can use Hugging Face Inference Endpoints.
|
||||||
|
|
||||||
|
<Steps>
|
||||||
|
|
||||||
|
### Step 1: Navigate to the HuggingFace Model Hub
|
||||||
|
|
||||||
|
Visit the [HuggingFace Model Hub](https://huggingface.co/models) (make sure you are logged in) and pick the model you want to use.
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
### Step 2: Configure HF Inference Endpoint and Deploy
|
||||||
|
|
||||||
|
After you have selected the model you want to use, click on the **Deploy** button and select a deployment method. We will select HF Inference Endpoints for this one.
|
||||||
|
|
||||||
|

|
||||||
|
<br/>
|
||||||
|
|
||||||
|
This will take you to the deployment set up page. For this example, we will leave the default settings as they are under the GPU tab and click on **Create Endpoint**.
|
||||||
|
|
||||||
|

|
||||||
|
<br/>
|
||||||
|
|
||||||
|
Once your endpoint is ready, test that it works on the **Test your endpoint** tab.
|
||||||
|
|
||||||
|

|
||||||
|
<br/>
|
||||||
|
|
||||||
|
If you get a response, you can click on **Copy** to copy the endpoint URL and API key.
|
||||||
|
|
||||||
|
<Callout type='info'>
|
||||||
|
You will need to be logged into the HuggingFace Inference Endpoints and have a credit card on file to deploy a model.
|
||||||
|
</Callout>
|
||||||
|
|
||||||
|
### Step 3: Configure Jan
|
||||||
|
|
||||||
|
If you do not have an API key you can create one under **Settings** > **Access Tokens** [here](https://huggingface.co/settings/tokens). Once you finish, copy the token and add it to Jan alongside your endpoint URL at **Settings** > **Model Providers** > **HuggingFace**.
|
||||||
|
|
||||||
|
**3.1 HF Token**
|
||||||
|

|
||||||
|
<br/>
|
||||||
|
|
||||||
|
**3.2 HF Endpoint URL**
|
||||||
|

|
||||||
|
<br/>
|
||||||
|
|
||||||
|
**3.3 Jan Settings**
|
||||||
|

|
||||||
|
|
||||||
|
<Callout type='warning'>
|
||||||
|
Make sure to add `/v1/` to the end of your endpoint URL. This is required by the OpenAI API.
|
||||||
|
</Callout>
|
||||||
|
|
||||||
|
**3.4 Add Model Details**
|
||||||
|

|
||||||
|
|
||||||
|
### Step 4: Start Using the Model
|
||||||
|
|
||||||
|
Now you can start using the model in any chat.
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
If you want to learn how to use Jan Nano with MCP, check out [the guide here](../jan-models/jan-nano-32).
|
||||||
|
<br/>
|
||||||
|
|
||||||
|
</Steps>
|
||||||
|
|
||||||
|
## Available Hugging Face Models
|
||||||
|
|
||||||
|
**Option 1 (HF Router):** Access to models from multiple providers as shown in the providers image above.
|
||||||
|
|
||||||
|
**Option 2 (Inference Endpoints):** You can follow the steps above with a large amount of models on Hugging Face and bring them to Jan. Check out other models in the [Hugging Face Model Hub](https://huggingface.co/models).
|
||||||
|
|
||||||
|
## Troubleshooting
|
||||||
|
|
||||||
|
Common issues and solutions:
|
||||||
|
|
||||||
|
**1. Started a chat but the model is not responding**
|
||||||
|
- Verify your API_KEY/HF_TOKEN is correct and not expired
|
||||||
|
- Ensure you have billing set up on your HF account
|
||||||
|
- For Inference Endpoints: Ensure the model you're trying to use is running again since, after a while, they go idle so that you don't get charged when you are not using it
|
||||||
|
|
||||||
|

|
||||||
|
|
||||||
|
**2. Connection Problems**
|
||||||
|
- Check your internet connection
|
||||||
|
- Verify Hugging Face's system status
|
||||||
|
- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
|
||||||
|
|
||||||
|
**3. Model Unavailable**
|
||||||
|
- Confirm your API key has access to the model
|
||||||
|
- Check if you're using the correct model ID
|
||||||
|
- Verify your Hugging Face account has the necessary permissions
|
||||||
|
|
||||||
|
Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the
|
||||||
|
[Hugging Face's documentation](https://docs.huggingface.co/en/inference-endpoints/index).
|
||||||