diff --git a/README.md b/README.md index 5b3655ba8..41a0e001a 100644 --- a/README.md +++ b/README.md @@ -12,16 +12,14 @@

- Getting Started - - Docs - - Changelog - - Bug reports + Getting Started + - Docs + - Changelog + - Bug reports - Discord

-Jan is a ChatGPT-alternative that runs 100% offline on your device. Our goal is to make it easy for a layperson to download and run LLMs and use AI with **full control** and **privacy**. - -**⚠️ Jan is in active development.** +Jan is an AI assistant that runs 100% offline on your device. Download and run LLMs with **full control** and **privacy**. ## Installation @@ -31,43 +29,32 @@ Because clicking a button is still the easiest way to get started: Platform Stable - Beta Nightly Windows jan.exe - jan.exe jan.exe macOS jan.dmg - jan.dmg jan.dmg Linux (deb) jan.deb - jan.deb jan.deb Linux (AppImage) jan.AppImage - jan.AppImage jan.AppImage Download from [jan.ai](https://jan.ai/) or [GitHub Releases](https://github.com/menloresearch/jan/releases). -## Demo - - ## Features diff --git a/docs/src/pages/docs/_assets/hf_hub.png b/docs/src/pages/docs/_assets/hf_hub.png new file mode 100644 index 000000000..ad059c49a Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_hub.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano.png b/docs/src/pages/docs/_assets/hf_jan_nano.png new file mode 100644 index 000000000..147a5c70e Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano_2.png b/docs/src/pages/docs/_assets/hf_jan_nano_2.png new file mode 100644 index 000000000..10c410240 Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano_2.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano_3.png b/docs/src/pages/docs/_assets/hf_jan_nano_3.png new file mode 100644 index 000000000..dac240d29 Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano_3.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano_4.png b/docs/src/pages/docs/_assets/hf_jan_nano_4.png new file mode 100644 index 000000000..552f07b06 Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano_4.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano_5.png b/docs/src/pages/docs/_assets/hf_jan_nano_5.png new file mode 100644 index 000000000..b322f0f93 Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano_5.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano_6.png b/docs/src/pages/docs/_assets/hf_jan_nano_6.png new file mode 100644 index 000000000..c8be2b707 Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano_6.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano_7.png b/docs/src/pages/docs/_assets/hf_jan_nano_7.png new file mode 100644 index 000000000..2a8ba8438 Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano_7.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano_8.png b/docs/src/pages/docs/_assets/hf_jan_nano_8.png new file mode 100644 index 000000000..4e1885a8e Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano_8.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_nano_9.png b/docs/src/pages/docs/_assets/hf_jan_nano_9.png new file mode 100644 index 000000000..09575c541 Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_nano_9.png differ diff --git a/docs/src/pages/docs/_assets/hf_jan_setup.png b/docs/src/pages/docs/_assets/hf_jan_setup.png new file mode 100644 index 000000000..2d917539b Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_jan_setup.png differ diff --git a/docs/src/pages/docs/_assets/hf_providers.png b/docs/src/pages/docs/_assets/hf_providers.png new file mode 100644 index 000000000..1f8e4daf7 Binary files /dev/null and b/docs/src/pages/docs/_assets/hf_providers.png differ diff --git a/docs/src/pages/docs/remote-models/_meta.json b/docs/src/pages/docs/remote-models/_meta.json index 39660be88..9ef524352 100644 --- a/docs/src/pages/docs/remote-models/_meta.json +++ b/docs/src/pages/docs/remote-models/_meta.json @@ -26,5 +26,9 @@ "openrouter": { "title": "OpenRouter", "href": "/docs/remote-models/openrouter" + }, + "huggingface": { + "title": "Hugging Face", + "href": "/docs/remote-models/huggingface" } } diff --git a/docs/src/pages/docs/remote-models/huggingface.mdx b/docs/src/pages/docs/remote-models/huggingface.mdx new file mode 100644 index 000000000..1808fd2a1 --- /dev/null +++ b/docs/src/pages/docs/remote-models/huggingface.mdx @@ -0,0 +1,152 @@ +--- +title: Hugging Face +description: Learn how to integrate Hugging Face models with Jan using the Router or Inference Endpoints. +keywords: + [ + Hugging Face, + Jan, + Jan AI, + Hugging Face Router, + Hugging Face Inference Endpoints, + Hugging Face API, + Hugging Face Integration, + Hugging Face API Integration + ] +--- + +import { Callout, Steps } from 'nextra/components' +import { Settings, Plus } from 'lucide-react' + +# Hugging Face + +Jan supports Hugging Face models through two methods: the new **HF Router** (recommended) and **Inference Endpoints**. Both methods require a Hugging Face token and **billing to be set up**. + +![HuggingFace Inference Providers](../_assets/hf_providers.png) + +## Option 1: HF Router (Recommended) + +The HF Router provides access to models from multiple providers (Replicate, Together AI, SambaNova, Fireworks, Cohere, and more) through a single endpoint. + + + +### Step 1: Get Your HF Token + +Visit [Hugging Face Settings > Access Tokens](https://huggingface.co/settings/tokens) and create a token. Make sure you have billing set up on your account. + +### Step 2: Configure Jan + +1. Go to **Settings** > **Model Providers** > **HuggingFace** +2. Enter your HF token +3. Use this URL: `https://router.huggingface.co/v1` + +![Jan HF Setup](../_assets/hf_jan_setup.png) + +You can find out more about the HF Router [here](https://huggingface.co/docs/inference-providers/index). + +### Step 3: Start Using Models + +Jan comes with three HF Router models pre-configured. Select one and start chatting immediately. + + + + +The HF Router automatically routes your requests to the best available provider for each model, giving you access to a wide variety of models without managing individual endpoints. + + +## Option 2: HF Inference Endpoints + +For more control over specific models and deployment configurations, you can use Hugging Face Inference Endpoints. + + + +### Step 1: Navigate to the HuggingFace Model Hub + +Visit the [HuggingFace Model Hub](https://huggingface.co/models) (make sure you are logged in) and pick the model you want to use. + +![HuggingFace Model Hub](../_assets/hf_hub.png) + +### Step 2: Configure HF Inference Endpoint and Deploy + +After you have selected the model you want to use, click on the **Deploy** button and select a deployment method. We will select HF Inference Endpoints for this one. + +![HuggingFace Deployment](../_assets/hf_jan_nano.png) +
+ +This will take you to the deployment set up page. For this example, we will leave the default settings as they are under the GPU tab and click on **Create Endpoint**. + +![HuggingFace Deployment](../_assets/hf_jan_nano_2.png) +
+ +Once your endpoint is ready, test that it works on the **Test your endpoint** tab. + +![HuggingFace Deployment](../_assets/hf_jan_nano_3.png) +
+ +If you get a response, you can click on **Copy** to copy the endpoint URL and API key. + + + You will need to be logged into the HuggingFace Inference Endpoints and have a credit card on file to deploy a model. + + +### Step 3: Configure Jan + +If you do not have an API key you can create one under **Settings** > **Access Tokens** [here](https://huggingface.co/settings/tokens). Once you finish, copy the token and add it to Jan alongside your endpoint URL at **Settings** > **Model Providers** > **HuggingFace**. + +**3.1 HF Token** +![Get Token](../_assets/hf_jan_nano_5.png) +
+ +**3.2 HF Endpoint URL** +![Endpoint URL](../_assets/hf_jan_nano_4.png) +
+ +**3.3 Jan Settings** +![Jan Settings](../_assets/hf_jan_nano_6.png) + + +Make sure to add `/v1/` to the end of your endpoint URL. This is required by the OpenAI API. + + +**3.4 Add Model Details** +![Add Model Details](../_assets/hf_jan_nano_7.png) + +### Step 4: Start Using the Model + +Now you can start using the model in any chat. + +![Start Using the Model](../_assets/hf_jan_nano_8.png) + +If you want to learn how to use Jan Nano with MCP, check out [the guide here](../jan-models/jan-nano-32). +
+ +
+ +## Available Hugging Face Models + +**Option 1 (HF Router):** Access to models from multiple providers as shown in the providers image above. + +**Option 2 (Inference Endpoints):** You can follow the steps above with a large amount of models on Hugging Face and bring them to Jan. Check out other models in the [Hugging Face Model Hub](https://huggingface.co/models). + +## Troubleshooting + +Common issues and solutions: + +**1. Started a chat but the model is not responding** +- Verify your API_KEY/HF_TOKEN is correct and not expired +- Ensure you have billing set up on your HF account +- For Inference Endpoints: Ensure the model you're trying to use is running again since, after a while, they go idle so that you don't get charged when you are not using it + +![Model Running](../_assets/hf_jan_nano_9.png) + +**2. Connection Problems** +- Check your internet connection +- Verify Hugging Face's system status +- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs) + +**3. Model Unavailable** +- Confirm your API key has access to the model +- Check if you're using the correct model ID +- Verify your Hugging Face account has the necessary permissions + +Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the +[Hugging Face's documentation](https://docs.huggingface.co/en/inference-endpoints/index).