From bd4d81842c1121c3d955c909e85e7d4aaba82549 Mon Sep 17 00:00:00 2001 From: Ashley Date: Tue, 7 Jan 2025 13:21:48 +0700 Subject: [PATCH] Updated Quickstart page --- docs/src/pages/docs/quickstart.mdx | 40 +++++++++++++++--------------- 1 file changed, 20 insertions(+), 20 deletions(-) diff --git a/docs/src/pages/docs/quickstart.mdx b/docs/src/pages/docs/quickstart.mdx index dcc7f3bd4..44ee2df92 100644 --- a/docs/src/pages/docs/quickstart.mdx +++ b/docs/src/pages/docs/quickstart.mdx @@ -22,6 +22,7 @@ keywords: import { Tabs } from 'nextra/components' import { Callout, Steps } from 'nextra/components' +import { Settings } from 'lucide-react' # Quickstart @@ -32,7 +33,7 @@ import { Callout, Steps } from 'nextra/components' 2. Install the application on your system ([Mac](/docs/desktop/mac), [Windows](/docs/desktop/windows), [Linux](/docs/desktop/linux)) 3. Launch Jan -Once installed, you'll see the Jan application interface with no local models pre-installed yet. You'll be able to: +Once installed, you'll see Jan application interface with no models pre-installed yet. You'll be able to: - Download and run local AI models - Connect to cloud AI providers if desired
@@ -48,7 +49,7 @@ Jan offers various local AI models, from smaller efficient models to larger more 3. Choose a model that fits your needs & hardware specifications 4. Click **Download** to begin installation -Local models run directly on your computer, which means they use your computer's memory (RAM) and processing power. Please choose models carefully based on your hardware specifications ([Mac](/docs/desktop/mac#minimum-requirements), [Windows](docs/desktop/windows#compatibility), [Linux](docs/desktop/linux#compatibility)). +Local models run directly on your computer, which means they use your computer's memory (RAM) and processing power. Please choose models carefully based on your hardware specifications ([Mac](/docs/desktop/mac#minimum-requirements), [Windows](/docs/desktop/windows#compatibility), [Linux](/docs/desktop/linux#compatibility)). For more model installation methods, please visit [Model Management](/docs/models/manage-models). @@ -58,9 +59,9 @@ For more model installation methods, please visit [Model Management](/docs/model
### Step 3: Turn on GPU Acceleration (Optional) -While the model downloads, let's optimize your hardware setup. If you have a compatible graphics card, you can significantly boost model performance by enabling GPU acceleration. -1. Navigate to **Settings** → **Hardware** -2. Enable your preferred GPU(s) +While the model downloads, let's optimize your hardware setup. If you're on **Windows** or **Linux** and have a compatible graphics card, you can significantly boost model performance by enabling GPU acceleration. +1. Navigate to **Settings** () > **Advanced Settings** +2. At **GPU Acceleration**, toggle on and select your preferred GPU(s) 3. App reload is required after the selection @@ -71,11 +72,11 @@ Ensure you have installed all required dependencies and drivers before enabling ![Turn on GPU acceleration](./_assets/trouble-shooting-01.png) ### Step 4: Customize Assistant Instructions -Once your model has downloaded and you're ready to start your first conversation with Jan, you can customize how it responds by setting specific instructions: -1. In any **Thread**, click the **Assistant** tab in the **right panel** -2. Enter your instructions in the **Instructions** field to define how Jan should respond +Once your model has been downloaded and you're ready to start your first conversation, you can customize how it responds by setting specific instructions: +1. In any **Thread**, click the **Assistant** tab in the **right sidebar** +2. Enter your instructions in **Instructions** field to define how Jan should respond -You can modify these instructions at any time during your conversation to adjust Jan's behavior for that specific thread. +You can modify these instructions at any time during your conversation to adjust Jan's behavior for that specific thread. See detailed guide at [Assistant](/docs/assistants).
![Assistant Instruction](./_assets/quick-start-02.png) @@ -85,9 +86,9 @@ You can modify these instructions at any time during your conversation to adjust Now that your model is downloaded and instructions are set, you can begin chatting with Jan. Type your message in the **input field** at the bottom of the thread to start the conversation. You can further customize your experience by: -- Adjusting [model parameters](/docs/models/model-parameters) in the **Model** tab in the **right panel** -- Trying different models for different tasks by clicking the **model selector** in **Model** tab or **input field** -- Creating new threads with different instructions and model configurations +- Adjust [model parameters](/docs/models/model-parameters) in the **Model** tab in the **right sidebar** +- Try different models for different tasks by clicking the **model selector** in **Model** tab or **input field** +- [Create new threads](/docs/threads#creating-new-thread) with different instructions and model configurations @@ -99,20 +100,19 @@ You can further customize your experience by: ### Step 6: Connect to cloud models (Optional) -Jan supports both local and remote AI models. You can connect to remote AI services that are OpenAI API-compatible, including: OpenAI (GPT-4, o1,...), Anthropic (Claude), Groq, Mistral, and more. +Jan supports both local and cloud AI models. You can connect to cloud AI services that are OpenAI API-compatible, including: OpenAI (GPT-4, o1,...), Anthropic (Claude), Groq, Mistral, and more. 1. Open any **Thread** -2. Click the **Model** tab in the **right panel** or the **model selector** in input field -3. Choose the **Cloud** tab -4. Choose your preferred provider (Anthropic, OpenAI, etc.) -5. Click the **Add (➕)** icon next to the provider -6. Obtain a valid API key from your chosen provider, ensure the key has sufficient credits & appropriate permissions -7. Copy & insert your **API Key** in Jan +2. Click **Model** tab in the **right sidebar** or **model selector** in input field +3. Once the selector is poped up, choose the **Cloud** tab +4. Select your preferred provider (Anthropic, OpenAI, etc.), click **Add (➕)** icon next to the provider +5. Obtain a valid API key from your chosen provider, ensure the key has sufficient credits & appropriate permissions +6. Copy & insert your **API Key** in Jan See [Remote APIs](/docs/remote-models/openai) for detailed configuration.
-![Connect Remote API](./_assets/quick-start-03.png) +![Connect Remote APIs](./_assets/quick-start-03.png)