diff --git a/docs/src/pages/docs/_assets/retrieval-01.png b/docs/src/pages/docs/_assets/retrieval-01.png new file mode 100644 index 000000000..12c8d832b Binary files /dev/null and b/docs/src/pages/docs/_assets/retrieval-01.png differ diff --git a/docs/src/pages/docs/_assets/retrieval-02.png b/docs/src/pages/docs/_assets/retrieval-02.png new file mode 100644 index 000000000..6c57e55a2 Binary files /dev/null and b/docs/src/pages/docs/_assets/retrieval-02.png differ diff --git a/docs/src/pages/docs/built-in/llama-cpp.mdx b/docs/src/pages/docs/built-in/llama-cpp.mdx index 6783eda08..8d71ff2ae 100644 --- a/docs/src/pages/docs/built-in/llama-cpp.mdx +++ b/docs/src/pages/docs/built-in/llama-cpp.mdx @@ -32,8 +32,6 @@ import { Callout, Steps } from 'nextra/components' Jan has [**Cortex**](https://github.com/janhq/cortex) - a default C++ inference server built on top of [llama.cpp](https://github.com/ggerganov/llama.cpp). This server provides an OpenAI-compatible API, queues, scaling, and additional features on top of the wide capabilities of `llama.cpp`. -## llama.cpp Engine - This guide shows you how to initialize the `llama.cpp` to download and install the required dependencies to start chatting with a model using the `llama.cpp` engine. ## Prerequisites diff --git a/docs/src/pages/docs/tools/retrieval.mdx b/docs/src/pages/docs/tools/retrieval.mdx index c0276d8ec..2f305d9e2 100644 --- a/docs/src/pages/docs/tools/retrieval.mdx +++ b/docs/src/pages/docs/tools/retrieval.mdx @@ -22,25 +22,35 @@ keywords: import { Callout, Steps } from 'nextra/components' # Knowledge Retrieval -This article lists the capabilities of the Jan platform and guides you through using RAG to chat with PDF documents. +Chat with your documents and images using Jan's RAG (Retrieval-Augmented Generation) capability. + -To access this feature, please enable Experimental mode in the [Advanced Settings](/guides/advanced/#enable-the-experimental-mode). + This feature is currently experimental and must be enabled through [Experimental Mode](/docs/settings#experimental-mode) in **Advanced Settings**. -## Enable the Knowledge Retrieval +## Enable File Search & Vision To chat with PDFs using RAG in Jan, follow these steps: -1. Create a **new thread**. -2. Click the **Tools** tab. +1. In any **Thread**, click the **Tools** tab in right sidebar +2. Enable **Retrieval** +
-![Retrieval](../_assets/tools.png) +![Retrieval](../_assets/retrieval-01.png)
-3. Enable the **Retrieval**. + +3. Once enabled, you should be able to **upload file & images** from thread input field + +Ensure that you are using a multimodal model. +- File Search: Jan currently supports PDF format +- Vision: only works with local models or [OpenAI](/docs/remote-models/openai) models for now + +
-![Retrieval](../_assets/retrieval1.png) +![Retrieval](../_assets/retrieval-02.png)
-4. Adjust the **Retrieval** settings as needed. These settings include the following: + +## Knowledge Retrieval Parameters | Feature | Description | |-----------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------| @@ -51,11 +61,4 @@ To chat with PDFs using RAG in Jan, follow these steps: | **Chunk Size** | - Sets the maximum number of tokens per data chunk, which is crucial for managing processing load and maintaining performance.

- Increase the chunk size for processing large blocks of text efficiently, or decrease it when dealing with smaller, more manageable texts to optimize memory usage. | | **Chunk Overlap** | - Specifies the overlap in tokens between adjacent chunks to ensure continuous context in split text segments.

- Adjust the overlap to ensure smooth transitions in text analysis, with higher overlap for complex texts where context is critical. | | **Retrieval Template**| - Defines the query structure using variables like `{CONTEXT}` and `{QUESTION}` to tailor searches to specific needs.

- Customize templates to closely align with your data's structure and the queries' nature, ensuring that retrievals are as relevant as possible. | -5. Select the model you want to use. - -To upload an image or GIF, ensure that you are using a multimodal model. If not, you are limited to uploading documents only. - -6. Click on the 📎 icon in the chat input field. -7. Select **Document** to upload a document file. -
-![Retrieval](../_assets/retrieval2.png) +