diff --git a/docs/public/sitemap-0.xml b/docs/public/sitemap-0.xml
index b74cbb237..2b4a04975 100644
--- a/docs/public/sitemap-0.xml
+++ b/docs/public/sitemap-0.xml
@@ -93,7 +93,7 @@
https://jan.ai/docs/desktop/windows2024-09-09T08:19:45.722Zdaily1
https://jan.ai/docs/error-codes2024-09-09T08:19:45.722Zdaily1
https://jan.ai/docs/extensions2024-09-09T08:19:45.722Zdaily1
-https://jan.ai/docs/installing-extension2024-09-09T08:19:45.722Zdaily1
+https://jan.ai/docs/install-extensions2024-09-09T08:19:45.722Zdaily1
https://jan.ai/docs/models2024-09-09T08:19:45.722Zdaily1
https://jan.ai/docs/models/manage-models2024-09-09T08:19:45.722Zdaily1
https://jan.ai/docs/models/model-parameters2024-09-09T08:19:45.722Zdaily1
diff --git a/docs/src/pages/docs/_assets/Anthropic-1.gif b/docs/src/pages/docs/_assets/Anthropic-1.gif
deleted file mode 100644
index 594c53172..000000000
Binary files a/docs/src/pages/docs/_assets/Anthropic-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Anthropic-2.gif b/docs/src/pages/docs/_assets/Anthropic-2.gif
deleted file mode 100644
index 63e76cc3b..000000000
Binary files a/docs/src/pages/docs/_assets/Anthropic-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Cohere-1.gif b/docs/src/pages/docs/_assets/Cohere-1.gif
deleted file mode 100644
index 6441bd8a9..000000000
Binary files a/docs/src/pages/docs/_assets/Cohere-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Cohere-2.gif b/docs/src/pages/docs/_assets/Cohere-2.gif
deleted file mode 100644
index 9488040c0..000000000
Binary files a/docs/src/pages/docs/_assets/Cohere-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Groq-1.gif b/docs/src/pages/docs/_assets/Groq-1.gif
deleted file mode 100644
index d95e577b1..000000000
Binary files a/docs/src/pages/docs/_assets/Groq-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Groq-2.gif b/docs/src/pages/docs/_assets/Groq-2.gif
deleted file mode 100644
index 46231696f..000000000
Binary files a/docs/src/pages/docs/_assets/Groq-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/LM-Studio-v1.gif b/docs/src/pages/docs/_assets/LM-Studio-v1.gif
deleted file mode 100644
index 65dfb5db1..000000000
Binary files a/docs/src/pages/docs/_assets/LM-Studio-v1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/LM-Studio-v2.gif b/docs/src/pages/docs/_assets/LM-Studio-v2.gif
deleted file mode 100644
index ad012df5d..000000000
Binary files a/docs/src/pages/docs/_assets/LM-Studio-v2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/LM-Studio-v3.gif b/docs/src/pages/docs/_assets/LM-Studio-v3.gif
deleted file mode 100644
index d5208c2bc..000000000
Binary files a/docs/src/pages/docs/_assets/LM-Studio-v3.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Martian-1.gif b/docs/src/pages/docs/_assets/Martian-1.gif
deleted file mode 100644
index 3da994cc4..000000000
Binary files a/docs/src/pages/docs/_assets/Martian-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Martian-2.gif b/docs/src/pages/docs/_assets/Martian-2.gif
deleted file mode 100644
index 27fa761b7..000000000
Binary files a/docs/src/pages/docs/_assets/Martian-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Mistral-1.gif b/docs/src/pages/docs/_assets/Mistral-1.gif
deleted file mode 100644
index 7d1e82108..000000000
Binary files a/docs/src/pages/docs/_assets/Mistral-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Mistral-2.gif b/docs/src/pages/docs/_assets/Mistral-2.gif
deleted file mode 100644
index 25c8d2fca..000000000
Binary files a/docs/src/pages/docs/_assets/Mistral-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Ollama-1.gif b/docs/src/pages/docs/_assets/Ollama-1.gif
deleted file mode 100644
index d9c7fa135..000000000
Binary files a/docs/src/pages/docs/_assets/Ollama-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Ollama-2.gif b/docs/src/pages/docs/_assets/Ollama-2.gif
deleted file mode 100644
index eadc6afdd..000000000
Binary files a/docs/src/pages/docs/_assets/Ollama-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/Ollama-3.gif b/docs/src/pages/docs/_assets/Ollama-3.gif
deleted file mode 100644
index 6b09000b9..000000000
Binary files a/docs/src/pages/docs/_assets/Ollama-3.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/OpenAi-1.gif b/docs/src/pages/docs/_assets/OpenAi-1.gif
deleted file mode 100644
index c98e2f2be..000000000
Binary files a/docs/src/pages/docs/_assets/OpenAi-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/OpenAi-2.gif b/docs/src/pages/docs/_assets/OpenAi-2.gif
deleted file mode 100644
index 8cd449f7f..000000000
Binary files a/docs/src/pages/docs/_assets/OpenAi-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/OpenRouter-1.gif b/docs/src/pages/docs/_assets/OpenRouter-1.gif
deleted file mode 100644
index cd6a7898f..000000000
Binary files a/docs/src/pages/docs/_assets/OpenRouter-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/OpenRouter-2.gif b/docs/src/pages/docs/_assets/OpenRouter-2.gif
deleted file mode 100644
index 307b07deb..000000000
Binary files a/docs/src/pages/docs/_assets/OpenRouter-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/amd.gif b/docs/src/pages/docs/_assets/amd.gif
deleted file mode 100644
index cc053b80c..000000000
Binary files a/docs/src/pages/docs/_assets/amd.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/anthropic.png b/docs/src/pages/docs/_assets/anthropic.png
new file mode 100644
index 000000000..108576903
Binary files /dev/null and b/docs/src/pages/docs/_assets/anthropic.png differ
diff --git a/docs/src/pages/docs/_assets/assistant-01.png b/docs/src/pages/docs/_assets/assistant-01.png
index 5b3546032..eabf2bba2 100644
Binary files a/docs/src/pages/docs/_assets/assistant-01.png and b/docs/src/pages/docs/_assets/assistant-01.png differ
diff --git a/docs/src/pages/docs/_assets/assistant-slider.png b/docs/src/pages/docs/_assets/assistant-slider.png
deleted file mode 100644
index 8289e0d51..000000000
Binary files a/docs/src/pages/docs/_assets/assistant-slider.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/assistant1.gif b/docs/src/pages/docs/_assets/assistant1.gif
deleted file mode 100644
index 75a56ba27..000000000
Binary files a/docs/src/pages/docs/_assets/assistant1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/assistant2.gif b/docs/src/pages/docs/_assets/assistant2.gif
deleted file mode 100644
index 44bfb2264..000000000
Binary files a/docs/src/pages/docs/_assets/assistant2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/chat.gif b/docs/src/pages/docs/_assets/chat.gif
deleted file mode 100644
index 94b8c2719..000000000
Binary files a/docs/src/pages/docs/_assets/chat.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/clear.gif b/docs/src/pages/docs/_assets/clear.gif
deleted file mode 100644
index 445708d22..000000000
Binary files a/docs/src/pages/docs/_assets/clear.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/cohere.png b/docs/src/pages/docs/_assets/cohere.png
new file mode 100644
index 000000000..87d3d0fa4
Binary files /dev/null and b/docs/src/pages/docs/_assets/cohere.png differ
diff --git a/docs/src/pages/docs/_assets/configure.gif b/docs/src/pages/docs/_assets/configure.gif
deleted file mode 100644
index 96daaaa61..000000000
Binary files a/docs/src/pages/docs/_assets/configure.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/data-folder.png b/docs/src/pages/docs/_assets/data-folder.png
deleted file mode 100644
index def0f38c7..000000000
Binary files a/docs/src/pages/docs/_assets/data-folder.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/default.gif b/docs/src/pages/docs/_assets/default.gif
deleted file mode 100644
index 3bc6a68f6..000000000
Binary files a/docs/src/pages/docs/_assets/default.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/delete-tensor.gif b/docs/src/pages/docs/_assets/delete-tensor.gif
deleted file mode 100644
index 86f45fc65..000000000
Binary files a/docs/src/pages/docs/_assets/delete-tensor.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/delete-threads.png b/docs/src/pages/docs/_assets/delete-threads.png
deleted file mode 100644
index b134e8e12..000000000
Binary files a/docs/src/pages/docs/_assets/delete-threads.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/delete.png b/docs/src/pages/docs/_assets/delete.png
deleted file mode 100644
index 6b3f669c2..000000000
Binary files a/docs/src/pages/docs/_assets/delete.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/disable-tensor.gif b/docs/src/pages/docs/_assets/disable-tensor.gif
deleted file mode 100644
index 91e9fe3d6..000000000
Binary files a/docs/src/pages/docs/_assets/disable-tensor.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download-button.png b/docs/src/pages/docs/_assets/download-button.png
deleted file mode 100644
index 2e8088d15..000000000
Binary files a/docs/src/pages/docs/_assets/download-button.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download-button2.png b/docs/src/pages/docs/_assets/download-button2.png
deleted file mode 100644
index 8f8c683c3..000000000
Binary files a/docs/src/pages/docs/_assets/download-button2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download-button3.png b/docs/src/pages/docs/_assets/download-button3.png
deleted file mode 100644
index cf05ec45a..000000000
Binary files a/docs/src/pages/docs/_assets/download-button3.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download-icon.png b/docs/src/pages/docs/_assets/download-icon.png
deleted file mode 100644
index f553e19f2..000000000
Binary files a/docs/src/pages/docs/_assets/download-icon.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download-llama2.gif b/docs/src/pages/docs/_assets/download-llama2.gif
deleted file mode 100644
index 18d671209..000000000
Binary files a/docs/src/pages/docs/_assets/download-llama2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download-model.gif b/docs/src/pages/docs/_assets/download-model.gif
deleted file mode 100644
index e10b293c2..000000000
Binary files a/docs/src/pages/docs/_assets/download-model.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download-model2.gif b/docs/src/pages/docs/_assets/download-model2.gif
deleted file mode 100644
index c69819a1e..000000000
Binary files a/docs/src/pages/docs/_assets/download-model2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download-phi.gif b/docs/src/pages/docs/_assets/download-phi.gif
deleted file mode 100644
index 5b2cd6034..000000000
Binary files a/docs/src/pages/docs/_assets/download-phi.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download.gif b/docs/src/pages/docs/_assets/download.gif
deleted file mode 100644
index 10cb4b466..000000000
Binary files a/docs/src/pages/docs/_assets/download.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/download.png b/docs/src/pages/docs/_assets/download.png
deleted file mode 100644
index 571e44dd3..000000000
Binary files a/docs/src/pages/docs/_assets/download.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/exp-mode.png b/docs/src/pages/docs/_assets/exp-mode.png
deleted file mode 100644
index 14ecf6bfe..000000000
Binary files a/docs/src/pages/docs/_assets/exp-mode.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/experimental.gif b/docs/src/pages/docs/_assets/experimental.gif
deleted file mode 100644
index 11c2d8622..000000000
Binary files a/docs/src/pages/docs/_assets/experimental.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/extensions-01.png b/docs/src/pages/docs/_assets/extensions-01.png
new file mode 100644
index 000000000..912943f83
Binary files /dev/null and b/docs/src/pages/docs/_assets/extensions-01.png differ
diff --git a/docs/src/pages/docs/_assets/extensions-02.png b/docs/src/pages/docs/_assets/extensions-02.png
new file mode 100644
index 000000000..e2e582b37
Binary files /dev/null and b/docs/src/pages/docs/_assets/extensions-02.png differ
diff --git a/docs/src/pages/docs/_assets/extensions-03.png b/docs/src/pages/docs/_assets/extensions-03.png
new file mode 100644
index 000000000..bbdd2c371
Binary files /dev/null and b/docs/src/pages/docs/_assets/extensions-03.png differ
diff --git a/docs/src/pages/docs/_assets/extensions-page.png b/docs/src/pages/docs/_assets/extensions-page.png
deleted file mode 100644
index 74a336769..000000000
Binary files a/docs/src/pages/docs/_assets/extensions-page.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/extensions-page2.png b/docs/src/pages/docs/_assets/extensions-page2.png
deleted file mode 100644
index 51863aa99..000000000
Binary files a/docs/src/pages/docs/_assets/extensions-page2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/gpt.gif b/docs/src/pages/docs/_assets/gpt.gif
deleted file mode 100644
index 3972e812f..000000000
Binary files a/docs/src/pages/docs/_assets/gpt.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/gpu-accel.png b/docs/src/pages/docs/_assets/gpu-accel.png
deleted file mode 100644
index 294013fcf..000000000
Binary files a/docs/src/pages/docs/_assets/gpu-accel.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/gpu.gif b/docs/src/pages/docs/_assets/gpu.gif
deleted file mode 100644
index 58d1da970..000000000
Binary files a/docs/src/pages/docs/_assets/gpu.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/gpu2.gif b/docs/src/pages/docs/_assets/gpu2.gif
deleted file mode 100644
index fcacc5e99..000000000
Binary files a/docs/src/pages/docs/_assets/gpu2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/groq.png b/docs/src/pages/docs/_assets/groq.png
new file mode 100644
index 000000000..d9927fe7a
Binary files /dev/null and b/docs/src/pages/docs/_assets/groq.png differ
diff --git a/docs/src/pages/docs/_assets/history.png b/docs/src/pages/docs/_assets/history.png
deleted file mode 100644
index 7112a0d24..000000000
Binary files a/docs/src/pages/docs/_assets/history.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/http.png b/docs/src/pages/docs/_assets/http.png
deleted file mode 100644
index afa0aeae0..000000000
Binary files a/docs/src/pages/docs/_assets/http.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/https.gif b/docs/src/pages/docs/_assets/https.gif
deleted file mode 100644
index e9ed05b85..000000000
Binary files a/docs/src/pages/docs/_assets/https.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/hub.gif b/docs/src/pages/docs/_assets/hub.gif
deleted file mode 100644
index fad29d5db..000000000
Binary files a/docs/src/pages/docs/_assets/hub.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/hub.png b/docs/src/pages/docs/_assets/hub.png
deleted file mode 100644
index 75b0f575d..000000000
Binary files a/docs/src/pages/docs/_assets/hub.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/import-folder.gif b/docs/src/pages/docs/_assets/import-folder.gif
deleted file mode 100644
index fc7822207..000000000
Binary files a/docs/src/pages/docs/_assets/import-folder.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/import-hf.gif b/docs/src/pages/docs/_assets/import-hf.gif
deleted file mode 100644
index 4ca1a8b2a..000000000
Binary files a/docs/src/pages/docs/_assets/import-hf.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/import.png b/docs/src/pages/docs/_assets/import.png
deleted file mode 100644
index 5ca7fff42..000000000
Binary files a/docs/src/pages/docs/_assets/import.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/import2.png b/docs/src/pages/docs/_assets/import2.png
deleted file mode 100644
index caf2c022d..000000000
Binary files a/docs/src/pages/docs/_assets/import2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/inf.gif b/docs/src/pages/docs/_assets/inf.gif
deleted file mode 100644
index aaef44380..000000000
Binary files a/docs/src/pages/docs/_assets/inf.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/install-engines-01.png b/docs/src/pages/docs/_assets/install-engines-01.png
new file mode 100644
index 000000000..85db02b98
Binary files /dev/null and b/docs/src/pages/docs/_assets/install-engines-01.png differ
diff --git a/docs/src/pages/docs/_assets/install-engines-02.png b/docs/src/pages/docs/_assets/install-engines-02.png
new file mode 100644
index 000000000..dcdd7e4df
Binary files /dev/null and b/docs/src/pages/docs/_assets/install-engines-02.png differ
diff --git a/docs/src/pages/docs/_assets/install-engines-03.png b/docs/src/pages/docs/_assets/install-engines-03.png
new file mode 100644
index 000000000..cb2ecde56
Binary files /dev/null and b/docs/src/pages/docs/_assets/install-engines-03.png differ
diff --git a/docs/src/pages/docs/_assets/install-ext.gif b/docs/src/pages/docs/_assets/install-ext.gif
deleted file mode 100644
index 4357ed236..000000000
Binary files a/docs/src/pages/docs/_assets/install-ext.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/install-ext.png b/docs/src/pages/docs/_assets/install-ext.png
deleted file mode 100644
index ee2ff8453..000000000
Binary files a/docs/src/pages/docs/_assets/install-ext.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/install-tensor.gif b/docs/src/pages/docs/_assets/install-tensor.gif
deleted file mode 100644
index 927ee0b9b..000000000
Binary files a/docs/src/pages/docs/_assets/install-tensor.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/install-tensor.png b/docs/src/pages/docs/_assets/install-tensor.png
deleted file mode 100644
index 050075d84..000000000
Binary files a/docs/src/pages/docs/_assets/install-tensor.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/linux.png b/docs/src/pages/docs/_assets/linux.png
deleted file mode 100644
index 72230191c..000000000
Binary files a/docs/src/pages/docs/_assets/linux.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/llama.cpp-01.png b/docs/src/pages/docs/_assets/llama.cpp-01.png
new file mode 100644
index 000000000..b794fe1df
Binary files /dev/null and b/docs/src/pages/docs/_assets/llama.cpp-01.png differ
diff --git a/docs/src/pages/docs/_assets/llama2.gif b/docs/src/pages/docs/_assets/llama2.gif
deleted file mode 100644
index 0c9502aa6..000000000
Binary files a/docs/src/pages/docs/_assets/llama2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/local-api1.png b/docs/src/pages/docs/_assets/local-api1.png
deleted file mode 100644
index d636d37ad..000000000
Binary files a/docs/src/pages/docs/_assets/local-api1.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/local-api2.png b/docs/src/pages/docs/_assets/local-api2.png
deleted file mode 100644
index 58652775c..000000000
Binary files a/docs/src/pages/docs/_assets/local-api2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/local-api3.png b/docs/src/pages/docs/_assets/local-api3.png
deleted file mode 100644
index aa8372640..000000000
Binary files a/docs/src/pages/docs/_assets/local-api3.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/local-api4.png b/docs/src/pages/docs/_assets/local-api4.png
deleted file mode 100644
index 2a66c9c1f..000000000
Binary files a/docs/src/pages/docs/_assets/local-api4.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/local-api5.png b/docs/src/pages/docs/_assets/local-api5.png
deleted file mode 100644
index 2dc80c6c7..000000000
Binary files a/docs/src/pages/docs/_assets/local-api5.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/mac2.png b/docs/src/pages/docs/_assets/mac2.png
deleted file mode 100644
index aeb4b6bbc..000000000
Binary files a/docs/src/pages/docs/_assets/mac2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/martian.png b/docs/src/pages/docs/_assets/martian.png
new file mode 100644
index 000000000..e00349b01
Binary files /dev/null and b/docs/src/pages/docs/_assets/martian.png differ
diff --git a/docs/src/pages/docs/_assets/mistralai.png b/docs/src/pages/docs/_assets/mistralai.png
new file mode 100644
index 000000000..70139c6ae
Binary files /dev/null and b/docs/src/pages/docs/_assets/mistralai.png differ
diff --git a/docs/src/pages/docs/_assets/model-anthropic.gif b/docs/src/pages/docs/_assets/model-anthropic.gif
deleted file mode 100644
index 8de07e625..000000000
Binary files a/docs/src/pages/docs/_assets/model-anthropic.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/model-cohere.gif b/docs/src/pages/docs/_assets/model-cohere.gif
deleted file mode 100644
index 2bf5007ff..000000000
Binary files a/docs/src/pages/docs/_assets/model-cohere.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/model-groq.gif b/docs/src/pages/docs/_assets/model-groq.gif
deleted file mode 100644
index 7c6d20280..000000000
Binary files a/docs/src/pages/docs/_assets/model-groq.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/model-management-01.png b/docs/src/pages/docs/_assets/model-management-01.png
index ea6c363e1..9020c66b2 100644
Binary files a/docs/src/pages/docs/_assets/model-management-01.png and b/docs/src/pages/docs/_assets/model-management-01.png differ
diff --git a/docs/src/pages/docs/_assets/model-management-02.png b/docs/src/pages/docs/_assets/model-management-02.png
index a0fe08646..bb446d056 100644
Binary files a/docs/src/pages/docs/_assets/model-management-02.png and b/docs/src/pages/docs/_assets/model-management-02.png differ
diff --git a/docs/src/pages/docs/_assets/model-management-04.png b/docs/src/pages/docs/_assets/model-management-04.png
index 6b3f82464..71d72dff2 100644
Binary files a/docs/src/pages/docs/_assets/model-management-04.png and b/docs/src/pages/docs/_assets/model-management-04.png differ
diff --git a/docs/src/pages/docs/_assets/model-management-05.png b/docs/src/pages/docs/_assets/model-management-05.png
index e7d4ea38d..891fa943a 100644
Binary files a/docs/src/pages/docs/_assets/model-management-05.png and b/docs/src/pages/docs/_assets/model-management-05.png differ
diff --git a/docs/src/pages/docs/_assets/model-management-06.png b/docs/src/pages/docs/_assets/model-management-06.png
index 5d5449c8b..ebd2de3d1 100644
Binary files a/docs/src/pages/docs/_assets/model-management-06.png and b/docs/src/pages/docs/_assets/model-management-06.png differ
diff --git a/docs/src/pages/docs/_assets/model-management-07.png b/docs/src/pages/docs/_assets/model-management-07.png
index 83a562ddb..612904ac7 100644
Binary files a/docs/src/pages/docs/_assets/model-management-07.png and b/docs/src/pages/docs/_assets/model-management-07.png differ
diff --git a/docs/src/pages/docs/_assets/model-martian.gif b/docs/src/pages/docs/_assets/model-martian.gif
deleted file mode 100644
index 8bf4dbbf5..000000000
Binary files a/docs/src/pages/docs/_assets/model-martian.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/model-mistral.gif b/docs/src/pages/docs/_assets/model-mistral.gif
deleted file mode 100644
index a26661903..000000000
Binary files a/docs/src/pages/docs/_assets/model-mistral.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/model-openai.gif b/docs/src/pages/docs/_assets/model-openai.gif
deleted file mode 100644
index b20aa9c0e..000000000
Binary files a/docs/src/pages/docs/_assets/model-openai.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/model-parameters.png b/docs/src/pages/docs/_assets/model-parameters.png
index beb26972b..a99deb75e 100644
Binary files a/docs/src/pages/docs/_assets/model-parameters.png and b/docs/src/pages/docs/_assets/model-parameters.png differ
diff --git a/docs/src/pages/docs/_assets/model-tab.png b/docs/src/pages/docs/_assets/model-tab.png
deleted file mode 100644
index ea595bf19..000000000
Binary files a/docs/src/pages/docs/_assets/model-tab.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/model.gif b/docs/src/pages/docs/_assets/model.gif
deleted file mode 100644
index e5abdf757..000000000
Binary files a/docs/src/pages/docs/_assets/model.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/modelparam.gif b/docs/src/pages/docs/_assets/modelparam.gif
deleted file mode 100644
index f33cf5ab5..000000000
Binary files a/docs/src/pages/docs/_assets/modelparam.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/models.gif b/docs/src/pages/docs/_assets/models.gif
deleted file mode 100644
index 8dc4d7925..000000000
Binary files a/docs/src/pages/docs/_assets/models.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/mymodels.png b/docs/src/pages/docs/_assets/mymodels.png
deleted file mode 100644
index d12664b86..000000000
Binary files a/docs/src/pages/docs/_assets/mymodels.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/nitro-network.png b/docs/src/pages/docs/_assets/nitro-network.png
deleted file mode 100644
index 5f1f7104e..000000000
Binary files a/docs/src/pages/docs/_assets/nitro-network.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/nvidia-1.gif b/docs/src/pages/docs/_assets/nvidia-1.gif
deleted file mode 100644
index df28abc1f..000000000
Binary files a/docs/src/pages/docs/_assets/nvidia-1.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/nvidia-2.gif b/docs/src/pages/docs/_assets/nvidia-2.gif
deleted file mode 100644
index 4c628884a..000000000
Binary files a/docs/src/pages/docs/_assets/nvidia-2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/nvidia-nim.png b/docs/src/pages/docs/_assets/nvidia-nim.png
new file mode 100644
index 000000000..8eb637c12
Binary files /dev/null and b/docs/src/pages/docs/_assets/nvidia-nim.png differ
diff --git a/docs/src/pages/docs/_assets/openai.png b/docs/src/pages/docs/_assets/openai.png
new file mode 100644
index 000000000..e6407d9ed
Binary files /dev/null and b/docs/src/pages/docs/_assets/openai.png differ
diff --git a/docs/src/pages/docs/_assets/openrouter.png b/docs/src/pages/docs/_assets/openrouter.png
new file mode 100644
index 000000000..d9e032700
Binary files /dev/null and b/docs/src/pages/docs/_assets/openrouter.png differ
diff --git a/docs/src/pages/docs/_assets/phi.gif b/docs/src/pages/docs/_assets/phi.gif
deleted file mode 100644
index e8cd587d2..000000000
Binary files a/docs/src/pages/docs/_assets/phi.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/preserve.png b/docs/src/pages/docs/_assets/preserve.png
deleted file mode 100644
index 459c940e2..000000000
Binary files a/docs/src/pages/docs/_assets/preserve.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/quick-ask.gif b/docs/src/pages/docs/_assets/quick-ask.gif
deleted file mode 100644
index f4081ca3c..000000000
Binary files a/docs/src/pages/docs/_assets/quick-ask.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/quick-ask.png b/docs/src/pages/docs/_assets/quick-ask.png
deleted file mode 100644
index 34620569b..000000000
Binary files a/docs/src/pages/docs/_assets/quick-ask.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/quick-start-01.png b/docs/src/pages/docs/_assets/quick-start-01.png
index e5aaeb88b..430b01fb0 100644
Binary files a/docs/src/pages/docs/_assets/quick-start-01.png and b/docs/src/pages/docs/_assets/quick-start-01.png differ
diff --git a/docs/src/pages/docs/_assets/quick-start-02.png b/docs/src/pages/docs/_assets/quick-start-02.png
index f1ecd7ee5..998e04dc8 100644
Binary files a/docs/src/pages/docs/_assets/quick-start-02.png and b/docs/src/pages/docs/_assets/quick-start-02.png differ
diff --git a/docs/src/pages/docs/_assets/quick-start-03.png b/docs/src/pages/docs/_assets/quick-start-03.png
index fe249f194..51d6de0d2 100644
Binary files a/docs/src/pages/docs/_assets/quick-start-03.png and b/docs/src/pages/docs/_assets/quick-start-03.png differ
diff --git a/docs/src/pages/docs/_assets/reset-jan.png b/docs/src/pages/docs/_assets/reset-jan.png
deleted file mode 100644
index 2c20ca667..000000000
Binary files a/docs/src/pages/docs/_assets/reset-jan.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/reset.gif b/docs/src/pages/docs/_assets/reset.gif
deleted file mode 100644
index 6ff171f25..000000000
Binary files a/docs/src/pages/docs/_assets/reset.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/retrieval-01.png b/docs/src/pages/docs/_assets/retrieval-01.png
index 12c8d832b..b29b964b5 100644
Binary files a/docs/src/pages/docs/_assets/retrieval-01.png and b/docs/src/pages/docs/_assets/retrieval-01.png differ
diff --git a/docs/src/pages/docs/_assets/retrieval-02.png b/docs/src/pages/docs/_assets/retrieval-02.png
index 6c57e55a2..37fc8bf99 100644
Binary files a/docs/src/pages/docs/_assets/retrieval-02.png and b/docs/src/pages/docs/_assets/retrieval-02.png differ
diff --git a/docs/src/pages/docs/_assets/retrieval.gif b/docs/src/pages/docs/_assets/retrieval.gif
deleted file mode 100644
index 38d3fbb58..000000000
Binary files a/docs/src/pages/docs/_assets/retrieval.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/retrieval1.png b/docs/src/pages/docs/_assets/retrieval1.png
deleted file mode 100644
index 8b5d63bb8..000000000
Binary files a/docs/src/pages/docs/_assets/retrieval1.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/retrieval2.png b/docs/src/pages/docs/_assets/retrieval2.png
deleted file mode 100644
index 2c9443c44..000000000
Binary files a/docs/src/pages/docs/_assets/retrieval2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/scheme.png b/docs/src/pages/docs/_assets/scheme.png
deleted file mode 100644
index fe50d4b6a..000000000
Binary files a/docs/src/pages/docs/_assets/scheme.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/search-bar.png b/docs/src/pages/docs/_assets/search-bar.png
deleted file mode 100644
index b6e559773..000000000
Binary files a/docs/src/pages/docs/_assets/search-bar.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/select-openai.gif b/docs/src/pages/docs/_assets/select-openai.gif
deleted file mode 100644
index 3196c4320..000000000
Binary files a/docs/src/pages/docs/_assets/select-openai.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-anthropic.gif b/docs/src/pages/docs/_assets/server-anthropic.gif
deleted file mode 100644
index da684dbd4..000000000
Binary files a/docs/src/pages/docs/_assets/server-anthropic.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-cohere.gif b/docs/src/pages/docs/_assets/server-cohere.gif
deleted file mode 100644
index 72c9bd883..000000000
Binary files a/docs/src/pages/docs/_assets/server-cohere.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-groq.gif b/docs/src/pages/docs/_assets/server-groq.gif
deleted file mode 100644
index dcb3e03f7..000000000
Binary files a/docs/src/pages/docs/_assets/server-groq.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-llama2.gif b/docs/src/pages/docs/_assets/server-llama2.gif
deleted file mode 100644
index b1ebafc84..000000000
Binary files a/docs/src/pages/docs/_assets/server-llama2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-martian.gif b/docs/src/pages/docs/_assets/server-martian.gif
deleted file mode 100644
index a33dd7f19..000000000
Binary files a/docs/src/pages/docs/_assets/server-martian.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-mistral.gif b/docs/src/pages/docs/_assets/server-mistral.gif
deleted file mode 100644
index 8fdb6ef0f..000000000
Binary files a/docs/src/pages/docs/_assets/server-mistral.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-openai.gif b/docs/src/pages/docs/_assets/server-openai.gif
deleted file mode 100644
index fd10aa75b..000000000
Binary files a/docs/src/pages/docs/_assets/server-openai.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-openai2.gif b/docs/src/pages/docs/_assets/server-openai2.gif
deleted file mode 100644
index b4338d122..000000000
Binary files a/docs/src/pages/docs/_assets/server-openai2.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/server-phi.gif b/docs/src/pages/docs/_assets/server-phi.gif
deleted file mode 100644
index 4abc916c1..000000000
Binary files a/docs/src/pages/docs/_assets/server-phi.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/set-model.gif b/docs/src/pages/docs/_assets/set-model.gif
deleted file mode 100644
index 87de85732..000000000
Binary files a/docs/src/pages/docs/_assets/set-model.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/set-tensor.gif b/docs/src/pages/docs/_assets/set-tensor.gif
deleted file mode 100644
index 22e824882..000000000
Binary files a/docs/src/pages/docs/_assets/set-tensor.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/set-url.gif b/docs/src/pages/docs/_assets/set-url.gif
deleted file mode 100644
index b675b3ea4..000000000
Binary files a/docs/src/pages/docs/_assets/set-url.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/settings-01.png b/docs/src/pages/docs/_assets/settings-01.png
index 8a2b734d7..7a8539d65 100644
Binary files a/docs/src/pages/docs/_assets/settings-01.png and b/docs/src/pages/docs/_assets/settings-01.png differ
diff --git a/docs/src/pages/docs/_assets/settings-02.png b/docs/src/pages/docs/_assets/settings-02.png
index f54d6bd44..b00924d5a 100644
Binary files a/docs/src/pages/docs/_assets/settings-02.png and b/docs/src/pages/docs/_assets/settings-02.png differ
diff --git a/docs/src/pages/docs/_assets/settings-03.png b/docs/src/pages/docs/_assets/settings-03.png
index f149ebb01..aced464be 100644
Binary files a/docs/src/pages/docs/_assets/settings-03.png and b/docs/src/pages/docs/_assets/settings-03.png differ
diff --git a/docs/src/pages/docs/_assets/settings-04.png b/docs/src/pages/docs/_assets/settings-04.png
index cb55f4cd0..36dc64fce 100644
Binary files a/docs/src/pages/docs/_assets/settings-04.png and b/docs/src/pages/docs/_assets/settings-04.png differ
diff --git a/docs/src/pages/docs/_assets/settings-05.png b/docs/src/pages/docs/_assets/settings-05.png
index 37aea156a..8fb76e0fc 100644
Binary files a/docs/src/pages/docs/_assets/settings-05.png and b/docs/src/pages/docs/_assets/settings-05.png differ
diff --git a/docs/src/pages/docs/_assets/settings-06.png b/docs/src/pages/docs/_assets/settings-06.png
index a25688f1e..3a582b42a 100644
Binary files a/docs/src/pages/docs/_assets/settings-06.png and b/docs/src/pages/docs/_assets/settings-06.png differ
diff --git a/docs/src/pages/docs/_assets/settings-07.png b/docs/src/pages/docs/_assets/settings-07.png
index 620ca2078..878c4ea66 100644
Binary files a/docs/src/pages/docs/_assets/settings-07.png and b/docs/src/pages/docs/_assets/settings-07.png differ
diff --git a/docs/src/pages/docs/_assets/settings-08.png b/docs/src/pages/docs/_assets/settings-08.png
index 9597831e0..a1c7d7dd1 100644
Binary files a/docs/src/pages/docs/_assets/settings-08.png and b/docs/src/pages/docs/_assets/settings-08.png differ
diff --git a/docs/src/pages/docs/_assets/settings-09.png b/docs/src/pages/docs/_assets/settings-09.png
index 58ce63cba..8d5e577e1 100644
Binary files a/docs/src/pages/docs/_assets/settings-09.png and b/docs/src/pages/docs/_assets/settings-09.png differ
diff --git a/docs/src/pages/docs/_assets/settings-10.png b/docs/src/pages/docs/_assets/settings-10.png
index c72aa7c0b..b89c9dace 100644
Binary files a/docs/src/pages/docs/_assets/settings-10.png and b/docs/src/pages/docs/_assets/settings-10.png differ
diff --git a/docs/src/pages/docs/_assets/settings-11.png b/docs/src/pages/docs/_assets/settings-11.png
index 833386ef4..b304f9f3b 100644
Binary files a/docs/src/pages/docs/_assets/settings-11.png and b/docs/src/pages/docs/_assets/settings-11.png differ
diff --git a/docs/src/pages/docs/_assets/settings-12.png b/docs/src/pages/docs/_assets/settings-12.png
index a41744fa1..a4d45dedd 100644
Binary files a/docs/src/pages/docs/_assets/settings-12.png and b/docs/src/pages/docs/_assets/settings-12.png differ
diff --git a/docs/src/pages/docs/_assets/settings-13.png b/docs/src/pages/docs/_assets/settings-13.png
index a1111d8e2..f2b41ba3f 100644
Binary files a/docs/src/pages/docs/_assets/settings-13.png and b/docs/src/pages/docs/_assets/settings-13.png differ
diff --git a/docs/src/pages/docs/_assets/settings-14.png b/docs/src/pages/docs/_assets/settings-14.png
index 049f0070a..b253cb0dc 100644
Binary files a/docs/src/pages/docs/_assets/settings-14.png and b/docs/src/pages/docs/_assets/settings-14.png differ
diff --git a/docs/src/pages/docs/_assets/settings-15.png b/docs/src/pages/docs/_assets/settings-15.png
index 6292523fa..f739f8513 100644
Binary files a/docs/src/pages/docs/_assets/settings-15.png and b/docs/src/pages/docs/_assets/settings-15.png differ
diff --git a/docs/src/pages/docs/_assets/settings-17.png b/docs/src/pages/docs/_assets/settings-17.png
index c8e74183b..83c10fa95 100644
Binary files a/docs/src/pages/docs/_assets/settings-17.png and b/docs/src/pages/docs/_assets/settings-17.png differ
diff --git a/docs/src/pages/docs/_assets/settings-18.png b/docs/src/pages/docs/_assets/settings-18.png
index fbc081e85..63718d0ef 100644
Binary files a/docs/src/pages/docs/_assets/settings-18.png and b/docs/src/pages/docs/_assets/settings-18.png differ
diff --git a/docs/src/pages/docs/_assets/settings.png b/docs/src/pages/docs/_assets/settings.png
deleted file mode 100644
index e53f71d1e..000000000
Binary files a/docs/src/pages/docs/_assets/settings.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/shortcut.png b/docs/src/pages/docs/_assets/shortcut.png
deleted file mode 100644
index b7912f838..000000000
Binary files a/docs/src/pages/docs/_assets/shortcut.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/specific-model.gif b/docs/src/pages/docs/_assets/specific-model.gif
deleted file mode 100644
index 2664bcc47..000000000
Binary files a/docs/src/pages/docs/_assets/specific-model.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/spell.png b/docs/src/pages/docs/_assets/spell.png
deleted file mode 100644
index 92e9a6d8c..000000000
Binary files a/docs/src/pages/docs/_assets/spell.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/ssl.gif b/docs/src/pages/docs/_assets/ssl.gif
deleted file mode 100644
index b18f075c1..000000000
Binary files a/docs/src/pages/docs/_assets/ssl.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/ssl.png b/docs/src/pages/docs/_assets/ssl.png
deleted file mode 100644
index 970051744..000000000
Binary files a/docs/src/pages/docs/_assets/ssl.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/start-anthropic.gif b/docs/src/pages/docs/_assets/start-anthropic.gif
deleted file mode 100644
index 9ca938655..000000000
Binary files a/docs/src/pages/docs/_assets/start-anthropic.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/start-cohere.gif b/docs/src/pages/docs/_assets/start-cohere.gif
deleted file mode 100644
index b28fde498..000000000
Binary files a/docs/src/pages/docs/_assets/start-cohere.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/start-groq.gif b/docs/src/pages/docs/_assets/start-groq.gif
deleted file mode 100644
index 8e5ced12a..000000000
Binary files a/docs/src/pages/docs/_assets/start-groq.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/start-martian.gif b/docs/src/pages/docs/_assets/start-martian.gif
deleted file mode 100644
index 31a79d16d..000000000
Binary files a/docs/src/pages/docs/_assets/start-martian.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/start-mistral.gif b/docs/src/pages/docs/_assets/start-mistral.gif
deleted file mode 100644
index 1012a651a..000000000
Binary files a/docs/src/pages/docs/_assets/start-mistral.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/start-router.gif b/docs/src/pages/docs/_assets/start-router.gif
deleted file mode 100644
index ca1737de0..000000000
Binary files a/docs/src/pages/docs/_assets/start-router.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/system-mili.png b/docs/src/pages/docs/_assets/system-mili.png
deleted file mode 100644
index 0a0f28592..000000000
Binary files a/docs/src/pages/docs/_assets/system-mili.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/system-mili2.png b/docs/src/pages/docs/_assets/system-mili2.png
deleted file mode 100644
index 6135c5a49..000000000
Binary files a/docs/src/pages/docs/_assets/system-mili2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/system-monitor.png b/docs/src/pages/docs/_assets/system-monitor.png
deleted file mode 100644
index 04d6749cb..000000000
Binary files a/docs/src/pages/docs/_assets/system-monitor.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/system-monitor2.png b/docs/src/pages/docs/_assets/system-monitor2.png
deleted file mode 100644
index e03dfbf63..000000000
Binary files a/docs/src/pages/docs/_assets/system-monitor2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/system-slider.png b/docs/src/pages/docs/_assets/system-slider.png
deleted file mode 100644
index 409338f70..000000000
Binary files a/docs/src/pages/docs/_assets/system-slider.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/system-slider2.png b/docs/src/pages/docs/_assets/system-slider2.png
deleted file mode 100644
index 9114d32c3..000000000
Binary files a/docs/src/pages/docs/_assets/system-slider2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/tensor.png b/docs/src/pages/docs/_assets/tensor.png
deleted file mode 100644
index 812640de3..000000000
Binary files a/docs/src/pages/docs/_assets/tensor.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/theme.png b/docs/src/pages/docs/_assets/theme.png
deleted file mode 100644
index ae14f2054..000000000
Binary files a/docs/src/pages/docs/_assets/theme.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/thread-settings.png b/docs/src/pages/docs/_assets/thread-settings.png
deleted file mode 100644
index a7cc60304..000000000
Binary files a/docs/src/pages/docs/_assets/thread-settings.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/thread-settings2.png b/docs/src/pages/docs/_assets/thread-settings2.png
deleted file mode 100644
index b0d728d43..000000000
Binary files a/docs/src/pages/docs/_assets/thread-settings2.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/threads-01.png b/docs/src/pages/docs/_assets/threads-01.png
index 0781c8347..564c753c2 100644
Binary files a/docs/src/pages/docs/_assets/threads-01.png and b/docs/src/pages/docs/_assets/threads-01.png differ
diff --git a/docs/src/pages/docs/_assets/threads-02.png b/docs/src/pages/docs/_assets/threads-02.png
index c30ce7a15..f51916571 100644
Binary files a/docs/src/pages/docs/_assets/threads-02.png and b/docs/src/pages/docs/_assets/threads-02.png differ
diff --git a/docs/src/pages/docs/_assets/threads-03.png b/docs/src/pages/docs/_assets/threads-03.png
index 5193bdc46..5a5fd3238 100644
Binary files a/docs/src/pages/docs/_assets/threads-03.png and b/docs/src/pages/docs/_assets/threads-03.png differ
diff --git a/docs/src/pages/docs/_assets/threads-04.png b/docs/src/pages/docs/_assets/threads-04.png
index 352133941..748282868 100644
Binary files a/docs/src/pages/docs/_assets/threads-04.png and b/docs/src/pages/docs/_assets/threads-04.png differ
diff --git a/docs/src/pages/docs/_assets/threads-05.png b/docs/src/pages/docs/_assets/threads-05.png
index c232c773f..4c11667d8 100644
Binary files a/docs/src/pages/docs/_assets/threads-05.png and b/docs/src/pages/docs/_assets/threads-05.png differ
diff --git a/docs/src/pages/docs/_assets/tools.png b/docs/src/pages/docs/_assets/tools.png
deleted file mode 100644
index 8104f6336..000000000
Binary files a/docs/src/pages/docs/_assets/tools.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/trouble-shooting-01.png b/docs/src/pages/docs/_assets/trouble-shooting-01.png
index fc965988b..a12ea97b0 100644
Binary files a/docs/src/pages/docs/_assets/trouble-shooting-01.png and b/docs/src/pages/docs/_assets/trouble-shooting-01.png differ
diff --git a/docs/src/pages/docs/_assets/trouble-shooting-02.png b/docs/src/pages/docs/_assets/trouble-shooting-02.png
index e8f7ee4a6..290c4a8ce 100644
Binary files a/docs/src/pages/docs/_assets/trouble-shooting-02.png and b/docs/src/pages/docs/_assets/trouble-shooting-02.png differ
diff --git a/docs/src/pages/docs/_assets/turn-off.png b/docs/src/pages/docs/_assets/turn-off.png
deleted file mode 100644
index a16a26a5b..000000000
Binary files a/docs/src/pages/docs/_assets/turn-off.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/ui.gif b/docs/src/pages/docs/_assets/ui.gif
deleted file mode 100644
index 946ad8fb9..000000000
Binary files a/docs/src/pages/docs/_assets/ui.gif and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/vulkan.png b/docs/src/pages/docs/_assets/vulkan.png
deleted file mode 100644
index e8104bcc7..000000000
Binary files a/docs/src/pages/docs/_assets/vulkan.png and /dev/null differ
diff --git a/docs/src/pages/docs/_assets/windows.png b/docs/src/pages/docs/_assets/windows.png
deleted file mode 100644
index 69ef6f9a8..000000000
Binary files a/docs/src/pages/docs/_assets/windows.png and /dev/null differ
diff --git a/docs/src/pages/docs/_meta.json b/docs/src/pages/docs/_meta.json
index 86cba0de6..bdd9be159 100644
--- a/docs/src/pages/docs/_meta.json
+++ b/docs/src/pages/docs/_meta.json
@@ -24,17 +24,20 @@
"threads": "Threads",
"settings": "Settings",
"inference-engines": {
- "title": "MODEL PROVIDER",
+ "title": "ENGINES",
"type": "separator"
},
- "built-in": "Built-in Models",
- "remote-models": "Remote APIs",
+ "local-engines": "Local Engines",
+ "remote-models": "Remote Engines",
+ "install-engines": "Install Engines",
"extensions-separator": {
"title": "EXTENSIONS",
"type": "separator"
},
"extensions": "Overview",
- "installing-extension": "Installing an Extension",
+ "extensions-settings": "Extension Settings",
+ "configure-extensions": "Configure Extensions",
+ "install-extensions": "Install Extensions",
"troubleshooting-separator": {
"title": "TROUBLESHOOTING",
"type": "separator"
diff --git a/docs/src/pages/docs/built-in/_meta.json b/docs/src/pages/docs/built-in/_meta.json
deleted file mode 100644
index 0b15c47f2..000000000
--- a/docs/src/pages/docs/built-in/_meta.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "llama-cpp": {
- "title": "llama.cpp",
- "href": "/docs/built-in/llama-cpp"
- },
- "tensorrt-llm": {
- "title": "TensorRT-LLM",
- "href": "/docs/built-in/tensorrt-llm"
- }
-}
diff --git a/docs/src/pages/docs/built-in/llama-cpp.mdx b/docs/src/pages/docs/built-in/llama-cpp.mdx
deleted file mode 100644
index 8d71ff2ae..000000000
--- a/docs/src/pages/docs/built-in/llama-cpp.mdx
+++ /dev/null
@@ -1,133 +0,0 @@
----
-title: llama.cpp
-description: A step-by-step guide on how to customize the llama.cpp engine.
-keywords:
- [
- Jan,
- Customizable Intelligence, LLM,
- local AI,
- privacy focus,
- free and open source,
- private and offline,
- conversational AI,
- no-subscription fee,
- large language models,
- Llama CPP integration,
- llama.cpp Engine,
- Intel CPU,
- AMD CPU,
- NVIDIA GPU,
- AMD GPU Radeon,
- Apple Silicon,
- Intel Arc GPU,
- ]
----
-
-import { Tabs } from 'nextra/components'
-import { Callout, Steps } from 'nextra/components'
-
-# llama.cpp (Cortex)
-
-## Overview
-
-Jan has [**Cortex**](https://github.com/janhq/cortex) - a default C++ inference server built on top of [llama.cpp](https://github.com/ggerganov/llama.cpp). This server provides an OpenAI-compatible API, queues, scaling, and additional features on top of the wide capabilities of `llama.cpp`.
-
-This guide shows you how to initialize the `llama.cpp` to download and install the required dependencies to start chatting with a model using the `llama.cpp` engine.
-
-## Prerequisites
-- Mac Intel:
- - Make sure you're using an Intel-based Mac. For a complete list of supported Intel CPUs, please see [here](https://en.wikipedia.org/wiki/MacBook_Pro_(Intel-based)).
- - For Mac Intel, it is recommended to utilize smaller models.
-- Mac Sillicon:
- - Make sure you're using a Mac Silicon. For a complete list of supported Apple Silicon CPUs, please see [here](https://en.wikipedia.org/wiki/Apple_Silicon).
- - Using an adequate model size based on your hardware is recommended for Mac Silicon.
-
- This can use Apple GPU with Metal by default for acceleration. Apple ANE is not supported yet.
-
-- Windows:
- - Ensure that you have **Windows with x86_64** architecture.
-- Linux:
- - Ensure that you have **Linux with x86_64** architecture.
-
-#### GPU Acceleration Options
-Enable the GPU acceleration option within the Jan application by following the [Installation Setup](/docs/desktop-installation) guide.
-## Step-by-step Guide
-
-### Step 1: Open the `model.json`
-1. Open [Jan Data Folder](/docs/data-folder#open-jan-data-folder)
-
-
-
-
-
-2. Select **models** folder > Click **model folder** that you want to modify > click `model.json`
-3. Once open, `model.json` file looks like below, use model "TinyLlama Chat 1.1B Q4" as an example:
-```json
-{
- "sources": [
- {
- "filename": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
- "url": "https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf"
- }
- ],
- "id": "tinyllama-1.1b",
- "object": "model",
- "name": "TinyLlama Chat 1.1B Q4",
- "version": "1.0",
- "description": "TinyLlama is a tiny model with only 1.1B. It's a good model for less powerful computers.",
- "format": "gguf",
- "settings": {
- "ctx_len": 4096,
- "prompt_template": "<|system|>\n{system_message}<|user|>\n{prompt}<|assistant|>",
- "llama_model_path": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf"
- },
- "parameters": {
- "temperature": 0.7,
- "top_p": 0.95,
- "stream": true,
- "max_tokens": 2048,
- "stop": [],
- "frequency_penalty": 0,
- "presence_penalty": 0
- },
- "metadata": {
- "author": "TinyLlama",
- "tags": [
- "Tiny",
- "Foundation Model"
- ],
- "size": 669000000
- },
- "engine": "nitro"
-}
-```
-### Step 2: Modify the `model.json`
-1. Modify the model's engine settings under the settings array. You can modify the settings with the following parameters:
-
-| Parameter | Type | Description |
-| --------------- | ----------- | ---------------------------------------------------------------------------------------------------------------------------------- |
-| `ctx_len` | **Integer** | Provides ample context for model operations like `GPT-3.5`. The default value is `2048` (_Maximum_: `4096`, _Minimum_: `1`). |
-| `prompt_template` | **String** | Defines the template used to format prompts |
-| `model_path` | **String** | Specifies the path to the model `.GGUF` file. |
-| `ngl` | **Integer** | Determines GPU layer usage. The default value is `100`. |
-| `cpu_threads` | **Integer** | Determines CPU inference threads, limited by hardware and OS. (_Maximum_ determined by system) |
-| `cont_batching` | **Integer** | Controls continuous batching, enhancing throughput for LLM inference. |
-| `embedding` | **Integer** | Enables embedding utilization for tasks like document-enhanced chat in RAG-based applications. |
-2. Save the `model.json` file.
-
- If you use a different model, you must set it up again. As this only affects the selected model.
-
-### Step 3: Start the Model
-1. Restart the Jan application to apply your settings.
-2. Navigate to the **Threads**.
-3. Chat with your model.
-
-
- - To utilize the embedding feature, include the JSON parameter `"embedding": true`. It will enable Nitro to process inferences with embedding capabilities. Please refer to the [Embedding in the Nitro documentation](https://nitro.jan.ai/features/embed) for a more detailed explanation.
- - To utilize the continuous batching feature for boosting throughput and minimizing latency in large language model (LLM) inference, include `cont_batching: true`. For details, please refer to the [Continuous Batching in the Nitro documentation](https://nitro.jan.ai/features/cont-batch).
-
-
-
-
- If you have questions, please join our [Discord community](https://discord.gg/Dt7MxDyNNZ) for support, updates, and discussions.
-
diff --git a/docs/src/pages/docs/built-in/tensorrt-llm.mdx b/docs/src/pages/docs/built-in/tensorrt-llm.mdx
deleted file mode 100644
index b165dedc9..000000000
--- a/docs/src/pages/docs/built-in/tensorrt-llm.mdx
+++ /dev/null
@@ -1,91 +0,0 @@
----
-title: TensorRT-LLM
-description: A step-by-step guide on customizing the TensorRT-LLM engine.
-keywords:
- [
- Jan,
- Customizable Intelligence, LLM,
- local AI,
- privacy focus,
- free and open source,
- private and offline,
- conversational AI,
- no-subscription fee,
- large language models,
- TensorRT-LLM Engine,
- TensorRT,
- tensorRT,
- engine,
- ]
----
-
-import { Callout, Steps } from 'nextra/components'
-
-# TensorRT-LLM
-
-## Overview
-
-This guide walks you through installing Jan's official [TensorRT-LLM Engine](https://github.com/janhq/nitro-tensorrt-llm). This engine uses [Cortex-TensorRT-LLM](https://github.com/janhq/cortex.tensorrt-llm) as the AI engine instead of the default [Cortex-Llama-CPP](https://github.com/janhq/cortex). It includes an efficient C++ server that executes the [TRT-LLM C++ runtime](https://nvidia.github.io/TensorRT-LLM/gpt_runtime.html) natively. It also includes features and performance improvements like OpenAI compatibility, tokenizer improvements, and queues.
-
-
- This feature is only available for Windows users. Linux is coming soon.
-
-
-### Pre-requisites
-
-- A **Windows** PC.
-- **Nvidia GPU(s)**: Ada or Ampere series (i.e. RTX 4000s & 3000s). More will be supported soon.
-- Sufficient disk space for the TensorRT-LLM models and data files (space requirements vary depending on the model size).
-
-
-
-### Step 1: Install TensorRT-Extension
-
-1. Click the **Gear Icon (⚙️)** on the bottom left of your screen.
-
-
-
-2. Select the **TensorRT-LLM** under the **Model Provider** section.
-
-
-
-3. Click **Install** to install the required dependencies to use TensorRT-LLM.
-
-
-
-3. Check that files are correctly downloaded.
-
-```bash
-ls ~/jan/data/extensions/@janhq/tensorrt-llm-extension/dist/bin
-# Your Extension Folder should now include `nitro.exe`, among other artifacts needed to run TRT-LLM
-```
-
-### Step 2: Download a Compatible Model
-
-TensorRT-LLM can only run models in `TensorRT` format. These models, aka "TensorRT Engines", are prebuilt for each target OS+GPU architecture.
-
-We offer a handful of precompiled models for Ampere and Ada cards that you can immediately download and play with:
-
-1. Restart the application and go to the Hub.
-2. Look for models with the `TensorRT-LLM` label in the recommended models list > Click **Download**.
-
-
- This step might take some time. 🙏
-
-
-
-
-3. Click **Download** to download the model.
-
-### Step 3: Configure Settings
-
-1. Navigate to the Thread section.
-2. Select the model that you have downloaded.
-3. Customize the default parameters of the model for how Jan runs TensorRT-LLM.
-
- Please see [here](/docs/models/model-parameters) for more detailed model parameters.
-
-
-
-
-
\ No newline at end of file
diff --git a/docs/src/pages/docs/configure-extensions.mdx b/docs/src/pages/docs/configure-extensions.mdx
new file mode 100644
index 000000000..c72d8c6a5
--- /dev/null
+++ b/docs/src/pages/docs/configure-extensions.mdx
@@ -0,0 +1,301 @@
+---
+title: Configure Extensions
+description: Learn about Jan's default extensions and explore how to configure them.
+ [
+ Jan,
+ Customizable Intelligence, LLM,
+ local AI,
+ privacy focus,
+ free and open source,
+ private and offline,
+ conversational AI,
+ no-subscription fee,
+ large language models,
+ Jan Extensions,
+ Extensions,
+ ]
+---
+
+# Configure Extensions
+Extensions in Jan are configured through a JSON file that defines their behavior and metadata. While most users won't need to modify this file directly, advanced users can customize extension settings for specific needs.
+
+## Access extensions.json
+
+1. Navigate to [Jan Data Folder](/docs/data-folder):
+2. Open the `extensions.json` file in any text editor
+
+## Configuration Options
+
+| Option | Description | Example |
+|--------|-------------|---------|
+| `_active` | Enable/disable the extension | `true` or `false` |
+| `listeners` | Event listeners configuration | `{}` for default |
+| `origin` | Installation path of the extension | `"C:\\Users\\...\\jan\\resources\\app.asar.unpacked\\..."` |
+| `installOptions` | Installation configuration | `{"version": false, "fullMetadata": true}` |
+| `name` | Internal extension identifier | `"@janhq/conversational-extension"` |
+| `productName` | Display name shown in UI | `"Conversational"` |
+| `version` | Extension version number | `"1.0.0"` |
+| `main` | Entry point file path | `"dist/index.js"` |
+| `description` | Extension description | `"This extension enables conversations..."` |
+| `url` | Extension repository URL | `"extension://@janhq/conversational-extension/dist/index.js"` |
+
+## Full Example Configuration
+
+```json title="~/jan/data/extensions/extensions.json"
+{
+ "@janhq/conversational-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-conversational-extension-1.0.0.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/conversational-extension",
+ "productName": "Conversational",
+ "version": "1.0.0",
+ "main": "dist/index.js",
+ "description": "This extension enables conversations and state persistence via your filesystem",
+ "url": "extension://@janhq/conversational-extension/dist/index.js"
+ },
+ "@janhq/inference-anthropic-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-anthropic-extension-1.0.2.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-anthropic-extension",
+ "productName": "Anthropic Inference Engine",
+ "version": "1.0.2",
+ "main": "dist/index.js",
+ "description": "This extension enables Anthropic chat completion API calls",
+ "url": "extension://@janhq/inference-anthropic-extension/dist/index.js"
+ },
+ "@janhq/inference-triton-trt-llm-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-triton-trt-llm-extension-1.0.0.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-triton-trt-llm-extension",
+ "productName": "Triton-TRT-LLM Inference Engine",
+ "version": "1.0.0",
+ "main": "dist/index.js",
+ "description": "This extension enables Nvidia's TensorRT-LLM as an inference engine option",
+ "url": "extension://@janhq/inference-triton-trt-llm-extension/dist/index.js"
+ },
+ "@janhq/inference-mistral-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-mistral-extension-1.0.1.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-mistral-extension",
+ "productName": "MistralAI Inference Engine",
+ "version": "1.0.1",
+ "main": "dist/index.js",
+ "description": "This extension enables Mistral chat completion API calls",
+ "url": "extension://@janhq/inference-mistral-extension/dist/index.js"
+ },
+ "@janhq/inference-martian-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-martian-extension-1.0.1.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-martian-extension",
+ "productName": "Martian Inference Engine",
+ "version": "1.0.1",
+ "main": "dist/index.js",
+ "description": "This extension enables Martian chat completion API calls",
+ "url": "extension://@janhq/inference-martian-extension/dist/index.js"
+ },
+ "@janhq/inference-openrouter-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-openrouter-extension-1.0.0.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-openrouter-extension",
+ "productName": "OpenRouter Inference Engine",
+ "version": "1.0.0",
+ "main": "dist/index.js",
+ "description": "This extension enables Open Router chat completion API calls",
+ "url": "extension://@janhq/inference-openrouter-extension/dist/index.js"
+ },
+ "@janhq/inference-nvidia-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-nvidia-extension-1.0.1.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-nvidia-extension",
+ "productName": "NVIDIA NIM Inference Engine",
+ "version": "1.0.1",
+ "main": "dist/index.js",
+ "description": "This extension enables NVIDIA chat completion API calls",
+ "url": "extension://@janhq/inference-nvidia-extension/dist/index.js"
+ },
+ "@janhq/inference-groq-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-groq-extension-1.0.1.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-groq-extension",
+ "productName": "Groq Inference Engine",
+ "version": "1.0.1",
+ "main": "dist/index.js",
+ "description": "This extension enables fast Groq chat completion API calls",
+ "url": "extension://@janhq/inference-groq-extension/dist/index.js"
+ },
+ "@janhq/inference-openai-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-openai-extension-1.0.2.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-openai-extension",
+ "productName": "OpenAI Inference Engine",
+ "version": "1.0.2",
+ "main": "dist/index.js",
+ "description": "This extension enables OpenAI chat completion API calls",
+ "url": "extension://@janhq/inference-openai-extension/dist/index.js"
+ },
+ "@janhq/inference-cohere-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-cohere-extension-1.0.0.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-cohere-extension",
+ "productName": "Cohere Inference Engine",
+ "version": "1.0.0",
+ "main": "dist/index.js",
+ "description": "This extension enables Cohere chat completion API calls",
+ "url": "extension://@janhq/inference-cohere-extension/dist/index.js"
+ },
+ "@janhq/model-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-model-extension-1.0.33.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/model-extension",
+ "productName": "Model Management",
+ "version": "1.0.33",
+ "main": "dist/index.js",
+ "description": "Model Management Extension provides model exploration and seamless downloads",
+ "url": "extension://@janhq/model-extension/dist/index.js"
+ },
+ "@janhq/monitoring-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-monitoring-extension-1.0.10.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/monitoring-extension",
+ "productName": "System Monitoring",
+ "version": "1.0.10",
+ "main": "dist/index.js",
+ "description": "This extension provides system health and OS level data",
+ "url": "extension://@janhq/monitoring-extension/dist/index.js"
+ },
+ "@janhq/assistant-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-assistant-extension-1.0.1.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/assistant-extension",
+ "productName": "Jan Assistant",
+ "version": "1.0.1",
+ "main": "dist/index.js",
+ "description": "This extension enables assistants, including Jan, a default assistant that can call all downloaded models",
+ "url": "extension://@janhq/assistant-extension/dist/index.js"
+ },
+ "@janhq/tensorrt-llm-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-tensorrt-llm-extension-0.0.3.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/tensorrt-llm-extension",
+ "productName": "TensorRT-LLM Inference Engine",
+ "version": "0.0.3",
+ "main": "dist/index.js",
+ "description": "This extension enables Nvidia's TensorRT-LLM for the fastest GPU acceleration. See the [setup guide](https://jan.ai/guides/providers/tensorrt-llm/) for next steps.",
+ "url": "extension://@janhq/tensorrt-llm-extension/dist/index.js"
+ },
+ "@janhq/inference-cortex-extension": {
+ "_active": true,
+ "listeners": {},
+ "origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-cortex-extension-1.0.15.tgz",
+ "installOptions": {
+ "version": false,
+ "fullMetadata": true
+ },
+ "name": "@janhq/inference-cortex-extension",
+ "productName": "Cortex Inference Engine",
+ "version": "1.0.15",
+ "main": "dist/index.js",
+ "description": "This extension embeds cortex.cpp, a lightweight inference engine written in C++. See https://nitro.jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.",
+ "url": "extension://@janhq/inference-cortex-extension/dist/index.js"
+ }
+}
+```
+
+## Common Use Cases
+
+1. **Disable an Extension**
+ ```json
+ {
+ "@janhq/example-extension": {
+ "_active": false
+ // other settings...
+ }
+ }
+ ```
+
+2. **Update Entry Point**
+ ```json
+ {
+ "@janhq/example-extension": {
+ "main": "new/path/index.js"
+ // other settings...
+ }
+ }
+ ```
+
+## Important Notes
+
+- Always backup the `extensions.json` file before making changes
+- Invalid JSON syntax can prevent Jan from starting properly
+- Most users should manage extensions through Jan's Settings UI
+- Changes require restarting Jan to take effect
diff --git a/docs/src/pages/docs/extensions-settings/_meta.json b/docs/src/pages/docs/extensions-settings/_meta.json
new file mode 100644
index 000000000..100d720dd
--- /dev/null
+++ b/docs/src/pages/docs/extensions-settings/_meta.json
@@ -0,0 +1,11 @@
+{
+ "model-management": {
+ "title": "Model Management",
+ "href": "/docs/extensions-settings/model-management"
+ },
+ "system-monitoring": {
+ "title": "System Monitoring",
+ "href": "/docs/extensions-settings/system-monitoring"
+ }
+}
+
\ No newline at end of file
diff --git a/docs/src/pages/docs/extensions-settings/model-management.mdx b/docs/src/pages/docs/extensions-settings/model-management.mdx
new file mode 100644
index 000000000..fb7ae6912
--- /dev/null
+++ b/docs/src/pages/docs/extensions-settings/model-management.mdx
@@ -0,0 +1,43 @@
+---
+title: Model Management
+description: Learn about Jan's default extensions and explore how to configure them.
+keywords:
+ [
+ Jan,
+ Customizable Intelligence, LLM,
+ local AI,
+ privacy focus,
+ free and open source,
+ private and offline,
+ conversational AI,
+ no-subscription fee,
+ large language models,
+ Jan Extensions,
+ Extensions,
+ ]
+---
+
+import { Callout } from 'nextra/components'
+import { Settings, EllipsisVertical, Plus, FolderOpen, Pencil } from 'lucide-react'
+
+# Model Management
+
+This extension configures how Jan handles model downloads and management:
+- Model discovery and browsing
+- Version control & configuration handling
+- Download management
+
+## Hugging Face Access Token
+
+Access tokens authenticate your identity to Hugging Face Hub for model downloads.
+1. Get your token from [Hugging Face Tokens](https://huggingface.co/docs/hub/en/security-tokens)
+2. Navigate to **Settings** () > **Core Extensions** > **Model Management**
+2. Enter your token in Jan: `hf_************************`
+
+
+Keep your access tokens secure and never share them.
+
+
+
+
+
\ No newline at end of file
diff --git a/docs/src/pages/docs/extensions-settings/system-monitoring.mdx b/docs/src/pages/docs/extensions-settings/system-monitoring.mdx
new file mode 100644
index 000000000..f506842e6
--- /dev/null
+++ b/docs/src/pages/docs/extensions-settings/system-monitoring.mdx
@@ -0,0 +1,54 @@
+---
+title: System Monitoring
+description: Learn about Jan's default extensions and explore how to configure them.
+keywords:
+ [
+ Jan,
+ Customizable Intelligence, LLM,
+ local AI,
+ privacy focus,
+ free and open source,
+ private and offline,
+ conversational AI,
+ no-subscription fee,
+ large language models,
+ Jan Extensions,
+ Extensions,
+ ]
+---
+import { Callout } from 'nextra/components'
+import { Settings, EllipsisVertical } from 'lucide-react'
+
+
+# System Monitoring
+Provides system health and OS level data:
+- Hardware utilization tracking
+- Performance monitoring
+- Error logging
+
+You can configure your logs in Jan in **Settings** () > **Core Extensions** > **System Monitoring**:
+
+
+## Enable App Logs
+Jan can save logs locally on your computer for:
+- Debugging model issues
+- Crash reports
+- Download troubleshooting
+
+To enable, toggle on **Enable App Logs**:
+
+
+
+
+
+
+## Log Cleaning Interval
+Set automatic log deletion interval in milliseconds:
+- Default: 120000 (2 minutes); however, there's no minimum or maximum intervals
+- Controls disk space usage
+- Prevents log accumulation
+
+
+You can clear Jan logs manually with [Clear logs](/docs/settings#log-management) in **Privacy**.
+
+
diff --git a/docs/src/pages/docs/extensions.mdx b/docs/src/pages/docs/extensions.mdx
index 8e7c23243..6fcfb7411 100644
--- a/docs/src/pages/docs/extensions.mdx
+++ b/docs/src/pages/docs/extensions.mdx
@@ -17,28 +17,17 @@ description: Learn about Jan's default extensions and explore how to configure t
---
import { Callout } from 'nextra/components'
+import { Settings, EllipsisVertical } from 'lucide-react'
-# Extensions Overview
+# Extensions
-The current Jan Desktop Client has some default extensions built on this framework to enhance the user experience. There are two types of extensions:
-- Model Providers: These are extensions that enable Jan to support cloud-based models and provide the engine to support additional local models.
-- Core Extensions: These are built-in tools for managing and monitoring the Jan's system.
+## Overview
+Extensions are modular components that add functionality to Jan. Each extension is designed to handle specific features.
-## List of Model Provider Extensions
-| Extension Name | Version | Description |
-|----------------|---------|-------------|
-| Anthropic | v1.0.2 | This extension enables Anthropic chat completion API calls |
-| Cohere | v1.0.0 | This extension enables Cohere chat completion API calls |
-| Groq | v1.0.1 | This extension enables fast Groq chat completion API calls |
-| Martian | v1.0.1 | This extension enables Martian chat completion API calls |
-| MistralAI | v1.0.1 | This extension enables Mistral chat completion API calls |
-| TensorRT-LLM | v0.0.3 | This extension enables Nvidia's TensorRT-LLM for the fastest GPU acceleration. See the setup guide for next steps |
-| NVIDIA NIM | v1.0.1 | This extension enables NVIDIA chat completion API calls |
-| OpenAI | v1.0.2 | This extension enables OpenAI chat completion API calls |
-| OpenRouter | v1.0.0 | This extension enables Open Router chat completion API calls |
-| Triton-TRT-LLM | v1.0.0 | This extension enables Nvidia's Triton-TRT-LLM as an inference engine option |
+Extensions can be managed through **Settings** () > **Extensions**:
+<<<<<<< HEAD
## List of Core Extensions
| Extension Name | Version | Description |
|---------------------|----------|-----------------------------------------------------------------------------|
@@ -303,60 +292,51 @@ To configure an extension settings:
Jan offers an Extensions settings menu for configuring extensions that have registered their settings within the application. Here, you can directly integrate Remote Inference Engines with Jan without inserting the URL and API Key directly in the `JSON` file. Additionally, you can turn the Logging extensions available on or off in Jan. To access the Extension settings, follow the steps below:
1. Navigate to the main dashboard.
2. Click the **gear icon (⚙️)** on the bottom left of your screen.
+=======
+>>>>>>> e5b6901b57c5b7edcd5895316f84d742a2772202
-
+
-3. Click **Extensions**.
-
-
-## Turn Off an Extension
+## Core Extensions
-To turn off the extension, follow the steps below:
+### Cortex
+The primary extension that manages both **local** and **remote engines** capabilities:
-1. Click the **Gear Icon (⚙️)** on the bottom left of your screen.
-
-
-
-2. Click the **Extensions** button.
-
-
-
-3. Click the slider button to turn off the extension.
+#### Local Engines
+[llama.cpp](/docs/local-engines/llama-cpp): Fast, efficient local inference engine that runs GGUF models directly on your device. Powers Jan's default local AI capabilities with support for multiple hardware configurations.
-4. Restart the app to see that the extension has been disabled.
+#### Remote Engines
+- [Anthropic](/docs/remote-models/anthropic): Access Claude models
+- [Cohere](/docs/remote-models/cohere): Access Cohere's models
+- [Groq](/docs/remote-models/groq): High-performance inference
+- [Martian](/docs/remote-models/martian): Specialized model access
+- [MistralAI](/docs/remote-models/mistralai): Access Mistral models
+- [NVIDIA NIM](/docs/remote-models/nvidia-nim) (NVIDIA Inference Microservices): Platform for deploying and serving GPU-accelerated AI models, providing enterprise-grade reliability and scalability.
+- [OpenAI](/docs/remote-models/openai): Access GPT models
+- [OpenRouter](/docs/remote-models/openrouter): Multi-provider model access
+- [Triton-TRT-LLM](/docs/remote-models/triton): High-performance inference backend using NVIDIA Triton Inference Server with TensorRT-LLM optimization, designed for large-scale model deployment.
-## Model Management
-The Model Management extension allows Jan's app to download specific models from Hugging Face repositories, which require an access token for access.
-1. Navigate to the main dashboard.
-2. Click the **Gear Icon (⚙️)** on the bottom left of your screen.
-
-
-
-3. Under the **Core Extensions** section, select the **Model Management** extension.
-4. Enter the HuggingFace access token.
-## System Monitor
-The System Monitor extension now offers enhanced customization for app logging. Users can toggle the application logging feature on or off and set a custom interval for clearing the app logs. To configure the app log feature, follow these steps:
-1. Navigate to the main dashboard.
-2. Click the **Gear Icon (⚙️)** on the bottom left of your screen.
-
-
-
-3. Under the **Core Extensions** section, select the **System Monitoring** extension.
-
-
-
-4. Use the **slider** to turn the app logging feature on or off.
-
-
-
-5. Specify the log cleaning interval in milliseconds.
-
-
-
-
- - You can clear the app logs manually by clicking the **Clear logs** button in the advanced settings.
- - There are no minimum or maximum intervals for setting the time. However, invalid inputs will default to `120000ms (2 minutes)`.
-
\ No newline at end of file
+### Jan Assistant
+Enables assistants functionality, including Jan - the default assistant that can utilize all downloaded models. This extension manages:
+- Default assistant configurations
+- Model selection
+- Conversation management
+
+### Conversational
+Manages all chat-related functionality and data persistence:
+
+
+### [Model Management](/docs/extensions-settings/model-management)
+Provides model exploration and seamless downloads:
+- Model discovery and browsing
+- Version control & configuration handling
+- Download management
+
+### [System Monitoring](/docs/extensions-settings/system-monitoring)
+Provides system health and OS level data:
+- Hardware utilization tracking
+- Performance monitoring
+- Error logging
diff --git a/docs/src/pages/docs/install-engines.mdx b/docs/src/pages/docs/install-engines.mdx
new file mode 100644
index 000000000..4e99d337d
--- /dev/null
+++ b/docs/src/pages/docs/install-engines.mdx
@@ -0,0 +1,193 @@
+---
+title: Install Engines
+description: Learn about Jan's default extensions and explore how to configure them.
+ [
+ Jan,
+ Customizable Intelligence, LLM,
+ local AI,
+ privacy focus,
+ free and open source,
+ private and offline,
+ conversational AI,
+ no-subscription fee,
+ large language models,
+ Jan Extensions,
+ Extensions,
+ ]
+---
+
+import { Callout } from 'nextra/components'
+import { Settings, EllipsisVertical } from 'lucide-react'
+
+# Install Engines
+
+## Install Local Engines
+Jan currently doesn't support installing a local engine yet.
+
+## Install Remote Engines
+
+### Step-by-step Guide
+You can add any OpenAI API-compatible providers like OpenAI, Anthropic, or others.
+To add a new remote engine:
+
+1. Navigate to **Settings** () > **Engines**
+1. At **Remote Engine** category, click **+ Install Engine**
+
+
+
+
+
+2. Fill in the following required information:
+
+
+
+
+
+| Field | Description | Required |
+|-------|-------------|----------|
+| Engine Name | Name for your engine (e.g., "OpenAI", "Claude") | ✓ |
+| API URL | The base URL of the provider's API | ✓ |
+| API Key | Your authentication key from the provider | ✓ |
+| Model List URL | URL for fetching available models | |
+| API Key Template | Custom authorization header format | |
+| Request Format Conversion | Function to convert Jan's request format to provider's format | |
+| Response Format Conversion | Function to convert provider's response format to Jan's format | |
+
+
+> - The conversion functions are only needed for providers that don't follow the OpenAI API format. For OpenAI-compatible APIs, you can leave these empty.
+> - For OpenAI-compatible APIs like OpenAI, Anthropic, or Groq, you only need to fill in the required fields. Leave optional fields empty.
+
+4. Click **Install**
+5. Once completed, you should see your engine in **Engines** page:
+ - You can rename or uninstall your engine
+ - You can navigate to its own settings page
+
+
+
+
+
+### Examples
+#### OpenAI-Compatible Setup
+Here's how to set up OpenAI as a remote engine:
+
+1. Engine Name: `OpenAI`
+2. API URL: `https://api.openai.com`
+3. Model List URL: `https://api.openai.com/v1/models`
+4. API Key: Your OpenAI API key
+5. Leave other fields as default
+
+
+#### Custom APIs Setup
+If you're integrating an API that doesn't follow OpenAI's format, you'll need to use the conversion functions.
+Let's say you have a custom API with this format:
+
+```javascript
+// Custom API Request Format
+{
+ "prompt": "What is AI?",
+ "max_length": 100,
+ "temperature": 0.7
+}
+
+// Custom API Response Format
+{
+ "generated_text": "AI is...",
+ "tokens_used": 50,
+ "status": "success"
+}
+```
+
+**Here's how to set it up in Jan:**
+```
+Engine Name: Custom LLM
+API Key: your_api_key_here
+Header template: your header template here
+Transform request template: your transform request template here
+Transform response template: your transform response template here
+```
+
+1. Header template
+```
+"Authorization: Bearer {{api_key}}"
+```
+2. Transform request template:
+Convert from Jan's OpenAI-style format to your API's format
+```
+"chat_completions": {
+ "url": "https://api.custom_endpoint.com/v1/messages",
+ "template": "{
+{% for key, value in input_request %}
+ {% if key == "messages" %}
+ "prompt": "{{ last(input_request.messages).content }}"
+ {% else if key == "max_tokens" or key == "temperature" %}
+ "{{ key }}": {{ tojson(value) }}
+ {% endif %}
+{% endfor %}
+}"
+}
+```
+
+
+3. Transform response template
+Convert from your API's format back to OpenAI-style format
+```
+"chat_completions": {
+"template": "{
+{
+ "choices": [{
+ "message": {
+ "role": "assistant",
+ "content": "{{ input_request.generated_text }}"
+ }
+ }],
+ "usage": {
+ "total_tokens": {{ input_request.tokens_used }}
+ }
+}
+}"
+}
+```
+
+
+
+**Expected Formats:**
+
+1. Jan's Request Format
+```
+{
+ "messages": [
+ {"role": "user", "content": "What is AI?"}
+ ],
+ "max_tokens": 100,
+ "temperature": 0.7
+}
+```
+
+2. Jan's Expected Response Format
+```
+{
+ "choices": [{
+ "message": {
+ "role": "assistant",
+ "content": "AI is..."
+ }
+ }],
+ "usage": {
+ "total_tokens": 50
+ }
+}
+```
+
+
+Make sure to test your conversion functions thoroughly. Incorrect conversions may cause errors or unexpected behavior.
+
+
+
+
+
+
+
+
+
+
+
diff --git a/docs/src/pages/docs/install-extensions.mdx b/docs/src/pages/docs/install-extensions.mdx
new file mode 100644
index 000000000..1e8078034
--- /dev/null
+++ b/docs/src/pages/docs/install-extensions.mdx
@@ -0,0 +1,89 @@
+---
+title: Install Extensions
+description: A step-by-step guide on installing an extension.
+keywords:
+ [
+ Jan,
+ Customizable Intelligence, LLM,
+ local AI,
+ privacy focus,
+ free and open source,
+ private and offline,
+ conversational AI,
+ no-subscription fee,
+ large language models,
+ Jan Extensions,
+ Extensions,
+ ]
+---
+
+import { Callout } from 'nextra/components'
+
+
+# Install Extensions
+
+Jan comes with several [pre-installed extensions](/docs/extensions#core-extensions) that provide core functionalities. You can manually add custom third-party extensions at your own risk.
+
+## Create Extensions
+
+
+Jan currently only accepts `.tgz` file format for extensions.
+
+
+> **Heads Up:**
+> - Please use the following structure and setup as a **reference** only.
+> - You're free to develop extensions using any approach or structure that works for your needs. As long as your extension can be packaged as a `.tgz` file, it can be installed in Jan. Feel free to experiment and innovate!
+> - If you already have your own `.tgz` extension file, please move forward to [install extension](/docs/install-extensions#install-extensions) step.
+
+#### Extension Structure
+Your extension should follow this basic structure:
+
+```
+my-extension/
+├── package.json # Extension metadata and dependencies
+├── dist/ # Compiled JavaScript files
+│ └── index.js # Main extension entry point
+├── src/ # Source code
+│ └── index.ts # TypeScript source
+└── README.md # Extension documentation
+```
+
+#### Required package.json Fields
+```json
+{
+ "name": "@your-org/extension-name",
+ "version": "1.0.0",
+ "main": "dist/index.js",
+ "types": "dist/index.d.ts",
+ "jan": {
+ "type": "extension",
+ "displayName": "Your Extension Name",
+ "description": "Description of what your extension does"
+ },
+ "scripts": {
+ "build": "tsc",
+ "package": "npm pack"
+ },
+ "dependencies": {
+ // List your dependencies
+ },
+ "devDependencies": {
+ "typescript": "^5.0.0"
+ }
+}
+```
+
+#### Example Extension Template
+
+You can find a template for creating Jan extensions in our [example repository](https://github.com/janhq/extension-template).
+## Install Extensions
+
+To install a custom extension in Jan:
+
+1. Navigate to [Jan Data Folder](/docs/data-folder):
+2. Copy `.tgz` extension file into the extensions directory
+3. Extract the `.tgz` file into its own folder under the extensions directory
+4. Restart Jan
+
+After restart, the `~/jan/data/extensions/extensions.json` file will be updated automatically to include your new extension.
+
diff --git a/docs/src/pages/docs/installing-extension.mdx b/docs/src/pages/docs/installing-extension.mdx
deleted file mode 100644
index 6471e9135..000000000
--- a/docs/src/pages/docs/installing-extension.mdx
+++ /dev/null
@@ -1,46 +0,0 @@
----
-title: Installing an Extension
-description: A step-by-step guide on installing an extension.
-keywords:
- [
- Jan,
- Customizable Intelligence, LLM,
- local AI,
- privacy focus,
- free and open source,
- private and offline,
- conversational AI,
- no-subscription fee,
- large language models,
- Jan Extensions,
- Extensions,
- ]
----
-
-import { Callout } from 'nextra/components'
-
-
-## Installing an Extension
-
-Jan automatically installs 11 default extensions when you set up the application. You can also add a custom third-party extension at your own risk.
-
-Here are the steps to install a custom extension:
-
-Jan only accepts the `.tgz` file format for installing a custom extension.
-
-
-1. Click the **Gear Icon (⚙️)** on the bottom left of your screen.
-
-
-
-2. Click the **Extensions** button.
-
-
-
-2. Select **Install Extension** on top right corner.
-
-
-
-3. Select a `.tgz` extension file.
-4. Restart the Jan application.
-5. Then, the `~/jan/data/extensions/extensions.json` file will be updated automatically.
\ No newline at end of file
diff --git a/docs/src/pages/docs/local-engines/_meta.json b/docs/src/pages/docs/local-engines/_meta.json
new file mode 100644
index 000000000..2e323af69
--- /dev/null
+++ b/docs/src/pages/docs/local-engines/_meta.json
@@ -0,0 +1,6 @@
+{
+ "llama-cpp": {
+ "title": "llama.cpp",
+ "href": "/docs/local-engines/llama-cpp"
+ }
+}
diff --git a/docs/src/pages/docs/local-engines/llama-cpp.mdx b/docs/src/pages/docs/local-engines/llama-cpp.mdx
new file mode 100644
index 000000000..df11c48ca
--- /dev/null
+++ b/docs/src/pages/docs/local-engines/llama-cpp.mdx
@@ -0,0 +1,171 @@
+---
+title: llama.cpp
+description: A step-by-step guide on how to customize the llama.cpp engine.
+keywords:
+ [
+ Jan,
+ Customizable Intelligence, LLM,
+ local AI,
+ privacy focus,
+ free and open source,
+ private and offline,
+ conversational AI,
+ no-subscription fee,
+ large language models,
+ Llama CPP integration,
+ llama.cpp Engine,
+ Intel CPU,
+ AMD CPU,
+ NVIDIA GPU,
+ AMD GPU Radeon,
+ Apple Silicon,
+ Intel Arc GPU,
+ ]
+---
+
+import { Tabs } from 'nextra/components'
+import { Callout, Steps } from 'nextra/components'
+import { Settings, EllipsisVertical, Plus, FolderOpen, Pencil } from 'lucide-react'
+
+# llama.cpp (Cortex)
+
+## Overview
+Jan uses **llama.cpp** for running local AI models. You can find its settings in **Settings** () > **Local Engine** > **llama.cpp**:
+
+
+
+
+
+These settings are for advanced users, you would want to check these settings when:
+- Your AI models are running slowly or not working
+- You've installed new hardware (like a graphics card)
+- You want to tinker & test performance with different [backends](/docs/local-engines/llama-cpp#available-backends)
+
+## Engine Version and Updates
+- **Engine Version**: View current version of llama.cpp engine
+- **Check Updates**: Verify if a newer version is available & install available updates when it's available
+
+
+## Available Backends
+
+Jan offers different backend variants for **llama.cpp** based on your operating system, you can:
+- Download different backends as needed
+- Switch between backends for different hardware configurations
+- View currently installed backends in the list
+
+
+Choose the backend that matches your hardware. Using the wrong variant may cause performance issues or prevent models from loading.
+
+
+
+
+
+### CUDA Support (NVIDIA GPUs)
+- `llama.cpp-avx-cuda-11-7`
+- `llama.cpp-avx-cuda-12-0`
+- `llama.cpp-avx2-cuda-11-7`
+- `llama.cpp-avx2-cuda-12-0`
+- `llama.cpp-avx512-cuda-11-7`
+- `llama.cpp-avx512-cuda-12-0`
+- `llama.cpp-noavx-cuda-11-7`
+- `llama.cpp-noavx-cuda-12-0`
+
+### CPU Only
+- `llama.cpp-avx`
+- `llama.cpp-avx2`
+- `llama.cpp-avx512`
+- `llama.cpp-noavx`
+
+### Other Accelerators
+- `llama.cpp-vulkan`
+
+
+- For detailed hardware compatibility, please visit our guide for [Windows](/docs/desktop/windows#compatibility).
+- AVX, AVX2, and AVX-512 are CPU instruction sets. For best performance, use the most advanced instruction set your CPU supports.
+- CUDA versions should match your installed NVIDIA drivers.
+
+
+
+
+
+### CUDA Support (NVIDIA GPUs)
+- `llama.cpp-avx-cuda-11-7`
+- `llama.cpp-avx-cuda-12-0`
+- `llama.cpp-avx2-cuda-11-7`
+- `llama.cpp-avx2-cuda-12-0`
+- `llama.cpp-avx512-cuda-11-7`
+- `llama.cpp-avx512-cuda-12-0`
+- `llama.cpp-noavx-cuda-11-7`
+- `llama.cpp-noavx-cuda-12-0`
+
+### CPU Only
+- `llama.cpp-avx`
+- `llama.cpp-avx2`
+- `llama.cpp-avx512`
+- `llama.cpp-noavx`
+
+### Other Accelerators
+- `llama.cpp-vulkan`
+- `llama.cpp-arm64`
+
+
+- For detailed hardware compatibility, please visit our guide for [Linux](docs/desktop/linux).
+- AVX, AVX2, and AVX-512 are CPU instruction sets. For best performance, use the most advanced instruction set your CPU supports.
+- CUDA versions should match your installed NVIDIA drivers.
+
+
+
+
+
+### Apple Silicon
+- `llama.cpp-mac-arm64`: For M1/M2/M3 Macs
+
+### Intel
+- `llama.cpp-mac-amd64`: For Intel-based Macs
+
+
+For detailed hardware compatibility, please visit our guide for [Mac](/docs/desktop/mac#compatibility).
+
+
+
+
+
+
+
+## Performance Settings
+
+
+| Setting | Description | Default |
+|---------|-------------|---------|
+| **Continuous Batching** | - Enables processing multiple requests simultaneously
- Improves throughput for concurrent operations | Enabled |
+| **Parallel Operations** | - Number of prompts to run in parallel
- Affects model inference speed | 4 |
+| **CPU Threads** | - Number of CPU cores to use when running without GPU
- Higher thread counts may improve performance but increase CPU usage | Auto-detected based on your system's capabilities |
+
+
+
+## Memory Settings
+
+| Setting | Description | Default |
+|---------|-------------|---------|
+| **Flash Attention** | - Optimizes attention computation
- Reduces memory usage
- Recommended for most cases | Enabled |
+| **Caching** | - Enable to store recent prompts and responses
- Improves response time for repeated prompts | Enabled |
+| **KV Cache Type** | - KV cache implementation type; controls memory usage and precision trade-off
- Options:
• f16 (most stable)
• q8_0 (balanced)
• q4_0 (lowest memory) | f16 |
+| **mmap** | - Enables memory-mapped model loading
- Reduces memory usage
- Recommended for large models | Enabled |
+
+
+## Best Practices
+- Start with default settings
+- Adjust based on your hardware capabilities
+- Monitor system performance
+- Test changes with your specific use case
+
+
+Performance impact varies by hardware, model size, and usage patterns.
+
+
+
+
+
+
+
+
diff --git a/docs/src/pages/docs/models/manage-models.mdx b/docs/src/pages/docs/models/manage-models.mdx
index dbbfa5d2d..01f0fc06f 100644
--- a/docs/src/pages/docs/models/manage-models.mdx
+++ b/docs/src/pages/docs/models/manage-models.mdx
@@ -17,7 +17,7 @@ keywords:
thread history,
]
---
-import { Callout, Steps } from 'nextra/components'
+import { Callout, Steps } from 'nextra/components'
import { Settings, EllipsisVertical, Plus, FolderOpen, Pencil } from 'lucide-react'
@@ -108,10 +108,13 @@ You need to own your **model configurations**, use at your own risk. Misconfigur
#### 4. Manual Setup
For advanced users who add a specific model that is not available within Jan **Hub**:
-1. Navigate to `~/jan/data/models/`
-2. Create a new **Folder** for your model
-3. Add a `model.json` file with your configuration:
-```
+
+##### Step 1: Create Model File
+1. Navigate to [Jan Data Folder]()
+2. Open `models` folder
+3. Create a new **Folder** for your model
+4. Add a `model.json` file with your configuration:
+```bash
"id": "",
"object": "",
"name": "",
@@ -130,9 +133,50 @@ For advanced users who add a specific model that is not available within Jan **H
"engine": "",
"source": ""
```
-Key fields to configure:
+Here's model "TinyLlama Chat 1.1B Q4" as an example:
+```json
+{
+ "sources": [
+ {
+ "filename": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
+ "url": "https://huggingface.co/TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF/resolve/main/tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf"
+ }
+ ],
+ "id": "tinyllama-1.1b",
+ "object": "model",
+ "name": "TinyLlama Chat 1.1B Q4",
+ "version": "1.0",
+ "description": "TinyLlama is a tiny model with only 1.1B. It's a good model for less powerful computers.",
+ "format": "gguf",
+ "settings": {
+ "ctx_len": 4096,
+ "prompt_template": "<|system|>\n{system_message}<|user|>\n{prompt}<|assistant|>",
+ "llama_model_path": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf"
+ },
+ "parameters": {
+ "temperature": 0.7,
+ "top_p": 0.95,
+ "stream": true,
+ "max_tokens": 2048,
+ "stop": [],
+ "frequency_penalty": 0,
+ "presence_penalty": 0
+ },
+ "metadata": {
+ "author": "TinyLlama",
+ "tags": [
+ "Tiny",
+ "Foundation Model"
+ ],
+ "size": 669000000
+ },
+ "engine": "nitro"
+}
+```
+##### Step 2: Modify Model Parameters
+Modify model parameters under the settings array. Key fields to configure:
1. **Settings** is where you can set your engine configurations.
-2. [**Parameters**](/docs/models#model-parameters) are the adjustable settings that affect how your model operates or processes the data. The fields in parameters are typically general and can be the same across models. Here is an example of model parameters:
+2. [**Parameters**](/docs/models/model-parameters) are the adjustable settings that affect how your model operates or processes the data. The fields in parameters are typically general and can be the same across models. Here is an example of model parameters:
```
"parameters":{
"temperature": 0.7,
@@ -140,9 +184,10 @@ Key fields to configure:
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
- "presence_penalty": 0
+ "presence_penalty": 0,
+}
```
-
+
### Delete Models
1. Go to **Settings** () > **My Models**
@@ -160,8 +205,8 @@ When using cloud models, be aware of any associated costs and rate limits from t
Jan supports connecting to various AI cloud providers that are OpenAI API-compatible, including: OpenAI (GPT-4, o1,...), Anthropic (Claude), Groq, Mistral, and more.
-1. Open **Settings**
-2. Under **Model Provider** section in left sidebar (OpenAI, Anthropic, etc.), choose a provider
+1. Navigate to **Settings** ()
+2. Under **Remote Engines** section in the left sidebar, choose your preferred engines (OpenAI, Anthropic, etc.)
3. Enter your API key
4. The activated cloud models will be available in your model selector in **Threads**
diff --git a/docs/src/pages/docs/quickstart.mdx b/docs/src/pages/docs/quickstart.mdx
index 44ee2df92..62a923b51 100644
--- a/docs/src/pages/docs/quickstart.mdx
+++ b/docs/src/pages/docs/quickstart.mdx
@@ -111,10 +111,9 @@ Jan supports both local and cloud AI models. You can connect to cloud AI service
See [Remote APIs](/docs/remote-models/openai) for detailed configuration.
-

-
+
## What's Next?
diff --git a/docs/src/pages/docs/remote-models/_meta.json b/docs/src/pages/docs/remote-models/_meta.json
index 97ad17bad..529047774 100644
--- a/docs/src/pages/docs/remote-models/_meta.json
+++ b/docs/src/pages/docs/remote-models/_meta.json
@@ -1,38 +1,39 @@
{
- "openai": {
- "title": "OpenAI API",
- "href": "/docs/remote-models/openai"
- },
- "azure": {
- "title": "Azure OpenAI API",
- "href": "/docs/remote-models/azure",
- "display": "hidden"
- },
- "groq": { "title": "Groq API", "href": "/docs/remote-models/groq" },
- "mistralai": {
- "title": "Mistral AI API",
- "href": "/docs/remote-models/mistralai"
- },
- "openrouter": { "title": "OpenRouter", "href": "/docs/remote-models/openrouter" },
- "generic-openai": { "title": "Any OpenAI Compatible API", "href": "/docs/remote-models/generic-openai", "display": "hidden"},
- "martian": {
- "title": "Martian API",
- "href": "/docs/remote-models/martian"
- },
- "cohere": {
- "title": "Cohere API",
- "href": "/docs/remote-models/cohere"
- },
"anthropic": {
- "title": "Anthropic API",
+ "title": "Anthropic",
"href": "/docs/remote-models/anthropic"
},
+ "cohere": {
+ "title": "Cohere",
+ "href": "/docs/remote-models/cohere"
+ },
+ "groq": {
+ "title": "Groq",
+ "href": "/docs/remote-models/groq"
+ },
+ "martian": {
+ "title": "Martian",
+ "href": "/docs/remote-models/martian"
+ },
+ "mistralai": {
+ "title": "Mistral AI",
+ "href": "/docs/remote-models/mistralai"
+ },
"nvidia-nim": {
- "title": "NVIDIA NIM API",
+ "title": "Nvidia NIM",
"href": "/docs/remote-models/nvidia-nim"
},
+ "openai": {
+ "title": "OpenAI",
+ "href": "/docs/remote-models/openai"
+ },
+ "openrouter": {
+ "title": "OpenRouter",
+ "href": "/docs/remote-models/openrouter"
+ },
"triton": {
"title": "Triton-TRT-LLM",
- "href": "/docs/remote-models/triton"
+ "href": "/docs/remote-models/triton",
+ "display": "hidden"
}
-}
+}
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/anthropic.mdx b/docs/src/pages/docs/remote-models/anthropic.mdx
index 33259bb3b..8b7fdbccd 100644
--- a/docs/src/pages/docs/remote-models/anthropic.mdx
+++ b/docs/src/pages/docs/remote-models/anthropic.mdx
@@ -16,46 +16,71 @@ keywords:
---
import { Callout, Steps } from 'nextra/components'
+import { Settings, Plus } from 'lucide-react'
# Anthropic
-## How to Integrate Anthropic with Jan
+Jan supports [Anthropic](https://anthropic.com/) API integration, allowing you to use Claude models (Claude 3, Claude 2.1, and more) through Jan's interface.
-This guide provides step-by-step instructions on integrating Anthropic with Jan, enabling users to chat with Claude's LLMs within Jan's conversational interface.
-
-Before proceeding, ensure you have the following:
-- Access to the Jan application
-- Anthropic API credentials
-
-## Integration Steps
+## Integrate Anthropic API with Jan
+### Step 1: Get Your API Key
+1. Visit [Anthropic Console](https://console.anthropic.com/settings/keys) and sign in
+2. Create & copy a new API key or copy your existing one
-### Step 1: Configure Anthropic API Key
-1. Obtain Anthropic API Keys from your [Anthropic Console](https://console.anthropic.com/).
-2. Copy your **Anthropic API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **Anthropic**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **Anthropic**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **Anthropic** section under Model Providers.
-4. Insert your **Anthropic API Key**.
-
-### Step 2: Start Chatting with the Model
-
-1. Select the Anthropic model you want to use.
-Anthropic is the default extension for the Jan application. All the Anthropic models are automatically installed when you install the Jan application.
+Ensure your API key has sufficient credits
-2. Specify the model's parameters.
-3. Start the conversation with the Anthropic model.
+### Step 2: Configure Jan
+There are two ways to add your Anthropic API keys in Jan:
+**Through Threads:**
+1. In Threads, click **Model** tab in the **right sidebar** or **model selector** in input field
+2. Once the selector is poped up, choose the **Cloud** tab
+3. Click **Add** () icon next to **Anthropic**
+4. Once you are directed to Anthropic settings, insert your **API Key**
+
+**Through Settings:**
+1. Navigate to **Settings** ()
+2. Under **Remote Engines**, select **Anthropic**
+3. Insert your **API Key**
+
+
+
+
+
+### Step 3: Start Using Anthropic's Models
+
+1. In any existing **Threads** or create a new one
+2. Select an Anthropic model from **model selector**
+3. Start chatting
+## Available Anthropic Models
+
+Jan automatically includes Anthropic's available models. In case you want to use a specific Anthropic model that you cannot find in **Jan**, follow instructions in [Manual Setup](/docs/models/manage-models#4-manual-setup) to add custom models:
+ - See list of available models in [Anthropic Models](https://docs.anthropic.com/claude/docs/models-overview).
+ - The `id` property must match the model name in the list. For example, `claude-3-opus-20240229`, `claude-3-sonnet-20240229`, or `claude-2.1`.
+
## Troubleshooting
-If you encounter any issues during the integration process or while using Anthropic with Jan, consider the following troubleshooting steps:
+Common issues and solutions:
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to Anthropic API support for assistance if needed.
\ No newline at end of file
+**1. API Key Issues**
+- Verify your API key is correct and not expired
+- Check if you have billing set up on your Anthropic account
+- Ensure you have access to the model you're trying to use
+
+**2. Connection Problems**
+- Check your internet connection
+- Verify Anthropic's system status
+- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
+
+**3. Model Unavailable**
+- Confirm your API key has access to the model
+- Check if you're using the correct model ID
+- Verify your Anthropic account has the necessary permissions
+
+Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the [Anthropic documentation](https://docs.anthropic.com/claude/docs).
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/azure.mdx b/docs/src/pages/docs/remote-models/azure.mdx
deleted file mode 100644
index af49fe89b..000000000
--- a/docs/src/pages/docs/remote-models/azure.mdx
+++ /dev/null
@@ -1,58 +0,0 @@
----
-title: Azure OpenAI
-description: A step-by-step guide on integrating Jan with Azure OpenAI.
-keywords:
- [
- Jan,
- Customizable Intelligence, LLM,
- local AI,
- privacy focus,
- free and open source,
- private and offline,
- conversational AI,
- no-subscription fee,
- large language models,
- integration,
- Azure OpenAI Service,
- ]
----
-
-import { Callout, Steps } from 'nextra/components'
-
-
-
-# Azure OpenAI API
-
-## How to Integrate Azure OpenAI API with Jan
-This guide provides step-by-step instructions for integrating the Azure OpenAI API with Jan, allowing users to utilize Azure's capabilities within Jan's conversational interface.
-
-## Integration Steps
-
-### Step 1: Configure OpenAI API Key
-1. Obtain OpenAI API Key from your [OpenAI Platform](https://platform.openai.com/api-keys) dashboard.
-2. Copy your **OpenAI API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **OpenAI**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **OpenAI**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **OpenAI** section under Model Providers.
-4. Insert your **OpenAI API Key**.
-
-
-The **OpenAI** fields can be used for any OpenAI-compatible API.
-
-
-### Step 2: Start Chatting with the Model
-
-1. Select the OpenAI model you want to use.
-2. Specify the model's parameters.
-3. Start the conversation with the OpenAI model.
-
-
-
-## Troubleshooting
-
-If you encounter any issues during the integration process or while using OpenAI with Jan, consider the following troubleshooting steps:
-
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to Azure OpenAI API support for assistance if needed.
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/cohere.mdx b/docs/src/pages/docs/remote-models/cohere.mdx
index 8f774a0de..41e60431d 100644
--- a/docs/src/pages/docs/remote-models/cohere.mdx
+++ b/docs/src/pages/docs/remote-models/cohere.mdx
@@ -16,46 +16,71 @@ keywords:
---
import { Callout, Steps } from 'nextra/components'
+import { Settings, Plus } from 'lucide-react'
# Cohere
-## How to Integrate Cohere with Jan
+Jan supports [Cohere](https://cohere.com/) API integration, allowing you to use Cohere's models (Command, Command-R and more) through Jan's interface.
-This guide provides step-by-step instructions on integrating Cohere with Jan, enabling users to chat with Cohere's LLMs within Jan's conversational interface.
-
-Before proceeding, ensure you have the following:
-- Access to the Jan application
-- Cohere API credentials
-
-## Integration Steps
+## Integrate Cohere API with Jan
+### Step 1: Get Your API Key
+1. Visit [Cohere Dashboard](https://dashboard.cohere.com/api-keys) and sign in
+2. Create & copy a new API key or copy your existing one
-### Step 1: Configure Cohere API Key
-1. Obtain Cohere API Keys from your [Cohere Dashboard](https://dashboard.cohere.com/).
-2. Copy your **Cohere API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **Cohere**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **Cohere**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **Cohere** section under Model Providers.
-4. Insert your **Cohere API Key**.
-
-
-### Step 2: Start Chatting with the Model
-
-1. Select the Cohere model you want to use.
-Cohere is the default extension for the Jan application. All the Cohere models are automatically installed when you install the Jan application.
+Ensure your API key has sufficient credits
-2. Specify the model's parameters.
-3. Start the conversation with the Cohere model.
+### Step 2: Configure Jan
+There are two ways to add your Cohere API keys in Jan:
+
+**Through Threads:**
+1. In Threads, click **Model** tab in the **right sidebar** or **model selector** in input field
+2. Once the selector is poped up, choose the **Cloud** tab
+3. Click **Add** () icon next to **Cohere**
+4. Once you are directed to Cohere settings, insert your **API Key**
+
+**Through Settings:**
+1. Navigate to **Settings** ()
+2. Under **Remote Engines**, select **Cohere**
+3. Insert your **API Key**
+
+
+
+
+
+### Step 3: Start Using Cohere's Models
+
+1. In any existing **Threads** or create a new one
+2. Select a Cohere model from **model selector**
+3. Start chatting
+## Available Cohere Models
+
+Jan automatically includes Cohere's available models. In case you want to use a specific Cohere model that you cannot find in **Jan**, follow instructions in [Manual Setup](/docs/models/manage-models#4-manual-setup) to add custom models:
+ - See list of available models in [Cohere Documentation](https://docs.cohere.com/v2/docs/models).
+ - The `id` property must match the model name in the list. For example, `command-nightly` or `command-light`.
+
## Troubleshooting
-If you encounter any issues during the integration process or while using Cohere with Jan, consider the following troubleshooting steps:
+Common issues and solutions:
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to Cohere API support for assistance if needed.
\ No newline at end of file
+**1. API Key Issues**
+- Verify your API key is correct and not expired
+- Check if you have billing set up on your Cohere account
+- Ensure you have access to the model you're trying to use
+
+**2. Connection Problems**
+- Check your internet connection
+- Verify Cohere's [system status](https://status.cohere.com/)
+- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
+
+**3. Model Unavailable**
+- Confirm your API key has access to the model
+- Check if you're using the correct model ID
+- Verify your Cohere account has the necessary permissions
+
+Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the [Cohere documentation](https://docs.cohere.com).
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/generic-openai.mdx b/docs/src/pages/docs/remote-models/generic-openai.mdx
deleted file mode 100644
index 22f483372..000000000
--- a/docs/src/pages/docs/remote-models/generic-openai.mdx
+++ /dev/null
@@ -1,69 +0,0 @@
----
-title: Any OpenAI Compatible API
-description: A step-by-step guide on how to set up Jan to connect with any remote or local API server.
-keywords:
- [
- Jan,
- Customizable Intelligence, LLM,
- local AI,
- privacy focus,
- free and open source,
- private and offline,
- conversational AI,
- no-subscription fee,
- large language models,
- import-models-manually,
- remote server,
- OAI compatible,
- ]
----
-
-import { Callout, Steps } from 'nextra/components'
-
-# Any OpenAI-compatible API
-This guide outlines the process for configuring Jan as a client for both remote and local API servers, using the `mistral-ins-7b-q4` model for illustration. We'll show how to connect to Jan's API-hosting servers.
-
-
- Currently, you can only connect to one OpenAI-compatible endpoint at a time.
-
-
-
-### Step 1: Configure a Client Connection
-
-1. Navigate to the **Jan app** > **Settings**.
-2. Select the **OpenAI**.
-
-
-The **OpenAI** fields can be used for any OpenAI-compatible API.
-
-
-3. Insert the **API Key** and the **endpoint URL** into their respective fields. For example, if you're going to communicate to Jan's API server, you can configure it as follows:
-```json
-"full_url": "https://:1337/v1/chat/completions"
-```
-
- Please note that currently, the code that supports any OpenAI-compatible endpoint only reads the `~/jan/data/extensions/@janhq/inference-openai-extension/settings.json` file, which is OpenAI Inference Engines in the extensions page. Thus, it will not search any other files in this directory.
-
-
-
-
-### Step 2: Start Chatting with the Model
-
-1. Navigate to the **Hub** section.
-2. Select the model you want to use.
-3. Specify the model's parameters.
-4. Start the conversation with the model.
-
-
-
-
- If you have questions or want more preconfigured GGUF models, please join our [Discord community](https://discord.gg/Dt7MxDyNNZ) for support, updates, and discussions.
-
-
-## Troubleshooting
-
-If you encounter any issues during the integration process or while using OpenAI with Jan, consider the following troubleshooting steps:
-
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to their API support for assistance if needed.
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/groq.mdx b/docs/src/pages/docs/remote-models/groq.mdx
index 32c192fc5..c0ccaf627 100644
--- a/docs/src/pages/docs/remote-models/groq.mdx
+++ b/docs/src/pages/docs/remote-models/groq.mdx
@@ -16,47 +16,72 @@ keywords:
---
import { Callout, Steps } from 'nextra/components'
+import { Settings, Plus } from 'lucide-react'
-# Groq API
+# Groq
-## How to Integrate Groq API with Jan
+Jan supports [Groq](https://groq.com/) API integration, allowing you to use Groq's high-performance LLM models (LLaMA 2, Mixtral and more) through Jan's interface.
-This guide provides step-by-step instructions on integrating the Groq API with Jan, enabling users to leverage Groq's capabilities within Jan's conversational interface.
-
-Before proceeding, ensure you have the following:
-- Access to the Jan application
-- Groq API credentials
-
-## Integration Steps
+## Integrate Groq API with Jan
-
-### Step 1: Configure Groq API Key
-1. Obtain Groq API Keys from your [Groq Console](https://console.groq.com/keys) dashboard.
-2. Copy your **Groq API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **Groq**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **Groq**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **Groq** section under Model Providers.
-4. Insert your **Groq API Key**.
-
-### Step 2: Start Chatting with the Model
-
-1. Select the Groq model you want to use.
+### Step 1: Get Your API Key
+1. Visit [Groq Console](https://console.groq.com/keys) and sign in
+2. Create & copy a new API key or copy your existing one
- The Groq Inference Engine is the default extension for the Jan application. All the Groq models are automatically installed when you install the Jan application.
+Ensure your API key has sufficient credits
-2. Specify the model's parameters.
-3. Start the conversation with the Groq model.
+
+### Step 2: Configure Jan
+There are two ways to add your Groq API keys in Jan:
+
+**Through Threads:**
+1. In Threads, click **Model** tab in the **right sidebar** or **model selector** in input field
+2. Once the selector is poped up, choose the **Cloud** tab
+3. Click **Add** () icon next to **Groq**
+4. Once you are directed to Groq settings, insert your **API Key**
+
+**Through Settings:**
+1. Navigate to **Settings** ()
+2. Under **Remote Engines**, select **Groq**
+3. Insert your **API Key**
+
+
+
+
+### Step 3: Start Using Groq's Models
+
+1. In any existing **Threads** or create a new one
+2. Select a Groq model from **model selector**
+3. Start chatting
+## Available Models Through Groq
+
+Jan automatically includes Groq's available models. In case you want to use a specific Groq model that you cannot find in **Jan**, follow instructions in [Manual Setup](/docs/models/manage-models#4-manual-setup) to add custom models:
+ - See list of available models in [Groq Documentation](https://console.groq.com/docs/models).
+ - The `id` property must match the model name in the list. For example, if you want to use Llama3.3 70B, you must set the `id` property to `llama-3.3-70b-versatile`.
+
## Troubleshooting
-If you encounter any issues during the integration process or while using Groq with Jan, consider the following troubleshooting steps:
+Common issues and solutions:
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to Groq API support for assistance if needed.
\ No newline at end of file
+**1. API Key Issues**
+- Verify your API key is correct and not expired
+- Check if you have billing set up on your Groq account
+- Ensure you have access to the model you're trying to use
+
+**2. Connection Problems**
+- Check your internet connection
+- Verify Groq's system status
+- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
+
+**3. Model Unavailable**
+- Confirm your API key has access to the model
+- Check if you're using the correct model ID
+- Verify your Groq account has the necessary permissions
+
+Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the [Groq documentation](https://console.groq.com/docs).
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/martian.mdx b/docs/src/pages/docs/remote-models/martian.mdx
index 03278709a..9b494419c 100644
--- a/docs/src/pages/docs/remote-models/martian.mdx
+++ b/docs/src/pages/docs/remote-models/martian.mdx
@@ -14,48 +14,71 @@ keywords:
API integration
]
---
-
import { Callout, Steps } from 'nextra/components'
+import { Settings, Plus } from 'lucide-react'
# Martian
-## How to Integrate Martian with Jan
+Jan supports [Martian](https://withmartian.com/) API integration, allowing you to use Martian's models through Jan's interface.
-This guide provides step-by-step instructions on integrating Martian with Jan, enabling users to leverage Martian's capabilities within Jan's conversational interface.
-
-Before proceeding, ensure you have the following:
-- Access to the Jan application
-- Martian API credentials
-
-## Integration Steps
+## Integrate Martian with Jan
+### Step 1: Get Your API Key
+1. Visit [Martian API Keys](https://www.withmartian.com/dashboard/undefined/api-keys) and sign in
+2. Create & copy a new API key or copy your existing one
-### Step 1: Configure Martian API Key
-1. Obtain Martian API Keys from your [Martian Dashboard](https://auth.withmartian.com/).
-2. Copy your **Martian API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **Martian**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **Martian**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **Martian** section under Model Providers.
-4. Insert your **Martian API Key**.
-
-
-### Step 2: Start Chatting with the Model
-
-1. Select the Martian model you want to use.
-Martian is the default extension for the Jan application. All the Martian models are automatically installed when you install the Jan application.
+Ensure your API key has sufficient credits
-2. Specify the model's parameters.
-3. Start the conversation with the Martian model.
+### Step 2: Configure Jan
+There are two ways to add your Martian key in Jan:
+
+**Through Threads:**
+1. In Threads, click **Model** tab in the **right sidebar** or **model selector** in input field
+2. Once the selector is poped up, choose the **Cloud** tab
+3. Click **Add** () icon next to **Martian**
+4. Once you are directed to Martian settings, insert your **API Key**
+
+**Through Settings:**
+1. Navigate to **Settings** ()
+2. Under **Remote Engines**, select **Martian**
+3. Insert your **API Key**
+
+
+
+
+
+### Step 3: Start Using Martian Models
+
+1. In any existing **Threads** or create a new one
+2. Select a Martian model from **model selector**
+3. Start chatting
+## Available Models
+
+Jan includes the Martian Model Router which automatically selects the best model for your use case. You can start using it right away after configuring your API key. See list of available models in [Martian Documentation](https://docs.withmartian.com/martian-model-router/getting-started/supported-models-gateway).
+
+
## Troubleshooting
-If you encounter any issues during the integration process or while using Martian with Jan, consider the following troubleshooting steps:
+Common issues and solutions:
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to Martian API support for assistance if needed.
\ No newline at end of file
+**1. API Key Issues**
+- Verify your API key is correct and not expired
+- Check if you have billing set up on your Martian account
+- Ensure you have access to the model you're trying to use
+
+**2. Connection Problems**
+- Check your internet connection
+- Verify Martian's system status
+- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
+
+**3. Model Unavailable**
+- Confirm your API key has access to the model
+- Check if you're using the correct model ID
+- Verify your Martian account has the necessary permissions
+
+Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the [Martian documentation](https://docs.withmartian.com/martian-model-router).
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/mistralai.mdx b/docs/src/pages/docs/remote-models/mistralai.mdx
index 5e153dbec..c1051a703 100644
--- a/docs/src/pages/docs/remote-models/mistralai.mdx
+++ b/docs/src/pages/docs/remote-models/mistralai.mdx
@@ -17,49 +17,71 @@ keywords:
---
import { Callout, Steps } from 'nextra/components'
+import { Settings, Plus } from 'lucide-react'
-# Mistral AI API
+# Mistral AI
-## How to Integrate Mistral AI with Jan
-This guide provides step-by-step instructions for integrating the Mistral API with Jan, enabling users to utilize Mistral's capabilities within Jan's conversational interface.
+Jan supports [Mistral AI](https://mistral.ai/) API integration, allowing you to use Mistral's powerful models (Mistral Large, Mistral Medium, Mistral Small and more) through Jan's interface.
-Before proceeding, ensure you have the following:
-- Access to the Jan Application
-- Mistral API credentials
+## Integrate Mistral AI with Jan
-## Integration Steps
-
-### Step 1: Configure Mistral API Key
-
-1. Obtain the Mistral API Key from your [Mistral](https://console.mistral.ai/user/api-keys/) dashboard.
-2. Copy your **Mistral API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **Mistral**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **Mistral**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **Mistral** section under Model Providers.
-4. Insert your **Mistral API Key**.
+### Step 1: Get Your API Key
+1. Visit [Mistral AI Platform](https://console.mistral.ai/api-keys/) and sign in
+2. Create & copy a new API key or copy your existing one
- - Mistral AI offers various endpoints. Refer to their [endpoint documentation](https://docs.mistral.ai/platform/endpoints/) to select the one that fits your requirements.
+Ensure your API key has sufficient credits
-### Step 2: Start Chatting with the Model
+### Step 2: Configure Jan
+There are two ways to add your Mistral AI keys in Jan:
-1. Select the Mistral model you want to use.
+**Through Threads:**
+1. In Threads, click **Model** tab in the **right sidebar** or **model selector** in input field
+2. Once the selector is poped up, choose the **Cloud** tab
+3. Click **Add** () icon next to **Mistral AI**
+4. Once you are directed to Mistral AI settings, insert your **API Key**
-
- The MistralAI is the default extension for the Jan application. All the Mistral models are automatically installed when you install the Jan application.
-
-2. Specify the model's parameters.
-3. Start the conversation with the Mistral model.
+**Through Settings:**
+1. Navigate to **Settings** ()
+2. Under **Remote Engines**, select **Mistral AI**
+3. Insert your **API Key**
+
+
+
+
+### Step 3: Start Using Mistral's Models
+
+1. In any existing **Threads** or create a new one
+2. Select a Mistral model from **model selector**
+3. Start chatting
+## Available Mistral Models
+
+Jan automatically includes Mistral's available models. In case you want to use a specific Mistral model that you cannot find in **Jan**, follow instructions in [Manual Setup](/docs/models/manage-models#4-manual-setup) to add custom models:
+ - See list of available models in [Mistral AI Documentation](https://docs.mistral.ai/platform/endpoints).
+ - The `id` property must match the model name in the list. For example, if you want to use Mistral Large, you must set the `id` property to `mistral-large-latest`
+
## Troubleshooting
-If you encounter any issues during the integration process or while using Mistral with Jan, consider the following troubleshooting steps:
+Common issues and solutions:
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to Mistral API support for assistance if needed.
\ No newline at end of file
+**1. API Key Issues**
+- Verify your API key is correct and not expired
+- Check if you have billing set up on your Mistral AI account
+- Ensure you have access to the model you're trying to use
+
+**2. Connection Problems**
+- Check your internet connection
+- Verify Mistral AI's system status
+- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
+
+**3. Model Unavailable**
+- Confirm your API key has access to the model
+- Check if you're using the correct model ID
+- Verify your Mistral AI account has the necessary permissions
+
+Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the [Mistral AI documentation](https://docs.mistral.ai/).
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/nvidia-nim.mdx b/docs/src/pages/docs/remote-models/nvidia-nim.mdx
index 2fafa96cd..f2e3b0d06 100644
--- a/docs/src/pages/docs/remote-models/nvidia-nim.mdx
+++ b/docs/src/pages/docs/remote-models/nvidia-nim.mdx
@@ -16,45 +16,76 @@ keywords:
---
import { Callout, Steps } from 'nextra/components'
+import { Settings, Plus } from 'lucide-react'
# NVIDIA NIM
-## How to Integrate NVIDIA NIM with Jan
-
-This guide provides step-by-step instructions on integrating NVIDIA NIM with Jan, enabling users to leverage Nvidia NIM's capabilities within Jan's conversational interface.
+Jan supports [NVIDIA NIM](https://www.nvidia.com/en-us/ai/) API integration, allowing you to use NVIDIA's Large Language Models through Jan's interface.
- Nvidia NIM extension is only supported on Jan version 0.5.1 or later.
+NVIDIA NIM extension is only supported on Jan version 0.5.1 or later.
-Before proceeding, ensure you have the following:
-- Access to the Jan application
-- NVIDIA NIM API credentials
+## Integrate Nvidia NIM API with Jan
-## Integration Steps
-### Step 1: Configure Nvidia API Key
-1. Obtain Nvidia API Keys from your [Nvidia dashboard](https://org.ngc.nvidia.com/setup/personal-keys).
-2. Copy your **Nvidia API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **Nvidia**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **Nvidia**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **NVIDIA NIM** section under Model Providers.
-4. Insert your **Nvidia API Key**.
+### Step 1: Get Your API Key
+1. Visit [NVIDIA Docs](https://docs.nvidia.com/nim/nemo-retriever/text-reranking/latest/getting-started.html#generate-an-api-key) and generate an API key
+2. Copy your API key
+
+Ensure your API key has sufficient credits
+
-### Step 2: Start Chatting with the Model
+### Step 2: Configure Jan
+There are two ways to add your Nvidia NIM API keys in Jan:
-1. Select the model you want to use.
-2. Specify the model's parameters.
-3. Start the conversation with the model.
+**Through Threads:**
+1. In Threads, click **Model** tab in the **right sidebar** or **model selector** in input field
+2. Once the selector is poped up, choose the **Cloud** tab
+3. Click **Add** () icon next to **Nvidia NIM**
+4. Once you are directed to Nvidia NIM settings, insert your **API Key**
+**Through Settings:**
+1. Navigate to **Settings** ()
+2. Under **Remote Engines**, select **Nvidia NIM**
+3. Insert your **API Key**
+
+
+
+
+
+### Step 3: Start Using Nvidia NIM Models
+
+1. In any existing **Threads** or create a new one
+2. Select a NVIDIA NIM model from **model selector**
+3. Start chatting
+## Available Nvidia NIM Models
+
+Jan automatically includes NVIDIA NIM's available models. In case you want to use a specific model that you cannot find in **Jan**, follow instructions in [Manual Setup](/docs/models/manage-models#4-manual-setup) to add custom models:
+ - See list of available models in [Nvidia NIM Documentation](https://build.nvidia.com/models).
+ - The `id` property must match the model name in the list.
+
## Troubleshooting
-If you encounter any issues during the integration process or while using Nvidia with Jan, consider the following troubleshooting steps:
+Common issues and solutions:
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to Nvidia API support for assistance if needed.
\ No newline at end of file
+**1. API Key Issues**
+- Verify your API key is correct and not expired
+- Check if you have billing set up on your NVIDIA account
+- Ensure you have access to the model you're trying to use
+
+**2. Connection Problems**
+- Check your internet connection
+- Verify NVIDIA's system status
+- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
+
+**3. Model Unavailable**
+- Confirm your API key has access to the model
+- Check if you're using the correct model ID
+- Verify your NVIDIA account has the necessary permissions
+- Make sure you're using Jan version 0.5.1 or later
+
+Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the [Nvidia documentation](https://docs.nvidia.com/nim/large-language-models/latest/getting-started.html).
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/openai.mdx b/docs/src/pages/docs/remote-models/openai.mdx
index cd7d0454e..502891f3d 100644
--- a/docs/src/pages/docs/remote-models/openai.mdx
+++ b/docs/src/pages/docs/remote-models/openai.mdx
@@ -16,57 +16,73 @@ keywords:
Azure OpenAI Service,
]
---
-
import { Callout, Steps } from 'nextra/components'
+import { Settings, Plus } from 'lucide-react'
+# OpenAI
+Jan supports [OpenAI](https://openai.com/) and OpenAI-compatible APIs, allowing you to use all models from OpenAI (GPT-4, GPT o1 and more) through Jan's interface.
-# OpenAI API
+## Integrate OpenAI API with Jan
-
-The OpenAI Extension can be used for any OpenAI-compatible API endpoints.
-
-
-## How to Integrate OpenAI API with Jan
-This guide provides step-by-step instructions for integrating the OpenAI API with Jan, allowing users to utilize OpenAI's capabilities within Jan's conversational interface.
-
-## Integration Steps
-### Step 1: Configure OpenAI API Key
-1. Obtain OpenAI API Key from your [OpenAI Platform](https://platform.openai.com/api-keys) dashboard.
-2. Copy your **OpenAI API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **OpenAI**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **OpenAI**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **OpenAI** section under Model Providers.
-4. Insert your **OpenAI API Key**.
-
-### Step 2: Start Chatting with the Model
-
-1. Select the OpenAI model you want to use.
+### Step 1: Get Your API Key
+1. Visit [OpenAI Platform](https://platform.openai.com/api-keys) and sign in
+2. Create & copy a new API key or copy your existing one
-The OpenAI is the default extension for the Jan application. All the OpenAI models are automatically installed when you install the Jan application.
+Ensure your API key has sufficient credits
-2. Specify the model's parameters.
-3. Start the conversation with the OpenAI model.
+### Step 2: Configure Jan
+There are two ways to add your OpenAI API keys in Jan:
+
+Through Threads:
+1. In Threads, click Model tab in the right sidebar or model selector in input field
+2. Once the selector is poped up, choose the Cloud tab
+3. Click Add () icon next to OpenAI
+4. Once you are directed to OpenAI settings, insert your API Key
+
+Through Settings:
+1. Navigate to Settings ()
+2. Under Remote Engines, select OpenAI
+3. Insert your API Key
+
+
+
+
+
+### Step 3: Start Using OpenAI's Models
+
+In any existing Threads or create a new one
+Select an OpenAI model from model selector
+Start chatting
+
-### OpenAI Models
+## Available OpenAI Models
-You can also use specific OpenAI models you cannot find in the **Hub** section by customizing the `model.yaml` file, which you can see in the `~/jan/data/models/`. Follow the steps in the [Manage Models](/docs/models/manage-models) to manually add a model.
-
-
-- You can find the list of available models in the [OpenAI Platform](https://platform.openai.com/docs/models/overview).
-- The `id` property must match the model name in the list.
- - For example, if you want to use the [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo), you must set the `id` property to `gpt-4-1106-preview`.
-
+Jan automatically includes popular OpenAI models. In case you want to use a specific OpenAI model that you cannot find in Jan, follow instructions in [Manual Setup](/docs/models/manage-models#4-manual-setup) to add custom models:
+ - See list of available models in [OpenAI Platform](https://platform.openai.com/docs/models/overview).
+ - The id property must match the model name in the list. For example, if you want to use the [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo), you must set the id property to gpt-4-1106-preview.
## Troubleshooting
-If you encounter any issues during the integration process or while using OpenAI with Jan, consider the following troubleshooting steps:
+Common issues and solutions:
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to OpenAI API support for assistance if needed.
\ No newline at end of file
+1. API Key Issues
+- Verify your API key is correct and not expired
+- Check if you have billing set up on your OpenAI account
+- Ensure you have access to the model you're trying to use
+
+2. Connection Problems
+- Check your internet connection
+- Verify OpenAI's [system status](https://status.openai.com)
+- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
+
+3. Model Unavailable
+- Confirm your API key has access to the model
+- Check if you're using the correct model ID
+- Verify your OpenAI account has the necessary permissions
+
+Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the [OpenAI documentation](https://platform.openai.com/docs).
\ No newline at end of file
diff --git a/docs/src/pages/docs/remote-models/openrouter.mdx b/docs/src/pages/docs/remote-models/openrouter.mdx
index 8efcd932a..a32de2423 100644
--- a/docs/src/pages/docs/remote-models/openrouter.mdx
+++ b/docs/src/pages/docs/remote-models/openrouter.mdx
@@ -18,38 +18,83 @@ keywords:
---
import { Callout, Steps } from 'nextra/components'
+import { Settings, Plus } from 'lucide-react'
# OpenRouter
## Integrate OpenRouter with Jan
-[OpenRouter](https://openrouter.ai/docs#quick-start) is a tool that gathers AI models. Developers can utilize its API to engage with diverse large language models, generative image models, and generative 3D object models.
+[OpenRouter](https://openrouter.ai/) is a tool that gathers AI models. Developers can utilize its API to engage with diverse large language models, generative image models, and generative 3D object models with a competitive pricing.
-To connect Jan with OpenRouter for accessing remote Large Language Models (LLMs) through OpenRouter, you can follow the steps below:
+Jan supports OpenRouter API integration, allowing you to use models from various providers (Anthropic, Google, Meta and more) through a single API.
+
+## Integrate OpenRouter with Jan
-### Step 1: Configure OpenRouter API Key
+### Step 1: Get Your API Key
+1. Visit [OpenRouter](https://openrouter.ai/keys) and sign in
+2. Create & copy a new API key or copy your existing one
-1. Find your API Key in the [OpenRouter API Key](https://openrouter.ai/keys).
-2. Copy your **OpenRouter API Key**.
-3. There are three ways to configure your API Key in Jan app:
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **My Models** tab > **Add Icon (➕)** next to **OpenRouter**.
- - Navigate to the **Jan app** > **Thread** > **Model** tab > **Add Icon (➕)** next to **OpenRouter**.
- - Navigate to the **Jan app** > **Gear Icon (⚙️)** > **OpenRouter** section under Model Providers.
-4. Insert your **OpenRouter API Key**.
-5. For **OpenRouter**, specify the model you want to use, or the system will default to the preset model linked to your **OpenRouter API Key**.
+
+Ensure your API key has sufficient credits. OpenRouter credits work across all available models.
+
-### Step 2: Start Chatting with the Model
+### Step 2: Configure Jan
+There are two ways to add your OpenRouter key in Jan:
-1. Select the OpenRouter model you want to use.
-2. Specify the model's parameters.
-3. Start the conversation with the OpenRouter model.
+**Through Threads:**
+1. In Threads, click **Model** tab in the **right sidebar** or **model selector** in input field
+2. Once the selector is poped up, choose the **Cloud** tab
+3. Click **Add** () icon next to **OpenRouter**
+4. Once you are directed to OpenRouter settings, insert your **API Key**
+
+**Through Settings:**
+1. Navigate to **Settings** ()
+2. Under **Remote Engines**, select **OpenRouter**
+3. Insert your **API Key**
+
+
+
+
+
+### Step 3: Start Using OpenRouter Models
+
+1. In any existing **Threads** or create a new one
+2. Select any model from **model selector** under OpenRouter
+3. Start chatting
+## Available Models Through OpenRouter
+
+Jan automatically use your default OpenRouter's available models. For custom configurations:
+
+**Model Field Settings:**
+- Leave empty to use your account's default model
+- Specify a model using the format: `organization/model-name`
+- Available options can be found in [OpenRouter's Model Reference](https://openrouter.ai/models)
+
+**Examples of Model IDs:**
+- Claude 3 Opus: `anthropic/claude-3-opus-20240229`
+- Google Gemini Pro: `google/gemini-pro`
+- Mistral Large: `mistralai/mistral-large`
+
## Troubleshooting
-If you encounter any issues during the integration process or while using OpenAI with Jan, consider the following troubleshooting steps:
+Common issues and solutions:
-- Double-check your API credentials to ensure they are correct.
-- Check for error messages or logs that may provide insight into the issue.
-- Reach out to OpenRouter API support for assistance if needed.
\ No newline at end of file
+**1. API Key Issues**
+- Verify your API key is correct and not expired
+- Check if you have sufficient credits in your OpenRouter account
+- Ensure you have access to the model you're trying to use
+
+**2. Connection Problems**
+- Check your internet connection
+- Verify OpenRouter's [system status](https://status.openrouter.ai)
+- Look for error messages in [Jan's logs](/docs/troubleshooting#how-to-get-error-logs)
+
+**3. Model Unavailable**
+- Confirm the model is currently available on OpenRouter
+- Check if you're using the correct model ID format
+- Verify the model provider is currently operational
+
+Need more help? Join our [Discord community](https://discord.gg/FTk2MvZwJH) or check the [OpenRouter documentation](https://openrouter.ai/docs).
\ No newline at end of file
diff --git a/docs/src/pages/docs/settings.mdx b/docs/src/pages/docs/settings.mdx
index fb99df518..9b5f3fe4d 100644
--- a/docs/src/pages/docs/settings.mdx
+++ b/docs/src/pages/docs/settings.mdx
@@ -29,10 +29,6 @@ import { Settings, EllipsisVertical, Plus, FolderOpen, Pencil } from 'lucide-rea
This guide explains how to customize your Jan application settings.
To access **Settings**, click icon in the bottom left corner of Jan.
-
-Settings are stored in a `cortex.db` file in [Jan Data Folder](/docs/data-folder), ensuring they persist across sessions.
-
-
## My models
Here's at **Settings** () > **My Models** you can manage all your installed AI models:
diff --git a/docs/src/pages/integrations/workflow-automation/n8n.mdx b/docs/src/pages/integrations/workflow-automation/n8n.mdx
new file mode 100644
index 000000000..28d838371
--- /dev/null
+++ b/docs/src/pages/integrations/workflow-automation/n8n.mdx
@@ -0,0 +1,81 @@
+---
+title: n8n
+keywords:
+ [
+ Jan,
+ Customizable Intelligence, LLM,
+ local AI,
+ privacy focus,
+ free and open source,
+ private and offline,
+ conversational AI,
+ no-subscription fee,
+ large language models,
+ n8n integration,
+ n8n,
+ ]
+description: A step-by-step guide on integrating Jan with n8n.
+---
+
+import { Steps } from 'nextra/components'
+
+# n8n
+
+## Integrate n8n with Jan
+
+[n8n](https://n8n.io/) is an open-source workflow automation tool that allows you to connect to more than 400+ integrations and services to automate repetitive tasks. With its visual interface, you can create complex workflows conveniently. To integrate n8n with Jan, follow the steps below:
+
+
+### Step 1: Run your preferred model with Jan server
+
+1. Open Jan app.
+2. Go to the **Hub** and download your preferred model
+3. Run the Jan server
+
+### Step 2: Start n8n service
+Start n8n immediately using npx:
+
+```
+npx n8n
+```
+
+Or deploy with Docker:
+
+```
+docker run -it --rm --name n8n -p 5678:5678 docker.n8n.io/n8nio/n8n
+```
+
+### Step 3: Integrate Jan with n8n service using HTTP Request
+
+Integrate Jan by selecting the HTTP Request node in n8n and importing the following cURL command:
+
+```bash
+curl -X 'POST' \
+ 'http://127.0.0.1:1337/v1/chat/completions' \
+ -H 'accept: application/json' \
+ -H 'Content-Type: application/json' \
+ -d '{
+ "messages": [
+ {
+ "content": "You are a helpful assistant.",
+ "role": "system"
+ },
+ {
+ "content": "Hello!",
+ "role": "user"
+ }
+ ],
+ "model": "tinyllama-1.1b",
+ "stream": true,
+ "max_tokens": 2048,
+ "stop": [
+ "hello"
+ ],
+ "frequency_penalty": 0,
+ "presence_penalty": 0,
+ "temperature": 0.7,
+ "top_p": 0.95
+}'
+```
+
+
\ No newline at end of file