+
+
diff --git a/website/src/components/YouTube.astro b/website/src/components/YouTube.astro
new file mode 100644
index 000000000..6459addcb
--- /dev/null
+++ b/website/src/components/YouTube.astro
@@ -0,0 +1,60 @@
+---
+export interface Props {
+ id: string;
+ title?: string;
+ class?: string;
+}
+
+const { id, title = 'YouTube video player', class: className } = Astro.props;
+
+// Extract video ID and handle both formats:
+// - Simple ID: "4mvHgLy_YV8"
+// - ID with params: "4mvHgLy_YV8?si=74cmdMmcH3gmpv0R"
+const videoId = id.split('?')[0];
+const params = id.includes('?') ? '?' + id.split('?')[1] : '';
+---
+
+
+
+
+
+
diff --git a/website/src/content.config.ts b/website/src/content.config.ts
index ee21da167..3c8b69d82 100644
--- a/website/src/content.config.ts
+++ b/website/src/content.config.ts
@@ -1,8 +1,38 @@
-import { defineCollection } from 'astro:content';
+import { defineCollection, z } from 'astro:content';
import { docsLoader } from '@astrojs/starlight/loaders';
import { docsSchema } from '@astrojs/starlight/schema';
import { videosSchema } from 'starlight-videos/schemas';
+const changelogSchema = z.object({
+ title: z.string(),
+ description: z.string(),
+ date: z.date(),
+ version: z.string().optional(),
+ image: z.string().optional(),
+ gif: z.string().optional(),
+ video: z.string().optional(),
+ featured: z.boolean().default(false),
+});
+
+const blogSchema = z.object({
+ title: z.string(),
+ description: z.string(),
+ date: z.date(),
+ tags: z.string().optional(),
+ categories: z.string().optional(),
+ author: z.string().optional(),
+ ogImage: z.string().optional(),
+ featured: z.boolean().default(false),
+});
+
export const collections = {
docs: defineCollection({ loader: docsLoader(), schema: docsSchema({ extend: videosSchema }) }),
+ changelog: defineCollection({
+ type: 'content',
+ schema: changelogSchema,
+ }),
+ blog: defineCollection({
+ type: 'content',
+ schema: blogSchema,
+ }),
};
diff --git a/website/src/content/blog/_assets/3090s.jpg b/website/src/content/blog/_assets/3090s.jpg
new file mode 100644
index 000000000..3a62b3f6f
Binary files /dev/null and b/website/src/content/blog/_assets/3090s.jpg differ
diff --git a/website/src/content/blog/_assets/4070s.jpg b/website/src/content/blog/_assets/4070s.jpg
new file mode 100644
index 000000000..3d8622347
Binary files /dev/null and b/website/src/content/blog/_assets/4070s.jpg differ
diff --git a/website/src/content/blog/_assets/4090s.png b/website/src/content/blog/_assets/4090s.png
new file mode 100644
index 000000000..2c49a3248
Binary files /dev/null and b/website/src/content/blog/_assets/4090s.png differ
diff --git a/website/src/content/blog/_assets/ai-locally-llama.cpp.jpg b/website/src/content/blog/_assets/ai-locally-llama.cpp.jpg
new file mode 100644
index 000000000..967b63bf7
Binary files /dev/null and b/website/src/content/blog/_assets/ai-locally-llama.cpp.jpg differ
diff --git a/website/src/content/blog/_assets/catastrophic-demo.png b/website/src/content/blog/_assets/catastrophic-demo.png
new file mode 100644
index 000000000..7c869fc0e
Binary files /dev/null and b/website/src/content/blog/_assets/catastrophic-demo.png differ
diff --git a/website/src/content/blog/_assets/chat-with-docs-prompt.jpg b/website/src/content/blog/_assets/chat-with-docs-prompt.jpg
new file mode 100644
index 000000000..df47dd4ef
Binary files /dev/null and b/website/src/content/blog/_assets/chat-with-docs-prompt.jpg differ
diff --git a/website/src/content/blog/_assets/chat-with-your-docs-offline-ai.jpg b/website/src/content/blog/_assets/chat-with-your-docs-offline-ai.jpg
new file mode 100644
index 000000000..efcda0f07
Binary files /dev/null and b/website/src/content/blog/_assets/chat-with-your-docs-offline-ai.jpg differ
diff --git a/website/src/content/blog/_assets/chat-with-your-docs2.jpg b/website/src/content/blog/_assets/chat-with-your-docs2.jpg
new file mode 100644
index 000000000..1577b3f5c
Binary files /dev/null and b/website/src/content/blog/_assets/chat-with-your-docs2.jpg differ
diff --git a/website/src/content/blog/_assets/deepseek-r1-locally-jan.jpg b/website/src/content/blog/_assets/deepseek-r1-locally-jan.jpg
new file mode 100644
index 000000000..2168b8986
Binary files /dev/null and b/website/src/content/blog/_assets/deepseek-r1-locally-jan.jpg differ
diff --git a/website/src/content/blog/_assets/download-jan.jpg b/website/src/content/blog/_assets/download-jan.jpg
new file mode 100644
index 000000000..f799260c7
Binary files /dev/null and b/website/src/content/blog/_assets/download-jan.jpg differ
diff --git a/website/src/content/blog/_assets/egpu.jpg b/website/src/content/blog/_assets/egpu.jpg
new file mode 100644
index 000000000..9f631d4fd
Binary files /dev/null and b/website/src/content/blog/_assets/egpu.jpg differ
diff --git a/website/src/content/blog/_assets/gradient-decent.gif b/website/src/content/blog/_assets/gradient-decent.gif
new file mode 100644
index 000000000..9828f2fe9
Binary files /dev/null and b/website/src/content/blog/_assets/gradient-decent.gif differ
diff --git a/website/src/content/blog/_assets/hugging-face-jan-model-download.jpg b/website/src/content/blog/_assets/hugging-face-jan-model-download.jpg
new file mode 100644
index 000000000..c6cfa8ea5
Binary files /dev/null and b/website/src/content/blog/_assets/hugging-face-jan-model-download.jpg differ
diff --git a/website/src/content/blog/_assets/jan-hf-model-download.jpg b/website/src/content/blog/_assets/jan-hf-model-download.jpg
new file mode 100644
index 000000000..929acf2ff
Binary files /dev/null and b/website/src/content/blog/_assets/jan-hf-model-download.jpg differ
diff --git a/website/src/content/blog/_assets/jan-hub-deepseek-r1.jpg b/website/src/content/blog/_assets/jan-hub-deepseek-r1.jpg
new file mode 100644
index 000000000..12c0c6640
Binary files /dev/null and b/website/src/content/blog/_assets/jan-hub-deepseek-r1.jpg differ
diff --git a/website/src/content/blog/_assets/jan-hub-download-deepseek-r1-2.jpg b/website/src/content/blog/_assets/jan-hub-download-deepseek-r1-2.jpg
new file mode 100644
index 000000000..24be4bd25
Binary files /dev/null and b/website/src/content/blog/_assets/jan-hub-download-deepseek-r1-2.jpg differ
diff --git a/website/src/content/blog/_assets/jan-hub-download-deepseek-r1.jpg b/website/src/content/blog/_assets/jan-hub-download-deepseek-r1.jpg
new file mode 100644
index 000000000..83d9ab370
Binary files /dev/null and b/website/src/content/blog/_assets/jan-hub-download-deepseek-r1.jpg differ
diff --git a/website/src/content/blog/_assets/jan-hub-for-ai-models.jpg b/website/src/content/blog/_assets/jan-hub-for-ai-models.jpg
new file mode 100644
index 000000000..a158499b4
Binary files /dev/null and b/website/src/content/blog/_assets/jan-hub-for-ai-models.jpg differ
diff --git a/website/src/content/blog/_assets/jan-library-deepseek-r1.jpg b/website/src/content/blog/_assets/jan-library-deepseek-r1.jpg
new file mode 100644
index 000000000..6a54082dc
Binary files /dev/null and b/website/src/content/blog/_assets/jan-library-deepseek-r1.jpg differ
diff --git a/website/src/content/blog/_assets/jan-local-ai.jpg b/website/src/content/blog/_assets/jan-local-ai.jpg
new file mode 100644
index 000000000..2c8c145ff
Binary files /dev/null and b/website/src/content/blog/_assets/jan-local-ai.jpg differ
diff --git a/website/src/content/blog/_assets/jan-model-download.jpg b/website/src/content/blog/_assets/jan-model-download.jpg
new file mode 100644
index 000000000..7e949403d
Binary files /dev/null and b/website/src/content/blog/_assets/jan-model-download.jpg differ
diff --git a/website/src/content/blog/_assets/jan-model-selection.jpg b/website/src/content/blog/_assets/jan-model-selection.jpg
new file mode 100644
index 000000000..b630c800e
Binary files /dev/null and b/website/src/content/blog/_assets/jan-model-selection.jpg differ
diff --git a/website/src/content/blog/_assets/jan-runs-deepseek-r1-distills.jpg b/website/src/content/blog/_assets/jan-runs-deepseek-r1-distills.jpg
new file mode 100644
index 000000000..02ce847f4
Binary files /dev/null and b/website/src/content/blog/_assets/jan-runs-deepseek-r1-distills.jpg differ
diff --git a/website/src/content/blog/_assets/jan-system-prompt-deepseek-r1.jpg b/website/src/content/blog/_assets/jan-system-prompt-deepseek-r1.jpg
new file mode 100644
index 000000000..f79e71af0
Binary files /dev/null and b/website/src/content/blog/_assets/jan-system-prompt-deepseek-r1.jpg differ
diff --git a/website/src/content/blog/_assets/jan.ai.jpg b/website/src/content/blog/_assets/jan.ai.jpg
new file mode 100644
index 000000000..d635d1ab9
Binary files /dev/null and b/website/src/content/blog/_assets/jan.ai.jpg differ
diff --git a/website/src/content/blog/_assets/local-ai-model-parameters.jpg b/website/src/content/blog/_assets/local-ai-model-parameters.jpg
new file mode 100644
index 000000000..1d26fc4a5
Binary files /dev/null and b/website/src/content/blog/_assets/local-ai-model-parameters.jpg differ
diff --git a/website/src/content/blog/_assets/offline-chatgpt-alternative-ai-without-internet.jpg b/website/src/content/blog/_assets/offline-chatgpt-alternative-ai-without-internet.jpg
new file mode 100644
index 000000000..6dffb1e95
Binary files /dev/null and b/website/src/content/blog/_assets/offline-chatgpt-alternative-ai-without-internet.jpg differ
diff --git a/website/src/content/blog/_assets/offline-chatgpt-alternatives-jan.jpg b/website/src/content/blog/_assets/offline-chatgpt-alternatives-jan.jpg
new file mode 100644
index 000000000..065b33636
Binary files /dev/null and b/website/src/content/blog/_assets/offline-chatgpt-alternatives-jan.jpg differ
diff --git a/website/src/content/blog/_assets/og-4090s.webp b/website/src/content/blog/_assets/og-4090s.webp
new file mode 100644
index 000000000..6db1b10b2
Binary files /dev/null and b/website/src/content/blog/_assets/og-4090s.webp differ
diff --git a/website/src/content/blog/_assets/open-source-ai-quantization.jpg b/website/src/content/blog/_assets/open-source-ai-quantization.jpg
new file mode 100644
index 000000000..fe605c3cd
Binary files /dev/null and b/website/src/content/blog/_assets/open-source-ai-quantization.jpg differ
diff --git a/website/src/content/blog/_assets/openchat-bench-0106.png b/website/src/content/blog/_assets/openchat-bench-0106.png
new file mode 100644
index 000000000..9fa37960f
Binary files /dev/null and b/website/src/content/blog/_assets/openchat-bench-0106.png differ
diff --git a/website/src/content/blog/_assets/qwen3-in-jan-hub.jpeg b/website/src/content/blog/_assets/qwen3-in-jan-hub.jpeg
new file mode 100644
index 000000000..e58c5beab
Binary files /dev/null and b/website/src/content/blog/_assets/qwen3-in-jan-hub.jpeg differ
diff --git a/website/src/content/blog/_assets/qwen3-settings-in-jan.jpeg b/website/src/content/blog/_assets/qwen3-settings-in-jan.jpeg
new file mode 100644
index 000000000..82d7540a7
Binary files /dev/null and b/website/src/content/blog/_assets/qwen3-settings-in-jan.jpeg differ
diff --git a/website/src/content/blog/_assets/qwen3-settings-jan-ai.jpeg b/website/src/content/blog/_assets/qwen3-settings-jan-ai.jpeg
new file mode 100644
index 000000000..7fc432e38
Binary files /dev/null and b/website/src/content/blog/_assets/qwen3-settings-jan-ai.jpeg differ
diff --git a/website/src/content/blog/_assets/replay.png b/website/src/content/blog/_assets/replay.png
new file mode 100644
index 000000000..8ada6ce84
Binary files /dev/null and b/website/src/content/blog/_assets/replay.png differ
diff --git a/website/src/content/blog/_assets/run-ai-locally-with-jan.jpg b/website/src/content/blog/_assets/run-ai-locally-with-jan.jpg
new file mode 100644
index 000000000..942ab38ba
Binary files /dev/null and b/website/src/content/blog/_assets/run-ai-locally-with-jan.jpg differ
diff --git a/website/src/content/blog/_assets/run-deepseek-r1-locally-in-jan.jpg b/website/src/content/blog/_assets/run-deepseek-r1-locally-in-jan.jpg
new file mode 100644
index 000000000..aa6980585
Binary files /dev/null and b/website/src/content/blog/_assets/run-deepseek-r1-locally-in-jan.jpg differ
diff --git a/website/src/content/blog/_assets/throughput_Comparison.png b/website/src/content/blog/_assets/throughput_Comparison.png
new file mode 100644
index 000000000..6bb63d03c
Binary files /dev/null and b/website/src/content/blog/_assets/throughput_Comparison.png differ
diff --git a/website/src/content/blog/benchmarking-nvidia-tensorrt-llm.mdx b/website/src/content/blog/benchmarking-nvidia-tensorrt-llm.mdx
new file mode 100644
index 000000000..75bb2099e
--- /dev/null
+++ b/website/src/content/blog/benchmarking-nvidia-tensorrt-llm.mdx
@@ -0,0 +1,321 @@
+---
+title: Benchmarking NVIDIA TensorRT-LLM
+description: This post compares the performance of TensorRT-LLM and llama.cpp on consumer NVIDIA GPUs, highlighting the trade-offs among speed, resource usage, and convenience.
+tags: Nvidia, TensorRT-LLM, llama.cpp, rtx3090, rtx4090, "inference engine"
+categories: research
+ogImage: assets/images/general/og-throughput-benchmark.png
+date: 2024-04-29
+---
+
+import { Aside } from '@astrojs/starlight/components'
+
+
+import throughputComparison from '@/assets/blog/throughput_Comparison.png';
+import img4090s from '@/assets/blog/4090s.png';
+import og4090s from '@/assets/blog/og-4090s.webp';
+import img3090s from '@/assets/blog/3090s.jpg';
+import img4070s from '@/assets/blog/4070s.jpg';
+import egpu from '@/assets/blog/egpu.jpg';
+
+
+Jan now supports [NVIDIA TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM) in addition to [llama.cpp](https://github.com/ggerganov/llama.cpp), making Jan multi-engine and ultra-fast for users with Nvidia GPUs.
+
+We've been excited for TensorRT-LLM for a while, and [had a lot of fun implementing it](https://github.com/menloresearch/nitro-tensorrt-llm). As part of the process, we've run some benchmarks, to see how TensorRT-LLM fares on consumer hardware (e.g. [4090s](https://www.nvidia.com/en-us/geforce/graphics-cards/40-series/), [3090s](https://www.nvidia.com/en-us/geforce/graphics-cards/30-series/)) we commonly see in the [Jan's hardware community](https://discord.com/channels/1107178041848909847/1201834752206974996).
+
+
+
+
+
+## Key Findings
+
+
+
+TensorRT-LLM was:
+
+- **30-70% faster** than llama.cpp on the same hardware
+- **Consumes less memory on consecutive runs** and **marginally more GPU VRAM utilization** than llama.cpp
+- **20%+ smaller compiled model sizes** than llama.cpp
+- **Less convenient** as models have to be compiled for a specific OS and GPU architecture, vs. llama.cpp's "Compile once, run everywhere" portability
+- **Less accessible** as it does not support older-generation NVIDIA GPUs
+
+## Why TensorRT-LLM?
+
+[TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM) is Nvidia's open-source inference library that incorporates Nvidia's proprietary optimizations beyond the open-source [cuBLAS](https://developer.nvidia.com/cublas) library.
+
+As compared to [llama.cpp](https://github.com/ggerganov/llama.cpp), which today dominates Desktop AI as a cross-platform inference engine, TensorRT-LLM is highly optimized for Nvidia GPUs. While llama.cpp compiles models into a [single, generalizable CUDA "backend"](https://github.com/ggerganov/llama.cpp/blob/master/ggml-cuda.cu) that can run on a wide range of Nvidia GPUs, TensorRT-LLM compiles models into a [GPU-specific execution graph](https://www.baseten.co/blog/high-performance-ml-inference-with-nvidia-tensorrt/) that is highly optimized for that specific GPU's Tensor Cores, CUDA cores, VRAM and memory bandwidth.
+
+TensorRT-LLM is typically used in datacenter-grade GPUs, where it produces a [face-melting 10,000 tokens/s](https://nvidia.github.io/TensorRT-LLM/blogs/H100vsA100.html) on [NVIDIA H100 Tensor Core GPUs](https://www.nvidia.com/en-us/data-center/h100/). We were curious for how TensorRT-LLM performs on consumer-grade GPUs, and gave it a spin.
+
+| Llama.cpp | TensorRT-LLM |
+| ------------------------------------------------------------------------------------------------------------------------- | --------------------------------------------------------------------------------------------------------------------------------- |
+| Baseline | Blazing fast (30-70% faster) |
+| Compile once, run cross-platform | Compiled and highly optimized for specific GPU architecture |
+| Generalizable and Portable | Highly-optimized |
+| Model compiles to [single, generalizable CUDA "backend"](https://github.com/ggerganov/llama.cpp/blob/master/ggml-cuda.cu) | Model compiles to [GPU-specific execution graph](https://www.baseten.co/blog/high-performance-ml-inference-with-nvidia-tensorrt/) |
+
+## Experiment Setup
+
+We ran the experiment using standardized inference requests in a sandboxed environment:
+- **Model**: Mistral 7b model, compiled and quantized at a comparable `int4` quantization.
+- **Test runs**: 5 batches of 10 runs each, per inference engine, on a bare metal PC with no other applications.
+- **Parameters**: User defaults, i.e. `batch_size 1`, `input_len 2048` and `output_len 512`
+- **Measurements**:
+ - CPU, memory from Jan system monitor
+ - GPU VRAM utilization metrics from `nvidia-smi`, and taken over an interval of 14 seconds.
+ - Throughput (token/sec) using [Jan's built-in Tokens/sec perf stat](https://github.com/search?q=repo%3Ajanhq%2Fjan%20timeDiffInSeconds&type=code).
+
+
+
+
+
+### Hardware Selection
+
+We chose the following GPUs based on our users' preferences:
+
+| NVIDIA GPU | VRAM (GB) | CUDA Cores | Tensor Cores | Memory Bus Width (bit) | Memory Bandwidth (GB/s) | Connection (GB/s) |
+| --------------------------------- | --------- | ---------- | ------------ | ---------------------- | ----------------------- | -------------------------------------------- |
+| GeForce RTX 4090 (Ada) | 24 | 16,384 | 512 | 384 | ~1000 | PCIe4.0 x16 (~32) |
+| GeForce RTX 3090 (Ampere) | 24 | 10,496 | 328 | 384 | 935.8 | PCIe4.0 x16 (~32) |
+| GeForce RTX 4070 Laptop GPU (Ada) | 8 | 7680 | 144 | 192 | 272 | PCIe4.0 x4 (~8) |
+| GeForce RTX 4090 eGPU (Ada) | 24 | 16,384 | 512 | 384 | ~1000 | Thunderbolt 3 connected to a USB4 USB-C port ([~1.25-5?](https://www.cablematters.com/Blog/Thunderbolt/usb4-vs-thunderbolt-3)) |
+
+### llama.cpp Setup
+
+- llama.cpp commit [15499eb](https://github.com/ggerganov/llama.cpp/commit/15499eb94227401bdc8875da6eb85c15d37068f7)
+- We used `Mistral-7b-q4_k_m` in `GGUF` with `ngl` at `100`
+
+
+
+### TensorRT-LLM Setup
+
+- TensorRT-LLM version [0.7.1](https://github.com/NVIDIA/TensorRT-LLM/releases/tag/v0.7.1) and build on Windows
+- For TensorRT-LLM, we used `Mistral-7b-int4 AWQ`
+- We ran TensorRT-LLM with `free_gpu_memory_fraction` to test it with the lowest VRAM consumption
+- Note: We picked AWQ for TensorRT-LLM to be a closer comparison to GGUF's Q4.
+
+## Results
+
+### NVIDIA GeForce RTX 4090 GPU
+
+
+*Jan is built on this Dual-4090 workstation, which recently got upgraded to a nice case*
+
+
+*The original case (or lack thereof) for our Dual-4090 cluster, as posted on [r/localllama](https://www.reddit.com/r/LocalLLaMA/comments/16lxt6a/case_for_dual_4090s/)*
+
+
+
+For this test, we used Jan's [Dual-4090 workstation](https://www.reddit.com/r/LocalLLaMA/comments/16lxt6a/case_for_dual_4090s/), which our engineers timeshare to build Jan.
+
+The [NVIDIA GeForce RTX 4090](https://www.nvidia.com/en-us/geforce/graphics-cards/40-series/) is the latest top-of-the-line desktop GPU, with an MSRP of $1,599, and uses the Ada architecture. It has a ~1000 GB/s memory bandwidth within VRAM, and a PCIe4 x16 lane (~32 GB/s) between the GPU and the CPU.
+
+| Metrics | GGUF (using CPU) | GGUF (using GPU) | TensorRT-LLM | How TensorRT-LLM Compares |
+| ------------------------ | ---------------- | ---------------- | ------------ | ------------------------- |
+| Throughput (token/s) | 14.0 | 100.43 | 170.63 | ✅ 69.89% faster |
+| Max GPU Utilization (%) | N/A | 83.50 | 88.50 | 5.99% more |
+| Max VRAM Utilization (%) | N/A | 64 | 72.1 | 12.66% more |
+| Avg RAM Used (GB) | 0.611 | 7.105 | 4.98 | ✅ 29.88% less |
+| Disk Size (GB) | 4.07 | 4.06 | 3.05 | ✅ 24.88% smaller |
+
+TensorRT-LLM was almost 70% faster than llama.cpp by building the model for the GeForce RTX 4090 GPU’s Ada architecture for optimal graph execution, fully utilizing the 512 Tensor Cores, 16,384 CUDA cores, and 1,000 GB/s of memory bandwidth.
+
+The intuition for why llama.cpp is slower is because it compiles a model into a [single, generalizable CUDA “backend”](https://github.com/ggerganov/llama.cpp/blob/master/ggml-cuda.cu) that can run on many NVIDIA GPUs. Doing so requires llama.cpp to sacrifice all the optimizations that TensorRT-LLM makes with its compilation to a GPU-specific execution graph.
+
+### NVIDIA GeForce RTX 3090 GPU
+
+
+*Our 3090 Machine, now used by one of our engineers to build Jan*
+
+
+
+The [NVIDIA's GeForce RTX 3090](https://www.nvidia.com/en-us/geforce/graphics-cards/30-series/rtx-3090-3090ti/) is a popular desktop GPU, and retails for approximately $1,500 (as of April 24). It uses the NVIDIA Ampere architecture. As compared to its successor GeForce RTX 4090, it has 33% fewer CUDA cores (10,496) and Tensor Cores (328) and 7% less memory bandwidth (~930 GB/s).
+
+| Metrics | GGUF (using CPU) | GGUF (using GPU) | TensorRT-LLM | How TensorRT-LLM Compares |
+| ------------------------ | ---------------- | ---------------- | ------------ | ------------------------- |
+| Throughput (token/s) | 11.42 | 88.70 | 144.19 | ✅ 62.57% faster |
+| Max GPU Utilization (%) | N/A | 80.40 | 89.10 | 10.82% more |
+| Max VRAM Utilization (%) | N/A | 66.80 | 76.20 | 14.07% more |
+| Avg RAM Used (GB) | 0.611 | 2.60 | 0.98 | 62.41%% less |
+| Disk Size (GB) | 4.07 | 4.06 | 3.05 | ✅ 24.88% smaller |
+
+Interestingly, the GeForce RTX 3090 was only 16.6% slower compared with the GeForce RTX 4090. On TPS, TensorRT-LLM outperformed llama.cpp by 62.57%. Curiously, it also used negligible RAM for subsequent inference requests after the initial model warmup.
+
+### NVIDIA GeForce RTX 4070 Laptop GPU
+
+
+
+
+
+We also benchmarked an NVIDIA GeForce RTX 4070 Laptop GPU with 8gb of VRAM, which is a popular configuration among Jan users. Laptop GPUs are less powerful than their desktop counterparts, as they trade portability for reduced energy consumption and thermal constraints.
+
+| Metrics | GGUF (using CPU) | GGUF (using GPU) | TensorRT-LLM | Difference on GPU |
+| ------------------------ | ---------------- | ---------------- | ------------ | ----------------- |
+| Throughput (token/s) | 11.57 | 39.70 | 51.57 | ✅ 29.9% faster |
+| Max GPU Utilization (%) | N/A | 80.00 | 84.67 | 5.83% more |
+| Max VRAM Utilization (%) | N/A | 72.78 | 81.22 | 11.60% more |
+| Avg RAM Used (GB) | 4.49 | 4.44 | 1.04 | ✅ 76.55%% less |
+| Disk Size (GB) | 4.07 | 4.06 | 3.05 | ✅ 24.88% smaller |
+
+TensorRT-LLM on the laptop dGPU was 29.9% faster in tokens per second throughput than llama.cpp, but significantly slower than the desktop GPUs.
+
+The intuition for this is fairly simple: the GeForce RTX 4070 Laptop GPU has 53.1% fewer CUDA cores and Tensor Cores (compared to the 4090), and less VRAM (8gb vs. 24gb). This reduces the surface area for GPU-specific optimizations for TensorRT-LLM.
+
+The GeForce RTX 4070 Laptop GPU is also ~70% slower than the GeForce RTX 4090 desktop GPU, showing the hardware effect of less electricity draw, less VRAM, and thermal constraints on inference speed.
+
+### Laptop with NVIDIA GeForce RTX 4090 eGPU
+
+
+
+
+
+Our last benchmark was to experiment with an [Asus RTX 4090 eGPU](https://www.gigabyte.com/Graphics-Card/GV-N4090IXEB-24GD), that was connected via a [Thunderbolt 3 port](https://www.gigabyte.com/Graphics-Card/GV-N4090IXEB-24GD) to the [Razer Blade 14's USB4 port](https://www.razer.com/sg-en/gaming-laptops/razer-blade-14). Theoretically, the results should be fairly similar to the GeForce RTX 4090 desktop GPU as they have identical underlying GPUs, but with very different connection speeds.
+
+We thought it would be an interesting to see how TensorRT-LLM handles a 68.4% reduction in communication bandwidth between the CPU and GPU:
+- Thunderbolt 3 connection (1.25-5 GB/s?) for eGPUs
+- PCIe 4.0 x4 (~8 GB/s) for "on device" desktops
+
+
+
+Overall, we used mid-to-high-end NVIDIA desktop GPUs for our tests, as TensorRT-LLM’s performance enhancements are most apparent on bigger VRAMs. For users with lower-spec machines, llama.cpp is better.
+
+| Metrics | GGUF (using CPU) | GGUF (using GPU) | TensorRT-LLM | Difference on GPU |
+| ------------------------ | ---------------- | ---------------- | ------------ | ----------------- |
+| Throughput (token/s) | 11.56 | 62.22 | 104.95 | ✅ 68.66% faster |
+| Max VRAM Utilization (%) | 0 | 65 | 99 | 52.31% more |
+| RAM Used (GB) | 0.611 | 5.38 | 4.11 | ✅ 23.61% less |
+| Disk Size (GB) | 4.07 | 4.06 | 3.05 | ✅ 24.88% smaller |
+
+The Thunderbolt 3 eGPU had a 38.5% lower tokens/s as compared to the PCIe4.0 x16 connected GPU. But the % speedup vs. llama.cpp was similar, at around 69%.
+
+Interestingly, the VRAM used with the eGPU was variably higher. Our hypothesis is that the slower communication bandwidth results in more VRAM being allocated, as memory is released mostly slowly as well.
+
+## Conclusion
+
+### Token Speed
+
+
+
+| Throughput (Higher is Better) | TensorRT-LLM | Llama.cpp | % Difference |
+| ---------------------------------- | --------------- | ----------- | ------------- |
+| GeForce RTX 4090 desktop GPU | ✅ 170.63t/s | 100.43t/s | 69.89% faster |
+| GeForce RTX 3090 desktop GPU | ✅ 144.19t/s | 88.70t/s | 62.57% faster |
+| GeForce RTX 4090 eGPU | ✅ 104.95t/s | 62.22t/s | 68.66% faster |
+| GeForce RTX 4070 Laptop GPU | ✅ 51.57t/s | 39.70t/s | 29.90% faster |
+| Laptop AMD Ryzen™ 9 8945HS, 8C/16T | (Not supported) | ✅ 11.57t/s | |
+
+- TensorRT-LLM is up to **70% faster** than llama.cpp on desktop GPUs (e.g. 3090 GPU, 4090 GPUs) while using less RAM & CPU (but more fully utilizing VRAM)
+- TensorRT-LLM is up to **30% faster** on laptop GPUs (e.g. 4070 GPUs) with smaller VRAM
+
+### Max VRAM Utilization
+
+| Average VRAM utilization % | TensorRT-LLM | Llama.cpp | % Difference |
+| ---------------------------- | ------------ | --------- | ------------ |
+| GeForce RTX 4090 desktop GPU | 72.10 | 64.00 | 12.66% more |
+| GeForce RTX 3090 desktop GPU | 76.20 | 66.80 | 14.07% more |
+| GeForce RTX 4070 Laptop GPU | 81.22 | 72.78 | 11.06% more |
+| GeForce RTX 4090 eGPU | N/A | N/A | N/A |
+
+- TensorRT-LLM used marginally more average VRAM utilization at peak utilization vs. llama.cpp (up to 14%). Though this could have interesting implications on consuming more electricity over time.
+- Note: we used comparable (but not identical) quantizations, and TensorRT-LLM’s `AWQ INT4` is implemented differently from llama.cpp’s `q4_k_m`
+
+### Max RAM Usage
+
+| Max RAM utilization | TensorRT-LLM | Llama.cpp | % Difference |
+| ---------------------------- | ------------ | --------- | ---------------- |
+| GeForce RTX 4090 desktop GPU | ✅ 4.98 | 7.11 | ✅ 29.88% less |
+| GeForce RTX 3090 desktop GPU | ✅ 0.98 | 2.60 | ✅ 62.41% less |
+| GeForce RTX 4070 Laptop GPU | ✅ 1.04 | 4.44 | ✅ 76.55%% less |
+| GeForce RTX 4090 eGPU | ✅ 4.11 | 5.38 | ✅ 23.61% less |
+
+TensorRT-LLM uses a lot less Max RAM vs. llama.cpp on slower connection (PCIe 3.0 or Thunderbolt 3) due to better memory management and efficient delegation to VRAM. On faster connection, it’s at least equal to llama.cpp.
+
+### Compiled Model Size and Number of Files
+- Contrary to popular belief, TensorRT-LLM prebuilt models turned out to not be that huge
+- Mistral 7b int4 was actually 25% smaller in TensorRT-LLM, at 3.05gb vs. 4.06gb
+- Note: These are approximate comparisons, as TensorRT-LLM’s AWQ INT4 is implemented differently from llama.cpp’s q4_k_m
+- The bigger takeaway is that the Compiled model sizes are roughly in the same ballpark, while the number of files for TensorRT-LLM is 7x the GGUF number of files.
+
+| Model size (Lower is better) | TensorRT-LLM AWQ int4 | Llama.cpp GGUF Q4 | % Difference |
+| ---------------------------- | --------------------- | ----------------- | ----------------- |
+| Mistral 7B | ✅ 3.05GB | 4.06GB | ✅ 24.88% smaller |
+
+### Convenience
+- Llama.cpp still wins on cross-platform versatility and convenience of a “compile once, run everywhere” approach
+- TensorRT-LLM still requires compilation to specific OS and architecture, though this could be solved by pre-compiling and publishing models on [Nvidia's NGC Model Catalog](https://catalog.ngc.nvidia.com/orgs/nvidia/teams/ai-foundation/collections/codellama)
+
+### Accessibility
+- Llama.cpp unsurprisingly beats TensorRT-LLM in terms of accessibility
+- TensorRT-LLM does not support older NVIDIA GPUs and won’t work well on smaller VRAM cards (e.g. 2-4gb VRAM)
+
+## Final Notes
+
+Our benchmarking is not perfect. We evaluated over a dozen tools ([llmperf](https://github.com/ray-project/llmperf), [psutil](https://pypi.org/project/psutil/), [gpustat](https://github.com/wookayin/gpustat), native utilities, and more) and found that everyone measures TPS, common metrics differently. We eventually settled on using our own tools in Jan, which are consistent across any inference engine and hardware. As for runtime parameters, we went with default settings, likely representing the typical user experience.
+
+We also did not overclock for this benchmark , as it is not a default setting for most users. But we've measured in our tests that TensorRT-LLM can go even faster with a few tweaks. We see this as a pretty exciting future direction.
+
+
+
+We're also publishing the underlying [raw experimental data](https://drive.google.com/file/d/1rDwd8XD8erKt0EgIKqOBidv8LsCO6lef/view?usp=sharing), and would encourage the community to scrutinize and help us improve.
+
+Special thanks to Asli Sabanci Demiroz, Annamalai Chockalingam, Jordan Dodge from Nvidia, and Georgi Gerganov from llama.cpp for feedback, review and suggestions.
diff --git a/website/src/content/blog/bitdefender.mdx b/website/src/content/blog/bitdefender.mdx
new file mode 100644
index 000000000..27a28e4b2
--- /dev/null
+++ b/website/src/content/blog/bitdefender.mdx
@@ -0,0 +1,149 @@
+---
+title: 'Bitdefender False Positive Flag'
+description: "10th January 2024, Jan's 0.4.4 Release on Windows triggered Bitdefender to incorrectly flag it as infected with Gen:Variant.Tedy.258323, leading to automatic quarantine warnings on users' computers."
+date: 2024-01-10
+tags: postmortem, bitdefender
+categories: building-jan
+keywords:
+ [
+ postmortem,
+ bitdefender,
+ false positive,
+ antivirus,
+ jan,
+ nitro,
+ incident,
+ incident response,
+ supply chain security,
+ user communication,
+ documentation,
+ antivirus compatibility,
+ cross-platform testing,
+ proactive incident response,
+ user education,
+ lessons learned,
+ ]
+---
+
+import CTABlog from '@/components/Blog/CTABlog.astro';
+
+
+# Bitdefender False Positive Flag
+
+Following the recent incident related to Jan version 0.4.4 triggering Bitdefender on Windows with Gen:Variant.Tedy.258323 on January 10, 2024, we wanted to provide a comprehensive postmortem and outline the necessary follow-up actions.
+
+## Incident Overview
+
+### Bug Description
+
+Jan 0.4.4 installation on Windows triggered Bitdefender to flag it as infected with Gen:Variant.Tedy.258323, leading to automatic quarantine.
+
+### Affected Antivirus
+
+- McAfee / Microsoft Defender was unaffected
+- Bitdefender consistently flagged the issue.
+
+### Incident Timeline
+
+- _10 Jan, 2:18 am SGT:_ Hawke flags up Malware antivirus errors for 0.4.4 installation on Windows computers.
+- _10 Jan, 2:21 am SGT:_ @0xSage responds in Discord.
+- _10 Jan, 2:35 am SGT:_ Hawke confirms multiple people have experienced this error on fresh installs.
+- _10 Jan, 2:41 am SGT:_ @louis-jan and @dan-jan revert 0.4.4 out of an abundance of caution.
+- _Incident ongoing:_ To triage and investigate the next day.
+- _10 Jan, 11:36 am SGT:_ @Hien has investigated all versions of Nitro and conducted scans using Bitdefender. Only the 2 latest versions raised warnings (0.2.7, 0.2.8).
+- _10 Jan, 12:44 pm SGT:_ @Hien tested again for the 0.2.6 and suggested using 0.2.6 for now, the 2 remaining Nitro version (0.2.7, 0.2.8) will under further investigation.
+- The team started testing on the fixed build.
+- _10 Jan, 3:22 pm SGT:_ Diagnosis found that it's most likely a false positive. @Hien has only found a solution by attempting to build Nitro Windows CPU on a GitHub-hosted runner and hasn't identified the root cause yet.
+- _10 Jan, 5:24 pm SGT:_ @Hien testing two scenarios and still trying to understand the workings of Bitdefender.
+- _11 Jan, 5:46 pm SGT:_ Postmortem meeting
+
+## Investigation Update
+
+- @Hien has investigated all versions of Nitro and conducted scans using Bitdefender. and only the 2 latest versions raised warnings from Bitdefender. Nitro 0.2.6, which is the highest version without the issue, was tested again, and it no longer triggers a warning from Bitdefender.
+- We have observed that Nitro versions up to 0.2.6 remain unaffected. However, Bitdefender flags versions 0.2.7 and 0.2.8 as infected, leading to the deletion. In order to proceed with the current release, Hien suggests downgrading Nitro to version 0.2.6 and conducting tests with this version. Simultaneously, he will investigate why Bitdefender is flagging versions 0.2.7 and 0.2.8.
+- It's essential to note that between versions 0.2.6, 0.2.7, and 0.2.8, only minor changes were made, which should not trigger a malicious code warning. We can refer to the changelog between 0.2.7 and 0.2.8 to pinpoint these changes.
+- Our primary message is to convey that we did not introduce malicious code into Jan (indicating a false positive), and the investigation aims to understand the root cause behind Bitdefender flagging versions 0.2.7 and 0.2.8.
+- The current diagnosis looks like a false positive but it's still under investigation. Reference link: [here](https://stackoverflow.com/questions/75886428/fake-positive-bit-defender-problem-genvariant-tedy-304469), [here](https://stackoverflow.com/questions/58010466/bitdefender-detects-my-console-application-as-genvariant-ursu-56053), and [here](https://www.cisa.gov/sites/default/files/2023-06/mar-10365227.r1.v1.clear_.pdf).
+- @Hien testing two scenarios and still trying to understand the workings of Bitdefender. Still under investigation: is the issue with the code or the CI?
+ - In Case 1, using the same CI agent for tags 0.2.6 and 0.2.8, after PRs by Alan and myself, Bitdefender flagged the Nitro CPU binary build. Naturally, one would conclude this is due to the code.
+ - However, I proceeded with a further experiment: for the 0.2.8 code, instead of using our CI agent, I used a GitHub hosted agent. This time, Bitdefender did not flag our binary build.
+- We've identified the Bitdefender warning was not an attack. There is no malicious code
+- We've isolated the event to originate from a CI agent, which resulted in a BitDefender false positive alert.
+
+## Follow-ups and Action Items
+
+1. **Reproduce Bitdefender Flag in Controlled Environment [Done]:**
+
+ - _Objective:_ To replicate the issue in a controlled environment to understand the triggers and specifics of Bitdefender's detection.
+
+2. **Investigate Malicious Code or False Positive:**
+
+ - _Objective:_ Determine whether the flagged issue is a result of actual malicious code or a false positive. If it's a false positive, work towards resolution while communicating with Bitdefender.
+
+3. **Supply Chain Attack Assessment:**
+
+ - _Objective:_ Evaluate the possibility of a supply chain attack. Investigate whether the Nitro 0.4.4 distribution was compromised or tampered with during the release process.
+
+4. **Testing after the Hotfix:**
+
+ - _Objective:_ In addition to verifying the issue after the fix, it is essential to conduct comprehensive testing across related areas, ensuring compatibility across different operating systems and antivirus software (latest version / free version only).
+
+5. **Process Improvement for Future Releases:**
+
+ - _Objective:_ Identify and implement improvements to our release process to prevent similar incidents in the future. This may include enhanced testing procedures, code analysis, and collaboration with antivirus software providers during the pre-release phase. Additionally, we should add verifying the latest antivirus software in the release checklist.
+
+6. **Documentation of Tested Antivirus Versions:**
+ - _Objective:_ Create a document that outlines the testing conducted, including a matrix that correlates Jan versions with the tested antivirus versions.
+ - _Sample list:_ for consideration purpose
+ - Bitdefender
+ - McAfee
+ - Avira
+ - Kaspersky
+ - Norton
+ - Microsoft defender
+ - AVG
+ - TotalAV
+
+## Next Steps
+
+- The team should follow up on each action item with clear ownership priority, and deadlines.
+- Communicate progress transparently with the community and clients through appropriate channels. If any insights or suggestions, share them within the dedicated channels.
+- Update internal documentation and procedures based on the lessons learned from this incident.
+
+## Lessons Learned
+
+1. **Antivirus Compatibility Awareness:**
+
+ - _Observation:_ The incident underscored the significance of recognizing and testing for antivirus compatibility, particularly with widely-used solutions like Bitdefender.
+ - _Lesson Learned:_ In the future, we will integrate comprehensive checks for compatibility with various antivirus software, including both antivirus and "Malicious Code Detection," into our CI or QA checklist. This proactive measure aims to minimize false positive detections during the release and testing processes.
+
+2. **Cross-Platform Testing:**
+
+ - _Observation:_ The problem did not occur on MacOS and Linux systems, implying a potential oversight in cross-platform testing during our release procedures.
+ - _Lesson Learned:_ Clarification — This observation is not directly related to antivirus testing. Instead, it underscores the necessity to improve our testing protocols, encompassing multiple operating systems. This ensures a thorough evaluation of potential issues on diverse platforms, considering the various antivirus software and differences in architectures on Mac and Linux systems.
+
+3. **User Communication and Documentation:**
+
+ - _Observation:_ Due to the timely response from Nicole, who was still active on Discord and Github at 2 am, this quick response facilitated our ability to assess the impact accurately.
+ - _Lesson Learned:_ While our communication with users was effective in this instance, it was mainly due to Nicole's presence during the incident. To improve our overall response capability, we should prioritize "24/7 rapid triage and response." This involves ensuring continuous availability or establishing a reliable rotation of team members for swift user communication and issue documentation, further enhancing our incident response efficiency.
+
+4. **Proactive Incident Response:**
+
+ - _Observation:_ The incident response, while involving a prompt version rollback, experienced a slight delay due to the release occurring at midnight. This delay postponed the initiation of the investigation until the next working hours.
+ - _Lesson Learned:_ Recognizing the importance of swift incident response, particularly in time-sensitive situations, we acknowledge that releasing updates during off-hours can impact the immediacy of our actions. Moving forward, we will strive to optimize our release schedules to minimize delays and ensure that investigations can commence promptly regardless of the time of day. This may involve considering alternative release windows or implementing automated responses to critical incidents, ensuring a more proactive and timely resolution.
+
+5. **Supply Chain Security Measures:**
+
+ - _Observation:_ While the incident prompted consideration of a potential supply chain attack, it's crucial to emphasize that this was not the case. Nonetheless, the incident underscored the importance of reviewing our supply chain security measures.
+ - _Lesson Learned:_ Going forward, we should strengthen supply chain security by introducing additional verification steps to uphold the integrity of our release process. Collaborating with distribution channels is essential for enhancing security checks and ensuring a robust supply chain.
+ - _Longer-term:_ Exploring options for checking Jan for malicious code and incorporating antivirus as part of our CI/CD pipeline should be considered for a more comprehensive and proactive approach.
+
+6. **User Education on False Positives:**
+ - _Observation:_ Users reported Bitdefender automatically "disinfecting" the flagged Nitro version without allowing any user actions.
+ - _Lesson Learned:_ Educate users about the possibility of false positives and guide them on how to whitelist or report such incidents to their antivirus provider (if possible). Provide clear communication on steps users can take in such situations.
+
+These lessons learned will serve as a foundation for refining our processes and ensuring a more resilient release and incident response framework in the future. Continuous improvement is key to maintaining the reliability and security of our software.
+
+Thank you for your dedication and cooperation in resolving this matter promptly.
+
+
\ No newline at end of file
diff --git a/website/src/content/blog/data-is-moat.mdx b/website/src/content/blog/data-is-moat.mdx
new file mode 100644
index 000000000..5e238103a
--- /dev/null
+++ b/website/src/content/blog/data-is-moat.mdx
@@ -0,0 +1,116 @@
+---
+title: "The Invisible Moat around Open-Source LLM"
+description: "Uncover the pivotal role of data ownership in training the next iteration of LLM."
+tags: OpenAI has a moat, Catastrophic forgetting, ChatGPT
+date: 2024-03-25
+unlisted: true
+categories: research
+---
+
+import CTABlog from '@/components/Blog/CTABlog.astro';
+
+import catastrophicDemo from '@/assets/blog/catastrophic-demo.png';
+import gradientDecent from '@/assets/blog/gradient-decent.gif';
+import replay from '@/assets/blog/replay.png';
+import openchatBench0106 from '@/assets/blog/openchat-bench-0106.png';
+
+
+# The Invisible Moat around Open-Source LLM
+
+In the crowded AI landscape, OpenAI's ChatGPT stands out, not just for its capabilities but for its unique access to the pre-trained dataset. This post explores the vital role of data in maintaining a competitive edge, focusing on OpenAI's strategic advantage through data ownership.
+
+## Data: The Secret Weapon
+OpenAI, with ChatGPT, has carved a distinct advantage. By harnessing user interactions, it gains invaluable insights into diverse use cases, enabling precise model refinements. The cornerstone of this advantage lies in the "pre-trained dataset." This treasure trove of data empowers OpenAI to cater to specific needs, ensuring sustained improvement and differentiation.
+
+## The rise of the opensource
+
+```
+- How they/Mistral/Llama make money?
+-> around having pretrained data -> finetuning
+First para:
+Rise of Open Source LLMs like Mistral, Llama2, Llama3
+People think they don't have a moat = everything is open source
+Second para:
+We actually think these guys have an "invisible moat"
+Pre-training data is not released, and makes a huge difference in fine-tuning efficacy
+```
+
+### Why pretrained data is important?
+
+> *Owning the pre-trained dataset is crucial as it represents the original distribution.*
+Access to the pre-trained dataset acts as a master key to address the critical issue of ["Catastrophic forgetting"](https://en.wikipedia.org/wiki/Catastrophic_interference) in Language Learning Models (LLMs). This phenomenon describes how LLMs lose hold of prior knowledge upon learning new information. Access to the foundational dataset allows for effective fine-tuning, balancing the introduction of new data with the retention of existing knowledge.
+
+
+
+**Figure 1.** Demonstrates the catastrophic forgetting issue: without mixing datasets, AI overfits on new tasks, impairing normal communication.
+
+### Illustrating Catastrophic Forgetting
+
+```
+What is fine-tuning
+Process of Finetuning (pretrain, instruct, finetune)
+Fine-tuning datasets
+Risk of catastrophic forgetting
+"Why is Pre-trained data important?"
+What is pre-training dataset
+How does fine-tuning with pre-training dataset differ from when you don't have it
+How does it avoid catastrophic forgetting
+```
+
+Catastrophic forgetting can be visualized as a ball in a multidimensional landscape, where moving towards new knowledge risks losing grasp on the old.
+Pre-trained data acts as a map, guiding fine-tuning in a way that incorporates new information while safeguarding existing knowledge.
+
+
+
+**Figure 2.** [Gradient decent demonstration](https://en.wikipedia.org/wiki/Gradient_descent)
+
+### Smoothing Distribution Shifts
+
+As described above, with the mixture of the pre-trained dataset ensures smoother distribution shifts when introducing new information, as it embodies a comprehensive spectrum of prior knowledge.
+
+This continuity in knowledge transition helps in maintaining the robustness of the model against sudden changes, akin to providing a more gradual learning curve where the new information is incrementally integrated with the existing knowledge base.
+
+This concept is supported by the [EleutherAI's research](https://arxiv.org/abs/2403.08763) highlighting the importance of how tasks are sequenced in the learning process, suggesting that introducing dissimilar tasks early on can expand the network's capacity for new information.
+
+**Table 1.** Final results for English-only 405M parameter models trained with different replay amounts show models with more replay perform better in balancing learning and forgetting (measured as AVG Loss). Notably, just 1% mix with a pre-trained dataset significantly lowers AVG loss, effectively shifting model knowledge from English (the Pile) to German.
+
+
+
+*Note:* **Replay** is the method involves combining the training dataset from the pre-trained model with new task datasets.
+
+### Acting as a Noise Mask
+
+The pre-trained data can also serve as a form of "noise masking", similar to techniques used in training [early computer vision models](https://arxiv.org/abs/1911.04252).
+
+This approach introduces a level of ["noise"](https://arxiv.org/abs/2310.05914) during training, which can prevent the model from overfitting to the new dataset. By retaining a mix of original and new data, the model is exposed to a broader range of scenarios, enhancing its generalization capabilities and robustness across tasks.
+
+## Solutions
+
+### Overwhelming approach
+
+Overcoming these challenges requires a balanced approach. One partial method involves inundating the model with extensive, curated data, allowing for comprehensive fine-tuning. While effective, this approach demands significant computational resources, a comprehensive filtering process for low-quality inputs, and an extraordinarily high cost associated with gathering millions of high-quality responses.
+
+In the open-source community, 2 notable examples of fine-tuning with Mistral as a base model on large datasets collected from top-rated GPT-4 and human responses demonstrate a distribution shift that enhances model performance, including [OpenChat](https://huggingface.co/openchat/openchat-3.5-0106) and [Hermes-Pro](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B).
+
+
+
+**Figure 2.** After fine-tuning with a large amount of data samples, the model's performance improved, outperforming ChatGPT and Grok-1 in some benchmarks.
+
+### Fully open source model
+
+- Example: Dolma + olma from allenai
+
+## Conclusion
+
+The ownership and strategic use of pre-trained data serve as an invisible moat. It not only enables the tackling of complex challenges like catastrophic forgetting but also provides a baseline for continuous, targeted improvements. Although there is a solution to decentralize, the cost remains reasonably high.
+
+Fully open pretrained + open weight
+
+## Reference
+- [Catastrophic forgetting](https://arxiv.org/abs/2308.08747)
+- [Simple and Scalable Strategies to Continually Pre-train Large Language Models](https://arxiv.org/abs/2403.08763)
+- [Gradient descent](https://en.wikipedia.org/wiki/Gradient_descent)
+- [Neftune](https://arxiv.org/abs/2310.05914)
+- [Self-training with Noisy Student improves ImageNet classification](https://arxiv.org/abs/1911.04252)
+
+
\ No newline at end of file
diff --git a/website/src/content/blog/deepseek-r1-locally.mdx b/website/src/content/blog/deepseek-r1-locally.mdx
new file mode 100644
index 000000000..5c7ae8f8d
--- /dev/null
+++ b/website/src/content/blog/deepseek-r1-locally.mdx
@@ -0,0 +1,141 @@
+---
+title: "Run DeepSeek R1 locally on your device (Beginner-Friendly Guide)"
+description: "A straightforward guide to running DeepSeek R1 locally regardless of your background."
+tags: DeepSeek, R1, local AI, Jan, GGUF, Qwen, Llama
+categories: guides
+date: 2025-01-31
+ogImage: assets/deepseek-r1-locally-jan.jpg
+twitter:
+ card: summary_large_image
+ site: "@jandotai"
+ title: "Run DeepSeek R1 locally on your device (Beginner-Friendly Guide)"
+ description: "A straightforward guide to running DeepSeek R1 locally regardless of your background."
+ image: assets/deepseek-r1-locally-jan.jpg
+---
+
+import CTABlog from '@/components/Blog/CTABlog.astro';
+import { Aside } from '@astrojs/starlight/components'
+
+
+import deepseekR1LocallyJan from '@/assets/blog/deepseek-r1-locally-jan.jpg';
+import downloadJan from '@/assets/blog/download-jan.jpg';
+import janLibraryDeepseekR1 from '@/assets/blog/jan-library-deepseek-r1.jpg';
+import janHubDeepseekR1 from '@/assets/blog/jan-hub-deepseek-r1.jpg';
+import janRunsDeepseekR1Distills from '@/assets/blog/jan-runs-deepseek-r1-distills.jpg';
+
+
+# Run DeepSeek R1 locally on your device (Beginner-Friendly Guide)
+
+
+
+DeepSeek R1 is one of the best open-source models in the market right now, and you can run DeepSeek R1 on your own computer!
+
+
+
+DeepSeek R1 requires data-center level computers to run at its full potential, and we'll use a smaller version that works great on regular computers.
+
+Why use an optimized version?
+- Efficient performance on standard hardware
+- Faster download and initialization
+- Optimized storage requirements
+- Maintains most of the original model's capabilities
+
+## Quick Steps at a Glance
+1. Download [Jan](https://jan.ai/)
+2. Select a model version
+3. Choose settings
+4. Set up the prompt template & start using DeepSeek R1
+
+Let's walk through each step with detailed instructions.
+
+## Step 1: Download Jan
+[Jan](https://jan.ai/) is an open-source application that enables you to run AI models locally. It's available for Windows, Mac, and Linux. For beginners, Jan is the best choice to get started.
+
+
+
+1. Visit [jan.ai](https://jan.ai)
+2. Download the appropriate version for your operating system
+3. Install the app
+
+## Step 2: Choose Your DeepSeek R1 Version
+
+To run AI models like DeepSeek R1 on your computer, you'll need something called VRAM (Video Memory). Think of VRAM as your computer's special memory for handling complex tasks like gaming or, in our case, running AI models. It's different from regular RAM - VRAM is part of your graphics card (GPU).
+
+
+
+Let's first check how much VRAM your computer has. Don't worry if it's not much - DeepSeek R1 has versions for all kinds of computers!
+
+Finding your VRAM is simple:
+- On Windows: Press `Windows + R`, type `dxdiag`, hit Enter, and look under the "Display" tab
+- On Mac: Click the Apple menu, select "About This Mac", then "More Info", and check under "Graphics/Displays"
+- On Linux: Open Terminal and type `nvidia-smi` for NVIDIA GPUs, or `lspci -v | grep -i vga` for other graphics cards
+
+
+
+Once you know your VRAM, here's what version of DeepSeek R1 will work best for you. If you have:
+- 6GB VRAM: Go for the 1.5B version - it's fast and efficient
+- 8GB VRAM: You can run the 7B or 8B versions, which offer great capabilities
+- 16GB or more VRAM: You have access to the larger models with enhanced features
+
+Available versions and basic requirements for DeepSeek R1 distills:
+
+| Version | Model Link | Required VRAM |
+|---------|------------|---------------|
+| Qwen 1.5B | [DeepSeek-R1-Distill-Qwen-1.5B-GGUF](https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-1.5B-GGUF) | 6GB+ |
+| Qwen 7B | [DeepSeek-R1-Distill-Qwen-7B-GGUF](https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-7B-GGUF) | 8GB+ |
+| Llama 8B | [DeepSeek-R1-Distill-Llama-8B-GGUF](https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-8B-GGUF) | 8GB+ |
+| Qwen 14B | [DeepSeek-R1-Distill-Qwen-14B-GGUF](https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-14B-GGUF) | 16GB+ |
+| Qwen 32B | [DeepSeek-R1-Distill-Qwen-32B-GGUF](https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF) | 16GB+ |
+| Llama 70B | [DeepSeek-R1-Distill-Llama-70B-GGUF](https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-70B-GGUF) | 48GB+ |
+
+To download your chosen model:
+
+Launch Jan and navigate to Jan Hub using the sidebar
+
+
+
+3. Input the model link in this field:
+
+
+
+## Step 3: Configure Model Settings
+When configuring your model, you'll encounter quantization options:
+
+
+
+## Step 4: Configure Prompt Template
+Final configuration step:
+
+1. Access Model Settings via the sidebar
+2. Locate the Prompt Template configuration
+3. Use this specific format:
+
+
+
+This template is for proper communication between you and the model.
+
+You're now ready to interact with DeepSeek R1:
+
+
+
+## Need Assistance?
+
+
+
+
diff --git a/website/src/content/blog/offline-chatgpt-alternative.mdx b/website/src/content/blog/offline-chatgpt-alternative.mdx
new file mode 100644
index 000000000..d6fa07e93
--- /dev/null
+++ b/website/src/content/blog/offline-chatgpt-alternative.mdx
@@ -0,0 +1,125 @@
+---
+title: "Offline ChatGPT: You can't run ChatGPT offline, do this instead"
+description: "Learn how to use AI offline with Jan - a free, open-source alternative to ChatGPT that works 100% offline on your computer."
+tags: AI, ChatGPT alternative, offline AI, Jan, local AI, privacy
+categories: guides
+date: 2025-02-08
+ogImage: _assets/offline-chatgpt-alternatives-jan.jpg
+twitter:
+ card: summary_large_image
+ site: "@jandotai"
+ title: "Offline ChatGPT: You can't run ChatGPT offline, do this instead"
+ description: "Want to use ChatGPT offline? Learn how to run AI models locally with Jan - free, open-source, and works without internet."
+ image: _assets/offline-chatgpt-alternatives-jan.jpg
+---
+
+import CTABlog from '@/components/Blog/CTABlog.astro';
+import { Aside } from '@astrojs/starlight/components'
+
+
+import offlineChatgptAlternativeAiWithoutInternet from '@/assets/blog/offline-chatgpt-alternative-ai-without-internet.jpg';
+
+
+# Offline ChatGPT: You can't run ChatGPT offline, do this instead
+
+ChatGPT is a cloud-based service that requires internet access. However, it's not the only way to use AI. You can run AI models offline on your device with [Jan](https://jan.ai/). It's completely free, open-source, and gives you 100% offline capability. You can even use AI on a plane!
+
+
+
+## Jan as an offline ChatGPT alternative
+
+
+*Jan lets you use AI offline - no internet connection needed*
+
+Here's how to get started with offline AI in 3 simple steps:
+
+### 1. Download Jan
+
+Go to [jan.ai](https://jan.ai) and download the version for your computer (Mac, Windows, or Linux). It's completely free.
+
+
+
+### 2. Download an AI model
+
+You'll need an AI model to use AI offline, so download a model from Jan. Once it's on your computer, you don't need internet anymore.
+
+
+*Select an AI model that matches your needs and computer capabilities*
+
+
+
+### 3. Start using AI offline
+
+
+*Use Jan's clean interface to chat with AI - no internet required*
+
+Once downloaded, you can use AI anywhere, anytime:
+
+- Chat like you do with ChatGPT
+- Work on documents offline
+- Get coding help without internet
+- Keep your conversations private
+- Use AI even when servers are down
+
+## How to chat with your docs in Jan?
+
+To chat with your docs in Jan, you need to activate experimental mode.
+
+
+*Turn on experimental mode in settings to chat with your docs*
+
+After activating experimental mode, simply add your files and ask questions about them.
+
+
+*Chat with your documents privately - no internet needed*
+
+I did this for you and got a reply from a 7B parameter model. If you'd like to learn what "7B" means and understand other local AI terms, check our [guide on running AI models locally](/blog/run-ai-models-locally).
+
+A response from AI, Qwen2.5 7B Instruct Q4:
+
+`This document appears to be about the benefits and advantages of running artificial intelligence (AI) models locally on your device rather than using cloud-based or remote AI services. The key points it highlights include data privacy, offline functionality, freedom from paywalls and restrictions, and giving users full control over their AI models. Additionally, the text mentions that local AI is becoming a new trend and provides a link to a guide for beginners who want to learn more about this topic.`
+
+Local AI makes possible offline AI use, so Jan is going to be your first step to get started.
+
+## Why choose Jan over ChatGPT?
+
+1. **True Offline Use:** Unlike ChatGPT, Jan works without internet
+2. **100% Private:** Your data never leaves your computer
+3. **Free Forever:** No subscriptions or API costs
+4. **No Server Issues:** No more "ChatGPT is at capacity"
+5. **Your Choice of Models:** Use newer models as they come out
+
+**"Is it really free? What's the catch?"**
+Yes, it's completely free and open source. Jan is built by developers who believe in making AI accessible to everyone.
+
+**"How does it compare to ChatGPT?"**
+Modern open-source models like DeepSeek and Mistral are very capable. While they might not match GPT-4, they're perfect for most tasks and getting better every month.
+
+**"Do I need a powerful computer?"**
+If your computer is from the last 5 years, it will likely work fine. You need about 8GB of RAM and 10GB of free space for comfortable usage.
+
+**"What about my privacy?"**
+Everything stays on your computer. Your conversations, documents, and data never leave your device unless you choose to share them.
+
+Want to learn more about the technical side? Check our detailed [guide on running AI models locally](/blog/run-ai-models-locally). It's not required to [use AI offline](https://jan.ai/) but helps understand how it all works.
+
+## Need help?
+
+
+
+
diff --git a/website/src/content/blog/qwen3-settings.mdx b/website/src/content/blog/qwen3-settings.mdx
new file mode 100644
index 000000000..0fee2ecc0
--- /dev/null
+++ b/website/src/content/blog/qwen3-settings.mdx
@@ -0,0 +1,131 @@
+---
+title: "Best Settings to Run Qwen3-30B-A3B Locally"
+description: "If you're running Qwen3-30B-A3B locally, don't guess your way through the settings. This guide tells you what actually works based on Qwen's own documentation and what we've seen hold up in practice."
+tags: Qwen3, local AI, model settings, Jan, offline AI
+categories: guides
+date: 2025-05-10
+ogImage: assets/images/general/qwen3-30b-settings.jpg
+---
+
+import CTABlog from '@/components/Blog/CTABlog.astro';
+import { Aside } from '@astrojs/starlight/components'
+
+
+import qwen3SettingsJanAi from '@/assets/blog/qwen3-settings-jan-ai.jpeg';
+import qwen3InJanHub from '@/assets/blog/qwen3-in-jan-hub.jpeg';
+import qwen3SettingsInJan from '@/assets/blog/qwen3-settings-in-jan.jpeg';
+
+
+# Best Settings to Run Qwen3-30B-A3B Locally
+
+If you're running Qwen3-30B-A3B locally, don't guess your way through the settings. This guide tells you what actually works based on Qwen's own documentation and what we've seen hold up in practice.
+
+
+
+Qwen3 comes with a unique toggle: `enable_thinking`. When it's on, the model "thinks", it breaks down problems, reasons step-by-step, and wraps part of its output in a `...` block. When it's off, the model skips all that and just gives you an answer.
+
+That changes how you configure it.
+
+---
+
+### Thinking mode (`enable_thinking=True`)
+
+This is the mode for reasoning, math, coding, logic — anything that benefits from step-by-step generation.
+
+**Use these generation settings:**
+
+```
+Temperature: 0.6
+TopP: 0.95
+TopK: 20
+Max tokens: 32,768
+Do not use greedy decoding
+```
+
+
+
+
+## Quick summary
+
+
+
+### Non-thinking mode (`enable_thinking=False`)
+
+This is for fast, general-purpose replies. Instruction following, chat, creative writing — no `` block, no extra steps.
+
+**Use these settings:**
+
+```makefile
+Temperature: 0.7
+TopP: 0.8
+TopK: 20
+```
+
+
+
+
+
+## Soft vs. hard switch
+
+You can toggle thinking dynamically in the prompt using:
+
+```
+/think # turns thinking ON
+/no_think # turns it OFF
+```
+
+This works only if `enable_thinking=True` is set in the code. If you set it to False, the soft switch won't do anything.
+
+
+
+### What most people miss
+
+- **Don't log the `think` block in chat history.** Qwen recommends keeping only the final answer. Otherwise, the next reply gets bloated and off-topic.
+- **Greedy decoding is a trap.** It's tempting to use for consistency, but Qwen3's output gets worse - and sometimes broken - without sampling.
+- **YaRN isn't always needed.** The model supports up to 32k context by default. Use YaRN only if you regularly go beyond that.
+
+---
+
+## Running Qwen3 locally with Jan
+
+The easiest way to run Qwen3-30B-A3B locally is through Jan.
+
+1. Download and install [Jan](https://jan.ai)
+2. Open Jan and navigate to Jan Hub
+3. Find `Qwen3` and `Qwen3-30B-A3B` in the model list
+4. Click "Download" to get the model
+
+### Qwen3 in Jan Hub
+
+You can easily find Qwen3 models in Jan Hub:
+
+
+
+Once downloaded, Jan handles all the technical setup, so you can focus on using the model rather than configuring it. The settings we covered in this guide are automatically applied when you use Qwen3 through Jan.
+
+### How to customize Qwen3-30B-A3B settings in Jan
+
+You can also customize these settings anytime by opening the right panel in Jan and adjusting the parameters to match your needs.
+
+
+
+## Bottom Line
+
+If you're running Qwen3-30B-A3B locally, treat it like two models in one. Flip the thinking mode based on the task, adjust the generation settings accordingly, and let it work how it was meant to.
+
+## Need help?
+
+
diff --git a/website/src/content/blog/rag-is-not-enough.mdx b/website/src/content/blog/rag-is-not-enough.mdx
new file mode 100644
index 000000000..c163b5588
--- /dev/null
+++ b/website/src/content/blog/rag-is-not-enough.mdx
@@ -0,0 +1,134 @@
+---
+title: "RAG is not enough: Lessons from Beating GPT-3.5 on Specialized Tasks with Mistral 7B"
+description: We present a straightforward approach to customizing small, open-source models using fine-tuning and RAG that outperforms GPT-3.5 for specialized use cases.
+tags: RAG, opensource chatgpt alternative, outperform ChatGPT, Mistral
+date: 2024-03-25
+unlisted: true
+categories: research
+---
+
+import CTABlog from '@/components/Blog/CTABlog.astro';
+
+
+# RAG is not enough: Lessons from Beating GPT-3.5 on Specialized Tasks with Mistral 7B
+
+## Abstract
+
+We present a straightforward approach to customizing small, open-source models using fine-tuning and RAG that outperforms GPT-3.5 for specialized use cases. With it, we achieved superior Q&A results of [technical documentation](https://nitro.jan.ai/docs) for a small codebase [codebase](https://github.com/menloresearch/nitro).
+
+In short, (1) extending a general foundation model like [Mistral](https://huggingface.co/mistralai/Mistral-7B-v0.1) with strong math and coding, and (2) training it over a high-quality, synthetic dataset generated from the intended corpus, and (3) adding RAG capabilities, can lead to significant accuracy improvements.
+
+Problems still arise with catastrophic forgetting in general tasks, commonly observed during specialized domain fine-tuning. In our case, this is likely exacerbated by our lack of access to Mistral’s original training dataset and various compression techniques used in our approach to keep the model small.
+
+## Selecting a strong foundation model
+
+[Mistral 7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) outshines both [Meta's Llama-2 7B](https://huggingface.co/meta-llama/Llama-2-7b) and [Google's Gemma 7B](https://huggingface.co/google/gemma-7b) in key benchmarks, making it our choice for a base model. Starting with a strong foundation like Mistral allowed us to achieve greater accuracy in our specialized adaptations.
+
+
+
+*Figure 1. Mistral 7B excels in benchmarks, ranking among the top foundational models.*
+
+*Note: We are not sponsored by the Mistral team, though lots of folks like to run Mistral locally using [Jan](https://jan.ai/)., our desktop client.*
+
+## Cost effectively improving the base model
+
+Our technical use case required excellent math capabilities, an area where Mistral can underperform. Thus, we tested a number of Mistral model variants, from foundation models to finetunes to model merges, to find a stronger base model before proceeding to finetuning.
+
+
+
+*Figure 2: The merged model, Stealth, doubles the mathematical capabilities of its foundational model while retaining the performance in other tasks.*
+
+We found merging models is quick and cost-effective, enabling fast adjustments based on the result of each iteration.
+
+We ended up with [Stealth 7B v1.1](https://huggingface.co/jan-hq/stealth-v1.1), a [SLERP](https://github.com/Digitous/LLM-SLERP-Merge) merge of Mistral with the following:
+
+- [WizardMath](https://huggingface.co/WizardLM/WizardMath-7B-V1.1) for its math capabilities.
+- [WizardCoder](https://huggingface.co/WizardLM/WizardCoder-Python-7B-V1.0) for its coding capabilities.
+- Our own [Trinity](https://huggingface.co/jan-hq/trinity-v1.2) model for its versatility across general tasks.
+
+This particular combination yielded the best tradeoff across mathematical & technical reasoning while retaining the most pre-merge performance on general tasks.
+
+## **DPO finetuning**
+
+Merging different LLMs can lead to a mixed answering style because each model was originally trained on different types of data.
+
+Thus, we applied Direct Preference Optimization ([DPO](https://arxiv.org/abs/2305.18290)) using the [Intel's Orca DPO pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs) dataset, chosen for its helpful answering style in general, math and coding concentration.
+
+This approach produced a final model - [Stealth 7B v1.2](https://huggingface.co/jan-hq/stealth-v1.2), aligned to our technical preferences and demonstrating minimal loss.
+
+## **Using our own technical documentation**
+
+With the base model ready, we started on our specific use case.
+
+Jan is an open-source project enjoying strong growth, but at one point we began receiving a new support ticket every minute, which quickly overwhelmed our bootstrapped resources.
+
+So, we directed our efforts toward training a model to answer user questions based on existing technical documentation.
+
+Specifically, we trained it on on the [Nitro documentation](https://nitro.jan.ai/docs). For context, Nitro is the default inference engine for Jan. It’s a enterprise-ready server implementation of LlamaCPP, written in C++, with multimodal, queues, and other production-level server capabilities.
+
+It made an interesting corpus because it was rife with post-2023 technical jargon, edge cases, and poor informational layout.
+
+## Generating training data
+
+The first step was to transform Nitro’s unstructured format into a synthetic Q&A dataset designed for [instruction tuning](https://arxiv.org/pdf/2109.01652.pdf).
+
+The text was split into chunks of 300-token segments with 30-token overlaps. This helped to avoid a [lost-in-the-middle](https://arxiv.org/abs/2307.03172) problem where LLM can’t use context efficiently to answer given questions.
+
+The chunks were then given to GPT-4 with 8k context length to generate 3800 Q&A pairs. The [training dataset](https://huggingface.co/datasets/jan-hq/nitro_binarized_v2) is available on HuggingFace.
+
+## **Training**
+
+Training was done with supervised finetuning (SFT) from the [Hugging Face's alignment-handbook](https://github.com/huggingface/alignment-handbook), per [Huggingface's Zephyr Beta](https://github.com/huggingface/alignment-handbook/tree/main/recipes/zephyr-7b-beta) guidelines.
+
+We used consumer-grade, dual Nvidia RTX 4090s for the training. The end-to-end training took 18 minutes. We found optimal hyperparameters in LoRA for this specific task to be `r = 256` and `alpha = 512`.
+
+This final model can be found [here on Huggingface](https://huggingface.co/jan-hq/nitro-v1.2-e3).
+
+
+
+*Figure 3. Using the new finetuned model in [Jan](https://jan.ai/)*
+
+## Improving results with RAG
+
+As an additional step, we also added [Retrieval Augmented Generation (RAG)](https://blogs.nvidia.com/blog/what-is-retrieval-augmented-generation/) as an experiment parameter.
+
+A simple RAG setup was done using **[Llamaindex](https://www.llamaindex.ai/)** and the **[bge-en-base-v1.5 embedding](https://huggingface.co/BAAI/bge-base-en-v1.5)** model for efficient documentation retrieval and question-answering. You can find the RAG implementation [here](https://github.com/menloresearch/open-foundry/blob/main/rag-is-not-enough/rag/nitro_rag.ipynb).
+
+## Benchmarking the Results
+
+We curated a new set of [50 multiple-choice questions](https://github.com/menloresearch/open-foundry/blob/main/rag-is-not-enough/rag/mcq_nitro.csv) (MCQ) based on the Nitro docs. The questions had varying levels of difficulty and had trick components that challenged the model's ability to discern misleading information.
+
+
+
+*Figure 4. Comparison between finetuned model and OpenAI's GPT*
+
+**Results**
+
+- GPT-3.5 with RAG: 56.7%
+- GPT-4 with RAG: 64.3%
+- Merged 7B Model ([Stealth 7B](https://huggingface.co/jan-hq/stealth-v1.3)) with RAG: 47.7%
+- Finetuned 7B Model (Nitro 7B) with RAG: 57.8%
+
+This indicates that with task-specific training, we can improve an open-source, Small Language Model to the level of GPT-3.5 on domain knowledge.
+
+Notably, the finetuned + RAG approach also demonstrated more consistency across benchmarking, as indicated by its lower standard deviation.
+
+## Conclusion
+
+We conclude that this combination of model merging + finetuning + RAG yields promise. This finding is relevant for teams and individuals that need specialized, technical small language models that need to run in resource-constrained or highly secured environments, where GPT may not be an option.
+
+Anecdotally, we’ve had some success using this model in practice to onboard new team members to the Nitro codebase.
+
+A full research report with more statistics can be found [here](https://github.com/menloresearch/open-foundry/blob/main/rag-is-not-enough/README.md).
+
+# References
+
+- [Catastrophic forgetting](https://arxiv.org/abs/2308.08747)
+- [Math specialization](https://arxiv.org/abs/2308.09583)
+- [Code specialization](https://arxiv.org/abs/2306.08568)
+- [Search specialization](https://github.com/SciPhi-AI/agent-search)
+- [Evol Instruct](https://github.com/nlpxucan/WizardLM)
+- [Lost in the middle](https://arxiv.org/abs/2307.03172)
+- [Instruction tuning](https://arxiv.org/pdf/2109.01652.pdf)
+
+
\ No newline at end of file
diff --git a/website/src/content/blog/run-ai-models-locally.mdx b/website/src/content/blog/run-ai-models-locally.mdx
new file mode 100644
index 000000000..99934dfab
--- /dev/null
+++ b/website/src/content/blog/run-ai-models-locally.mdx
@@ -0,0 +1,228 @@
+---
+title: "How to run AI models locally as a beginner?"
+description: "A straightforward guide to running AI models locally on your computer, regardless of your background."
+tags: AI, local models, Jan, GGUF, privacy, local AI
+categories: guides
+date: 2025-01-31
+ogImage: assets/run-ai-locally-with-jan.jpg
+twitter:
+ card: summary_large_image
+ site: "@jandotai"
+ title: "How to run AI models locally as a beginner?"
+ description: "Learn how to run AI models locally on your computer for enhanced privacy and control. Perfect for beginners!"
+ image: assets/run-ai-locally-with-jan.jpg
+---
+
+import CTABlog from '@/components/Blog/CTABlog.astro';
+
+import { Aside } from '@astrojs/starlight/components'
+
+
+# How to run AI models locally as a beginner?
+
+Most people think running AI models locally is complicated. It's not. Anyone can run powerful AI models like DeepSeek, Llama, and Mistral on their own computer. This guide will show you how, even if you've never written a line of code.
+
+## Quick steps:
+### 1. Download [Jan](https://jan.ai)
+
+
+*Download Jan from [jan.ai](https://jan.ai) - it's free and open source.*
+
+### 2. Choose a model that fits your hardware
+
+
+*Jan helps you pick the right AI model for your computer.*
+
+### 3. Start using AI locally
+
+That's all to run your first AI model locally!
+
+
+*Jan's easy-to-use chat interface after installation.*
+
+Keep reading to learn key terms of local AI and the things you should know before running AI models locally.
+
+## How Local AI Works
+
+Before diving into the details, let's understand how AI runs on your computer:
+
+
+
+
+*llama.cpp helps millions of people run AI locally on their computers.*
+
+
+
+## Understanding AI Models
+
+Think of AI models like apps on your computer - some are light and quick to use, while others are bigger but can do more things. When you're choosing an AI model to run on your computer, you'll see names like "Llama-3-8B" or "Mistral-7B". Let's break down what this means in simple terms.
+
+
+
+
+*Jan Hub makes it easy to understand different model sizes and versions*
+
+**Good news:** Jan helps you pick the right model size for your computer automatically! You don't need to worry about the technical details - just choose a model that matches what Jan recommends for your computer.
+
+## What You Can Do with Local AI
+
+
+
+## Hardware Requirements
+
+Before downloading an AI model, consider checking if your computer can run it. Here's a basic guide:
+
+**The basics your computer needs:**
+- A decent processor (CPU) - most computers from the last 5 years will work fine
+- At least 8GB of RAM - 16GB or more is better
+- Some free storage space - at least 5GB recommended
+
+### What Models Can Your Computer Run?
+
+| | | |
+|---|---|---|
+| Regular Laptop | 3B-7B models | Good for chatting and writing. Like having a helpful assistant |
+| Gaming Laptop | 7B-13B models | More capable. Better at complex tasks like coding and analysis |
+| Powerful Desktop | 13B+ models | Better performance. Great for professional work and advanced tasks |
+
+
+
+## Getting Started with Models
+
+### Model Versions
+
+When browsing models in Jan, you'll see terms like "Q4", "Q6", or "Q8". Here's what that means in simple terms:
+
+
+
+**Pro tip**: Start with Q4 versions - they work great for most people and run smoothly on regular computers!
+
+### Getting Models from Hugging Face
+
+You'll often see links to "Hugging Face" when downloading AI models. Think of Hugging Face as the "GitHub for AI" - it's where the AI community shares their models. Jan makes it super easy to use:
+
+1. Jan has a built-in connection to Hugging Face
+2. You can download models right from Jan's interface
+3. No need to visit the Hugging Face website unless you want to explore more options
+
+## Setting up your local AI
+
+### Getting Models from Hugging Face
+
+You'll often see links to "Hugging Face" when downloading AI models. Think of Hugging Face as the "GitHub for AI" - it's where the AI community shares their models. This sounds technical, but Jan makes it super easy to use:
+
+1. Jan has a built-in connection to Hugging Face
+2. You can download models right from Jan's interface
+3. No need to visit the Hugging Face website unless you want to explore more options
+
+
+
+### 1. Get Started
+Download Jan from [jan.ai](https://jan.ai) - it sets everything up for you.
+
+### 2. Get an AI Model
+
+You can get models two ways:
+
+#### 1. Use Jan Hub (Recommended):
+ - Click "Download Model" in Jan
+ - Pick a recommended model
+ - Choose one that fits your computer
+
+
+*Use Jan Hub to download AI models*
+
+#### 2. Use Hugging Face:
+
+
+
+##### Step 1: Get the model link
+Find and copy a GGUF model link from [Hugging Face](https://huggingface.co)
+
+
+*Look for models with "GGUF" in their name*
+
+##### Step 2: Open Jan
+Launch Jan and go to the Models tab
+
+
+*Navigate to the Models section in Jan*
+
+##### Step 3: Add the model
+Paste your Hugging Face link into Jan
+
+
+*Paste your GGUF model link here*
+
+##### Step 4: Download
+Select your quantization and start the download
+
+
+*Choose your preferred model size and download*
+
+### Common Questions
+
+**"My computer doesn't have a graphics card - can I still use AI?"**
+
+Yes! It will run slower but still work. Start with 7B models.
+
+**"Which model should I start with?"**
+
+Try a 7B model first - it's the best balance of smart and fast.
+
+**"Will it slow down my computer?"**
+
+Only while you're using the AI. Close other big programs for better speed.
+
+## Need help?
+
+
diff --git a/website/src/content/changelog/2023-12-21-faster-inference-across-platform.mdx b/website/src/content/changelog/2023-12-21-faster-inference-across-platform.mdx
new file mode 100644
index 000000000..00515bd5b
--- /dev/null
+++ b/website/src/content/changelog/2023-12-21-faster-inference-across-platform.mdx
@@ -0,0 +1,17 @@
+---
+title: "Faster inference across: Mac, Windows, Linux, and GPUs"
+version: 0.4.3
+description: ""
+date: 2023-12-21
+ogImage: "/assets/images/changelog/Jan_v0.4.3.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- Custom models: `Trinity`, `Pandora` (great for general use).
+- Faster inference across: Mac, Windows, Linux, and GPUs.
+- Connect to remote OpenAI models like GPT4 via API key.
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-01-16-settings-options-right-panel.mdx b/website/src/content/changelog/2024-01-16-settings-options-right-panel.mdx
new file mode 100644
index 000000000..b48fd4601
--- /dev/null
+++ b/website/src/content/changelog/2024-01-16-settings-options-right-panel.mdx
@@ -0,0 +1,20 @@
+---
+title: "Thread settings options in the right panel"
+version: 0.4.4
+description: ""
+date: 2024-01-16
+ogImage: ""
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- You can now see whether the model is compatible with running on your device.
+- You can switch model mid-threads.
+- More thread settings options in the right panel.
+- CI automation, anti-virus checks.
+- Social media access to Jan's Discord & Github from the app for further user support.
+- Fixed major bugs, more stability.
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-01-29-local-api-server.mdx b/website/src/content/changelog/2024-01-29-local-api-server.mdx
new file mode 100644
index 000000000..d90ae00b2
--- /dev/null
+++ b/website/src/content/changelog/2024-01-29-local-api-server.mdx
@@ -0,0 +1,23 @@
+---
+title: "Local API server"
+version: 0.4.5
+description: ""
+date: 2024-01-29
+ogImage: "/assets/images/changelog/Jan_v0.4.5.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- Local API Server: Experience Jan's dashboard for the local API server to make your data processing smoother and more efficient.
+- HTTP Proxy Support: Now, downloading and connecting are seamless, just like browsing Hugging Face in your browser.
+- Updated Settings Page: Find what you need faster! We've updated the settings page.
+
+### Fixes 💫
+
+- Auto Update: Enjoy smoother updates. We've fixed the glitches.
+- Swagger API Page: Full documentation, no more blanks.
+- GPU for Models: Your imported models now fully leverage GPU power.
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-02-05-jan-data-folder.mdx b/website/src/content/changelog/2024-02-05-jan-data-folder.mdx
new file mode 100644
index 000000000..caf9dec98
--- /dev/null
+++ b/website/src/content/changelog/2024-02-05-jan-data-folder.mdx
@@ -0,0 +1,22 @@
+---
+title: "Jan Data Folder"
+version: 0.4.6
+description: ""
+date: 2024-02-05
+ogImage: "/assets/images/changelog/jan_product_update_feature.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- `Changing Jan Data Folder`: Now, moving your data across different folders is just a click away, making organization simpler. So you can even use an external drive.
+- Factory Settings: You can reset all of Jan's usage data for a fresh start.
+
+### Fixes 💫
+
+- Easily see each thread's last update time, like WhatsApp, keeping you informed.
+- A new loading screen during data migration ensures the app is responsive.
+- Enhanced notifications for clearer feedback on model runs or errors.
diff --git a/website/src/content/changelog/2024-02-10-jan-is-more-stable.mdx b/website/src/content/changelog/2024-02-10-jan-is-more-stable.mdx
new file mode 100644
index 000000000..35e96f35f
--- /dev/null
+++ b/website/src/content/changelog/2024-02-10-jan-is-more-stable.mdx
@@ -0,0 +1,27 @@
+---
+title: "Jan is more stable 👋"
+version: 0.5.5
+description: "Jan supports Llama 3.2 and Qwen 2.5"
+date: 2024-10-02
+ogImage: "/assets/images/changelog/jan-v0.5.5.jpeg"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Highlights 🎉
+
+- Meta's Llama 3.2 and Alibaba's Qwen 2.5 added to the hub
+- Improved starter screen
+- Better local vs. cloud model navigation
+
+Fixes 💫
+
+- Solved GPU acceleration for GGUF models
+- Improved model caching & threading
+- Resolved input & toolbar overlaps
+
+Update your product or download the latest: https://jan.ai
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.5).
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-02-26-home-servers-with-helm.mdx b/website/src/content/changelog/2024-02-26-home-servers-with-helm.mdx
new file mode 100644
index 000000000..768c3f7bd
--- /dev/null
+++ b/website/src/content/changelog/2024-02-26-home-servers-with-helm.mdx
@@ -0,0 +1,24 @@
+---
+title: "Run Jan on your home-servers with Helm"
+version: 0.4.7
+description: ""
+date: 2024-02-26
+ogImage: ""
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- Run Jan on your `home-servers` with `Helm`
+- Use Jan headless or with a Web UI
+- `Intel Arc` & `AMD GPU` support through `Vulkan` & `LlamaCPP`
+
+
+### Features & Fixes 💫
+
+- 48 fixes, refactoring and stability improvements.
+- Conversation threads are auto-summarized & messages are editable.
+- Encountering an error? We've replaced vague alerts with a troubleshooting assistant.
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-03-06-ui-revamp-settings.mdx b/website/src/content/changelog/2024-03-06-ui-revamp-settings.mdx
new file mode 100644
index 000000000..dd48c7fbb
--- /dev/null
+++ b/website/src/content/changelog/2024-03-06-ui-revamp-settings.mdx
@@ -0,0 +1,27 @@
+---
+title: "New UI & Codestral Support"
+version: 0.5.0
+description: "Revamped Jan's UI to make it clearer and more user-friendly"
+date: 2024-06-03
+ogImage: "/assets/images/changelog/jan_v0.5.0.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Revamped Jan's UI to make it clearer and more user-friendly.
+
+- Updated Assistant, Model, and Tools sections
+- Categorized customization options for easier control
+- New settings for models, APIs, and experimental features
+
+## New Model: Codestral
+
+Jan now supports Mistral's new model Codestral. Thanks [Bartowski](https://huggingface.co/bartowski) for the GGUF model. You can download the model from the hub.
+
+## More GGUF models
+
+More GGUF models can run in Jan - we rebased to llama.cpp b3012.Big thanks to [ggerganov](https://github.com/ggerganov)
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.0).
diff --git a/website/src/content/changelog/2024-03-11-import-models.mdx b/website/src/content/changelog/2024-03-11-import-models.mdx
new file mode 100644
index 000000000..1d51fe087
--- /dev/null
+++ b/website/src/content/changelog/2024-03-11-import-models.mdx
@@ -0,0 +1,22 @@
+---
+title: "Import models directly using the UI"
+version: 0.4.8
+description: ""
+date: 2024-03-11
+ogImage: ""
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- Import models directly using the UI
+- A revamped system monitor
+
+### Features & Fixes 💫
+
+- Install Jan using Homebrew on Mac Silicon (thanks to https://github.com/chenrui333 (I'll tag you when I find your Discord handle! :D)).
+- Fixed an HTTP Proxy issue causing download errors.
+- UI Improvements and more.
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-03-19-nitro-tensorrt-llm-extension.mdx b/website/src/content/changelog/2024-03-19-nitro-tensorrt-llm-extension.mdx
new file mode 100644
index 000000000..817c5a091
--- /dev/null
+++ b/website/src/content/changelog/2024-03-19-nitro-tensorrt-llm-extension.mdx
@@ -0,0 +1,18 @@
+---
+title: "Nitro-Tensorrt-LLM Extension"
+version: 0.4.9
+description: ""
+date: 2024-03-19
+ogImage: ""
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- Nitro-Tensorrt-LLM Extension.
+- Update models.json.
+- Move tensorrt executable to the engine.
+
diff --git a/website/src/content/changelog/2024-04-02-groq-api-integration.mdx b/website/src/content/changelog/2024-04-02-groq-api-integration.mdx
new file mode 100644
index 000000000..aabcb0ef1
--- /dev/null
+++ b/website/src/content/changelog/2024-04-02-groq-api-integration.mdx
@@ -0,0 +1,22 @@
+---
+title: "Groq API Integration"
+version: 0.4.10
+description: ""
+date: 2024-04-02
+ogImage: "/assets/images/changelog/jan_update_groq.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- Groq API Integration
+- Enhanced hardware troubleshooting guide
+
+### Features & Fixes 💫
+
+- Improved Jan data folder's functionality.
+- Fixed URI malformed and `stop` parameter error.
+- VRAM-aware model recommendations.
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-04-15-new-mistral-extension.mdx b/website/src/content/changelog/2024-04-15-new-mistral-extension.mdx
new file mode 100644
index 000000000..97fd9541e
--- /dev/null
+++ b/website/src/content/changelog/2024-04-15-new-mistral-extension.mdx
@@ -0,0 +1,18 @@
+---
+title: "New Mistral Extension"
+version: 0.4.11
+description: "Jan has a new Mistral Extension letting you chat with larger Mistral models via Mistral API"
+date: 2024-04-15
+ogImage: "/assets/images/changelog/jan_mistral_api.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+### Highlights 🎉
+
+- Jan has a new `Mistral Extension` letting you chat with larger Mistral models via Mistral API. You can still run smaller Mistral models locally, but now there's a remote option.
+- 3rd party extensions can register their own settings pages in the app without any code.
+- You can now change set API Prefix for the local API server.
+- You can now customize your Assistant's name in Thread Settings.
diff --git a/website/src/content/changelog/2024-04-25-llama3-command-r-hugginface.mdx b/website/src/content/changelog/2024-04-25-llama3-command-r-hugginface.mdx
new file mode 100644
index 000000000..a2cbdc5b3
--- /dev/null
+++ b/website/src/content/changelog/2024-04-25-llama3-command-r-hugginface.mdx
@@ -0,0 +1,31 @@
+---
+title: 'Jan now supports Llama3 and Command R+'
+version: 0.4.12
+description: "Jan has added compatibility with Llama3 & Command R+"
+date: 2024-04-25
+ogImage: "/assets/images/changelog/jan_llama3.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Jan has added compatibility with Meta’s open-source language model, `Llama3`, through the integration with `llamacpp` (thanks to [@ggerganov](https://github.com/ggerganov)).
+
+Additionally, `Command R+` is now supported. It is the first open-source model to surpass GPT-4 on the [LMSys leaderboard](https://chat.lmsys.org/?leaderboard).
+
+
+
+## Import Huggingface models directly
+
+Users can now import Huggingface models into Jan. Simply copy the model’s link from Huggingface and paste it into the search bar on Jan Hub.
+
+
+
+## Enhanced LaTeX understanding
+
+Jan now understands LaTeX, allowing users to process and understand complex mathematical expressions more effectively.
+
+
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.4.12).
diff --git a/website/src/content/changelog/2024-05-20-llamacpp-upgrade-new-remote-models.mdx b/website/src/content/changelog/2024-05-20-llamacpp-upgrade-new-remote-models.mdx
new file mode 100644
index 000000000..531e5948e
--- /dev/null
+++ b/website/src/content/changelog/2024-05-20-llamacpp-upgrade-new-remote-models.mdx
@@ -0,0 +1,31 @@
+---
+title: "Jan now supports more GGUF models"
+version: 0.4.13
+description: "We rebased to llamacpp b2865."
+date: 2024-05-20
+ogImage: "/assets/images/changelog/jan_v0.4.13_update.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+With this release, more GGUF models should work now! We rebased to llamacpp b2865!
+
+## New remote models: Anthropic & Cohere APIs
+
+Jan now supports `Anthropic API` models `Command R` and `Command R+`, along with `Cohere`'s `Claude 3 Opus`, `Claude 3 Sonnet`, and `Claude 3 Haiku`.
+
+## New integrations: Martian and OpenRouter
+
+Jan supports `Martian`, a dynamic LLM router that routes between multiple models and allows users to reduce costs by 20% to 97%. Jan also supports `OpenRouter`, helping users select the best model for each query.
+
+
+
+## GPT-4o Access
+
+Users can now connect to OpenAI's new model GPT-4o.
+
+
+
+For more details, see the [GitHub release notes.](https://github.com/menloresearch/jan/releases/tag/v0.4.13)
diff --git a/website/src/content/changelog/2024-05-28-cohere-aya-23-8b-35b-phi-3-medium.mdx b/website/src/content/changelog/2024-05-28-cohere-aya-23-8b-35b-phi-3-medium.mdx
new file mode 100644
index 000000000..8c976106a
--- /dev/null
+++ b/website/src/content/changelog/2024-05-28-cohere-aya-23-8b-35b-phi-3-medium.mdx
@@ -0,0 +1,19 @@
+---
+title: "Jan now compatible with Aya 23 8B & 35B and Phi-3-Medium"
+version: 0.4.14
+description: "Jan now supports Cohere's Aya 23 8B & 35B and Microsoft's Phi-3-Medium."
+date: 2024-05-28
+ogImage: "/assets/images/changelog/jan-v0-4-14-phi3.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Jan now supports `Cohere`'s new models `Aya 23 (8B)` & `Aya 23 (35B)` and `Microsoft`'s `Phi-3-Medium`.
+
+More GGUF models can run in Jan - we rebased to llama.cpp b2961.
+
+Huge shoutouts to [ggerganov](https://github.com/ggerganov) and contributors for llama.cpp, and [Bartowski](https://huggingface.co/bartowski) for GGUF models.
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.4.14).
diff --git a/website/src/content/changelog/2024-06-21-nvidia-nim-support.mdx b/website/src/content/changelog/2024-06-21-nvidia-nim-support.mdx
new file mode 100644
index 000000000..13191a869
--- /dev/null
+++ b/website/src/content/changelog/2024-06-21-nvidia-nim-support.mdx
@@ -0,0 +1,29 @@
+---
+title: "Jan supports NVIDIA NIM"
+version: 0.5.1
+description: "Jan has integrated NVIDIA NIM and supports Qwen 2 7B"
+date: 2024-06-21
+ogImage: "/assets/images/changelog/jan_nvidia_nim_support.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+## NVIDIA NIM
+
+We've integrated NVIDIA NIM support.
+
+## Qwen 2 7B
+
+You can now access Qwen 2 7B directly in the Jan Hub.
+
+We've updated to llama.cpp b3088 for better performance - thanks to [GG](https://github.com/ggerganov)
+
+## Fixes
+
+- Fixed Anthropic API error
+- Reduced chat font weight (back to normal!)
+- Restored the maximize button
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.1).
diff --git a/website/src/content/changelog/2024-07-15-claude-3-5-support.mdx b/website/src/content/changelog/2024-07-15-claude-3-5-support.mdx
new file mode 100644
index 000000000..d13228a43
--- /dev/null
+++ b/website/src/content/changelog/2024-07-15-claude-3-5-support.mdx
@@ -0,0 +1,35 @@
+---
+title: "Jan supports Claude 3.5 Sonnet"
+version: 0.5.2
+description: "You can run Claude 3.5 Sonnet in Jan"
+date: 2024-07-15
+ogImage: "/assets/images/changelog/jan_supports_claude_3_5.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+## Claude 3.5 Sonnet
+
+We've integrated support for Claude 3.5 Sonnet, you can run Anthropic's latest model in Jan.
+
+Plus, You can now use optional spell check for chats. There's also a new shortcut for app logs in System Monitor.
+
+## Fixes
+
+In this hotfix, we've addressed several issues to improve your Jan experience:
+
+### Gemma 2B Stability
+
+Gemma 2B now runs without any issues.
+
+### Tooltip Hover Functionality
+
+We've restored the tooltip hover functionality, which makes it easier to access helpful information without any glitches.
+
+### Right-click Options for Thread Settings
+
+The right-click options for thread settings are now fully operational again. You can now manage your threads with this fix.
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.2).
diff --git a/website/src/content/changelog/2024-09-01-llama3-1-gemma2-support.mdx b/website/src/content/changelog/2024-09-01-llama3-1-gemma2-support.mdx
new file mode 100644
index 000000000..debe207bf
--- /dev/null
+++ b/website/src/content/changelog/2024-09-01-llama3-1-gemma2-support.mdx
@@ -0,0 +1,26 @@
+---
+title: "v0.5.3 is out with stability improvements!"
+version: 0.5.3
+description: "You can run Llama 3.1 and Gemma 2 in Jan"
+date: 2024-08-29
+ogImage: "/assets/images/changelog/janv0.5.3.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+## Llama 3.1 and Gemma 2 Support
+
+Jan finally supports Meta's Llama 3.1 and Google's Gemma 2. Thanks for the patience folks!
+
+We've been working on stability issues over the last few weeks. Jan is now more stable.
+
+### Additional Notes
+
+- Upgraded the inference engine for better performance
+- Model settings now persist across new threads
+- Fixed the GPU memory utilization bar
+- Some UX and copy improvements
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.3).
diff --git a/website/src/content/changelog/2024-09-17-improved-cpu-performance.mdx b/website/src/content/changelog/2024-09-17-improved-cpu-performance.mdx
new file mode 100644
index 000000000..6951c8647
--- /dev/null
+++ b/website/src/content/changelog/2024-09-17-improved-cpu-performance.mdx
@@ -0,0 +1,35 @@
+---
+title: "Jan is faster now"
+version: 0.5.4
+description: "Jan has faster CPU inference."
+date: 2024-09-17
+ogImage: "/assets/images/changelog/jan-v0.5.4.jpg"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Over the last few weeks, we've been working on improving Jan's stability. Every update helps us make Jan's experience faster and smoother.
+With version 0.5.4, you’ll notice AI running quicker on CPU - better performance all around.
+
+### Model Downloads
+You can now download models directly from the model selector in Threads. No more jumping between different tabs – just pick, download, and get started.
+
+### Fast CPU Inference
+We've addressed the slower inference speeds on CPU, so you'll notice faster processing times, especially when using larger models.
+
+### Model Starts
+We tackled the notorious "model can't start / The specified module could not be found" error.
+Plus, Phi-3 models are now working smoothly even if you're using an outdated version.
+
+### Consistent Warnings
+Performance warnings are now aligned between Model Hub and Threads, giving you more reliable insights no matter where you're working.
+
+### Persistent Thread Settings
+Switching between threads used to reset your instruction settings. That’s fixed now! Your settings will stay intact as you jump between old and new threads.
+
+### Minor UI Tweaks & Bug Fixes
+We’ve also resolved issues with the input slider on the right panel and tackled several smaller bugs to keep everything running smoothly.
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.4).
diff --git a/website/src/content/changelog/2024-10-24-jan-stable.mdx b/website/src/content/changelog/2024-10-24-jan-stable.mdx
new file mode 100644
index 000000000..9acc7f31e
--- /dev/null
+++ b/website/src/content/changelog/2024-10-24-jan-stable.mdx
@@ -0,0 +1,26 @@
+---
+title: "Jan has Stable, Beta and Nightly versions"
+version: 0.5.7
+description: "This release is mostly focused on bug fixes."
+date: 2024-10-24
+ogImage: "/assets/images/changelog/jan-v0.5.7.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Highlights 🎉
+
+- Jan has Stable, Beta and Nightly versions
+- Saving instructions for new threads is now stable
+
+Fixes 💫
+
+- Fixed broken links, hardware issues, and multi-modal download
+- Resolved text overlap, scrolling, and multi-monitor reset problems
+- Adjusted LLava model EOS token and context input
+
+Update your product or download the latest: https://jan.ai
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.7).
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-11-22-jan-bugs.mdx b/website/src/content/changelog/2024-11-22-jan-bugs.mdx
new file mode 100644
index 000000000..3d1020217
--- /dev/null
+++ b/website/src/content/changelog/2024-11-22-jan-bugs.mdx
@@ -0,0 +1,25 @@
+---
+title: "Model downloads & running issues fixed"
+version: 0.5.9
+description: "Jan v0.5.9 is here: fixing what needed fixing."
+date: 2024-11-22
+ogImage: "/assets/images/changelog/jan-v0.5.9.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Jan v0.5.9 is here: fixing what needed fixing
+
+### Highlights 🎉
+
+- Model downloads & running issues fixed
+- Document upload bugs resolved
+- System glitches addressed: Factory Reset, HTTP Proxy, Hugging Face tokens
+- Fixed issues with code blocks in streaming responses
+- Improved the UX of the Local API Server page
+
+Update your product or download the latest: https://jan.ai
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.9).
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-11.14-jan-supports-qwen-coder.mdx b/website/src/content/changelog/2024-11.14-jan-supports-qwen-coder.mdx
new file mode 100644
index 000000000..430316da9
--- /dev/null
+++ b/website/src/content/changelog/2024-11.14-jan-supports-qwen-coder.mdx
@@ -0,0 +1,25 @@
+---
+title: "Jan supports Qwen2.5-Coder 14B & 32B"
+version: 0.5.8
+description: "Jan v0.5.8 is out: Jan supports Qwen2.5-Coder 14B & 32B through Cortex"
+date: 2024-11-14
+ogImage: "/assets/images/changelog/jan-v0.5.8.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Jan v0.5.8 is out: Jan supports Qwen2.5-Coder 14B & 32B through Cortex
+
+### Highlights 🎉
+
+- A new engine: Jan now runs models via [Cortex](https://cortex.so)
+- SupportsAlibaba_Qwen's Coder 14B & 32B Support
+- Supports markdown rendering on user messages
+
+and various UI/UX enhancements 💫
+
+Update your product or download the latest: https://jan.ai
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.8).
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-12-03-jan-is-faster.mdx b/website/src/content/changelog/2024-12-03-jan-is-faster.mdx
new file mode 100644
index 000000000..79f05b12d
--- /dev/null
+++ b/website/src/content/changelog/2024-12-03-jan-is-faster.mdx
@@ -0,0 +1,22 @@
+---
+title: "Jan v0.5.10 is live"
+version: 0.5.10
+description: "Jan is faster, smoother, and more reliable."
+date: 2024-12-03
+ogImage: "/assets/images/changelog/jan-v0.5.10.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Jan v0.5.10 is live: Jan is faster, smoother, and more reliable.
+
+### Highlights 🎉
+
+- Resolved model startup issues, memory leaks, and improved token limits
+- Clearer error messages and subtle UX improvements
+
+Update your product or download the latest: https://jan.ai
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.10).
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-12-05-jan-hot-fix-mac.mdx b/website/src/content/changelog/2024-12-05-jan-hot-fix-mac.mdx
new file mode 100644
index 000000000..d79c77b03
--- /dev/null
+++ b/website/src/content/changelog/2024-12-05-jan-hot-fix-mac.mdx
@@ -0,0 +1,26 @@
+---
+title: "Jan v0.5.11 is here!"
+version: 0.5.11
+description: "Critical issues fixed, Mac installation updated."
+date: 2024-12-05
+ogImage: "/assets/images/changelog/jan-v0.5.11.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Jan v0.5.11 is here - critical issues fixed, Mac installation updated.
+
+### Highlights 🎉
+
+- Crashes (markdown & code highlighting)
+- Thread switching & auto-scroll
+- Syntax highlighting bugs
+- API issues (Anthropic, OpenRouter)
+- Title glitches with special characters
+- Model settings inconsistencies
+
+Update your product or download the latest: https://jan.ai
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.11).
\ No newline at end of file
diff --git a/website/src/content/changelog/2024-12-30-jan-new-privacy.mdx b/website/src/content/changelog/2024-12-30-jan-new-privacy.mdx
new file mode 100644
index 000000000..3f29dd245
--- /dev/null
+++ b/website/src/content/changelog/2024-12-30-jan-new-privacy.mdx
@@ -0,0 +1,28 @@
+---
+title: "Jan gives you full control over your privacy"
+version: 0.5.12
+description: "Improved Privacy settings to give full control over analytics"
+date: 2024-12-30
+ogImage: "/assets/images/changelog/jan-v0.5.12.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+Jan v0.5.12 is here - critical issues fixed, Mac installation updated.
+
+### Highlights 🎉
+
+- Updated privacy settings with opt-in/out options for Jan Analytics
+- Adjustable chat width
+- The right sidebar and input box are now optimized for new users
+
+### Fixes 💫
+- Updated privacy settings with opt-in/out options for Jan Analytics
+- Adjustable chat width
+- The right sidebar and input box are now optimized for new users
+
+Update your product or download the latest: https://jan.ai
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.12).
\ No newline at end of file
diff --git a/website/src/content/changelog/2025-01-06-key-issues-resolved.mdx b/website/src/content/changelog/2025-01-06-key-issues-resolved.mdx
new file mode 100644
index 000000000..40e541c2a
--- /dev/null
+++ b/website/src/content/changelog/2025-01-06-key-issues-resolved.mdx
@@ -0,0 +1,23 @@
+---
+title: "A few key issues have been solved!"
+version: 0.5.13
+description: "Jan v0.5.13 is here: A few key issues have been solved."
+date: 2025-01-06
+ogImage: "/assets/images/changelog/jan-v0-5-13.jpg"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+👋 Jan v0.5.13 is here: A few key issues have been solved!
+
+### Highlights 🎉
+
+- Resolved model loading issues on MacOS Intel
+- Fixed app resetting max_tokens to 8192 on new threads - now uses model settings
+- Fixed Vulkan settings visibility for some users
+
+Update your product or download the latest: https://jan.ai
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.13).
diff --git a/website/src/content/changelog/2025-01-23-deepseek-r1-jan.mdx b/website/src/content/changelog/2025-01-23-deepseek-r1-jan.mdx
new file mode 100644
index 000000000..3006588b8
--- /dev/null
+++ b/website/src/content/changelog/2025-01-23-deepseek-r1-jan.mdx
@@ -0,0 +1,36 @@
+---
+title: "Run DeepSeek R1 Distills error-free!"
+version: 0.5.14
+description: "Jan v0.5.14 is out: Run DeepSeek R1 Distills error-free!"
+date: 2025-01-23
+ogImage: "/assets/images/changelog/jan-v0-5-14-deepseek-r1.jpg"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+👋 Jan v0.5.14 is out: Run DeepSeek R1 Distills error-free!
+
+You can run DeepSeek R1 distills in Jan error-free. Follow our [step-by-step guide to run DeepSeek R1 locally](/blog/deepseek-r1-locally) and get this AI model running on your device in minutes.
+
+llama.cpp version updated via Cortex—thanks to GG & llama.cpp community!
+
+- Paste GGUF links into Jan Hub to download
+- Already downloaded the model but facing issues? Update Jan.
+
+Models:
+
+Qwen
+- DeepSeek-R1-Distill-Qwen-1.5B-GGUF: https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-1.5B-GGUF
+- DeepSeek-R1-Distill-Qwen-7B-GGUF: https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-7B-GGUF
+- DeepSeek-R1-Distill-Qwen-14B-GGUF: https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-14B-GGUF
+- DeepSeek-R1-Distill-Qwen-32B-GGUF: https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF
+
+Llama
+- DeepSeek-R1-Distill-Llama-8B-GGUF: https://huggingface.co/bartowski/DeepSeek-R1-Distill-Llama-8B-GGUF
+- DeepSeek-R1-Distill-Llama-70B-GGUF: https://huggingface.co/bartowski/DeepSeek-R1-Distill-Llama-70B-GGUF
+
+Update your Jan or [download the latest](https://jan.ai/).
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.14).
diff --git a/website/src/content/changelog/2025-02-14-enterprise-security.mdx b/website/src/content/changelog/2025-02-14-enterprise-security.mdx
new file mode 100644
index 000000000..00d973bb2
--- /dev/null
+++ b/website/src/content/changelog/2025-02-14-enterprise-security.mdx
@@ -0,0 +1,134 @@
+---
+title: "Jan Enterprise: Security & Compliance"
+version: 0.6.2
+description: "Enterprise-grade security features, compliance tools, and team management capabilities"
+date: 2025-02-14
+image: "/assets/images/changelog/enterprise-security.png"
+featured: false
+---
+
+## Enterprise Ready 🏢
+
+Jan v0.6.2 introduces comprehensive enterprise features designed for organizations that need advanced security, compliance, and management capabilities. Deploy AI with confidence across your entire organization.
+
+### 🔐 Advanced Security Framework
+
+Military-grade security for your AI infrastructure:
+- **Zero-Trust Architecture**: Verify every request, trust nothing by default
+- **End-to-End Encryption**: Data encrypted in transit and at rest
+- **Certificate Management**: Full PKI support with automatic rotation
+- **Network Isolation**: Separate AI workloads from corporate networks
+- **Audit Logging**: Comprehensive logs for security analysis and compliance
+
+### 👥 Team Management
+
+Sophisticated user and access control:
+- **Role-Based Access**: Granular permissions for different user types
+- **Single Sign-On**: Integration with Active Directory, SAML, and OAuth
+- **Multi-Factor Authentication**: Required 2FA with hardware token support
+- **Session Management**: Control session duration and concurrent logins
+- **User Provisioning**: Automated user creation and deactivation
+
+### 📊 Compliance Dashboard
+
+Meet regulatory requirements with confidence:
+- **SOC 2 Compliance**: Built-in controls for SOC 2 Type II certification
+- **GDPR Tools**: Data subject rights management and privacy controls
+- **HIPAA Ready**: Healthcare-specific security and privacy features
+- **ISO 27001**: Information security management system alignment
+- **Custom Frameworks**: Adapt to your specific compliance requirements
+
+### 🏗️ Infrastructure Management
+
+Deploy and scale AI across your organization:
+- **Containerized Deployment**: Docker and Kubernetes ready
+- **Load Balancing**: Distribute AI workloads across multiple instances
+- **Auto-Scaling**: Automatically scale based on demand
+- **Health Monitoring**: Real-time system health and performance tracking
+- **Disaster Recovery**: Automated backups and failover capabilities
+
+### 🎯 Data Governance
+
+Complete control over your data:
+- **Data Classification**: Automatically tag and classify sensitive information
+- **Retention Policies**: Automated data lifecycle management
+- **Data Loss Prevention**: Prevent sensitive data from leaving your environment
+- **Geographic Controls**: Control where data is processed and stored
+- **Right to Deletion**: Complete data removal on request
+
+### 🔍 Advanced Analytics
+
+Insights into AI usage across your organization:
+- **Usage Analytics**: Understand how teams use AI capabilities
+- **Cost Analysis**: Track AI costs by department, team, or user
+- **Performance Metrics**: Monitor AI performance and quality
+- **Adoption Reports**: Measure AI adoption across the organization
+- **Custom Dashboards**: Create reports tailored to your needs
+
+### 🛡️ Threat Protection
+
+Advanced protection against AI-specific threats:
+- **Prompt Injection Detection**: Identify and block malicious prompts
+- **Content Filtering**: Prevent inappropriate content generation
+- **Rate Limiting**: Protect against abuse and resource exhaustion
+- **Anomaly Detection**: Identify unusual usage patterns
+- **Incident Response**: Automated response to security events
+
+### 🌐 Integration Capabilities
+
+Connect Jan to your existing enterprise systems:
+- **API Gateway**: Secure API access with rate limiting and authentication
+- **Webhook Support**: Real-time notifications to external systems
+- **Database Connections**: Direct integration with enterprise databases
+- **Workflow Integration**: Connect to existing business processes
+- **Custom Connectors**: Build integrations specific to your needs
+
+### 📋 Policy Management
+
+Implement and enforce AI governance policies:
+- **Usage Policies**: Define acceptable use of AI capabilities
+- **Content Policies**: Control what types of content can be generated
+- **Model Policies**: Restrict access to specific AI models
+- **Data Policies**: Control how data is processed and stored
+- **Approval Workflows**: Require approval for sensitive operations
+
+### 🔧 Administrative Tools
+
+Powerful tools for IT administrators:
+- **Centralized Configuration**: Manage settings across all deployments
+- **Bulk Operations**: Perform actions across multiple users or systems
+- **Migration Tools**: Move data and settings between environments
+- **Backup Management**: Automated and manual backup capabilities
+- **System Diagnostics**: Comprehensive troubleshooting tools
+
+### 📞 Enterprise Support
+
+Dedicated support for mission-critical deployments:
+- **24/7 Support**: Round-the-clock assistance for critical issues
+- **Dedicated Success Manager**: Personal point of contact for your organization
+- **Training Programs**: Comprehensive training for administrators and users
+- **Implementation Services**: Professional deployment and configuration
+- **Custom Development**: Tailored features for unique requirements
+
+## Deployment Options
+
+Choose the deployment model that fits your needs:
+- **On-Premises**: Complete control with on-site deployment
+- **Private Cloud**: Dedicated cloud environment just for you
+- **Hybrid**: Combine on-premises and cloud capabilities
+- **Multi-Region**: Deploy across multiple geographic regions
+- **Air-Gapped**: Completely isolated environments for maximum security
+
+## Getting Started
+
+Ready to deploy Jan Enterprise?
+
+1. **Assessment**: Our team evaluates your requirements
+2. **Pilot Program**: Start with a small-scale deployment
+3. **Training**: Comprehensive training for your team
+4. **Full Deployment**: Roll out to your entire organization
+5. **Ongoing Support**: Continuous support and optimization
+
+Transform how your organization uses AI. Contact our enterprise team to learn more.
+
+[Contact Enterprise Sales](mailto:enterprise@jan.ai) • [Enterprise Documentation](/docs/enterprise) • [Security Whitepaper](/security-whitepaper)
\ No newline at end of file
diff --git a/website/src/content/changelog/2025-02-18-advanced-llama.cpp-settings.mdx b/website/src/content/changelog/2025-02-18-advanced-llama.cpp-settings.mdx
new file mode 100644
index 000000000..b24c72908
--- /dev/null
+++ b/website/src/content/changelog/2025-02-18-advanced-llama.cpp-settings.mdx
@@ -0,0 +1,28 @@
+---
+title: "You can now tweak llama.cpp settings, and add any cloud model!"
+version: 0.5.15
+description: "Jan v0.5.15 is out: Advanced llama.cpp settings and cloud model support"
+date: 2025-02-18
+ogImage: "/assets/images/changelog/jan-v0-5-15-llamacpp-settings.jpg"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+👋 Jan v0.5.15 is out with exciting new features and improvements!
+
+## Highlights 🎉
+
+- Improved llama.cpp settings for better model control and performance
+- Install and use any cloud model seamlessly
+- Enhanced hardware controls for better resource management
+- New models supported:
+ - DeepSeek AI
+ - Google Gemini
+ - OpenAI o3-mini
+ - R1 distills
+
+Update your Jan or [download the latest](https://jan.ai/).
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.15).
diff --git a/website/src/content/changelog/2025-03-05-ui-redesign.mdx b/website/src/content/changelog/2025-03-05-ui-redesign.mdx
new file mode 100644
index 000000000..fb47b7b21
--- /dev/null
+++ b/website/src/content/changelog/2025-03-05-ui-redesign.mdx
@@ -0,0 +1,122 @@
+---
+title: "Jan v0.6.3: UI Renaissance"
+version: 0.6.3
+description: "Complete interface redesign with improved UX, dark mode, and accessibility features"
+date: 2025-03-05
+image: "/assets/images/changelog/ui-redesign.png"
+featured: false
+---
+
+## A Beautiful New Beginning 🎨
+
+Jan v0.6.3 introduces our most comprehensive UI overhaul yet. Every pixel has been reconsidered, every interaction refined. The result? A beautiful, intuitive interface that gets out of your way and lets you focus on what matters - conversations with AI.
+
+### 🎯 Design Philosophy
+
+Our new design principles:
+- **Clarity First**: Remove visual noise, highlight what's important
+- **Consistent Language**: Unified design patterns throughout the app
+- **Accessibility Focus**: Usable by everyone, regardless of ability
+- **Performance Minded**: Beautiful interfaces that are also fast
+- **Future Ready**: Scalable design system for upcoming features
+
+### 🌙 Enhanced Dark Mode
+
+Dark mode, reimagined:
+- **True Black Option**: Perfect for OLED displays and low-light use
+- **Adaptive Contrast**: Automatically adjusts based on ambient light
+- **Custom Accent Colors**: Choose your preferred highlight color
+- **Smart Switching**: Follows system preferences or custom schedule
+- **Reduced Eye Strain**: Carefully calibrated colors for long usage sessions
+
+### 💬 Conversation Experience
+
+Completely redesigned chat interface:
+- **Improved Message Bubbles**: Better readability and visual hierarchy
+- **Smart Typography**: Optimal font sizes and spacing for every device
+- **Code Highlighting**: Syntax highlighting for 200+ programming languages
+- **Math Rendering**: Beautiful LaTeX math equation display
+- **Image Gallery**: Enhanced image viewing with zoom and navigation
+
+### 🎛️ Settings Overhaul
+
+Settings that make sense:
+- **Organized Categories**: Logical grouping of related options
+- **Search Settings**: Find any setting instantly
+- **Visual Previews**: See changes before applying them
+- **Quick Actions**: Common tasks accessible with fewer clicks
+- **Import/Export**: Backup and restore your entire configuration
+
+### 📱 Responsive Design
+
+Perfect on every screen:
+- **Mobile Optimized**: Touch-friendly interface for tablets and phones
+- **Desktop Polish**: Take advantage of larger screens and precise input
+- **Window Management**: Better handling of multiple windows and panels
+- **Flexible Layouts**: Adapt to any screen size or orientation
+- **High DPI Support**: Crisp on retina and 4K displays
+
+### ♿ Accessibility Improvements
+
+Jan for everyone:
+- **Screen Reader Support**: Full compatibility with assistive technologies
+- **Keyboard Navigation**: Complete interface control without a mouse
+- **High Contrast Mode**: Enhanced visibility for low-vision users
+- **Font Scaling**: Respect system font size preferences
+- **Motion Controls**: Reduced motion options for sensitive users
+
+### 🎨 Theming System
+
+Express your style:
+- **Built-in Themes**: 12 carefully crafted color schemes
+- **Custom Themes**: Create your own with our theme editor
+- **Theme Sharing**: Import themes created by the community
+- **Seasonal Themes**: Special themes for holidays and events
+- **Auto-Theming**: Themes that change based on time of day
+
+### 🔍 Improved Navigation
+
+Find everything faster:
+- **Global Search**: Search conversations, settings, and help instantly
+- **Breadcrumbs**: Always know where you are in the app
+- **Quick Switcher**: Jump between conversations with keyboard shortcuts
+- **Recent Items**: Quick access to your most-used features
+- **Favorites System**: Pin important conversations and tools
+
+### 🎪 Animation & Transitions
+
+Delightful micro-interactions:
+- **Smooth Transitions**: Fluid movement between screens and states
+- **Loading Animations**: Engaging feedback during wait times
+- **Hover Effects**: Subtle responses to mouse interaction
+- **Focus Indicators**: Clear visual feedback for keyboard users
+- **Performance Optimized**: 60fps animations that don't drain battery
+
+### 📊 Visual Data
+
+Information design that informs:
+- **Usage Charts**: Beautiful visualizations of your AI usage
+- **Performance Graphs**: Real-time system performance monitoring
+- **Progress Indicators**: Clear feedback for long-running operations
+- **Status Displays**: At-a-glance system health information
+- **Comparison Views**: Side-by-side analysis of models and settings
+
+### 🚀 Performance Improvements
+
+Beauty with brains:
+- **Faster Rendering**: 40% improvement in interface responsiveness
+- **Memory Efficiency**: Reduced RAM usage for smoother operation
+- **Bundle Optimization**: Smaller app size, faster loading
+- **Asset Loading**: Progressive loading for smoother startup
+- **Animation Performance**: Hardware-accelerated animations
+
+## Migration Guide
+
+Your existing data and settings are automatically preserved. Some visual elements may look different, but all functionality remains the same or improved.
+
+**New Users**: Welcome to the most beautiful Jan yet!
+**Existing Users**: Your themes and customizations will be migrated automatically.
+
+Experience the new Jan. Clean, beautiful, and more powerful than ever.
+
+[Download Jan v0.6.3](https://jan.ai/) • [UI Guide](/docs/interface) • [Accessibility Documentation](/docs/accessibility)
\ No newline at end of file
diff --git a/website/src/content/changelog/2025-03-14-jan-security-patch.mdx b/website/src/content/changelog/2025-03-14-jan-security-patch.mdx
new file mode 100644
index 000000000..19d5bbf9e
--- /dev/null
+++ b/website/src/content/changelog/2025-03-14-jan-security-patch.mdx
@@ -0,0 +1,29 @@
+---
+title: "Security fixes and UI improvements"
+version: 0.5.16
+description: "Jan v0.5.16 is out: Security fixes and major improvements to Model Hub and chat experience"
+date: 2025-03-14
+ogImage: "/assets/images/changelog/jan-v0-5-16-security-patch.jpg"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+👋 Jan v0.5.16 is out: Security fixes (found in Cortex) and a few improvements!
+
+## Highlights 🎉
+
+- Redesigned Model Hub for better user experience
+- Faster chat response times
+- Cleaner layout and improved model picker
+- New model support:
+ - GPT-4.5 preview
+ - Claude 3.7 Sonnet
+ - Gemma 3
+
+⚠️ Important: This release includes critical security fixes. We'll share more details about these security improvements soon.
+
+Update your Jan or [download the latest](https://jan.ai/).
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.16).
diff --git a/website/src/content/changelog/2025-04-10-mobile-launch.mdx b/website/src/content/changelog/2025-04-10-mobile-launch.mdx
new file mode 100644
index 000000000..1f75b15a9
--- /dev/null
+++ b/website/src/content/changelog/2025-04-10-mobile-launch.mdx
@@ -0,0 +1,119 @@
+---
+title: "Jan Mobile: AI in Your Pocket"
+version: 1.0.0
+description: "Introducing Jan Mobile - full-featured AI assistant for iOS and Android with cloud sync"
+date: 2025-04-10
+image: "/assets/images/changelog/jan-mobile.png"
+featured: true
+---
+
+## AI Goes Mobile 📱
+
+After months of development, we're thrilled to announce Jan Mobile - bringing the full power of Jan's AI capabilities to your smartphone. Chat with AI anywhere, sync across devices, and never miss a conversation.
+
+### 📱 Native Mobile Experience
+
+Built from the ground up for mobile:
+- **Native Performance**: Smooth, responsive interface optimized for touch
+- **Offline Capable**: Continue conversations even without internet
+- **Battery Optimized**: Efficient background processing preserves battery life
+- **Dark Mode**: Beautiful dark theme that's easy on the eyes
+- **Haptic Feedback**: Tactile responses for better interaction
+
+### ☁️ Seamless Cloud Sync
+
+Your conversations follow you everywhere:
+- **Real-time Sync**: Conversations update instantly across all devices
+- **Conflict Resolution**: Smart merging when editing on multiple devices
+- **Selective Sync**: Choose which conversations to sync to mobile
+- **End-to-End Encryption**: Your data remains private and secure
+- **Offline Queue**: Messages sync when connection returns
+
+### 🎯 Mobile-First Features
+
+Designed for how you use your phone:
+- **Voice Input**: Speak your messages instead of typing
+- **Voice Output**: AI responses read aloud with natural voices
+- **Quick Actions**: Swipe gestures for common tasks
+- **Share Integration**: Share content to Jan from any app
+- **Widget Support**: Quick access from your home screen
+
+### 🔒 Privacy & Security
+
+Your privacy, protected:
+- **Local Processing**: Sensitive conversations can stay on-device
+- **Biometric Lock**: Secure Jan with fingerprint or face recognition
+- **Auto-Lock**: Automatically locks after inactivity
+- **Private Mode**: Conversations that don't sync to cloud
+- **Data Controls**: Full control over what data is stored where
+
+### 🤖 Full Model Support
+
+Access all your favorite models:
+- **Cloud Models**: GPT-4, Claude, Gemini, and more
+- **Local Models**: Run smaller models directly on your phone
+- **Model Switching**: Change models mid-conversation
+- **Smart Routing**: Automatically choose the best model for your query
+- **Offline Models**: Basic AI capabilities without internet
+
+### 📸 Rich Media Support
+
+Beyond just text:
+- **Image Analysis**: Upload photos for AI to analyze and discuss
+- **Camera Integration**: Take photos directly in Jan for analysis
+- **Voice Messages**: Send and receive voice messages
+- **File Sharing**: Share documents, PDFs, and more
+- **Link Previews**: Rich previews for shared links
+
+### 🎨 Personalization
+
+Make Jan your own:
+- **Custom Themes**: Choose colors and appearance
+- **Chat Backgrounds**: Personalize your conversation view
+- **Notification Settings**: Control when and how you're notified
+- **Assistant Personalities**: Different AI personalities for different contexts
+- **Quick Replies**: Set up common responses
+
+### 🔄 Cross-Platform Features
+
+Unified experience across desktop and mobile:
+- **Universal Search**: Find conversations across all devices
+- **Shared Assistants**: Use the same AI assistants everywhere
+- **Unified Settings**: Preferences sync between devices
+- **File Access**: Access files shared from desktop
+- **Continuous Conversations**: Start on desktop, continue on mobile
+
+### 📊 Usage Analytics
+
+Understand your AI usage:
+- **Conversation Stats**: See your most active conversations
+- **Model Usage**: Track which models you use most
+- **Time Analytics**: Understand your usage patterns
+- **Export Data**: Download your conversation history
+- **Privacy Dashboard**: See exactly what data is stored
+
+### 🌟 Launch Features
+
+Available from day one:
+- **Free Tier**: Full functionality with generous usage limits
+- **Pro Features**: Enhanced models and advanced features
+- **Family Sharing**: Share Pro features with family members
+- **Student Discount**: Special pricing for students
+- **Enterprise Options**: Advanced security and management
+
+## Platform Availability
+
+- **iOS**: Available on the App Store (iOS 15.0+)
+- **Android**: Available on Google Play (Android 8.0+)
+- **Cross-Platform**: Full feature parity between platforms
+
+## Getting Started
+
+1. Download Jan Mobile from your app store
+2. Sign in with your Jan account (or create one)
+3. Your desktop conversations automatically appear
+4. Start chatting with AI on the go!
+
+Your AI assistant is now truly everywhere. Download Jan Mobile today and experience AI without boundaries.
+
+[Download for iOS](https://apps.apple.com/app/jan-ai) • [Download for Android](https://play.google.com/store/apps/jan) • [Mobile Guide](/docs/mobile)
\ No newline at end of file
diff --git a/website/src/content/changelog/2025-05-14-jan-qwen3-patch.mdx b/website/src/content/changelog/2025-05-14-jan-qwen3-patch.mdx
new file mode 100644
index 000000000..24d784153
--- /dev/null
+++ b/website/src/content/changelog/2025-05-14-jan-qwen3-patch.mdx
@@ -0,0 +1,23 @@
+---
+title: "Qwen3 support is now more reliable."
+version: 0.5.17
+description: "Jan v0.5.17 is out: Qwen3 support is now more reliable"
+date: 2025-05-14
+ogImage: "/assets/images/changelog/jan-v0-5-17-gemm3-patch.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+👋 Jan v0.5.17 is out: Qwen3 support is now more reliable
+
+## Highlights 🎉
+
+- Improved Qwen3 support with cleaner token output
+- Clearer install and quickstart docs
+- UI polish and bug fixes throughout
+
+Update your Jan or [download the latest](https://jan.ai/).
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.5.17).
diff --git a/website/src/content/changelog/2025-05-20-performance-boost.mdx b/website/src/content/changelog/2025-05-20-performance-boost.mdx
new file mode 100644
index 000000000..489bb8eea
--- /dev/null
+++ b/website/src/content/changelog/2025-05-20-performance-boost.mdx
@@ -0,0 +1,101 @@
+---
+title: "Jan v0.6.4: Performance Powerhouse"
+version: 0.6.4
+description: "Massive speed improvements, GPU optimization, and streamlined model management"
+date: 2025-05-20
+image: "/assets/images/changelog/performance-boost.png"
+featured: false
+---
+
+## Speed Like Never Before ⚡
+
+Jan v0.6.4 delivers our biggest performance update yet. Models load faster, inference is smoother, and memory usage is dramatically reduced. This is the Jan you've been waiting for.
+
+### 🚀 Inference Speed Improvements
+
+Dramatic performance gains across the board:
+- **3x Faster Model Loading**: Optimized model initialization reduces wait times
+- **50% Faster Inference**: Improved CUDA kernels and memory management
+- **Instant Model Switching**: Switch between models with near-zero delay
+- **Background Preloading**: Frequently used models stay ready in memory
+- **Smart Caching**: Intelligent context caching reduces repeated work
+
+### 🎯 GPU Optimization Revolution
+
+Completely rewritten GPU acceleration:
+- **Auto-GPU Detection**: Automatically finds and uses your best GPU
+- **Multi-GPU Support**: Distribute model layers across multiple GPUs
+- **Memory Optimization**: 40% reduction in VRAM usage
+- **Dynamic Offloading**: Automatically balance between GPU and CPU
+- **CUDA 12 Support**: Latest NVIDIA drivers and optimizations
+
+### 🧠 Smarter Memory Management
+
+Revolutionary memory handling:
+- **Adaptive Memory**: Automatically adjusts to available system memory
+- **Memory Pressure Detection**: Gracefully handles low-memory situations
+- **Efficient Model Unloading**: Frees memory when models aren't needed
+- **Context Length Optimization**: Handle longer conversations without slowdown
+- **Memory Usage Dashboard**: Real-time visibility into memory consumption
+
+### 📱 Startup Speed Breakthrough
+
+Jan now starts in seconds, not minutes:
+- **Cold Start Optimization**: 5x faster first launch
+- **Background Services**: Core services start in parallel
+- **Lazy Loading**: Only load components when needed
+- **Configuration Caching**: Settings load instantly
+- **Progressive Initialization**: UI appears immediately, features load progressively
+
+### 🔧 Model Management Overhaul
+
+Streamlined model experience:
+- **One-Click Downloads**: Simplified model acquisition
+- **Download Resume**: Interrupted downloads continue automatically
+- **Parallel Downloads**: Download multiple models simultaneously
+- **Storage Optimization**: Automatic cleanup of unused model files
+- **Model Recommendations**: AI suggests optimal models for your hardware
+
+### 💾 Storage Efficiency
+
+Dramatic reduction in disk usage:
+- **Model Compression**: 30% smaller model files without quality loss
+- **Duplicate Detection**: Automatically removes duplicate models
+- **Incremental Updates**: Only download model changes, not entire files
+- **Smart Cleanup**: Removes temporary files and caches automatically
+- **Storage Analytics**: See exactly what's using your disk space
+
+### 🌐 Network Optimizations
+
+Faster downloads and better connectivity:
+- **CDN Integration**: Download models from the closest server
+- **Connection Pooling**: Efficient network resource usage
+- **Retry Logic**: Automatic recovery from network interruptions
+- **Bandwidth Adaptation**: Adjusts download speed to network conditions
+- **Offline Mode**: Better handling when internet is unavailable
+
+### 🔍 Performance Monitoring
+
+New tools to understand performance:
+- **Real-time Metrics**: See inference speed, memory usage, GPU utilization
+- **Performance History**: Track performance over time
+- **Bottleneck Detection**: Identify what's slowing down your system
+- **Benchmark Tools**: Compare performance across different configurations
+- **Performance Profiles**: Save optimal settings for different use cases
+
+### 🐛 Critical Fixes
+
+Major stability improvements:
+- Fixed memory leaks during long conversations
+- Resolved GPU driver compatibility issues
+- Eliminated random crashes during model switching
+- Fixed model corruption during interrupted downloads
+- Resolved race conditions in multi-threaded operations
+
+## Technical Details
+
+This release includes fundamental changes to our inference engine, memory management, and GPU acceleration systems. While backwards compatible, you may notice different memory usage patterns and significantly improved performance.
+
+Experience the fastest Jan ever. Download v0.6.4 and feel the difference.
+
+[Download Jan v0.6.4](https://jan.ai/) • [Performance Guide](/docs/performance) • [Release Notes](https://github.com/menloresearch/jan/releases/tag/v0.6.4)
\ No newline at end of file
diff --git a/website/src/content/changelog/2025-06-15-mcp-revolution.mdx b/website/src/content/changelog/2025-06-15-mcp-revolution.mdx
new file mode 100644
index 000000000..3bf1b03e4
--- /dev/null
+++ b/website/src/content/changelog/2025-06-15-mcp-revolution.mdx
@@ -0,0 +1,88 @@
+---
+title: "Jan v0.6.5: MCP Revolution - Connect AI to Everything"
+version: 0.6.5
+description: "Introducing Model Context Protocol support, browser automation, and powerful tool integrations"
+date: 2025-06-15
+image: "../../assets/images/changelog/mcp-revolution.gif"
+featured: true
+---
+
+## The MCP Era Begins 🚀
+
+Jan v0.6.5 introduces groundbreaking Model Context Protocol (MCP) support, transforming Jan from a simple chat interface into a powerful AI automation platform. Connect your AI to browsers, APIs, databases, and countless tools.
+
+### 🔗 Model Context Protocol Integration
+
+MCP opens up infinite possibilities:
+- **Universal Tool Access**: Connect to any service that supports MCP
+- **Real-time Data**: Access live information from APIs and databases
+- **Browser Automation**: Control web browsers directly through AI commands
+- **File System Access**: Read, write, and manipulate files with AI assistance
+- **Custom Tools**: Build your own MCP servers for specialized workflows
+
+### 🌐 Built-in MCP Servers
+
+Launch with powerful integrations:
+- **Browser Control**: Automate web tasks, scrape data, fill forms
+- **File Management**: AI-powered file operations and organization
+- **API Integration**: Connect to REST APIs, GraphQL endpoints
+- **Database Access**: Query and update databases through natural language
+- **Git Operations**: Manage repositories with AI assistance
+
+### 🎯 Smart Tool Discovery
+
+Jan automatically discovers and configures MCP tools:
+- **Auto-detection**: Finds available MCP servers on your system
+- **One-click Setup**: Enable tools with simple toggle switches
+- **Permission Control**: Fine-grained control over tool access
+- **Usage Analytics**: Track which tools your AI uses most
+
+### 🛡️ Enhanced Security Framework
+
+Built with security as a priority:
+- **Sandboxed Execution**: Tools run in isolated environments
+- **Permission System**: Explicit approval for sensitive operations
+- **Audit Logging**: Complete history of tool usage and permissions
+- **Safe Defaults**: Conservative permissions that you can expand
+
+### 🎨 Redesigned Tool Interface
+
+Beautiful new interface for tool management:
+- **Visual Tool Cards**: See available tools at a glance
+- **Real-time Status**: Know when tools are active or inactive
+- **Interactive Setup**: Guided configuration for complex tools
+- **Usage Insights**: Understand how your AI uses different tools
+
+### 🔧 Developer Experience
+
+For MCP server developers:
+- **Local Development**: Test MCP servers directly in Jan
+- **Debug Tools**: Built-in logging and error reporting
+- **Hot Reload**: Changes to MCP servers update instantly
+- **Protocol Validation**: Ensure your servers follow MCP standards
+
+### 🚀 Performance Improvements
+
+Under the hood optimizations:
+- **Faster Tool Loading**: MCP servers start 3x faster
+- **Memory Efficiency**: Reduced memory usage for tool operations
+- **Connection Pooling**: Efficient management of tool connections
+- **Async Operations**: Non-blocking tool execution
+
+### 🌟 Coming Next
+
+The MCP ecosystem is just getting started:
+- More built-in integrations (Slack, Discord, GitHub)
+- Visual workflow builder for complex automations
+- Community marketplace for sharing MCP servers
+- Enterprise-grade security and compliance features
+
+## Breaking Changes
+
+- Tool permissions now require explicit user approval
+- Some legacy integrations have been migrated to MCP
+- Configuration format updated for better security
+
+Transform how you work with AI. Download Jan v0.6.5 and enter the MCP era.
+
+For technical details, see our [MCP documentation](/docs/mcp) and [GitHub release](https://github.com/menloresearch/jan/releases/tag/v0.6.5).
diff --git a/website/src/content/changelog/2025-06-19-jan-ui-revamp.mdx b/website/src/content/changelog/2025-06-19-jan-ui-revamp.mdx
new file mode 100644
index 000000000..ae9e2d4f0
--- /dev/null
+++ b/website/src/content/changelog/2025-06-19-jan-ui-revamp.mdx
@@ -0,0 +1,21 @@
+---
+title: "Jan v0.6.1 is here: It's a whole new vibe!"
+version: 0.6.1
+description: "Are you ready for the sexiest UI ever?"
+date: 2025-06-19
+ogImage: "/assets/images/changelog/jan-v0.6.1-ui-revamp.png"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+## Highlights 🎉
+
+- Jan's been redesigned to be faster, cleaner, and easier to use.
+- You can now create assistants with custom instructions and settings from a dedicated tab.
+- You can now use Jan with Menlo's models.
+
+Update your Jan or [download the latest](https://jan.ai/).
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.6.1).
\ No newline at end of file
diff --git a/website/src/content/changelog/2025-06-26-jan-nano-mcp.mdx b/website/src/content/changelog/2025-06-26-jan-nano-mcp.mdx
new file mode 100644
index 000000000..9380c2272
--- /dev/null
+++ b/website/src/content/changelog/2025-06-26-jan-nano-mcp.mdx
@@ -0,0 +1,21 @@
+---
+title: "Jan v0.6.3 brings new features and models!"
+version: 0.6.3
+description: "Unlocking MCP for everyone and bringing our latest model to Jan!"
+date: 2025-06-26
+ogImage: "/assets/images/changelog/jn128.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+## Highlights 🎉
+
+- We have added Model Context Protocol (MCP) support to the stable build of Jan. It needs to be enabled in the General Settings tab.
+- Jan now supports Menlo's latest model, Jan-Nano-128k.
+- Some hot fixes and improvements.
+
+Update your Jan or [download the latest](https://jan.ai/).
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.6.3).
\ No newline at end of file
diff --git a/website/src/content/changelog/2025-07-17-responsive-ui.mdx b/website/src/content/changelog/2025-07-17-responsive-ui.mdx
new file mode 100644
index 000000000..ac27d3630
--- /dev/null
+++ b/website/src/content/changelog/2025-07-17-responsive-ui.mdx
@@ -0,0 +1,26 @@
+---
+title: "Jan v0.6.5 brings responsive UI and MCP examples!"
+version: 0.6.5
+description: "New MCP examples, updated pages, and bug fixes!"
+date: 2025-07-17
+ogImage: "/assets/images/changelog/release_v0_6_5.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+## Highlights 🎉
+
+Jan v0.6.5 brings responsive UI improvements, enhanced model provider management, and better Linux compatibility alongside
+new MCP examples.
+
+- Support responsive UI on Jan
+- Rework of Model Providers UI
+- Bump version of llama.cpp
+- Fix the bug where fetching models from custom provider can cause app to crash
+- AppImage can now render on wayland + mesa
+
+Update your Jan or [download the latest](https://jan.ai/).
+
+For more details, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.6.5).
\ No newline at end of file
diff --git a/website/src/content/changelog/2025-07-31-llamacpp-tutorials.mdx b/website/src/content/changelog/2025-07-31-llamacpp-tutorials.mdx
new file mode 100644
index 000000000..fc0310174
--- /dev/null
+++ b/website/src/content/changelog/2025-07-31-llamacpp-tutorials.mdx
@@ -0,0 +1,119 @@
+---
+title: "Jan v0.6.6: Enhanced llama.cpp integration and smarter model management"
+version: 0.6.6
+description: "Major llama.cpp improvements, Hugging Face provider support, and refined MCP experience"
+date: 2025-07-31
+ogImage: "/assets/images/changelog/changelog0.6.6.gif"
+---
+
+import ChangelogHeader from '@/components/Changelog/ChangelogHeader.astro'
+
+
+
+## Highlights 🎉
+
+Jan v0.6.6 delivers significant improvements to the llama.cpp backend, introduces Hugging Face as a
+built-in provider, and brings smarter model management with auto-unload capabilities. This release
+also includes numerous MCP refinements and platform-specific enhancements.
+
+### 🚀 Major llama.cpp Backend Overhaul
+
+We've completely revamped the llama.cpp integration with:
+- **Smart Backend Management**: The backend now auto-updates and persists your settings properly
+- **Device Detection**: Jan automatically detects available GPUs and hardware capabilities
+- **Direct llama.cpp Access**: Models now interface directly with llama.cpp (previously hidden behind Cortex)
+- **Automatic Migration**: Your existing models seamlessly move from Cortex to direct llama.cpp management
+- **Better Error Handling**: Clear error messages when models fail to load, with actionable solutions
+- **Per-Model Overrides**: Configure specific settings for individual models
+
+### 🤗 Hugging Face Cloud Router Integration
+
+Connect to Hugging Face's new cloud inference service:
+- Access pre-configured models running on various providers (Fireworks, Together AI, and more)
+- Hugging Face handles the routing to the best available provider
+- Simplified setup with just your HF token
+- Non-deletable provider status to prevent accidental removal
+- Note: Direct model ID search in Hub remains available as before
+
+### 🧠 Smarter Model Management
+
+New intelligent features to optimize your system resources:
+- **Auto-Unload Old Models**: Automatically free up memory by unloading unused models
+- **Persistent Settings**: Your model capabilities and settings now persist across app restarts
+- **Zero GPU Layers Support**: Set N-GPU Layers to 0 for CPU-only inference
+- **Memory Calculation Improvements**: More accurate memory usage reporting
+
+### 🎯 MCP Refinements
+
+Enhanced MCP experience with:
+- Tool approval dialog improvements with scrollable parameters
+- Better experimental feature edge case handling
+- Fixed tool call button disappearing issue
+- JSON editing tooltips for easier configuration
+- Auto-focus on "Always Allow" action for smoother workflows
+
+### 📚 New MCP Integration Tutorials
+
+Comprehensive guides for powerful MCP integrations:
+- **Canva MCP**: Create and manage designs through natural language - generate logos, presentations, and marketing materials directly from chat
+- **Browserbase MCP**: Control cloud browsers with AI - automate web tasks, extract data, and monitor sites without complex scripting
+- **Octagon Deep Research MCP**: Access finance-focused research capabilities - analyze markets, investigate companies, and generate investment insights
+
+### 🖥️ Platform-Specific Improvements
+
+**Windows:**
+- Fixed terminal windows popping up during model loading
+- Better process termination handling
+- VCRuntime included in installer for compatibility
+- Improved NSIS installer with app running checks
+
+**Linux:**
+- AppImage now works properly with newest Tauri version and it went from almost 1GB to less than 200MB
+- Better Wayland compatibility
+
+**macOS:**
+- Improved build process and artifact naming
+
+### 🎨 UI/UX Enhancements
+
+Quality of life improvements throughout:
+- Fixed rename thread dialog showing incorrect thread names
+- Assistant instructions now have proper defaults
+- Download progress indicators remain visible when scrolling
+- Better error pages with clearer messaging
+- GPU detection now shows accurate backend information
+- Improved clickable areas for better usability
+
+### 🔧 Developer Experience
+
+Behind the scenes improvements:
+- New automated QA system using CUA (Computer Use Automation)
+- Standardized build process across platforms
+- Enhanced error stream handling and parsing
+- Better proxy support for the new downloader
+- Reasoning format support for advanced models
+
+### 🐛 Bug Fixes
+
+Notable fixes include:
+- Factory reset no longer fails with access denied errors
+- OpenRouter provider stays selected properly
+- Model search in Hub shows latest data only
+- Temporary download files are cleaned up on cancel
+- Legacy threads no longer appear above new threads
+- Fixed encoding issues on various platforms
+
+## Breaking Changes
+
+- Models previously managed by Cortex now interface directly with llama.cpp (automatic migration included)
+- Some sampling parameters have been removed from the llama.cpp extension for consistency
+- Cortex extension is deprecated in favor of direct llama.cpp integration
+
+## Coming Next
+
+We're working on expanding MCP capabilities, improving model download speeds, and adding more provider
+integrations. Stay tuned!
+
+Update your Jan or [download the latest](https://jan.ai/).
+
+For the complete list of changes, see the [GitHub release notes](https://github.com/menloresearch/jan/releases/tag/v0.6.6).
diff --git a/website/src/content/docs/handbook/brand/index.mdx b/website/src/content/docs/handbook/brand/index.mdx
new file mode 100644
index 000000000..74f723dc7
--- /dev/null
+++ b/website/src/content/docs/handbook/brand/index.mdx
@@ -0,0 +1,4 @@
+---
+title: Brand & Identity
+asIndexPage: true
+---
\ No newline at end of file
diff --git a/website/src/content/docs/handbook/comp/bonus.mdx b/website/src/content/docs/handbook/comp/bonus.mdx
new file mode 100644
index 000000000..e30321579
--- /dev/null
+++ b/website/src/content/docs/handbook/comp/bonus.mdx
@@ -0,0 +1,5 @@
+---
+title: "Handbook"
+description: "Jan Team Handbook"
+---
+
diff --git a/website/src/content/docs/handbook/comp/esop.mdx b/website/src/content/docs/handbook/comp/esop.mdx
new file mode 100644
index 000000000..53f719ee7
--- /dev/null
+++ b/website/src/content/docs/handbook/comp/esop.mdx
@@ -0,0 +1,21 @@
+---
+title: "ESOP"
+description: "Employee Stock Options"
+---
+
+
+## Shares and Equity
+
+Jan is [employee-owned](/handbook/why/ownership). The people building Jan own a meaningful percentage of the company over time.
+
+### Distributions
+
+Every 6 months, Jan distributes 1% of company equity among its team members:
+
+- Dec 2023: 1% distributed among 10 team members (1 part-time)
+- June 2024: 1% distributed among 15 team members (4 part-time)
+- Dec 2024: 1% distributed among 18 team members (5 part-time)
+
+Distributions are performance-based, and cover both full-time and part-time team members and open source contributors.
+
+This schedule is subject to change based on the discretion of the board.
diff --git a/website/src/content/docs/handbook/comp/index.mdx b/website/src/content/docs/handbook/comp/index.mdx
new file mode 100644
index 000000000..63e59206a
--- /dev/null
+++ b/website/src/content/docs/handbook/comp/index.mdx
@@ -0,0 +1,34 @@
+---
+title: "Compensation"
+description: "Payment and benefits schemes."
+asIndexPage: true
+---
+
+
+### Pay
+
+Everyone at Jan is on the same payscale, with cost of living adjustments based on [location](#location).
+
+Your base cash pay is determined based on realized impact, not titles. Workstream owners, responsible for the success/failure of a project & team, are compensated more than individual contributors. Folks at the same contribution levels receive equitable base pay.
+
+Based on your [progression](progression) speed, we have been known to adjust pay frequently and aggressively.
+
+### Location
+
+We provide a location adjustment for folks living in a high cost of living area. We use the [Numbeo](https://www.numbeo.com/cost-of-living/) index to arrive at a fair & equitable total.
+
+### Bonus
+
+Depending on location, we pay a discretionary cash bonus twice per year. The bonus typically ranges from 5-17% of your base pay, and is based on the company's performance and your individual contribution.
+
+### ESOP
+
+We offer an employee stock ownership plan to all full-time employees. The ESOP is a key part of our compensation package and is designed to align everyone's incentives with the long-term success of the company. Read our [ESOP policy](esop).
+
+### Benefits
+
+- **Equipment**: After the probation period, you get a $1000-$1500 USD budget for equipment. Eligible items include: laptop, monitors, keyboard, mouse, and noise cancelling headphones. Please see our [spending policy](spending) for more details.
+
+- **AI Subscriptions**: We are an AI native team - the expectation is to use AI to 100x your productivity. You get $100/month towards AI subscriptions and tools. Search `AI API Key Access Instructions` in Discord to get access.
+
+- **Language & Presentation Skills**: We grant a $100 yearly budget to improve in language skills (eg [Italki](https://www.italki.com/), [Elsa](https://elsaspeak.com/en/), [Cambly](https://www.cambly.com/english?lang=en) tutors, classes and books)
diff --git a/website/src/content/docs/handbook/culture/communicate.mdx b/website/src/content/docs/handbook/culture/communicate.mdx
new file mode 100644
index 000000000..c651ad38f
--- /dev/null
+++ b/website/src/content/docs/handbook/culture/communicate.mdx
@@ -0,0 +1,219 @@
+---
+title: How We Communicate
+description: Building a global AI platform requires clear, open communication
+---
+
+We're a fully-remote team building open superintelligence across continents
+and cultures. Clear communication isn't just nice to have—it's how we ship fast,
+stay aligned, and build trust with millions of users.
+
+## Core Communication Principles
+
+### Default to Open
+
+- **Build in Public**: Development discussions happen in open GitHub issues
+- **Share Progress**: Daily updates in Discord keep everyone informed
+- **Document Decisions**: Important choices are recorded for future reference
+- **Learn Together**: Share discoveries, failures, and insights with the team
+
+### Async First, Sync When Needed
+
+As a global team, we optimize for asynchronous communication:
+
+- **GitHub**: Technical discussions, code reviews, feature planning
+- **Discord**: Quick questions, daily updates, community engagement
+- **Documentation**: Long-form thinking, architectural decisions
+- **Meetings**: Only when real-time collaboration adds value
+
+### Write for Clarity
+
+With team members from different linguistic backgrounds:
+
+- **Simple English**: Clear over clever, direct over diplomatic
+- **Context Included**: Assume readers lack your specific context
+- **Examples Help**: Show, don't just tell
+- **Visual Aids**: Diagrams, screenshots, and code samples
+
+> Good: "The model fails to load on M1 Macs. Here's the error log and steps to reproduce..."
+> Bad: "It's broken on Mac."
+
+## Communication Channels
+
+### Where to Communicate What
+
+| Type | Channel | Examples |
+|------|---------|----------|
+| **Feature Development** | GitHub Issues | New features, bug reports, technical discussions |
+| **Daily Updates** | Discord #daily-updates | What you worked on, blockers, discoveries |
+| **Quick Questions** | Discord team channels | "Anyone know why X happens?" |
+| **Long-form Thinking** | GitHub Discussions / Docs | Architecture proposals, post-mortems |
+| **User Support** | Discord #support | Helping users with Jan |
+| **Urgent Issues** | Discord + @mention | Production bugs, security issues |
+
+### Response Time Expectations
+
+- **GitHub Issues**: Within 24-48 hours
+- **Discord Questions**: Best effort, timezone dependent
+- **User Bug Reports**: Acknowledge within 24 hours
+- **Security Issues**: Immediate escalation
+
+## Communication Best Practices
+
+### For Engineers
+
+```
+✅ DO:
+- Comment your code like someone will read it at 3am
+- Update documentation when you change behavior
+- Share WIP early for feedback
+- Close the loop: report back on solutions
+
+❌ DON'T:
+- Assume context that isn't written down
+- DM technical discussions (keep them public)
+- Wait until perfect to share progress
+- Leave PRs without description
+```
+
+### For Everyone
+
+**Assume Positive Intent**
+- We're all working toward the same goal
+- Language barriers can cause misunderstandings
+- Ask for clarification before assuming
+
+**Be Specific**
+- "The download button doesn't work" → "The download button on Windows returns 404 for model X"
+- "It's slow" → "Model loading takes 45 seconds on 8GB RAM"
+- "Users want this" → "15 users requested this in issue #123"
+
+**Close Loops**
+- Follow up on questions you ask
+- Update issues with solutions
+- Thank people who help you
+- Share outcomes of discussions
+
+## Meetings That Work
+
+We minimize meetings but when we have them:
+
+### Before the Meeting
+- Share agenda 24 hours prior
+- Include pre-read materials
+- State desired outcome clearly
+- Invite only necessary people
+
+### During the Meeting
+- Start with 5-minute silent reading of materials
+- Stick to agenda
+- Assign action items with owners
+- End 5 minutes early
+
+### After the Meeting
+- Post summary in relevant channel
+- Create GitHub issues for action items
+- Share recording if applicable
+
+## Writing Culture
+
+### Pull Requests
+
+```markdown
+## What
+Brief description of changes
+
+## Why
+Context and motivation
+
+## How
+Technical approach
+
+## Testing
+How to verify it works
+
+## Screenshots
+If UI changes
+```
+
+### Daily Updates
+
+Keep them brief but informative:
+
+```
+**Yesterday**: Shipped GGUF loader optimization (30% faster)
+**Today**: Working on Windows installer bug #456
+**Blockers**: Need review on PR #789
+**TIL**: Quantization affects different models differently
+```
+
+### Documentation
+
+- Write docs as you code
+- Include examples users can copy
+- Explain the why, not just the how
+- Keep it up to date or delete it
+
+## Community Communication
+
+When engaging with our open-source community:
+
+### Be Helpful
+- No question is too basic
+- Provide context and examples
+- Point to relevant documentation
+- Follow up on issues
+
+### Be Humble
+- We don't have all the answers
+- Users often have great ideas
+- Mistakes happen, own them
+- Thank contributors publicly
+
+### Be Human
+- Personality is welcome
+- Celebrate wins together
+- Share the journey
+- Build relationships
+
+## Global Team Considerations
+
+### Time Zones
+- Post updates at consistent times
+- Record important discussions
+- Rotate meeting times fairly
+- Respect off hours
+
+### Cultural Awareness
+- Direct feedback styles vary by culture
+- Silence doesn't mean agreement
+- Ask if unsure about interpretation
+- Celebrate diverse perspectives
+
+### Language
+- English is second language for many
+- Avoid idioms and slang
+- Use simple, clear language
+- Be patient with communication
+
+## Red Flags to Avoid
+
+- **Information Hoarding**: Share knowledge freely
+- **Private Discussions**: Keep technical talk public
+- **Assuming Context**: Document everything
+- **Delayed Responses**: Acknowledge even if you can't solve immediately
+- **Unclear Communication**: If confused, ask for clarification
+
+## The Jan Way
+
+Our communication style reflects our values:
+- **Open**: Like our code
+- **Inclusive**: Like our community
+- **Clear**: Like our mission
+- **Async**: Like our architecture
+- **Global**: Like our vision
+
+Good communication at Jan isn't about perfection—it's about clarity, openness, and building together across any distance.
+
+---
+
+*"The best code is documented code. The best decisions are documented decisions. The best team is one that communicates openly."*
diff --git a/website/src/content/docs/handbook/culture/fight.mdx b/website/src/content/docs/handbook/culture/fight.mdx
new file mode 100644
index 000000000..86fae1b20
--- /dev/null
+++ b/website/src/content/docs/handbook/culture/fight.mdx
@@ -0,0 +1,3 @@
+---
+title: How We Fight
+---
\ No newline at end of file
diff --git a/website/src/content/docs/handbook/how/sprints.mdx b/website/src/content/docs/handbook/how/sprints.mdx
new file mode 100644
index 000000000..f5588ee7d
--- /dev/null
+++ b/website/src/content/docs/handbook/how/sprints.mdx
@@ -0,0 +1,9 @@
+---
+title: "Sprints"
+description: "Jan Team Handbook"
+---
+
+[Jan sprints](https://github.com/orgs/menloresearch/projects/5/views/25) are every 2 weeks.
+
+Releases are QA'd prior to release. We never release on Fridays, unless ya'll
+wanna come back on Saturday.
diff --git a/website/src/content/docs/handbook/how/tools.mdx b/website/src/content/docs/handbook/how/tools.mdx
new file mode 100644
index 000000000..b54947305
--- /dev/null
+++ b/website/src/content/docs/handbook/how/tools.mdx
@@ -0,0 +1,37 @@
+---
+title: "Internal Tools"
+description: "Jan Team Handbook"
+---
+
+We like to make AI do the boring stuff. We also automate and dogfood our own tools as much as possible.
+
+### Tools
+
+- **Github**: For engineering specs, sprint planning, and documentation.
+- **Discord**: For chat, voice, and video calls.
+- **Google Hangouts**: For a short, daily standup.
+- **Gmail**: For external communication.
+- **HackMD/Obsidian**: For ad hoc ideas and meeting notes.
+- **Excalidraw**: For whiteboarding.
+- **Password Manager**: You will be invited.
+- **Jan/Cortex**: A personal, private copilot.
+
+### Infrastructure
+
+We use a mix of cloud providers and build our on-premises hardware to cut costs and optimize performance.
+
+- Singapore Cluster: For training and inference.
+- Hanoi Cluster: For CI/CD and data pipelines.
+- Burst Compute: We like Runpod for larger training runs.
+
+See [Infra](/handbook/infra/infra) for more details.
+
+### CI/CD
+
+We have a dedicated devops/mlops team. Ping in the `infra-internal` channel for any requests.
+
+### Automations
+
+We have a dedicated automations engineer. Her goal is to automate yesterday's tasks today, so that you are doing something different tomorrow.
+
+Ping in the `#automations` channel for any requests.
diff --git a/website/src/content/docs/handbook/hr/1-on-1s.mdx b/website/src/content/docs/handbook/hr/1-on-1s.mdx
new file mode 100644
index 000000000..e19e83fd1
--- /dev/null
+++ b/website/src/content/docs/handbook/hr/1-on-1s.mdx
@@ -0,0 +1,20 @@
+---
+title: "One on Ones"
+description: "One on Ones."
+---
+
+We don't schedule recurring 1-on-1s, because in our experience, engineers and researchers [hate it](https://www.reddit.com/r/cscareerquestions/comments/14rkbwv/does_anybody_else_hate_their_11_meetings/). We believe it's better to address issues and provide feedback as they arise.
+
+If you need to discuss something with us, simply book a time on the founder or team lead's calendar. **We're always available to chat.**
+
+✅ Do:
+- Schedule check-ins as soon as you need them
+- Provide honest feedback and criticism (we appreciate it!)
+
+❌ Don't:
+- Wait for a scheduled one-on-one to raise issues
+- Ask for confirmation before scheduling a check-in
+
+### Exceptions
+
+If you'd like to establish regular check-ins, please let us know and we're very happy to set one up.
diff --git a/website/src/content/docs/handbook/hr/interviewing.mdx b/website/src/content/docs/handbook/hr/interviewing.mdx
new file mode 100644
index 000000000..696f44f8c
--- /dev/null
+++ b/website/src/content/docs/handbook/hr/interviewing.mdx
@@ -0,0 +1,46 @@
+---
+title: "Interviewing"
+description: "How to interview and hire outliers."
+---
+
+Our talent pipeline is an organizational moat. This guide walks you through our interviewing process.
+
+### Process
+
+Jan's interview process is designed to be self-serve.
+
+1. **Application**: Candidates apply through our [careers page](https://homebrew.bamboohr.com/careers). We use a combination of AI and human review to shortlist candidates.
+2. **Initial Screening**: Shortlisted candidates are invited to a 30-minute initial screening, usually with HR/founder.
+3. **2-3 Interviews**: Candidates are interviewed by 2-3 team members. Each interview is 30-45 minutes long.
+4. **Founder interview**: The final interview is with one of the founders. This is a 30-minute interview.
+5. **Deliberation**: Candidates are evaluated based on their interviews and portfolio. We aim to make a decision within 1 week.
+
+We aim to be data-driven, and each open role is tracked in our ATS (BambooHR). We use this data to improve our hiring process.
+
+> Historically, our acceptance rate has been less than 1% of all applications. For direct referrals, this rate is higher at >10%.
+
+### Preparing for Interviews
+
+To start interviewing, please follow these steps:
+
+- [ ] Indicate your interest in helping with interviews to HR.
+- [ ] Shadow existing interviews (2-3) to understand our process and ask questions.
+- [ ] Create a [Google Appointments link](https://workspace.google.com/intl/en_sg/resources/appointment-scheduling/) for your interviews. Make sure to set a default Google Hangouts link. Share the link with HR.
+- [ ] New candidates will automatically schedule interviews with you using the Google Appointments link.
+- [ ] If you have an upcoming interview, review their resume and prepare thoughtful questions.
+- [ ] Input **detailed evaluations in BambooHR ATS** after each interview.
+
+### Evaluation
+
+We do not use a scoring system for interviews and prefer to encourage natural conversation.
+
+However, if you do need an evaluation template, you can use the following:
+
+| Criteria | Description | Response | Notes |
+|--------------|----------------------------------------|--------------|--------------------------------------------|
+| Technical Skills | Proficiency in relevant technical areas (AI, robotics, programming, etc.) indicated in the resume |||
+| Fundamentals | Understanding of core concepts in the candidate's field of expertise (math, statistics, ML, physics, etc.) | | |
+| Problem-Solving Ability | Ability to approach and solve complex problems |||
+| Communication Skills | Clarity and effectiveness in communication |||
+| Cultural Fit | Alignment with company [values](/handbook/philosophy/humanity) and [culture](/handbook/who/curiosity) |||
+| Overall Impression | General impression of the candidate's suitability for the role |||
diff --git a/website/src/content/docs/handbook/hr/leave.mdx b/website/src/content/docs/handbook/hr/leave.mdx
new file mode 100644
index 000000000..ffbe11cb2
--- /dev/null
+++ b/website/src/content/docs/handbook/hr/leave.mdx
@@ -0,0 +1,23 @@
+---
+title: "Time Off"
+description: "Vacation and unplanned leave policy."
+---
+
+We offer 14 base vacation days, unlimited sick leave, and additional public holidays based on your respective region.
+
+### Planning Leave
+
+1. Submit the days in BambooHR at least 1-2 weeks in advance
+2. Inform your team as early as possible
+
+### Public Holidays
+
+We observe public holidays in our respective countries of residence. You do not need to use vacation days.
+
+### Illness & Unplanned Leave
+
+Please submit all sick leave requests in BambooHR with a medical certificate (when applicable). For sick leave longer than 1 week, please inform HR and your team.
+
+### Parental
+
+We offer parental leave for eligible staff. Please inform HR to coordinate parental leave.
diff --git a/website/src/content/docs/handbook/hr/progression.mdx b/website/src/content/docs/handbook/hr/progression.mdx
new file mode 100644
index 000000000..d4105e07b
--- /dev/null
+++ b/website/src/content/docs/handbook/hr/progression.mdx
@@ -0,0 +1,20 @@
+---
+title: "One on Ones"
+description: "One on Ones."
+---
+
+### Promotions
+
+We don't believe in titles. We believe in fair compensation and fast progression.
+
+> Previously, we have been known to organically promote team members 2x within 6 months of joining.
+
+### Perks
+
+- Your contributions are Open Source and public
+- Speaking opportunities at conferences (we get invited often)
+- Direct access to a large network of advisors
+- No shortage of hard problems (you don't need to "prove" yourself before working on something interesting)
+- Visibility across our hardware, software, and research projects
+- Real, blunt feedback, at scale (our users are not shy!)
+- Hard, tough, evergreen problems
diff --git a/website/src/content/docs/handbook/hr/retreats.mdx b/website/src/content/docs/handbook/hr/retreats.mdx
new file mode 100644
index 000000000..2560a7bd8
--- /dev/null
+++ b/website/src/content/docs/handbook/hr/retreats.mdx
@@ -0,0 +1,17 @@
+---
+title: "Retreats"
+description: "Annual Team Retreat."
+---
+
+## Retreats
+
+We host an annual team retreat to bring the team together, bond, and plan for the year ahead.
+
+Previous retreats have been held in:
+- 2023 (8 people): Ho Chi Minh City, Vietnam
+- 2024 (16 people): Singapore
+- 2025: TBA!
+
+## Meetups
+
+We encourage team members to meet up in person whenever possible. Remote work can get lonely, and we offer a small budget for regional get-togethers.
diff --git a/website/src/content/docs/handbook/hr/side-gigs.mdx b/website/src/content/docs/handbook/hr/side-gigs.mdx
new file mode 100644
index 000000000..a5e9a73dc
--- /dev/null
+++ b/website/src/content/docs/handbook/hr/side-gigs.mdx
@@ -0,0 +1,26 @@
+---
+title: "Side Gigs"
+description: "Best practices around external engagements."
+---
+
+We have a permissive policy regarding side gigs, similar to [Posthog's](https://posthog.com/handbook/people/side-gigs).
+
+However, our main ask is for Jan to be your "main gig", as you will be taking a crucial role in success/failure of Jan. Manage your time in a way that is fair to the company and your colleagues.
+
+At the end of the day, you all own (or will own) shares of the company. Handle company IP as you would your own property.
+
+### Declaration
+
+**For work in related industries, we do ask for an upfront declaration**, so that we can clear potential conflicts of interest.
+
+Please fill out the following form and submit it to: `hr@menlo.ai`
+
+| Dates | Organization | Nature of Work | Hours/week |
+|-------------|--------------|----------------|------------|
+| | | | |
+| | | | |
+| | | | |
+
+### A Note on IP
+
+Startups & careers die because of IP disputes from moonlighting employees. Please respect all party's IP - never intermingle assets (like laptops and logins) between engagements.
diff --git a/website/src/content/docs/handbook/hr/spending.mdx b/website/src/content/docs/handbook/hr/spending.mdx
new file mode 100644
index 000000000..48e72dabf
--- /dev/null
+++ b/website/src/content/docs/handbook/hr/spending.mdx
@@ -0,0 +1,36 @@
+---
+title: "Spending"
+description: "Spending and reimbursement policy."
+---
+
+
+We are incredibly frugal and intend to be around in 50 years. However, we avoid being "penny-wise & pound-foolish" and spend when it matters.
+
+### Expenses
+
+#### Software
+
+For software subscriptions and other expenses incurred as a part of [benefits](compensation#benefits), please save receipts and submit it as a reimbursement [request](#reimbursements).
+
+#### Large Purchases
+
+For larger purchases like equipment, please first submit a purchase request through `#admin-requests` in Discord.
+
+We can directly order for you or reimburse you for the purchase.
+
+#### Incidentals
+
+We don't have a policy for incidentals. We trust you to make the right decisions, while keeping in mind that we are a startup and every dollar counts.
+
+### Reimbursements
+
+Reimbursements are processed at the end of every month with payroll.
+
+Email all receipts (required) & the following form to `finance@menlo.ai`:
+
+| Date | Description | Amount | Currency | Exchange Rate | Total (USD) |
+|------------|-------------|--------|----------|---------------|-------------|
+| | | | | | |
+| | | | | | |
+| | | | | | |
+| | | | | Grand Total | |
diff --git a/website/src/content/docs/handbook/hr/travel.mdx b/website/src/content/docs/handbook/hr/travel.mdx
new file mode 100644
index 000000000..a446435cc
--- /dev/null
+++ b/website/src/content/docs/handbook/hr/travel.mdx
@@ -0,0 +1,19 @@
+---
+title: "Travel"
+description: "Travel policy."
+---
+
+
+We encourage team members to represent Jan at conferences and travel for in-person alignment
+
+For all work travel, we ask that you get approval from HR before confirming your trip & booking flights/hotels.
+
+### Bookings
+
+We prefer to directly book flights & hotels on your behalf, through our travel portals.
+
+If you need to arrange travel by yourself, please obtain a travel budget first. Save all receipts for reimbursement.
+
+### Per Diem
+
+Depending on the location, we provide a $20 USD per diem for meals and incidentals. This amount is adjustable based on the location.
diff --git a/website/src/content/docs/handbook/index.mdx b/website/src/content/docs/handbook/index.mdx
new file mode 100644
index 000000000..f023d7bc3
--- /dev/null
+++ b/website/src/content/docs/handbook/index.mdx
@@ -0,0 +1,82 @@
+---
+title: "Jan Team Handbook"
+description: "Building superintelligence that you can own and run anywhere."
+---
+
+{/* TODO: Replace with Astro-compatible icons: FaDiscord, FaMap */};
+import Steps from '@/components/Steps.astro';
+
+## Jan Handbook
+
+> Jan's Handbook is inspired by [Posthog](https://posthog.com/handbook) and [Gitlab](https://handbook.gitlab.com/).
+> Thank you for showing us the way 🙏
+
+### Welcome!
+
+This handbook explains how [Jan](https://jan.ai) works, and is public.
+
+We're building superintelligence that you can self-host and use locally. Not as a limitation, but as a feature. Your AI should work wherever you need it - on your laptop during a flight, on your company's servers for compliance, or in the cloud for scale.
+
+Jan's Handbook is a [living document](https://en.wikipedia.org/wiki/Living_document), constantly evolving as we build the future of AI ownership.
+
+## 1. Chapter 1: Why does Jan exist?
+
+- [What problem are we solving?](/handbook/why/problem) - Why AI needs to be ownable, not just rentable
+- [Who we are building for](/handbook/why/icp) - People who want flexibility and control
+- [Our vision for open superintelligence](/handbook/why/ownership) - An ecosystem you can own
+
+## 2. Chapter 2: How we make money
+
+- [How we make money](/handbook/money/how) - Open core + optional services
+- [What is our moat](/handbook/money/moat) - Community, trust, and aligned incentives
+
+## 3. Chapter 3: Who we hire
+
+- [The Fast and the Curious](/handbook/who/curiosity) - Ship today, learn for tomorrow
+- [Underdogs Welcome](/handbook/who/underdogs) - Talent is everywhere, opportunity isn't
+- [dy/dx > y-intercept](/handbook/who/dydx) - Growth rate beats starting position
+- [Obsession](/handbook/who/obsession) - We seek those who can't not build this
+- [Team, not family](/handbook/who/team) - Pro sports team approach
+
+## 4. Chapter 4: How we work
+
+- [Remote-first](/handbook/how/remote) - Global team, local impact
+- [Open-source](/handbook/how/open-source) - Build in public
+- [How we coordinate](/handbook/how/tools) - Tools and processes
+- [Sprints](/handbook/how/sprints) - Ship every week
+- [How we communicate](/handbook/culture/communicate) - Clear, open, async-first
+
+## 5. Chapter 5: Jan's Philosophy
+
+- [Open Superintelligence Platform](/handbook/philosophy/open-superintelligence) - Not one model, but an ecosystem
+- [Lateral Thinking with Proven Technology](/handbook/philosophy/lateral) - Innovation through integration
+- [Humanity-aligned](/handbook/philosophy/humanity) - Technology that unites
+- [Perpetual Optimism](/handbook/philosophy/optimism) - Building the future we want to live in
+
+## 6. Chapter 6: Team & Operations
+
+- [Team Structure](/handbook/team) - How we're organized
+- [Compensation](/handbook/comp) - Fair pay, meaningful equity
+- [HR & Culture](/handbook/hr/onboarding) - Joining and growing with Jan
+
+---
+
+## Quick Links
+
+- **For new team members**: Start with [Onboarding](/handbook/hr/onboarding)
+- **For contributors**: Check out our [GitHub](https://github.com/janhq) and [Discord](https://discord.gg/FTk2MvZwJH)
+- **For the curious**: Read about [our vision](/handbook/why/ownership)
+
+## Our North Star
+
+We're building superintelligence that:
+- **Works anywhere**: From your laptop to your data center
+- **Belongs to you**: Download it, own it, modify it
+- **Scales infinitely**: One person or ten thousand, same platform
+- **Improves constantly**: Community-driven development
+
+This isn't just about making AI accessible. It's about ensuring the most transformative technology in human history can be owned by those who use it.
+
+---
+
+*"The future of AI isn't about choosing between local or cloud. It's about having both, and everything in between, working perfectly together."*
\ No newline at end of file
diff --git a/website/src/content/docs/handbook/lifecycle/onboarding.mdx b/website/src/content/docs/handbook/lifecycle/onboarding.mdx
new file mode 100644
index 000000000..9bf1bbcea
--- /dev/null
+++ b/website/src/content/docs/handbook/lifecycle/onboarding.mdx
@@ -0,0 +1,73 @@
+---
+title: "Onboarding to Jan"
+description: "Onboarding Checklist for New Hires."
+asIndexPage: true
+---
+
+Welcome to Jan! We are excited to have you join our team. This guide walks you through the onboarding process.
+
+> You will receive an onboarding invitation morning of your first day. Prior to the scheduled onboarding call, please send your Github & Discord handles to `hr@menlo.ai`.
+
+### Expectations
+
+| Expectation | Description |
+|-------------|-------------|
+| **Take Initiative** | Take ownership of an area. If you see a problem, take it and own it to completion. Your work will often not be defined. Take the initiative to figure out what needs to be done, seek clarification, and then communicate what your plan to the team. |
+| **Bias to Action** | There are many problems to solve. Don't ask for permission or try to build consensus - just take action. After 2-3 months, you should be able to show clear results having turned around a problem-filled area. |
+| **Communication** | Clear and timely communication is key. If unsure, ask questions. We are a global team and respect is paramount. Disrespectful behavior is not tolerated. Focus on problem-solving, not personal attacks. Assume Hanlon’s Razor: “Never attribute to malice that which is adequately explained by lack of ~~stupidity~~ communication or ambiguity.” |
+| **Mastery** | In this frontier industry, expertise comes from doing. Own your area and build mastery. |
+
+For more details, see our guides on [remote work](/handbook/how/remote).
+
+### Code of Conduct
+
+- **Availability and Communication**: Ensure you are available and engaged during designated work hours and scheduled meetings.
+- **Work Environment**: Strive to create a quiet and distraction-free workspace whenever possible, especially on video calls.
+- **Video Participation**: Video participation is expected unless there are exceptional circumstances.
+- **Dress Code**: Casual attire is perfectly acceptable for meetings, but please exercise good judgment (e.g. no pajamas).
+- **Vacations:** Communicate leave at least 2 weeks (1 sprint) in advance via Bamboo.
+- **Emergency Leave:** Please inform Daniel, Nicole and HR in your #hr-channel if you require emergency leave.
+- **Side Gigs Policy**: See [side-gigs](side-gigs). Please consult with HR on engagements with potential legal & IP implications.
+- **Sexual harassment:** We have a zero tolerance policy against behavior of a sexual nature that could reasonably be expected to cause offense or humiliation, e.g. verbal, nonverbal, or physical conduct, via written and electronic communications.
+
+### Tools
+
+As much as possible, we build-in-public and use the following tools to asynchronously collaborate:
+
+- [Github](https://github.com/menloresearch)
+- [Jan Discord](https://discord.gg/VSbRN3vwCD)
+- [Google Workspace](https://workspace.google.com)
+- [Hugging Face](https://huggingface.co/menloresearch)
+- Password Manager: You will be invited
+- AI Tools and API keys: Coding with AI is heavily encouraged
+
+### Checklists
+
+#### Day 1
+
+- Sign all HR documents.
+- Download and access all tools.
+- Check calendar invites: daily standups and TGIF.
+- Introduce yourself in the [`#welcome`](https://discord.gg/VSbRN3vwCD) Discord channel.
+- Set up your [BambooHR](https://homebrew.bamboohr.com/home) account.
+- Set up VPN. Search `VPN access instructions` in Discord for the latest instructions.
+- Check out the current sprint in [Github](https://github.com/orgs/menloresearch/projects/5)
+- Ask questions in your private `#hr-NAME` channel.
+
+
+Import **Who's out** (on leave) calendar from BambooHR
+
+- Go to https://homebrew.bamboohr.com/calendar. Login if needed.
+- Click on the gear icon, select **iCal Feeds...**
+- Select **Create Calendar Feed** under **Who's Out**. Copy the generated link.
+- In Google Calendar, you can import the new calendar from URL.
+
+
+
+#### Week 1
+
+- Post daily updates in the [`#daily-updates`](https://discord.gg/AxypHJRQxd) channel prior to standup.
+- Review this [Jan Handbook](https://menlo.ai/handbook).
+- Push 1 PR into this Handbook. This a living document!
+- Disclose side gigs with potential legal & IP implications to HR.
+- Attend TGIF demos on Friday afternoon (6PM GMT+8).
diff --git a/website/src/content/docs/handbook/money/how.mdx b/website/src/content/docs/handbook/money/how.mdx
new file mode 100644
index 000000000..a49701b1c
--- /dev/null
+++ b/website/src/content/docs/handbook/money/how.mdx
@@ -0,0 +1,144 @@
+---
+title: "How we make money"
+description: "Building a sustainable business around open superintelligence"
+---
+
+Jan is building a sustainable business that aligns our incentives with our users'
+success. We believe the best way to build beneficial AI is to ensure our business
+model reinforces our values: privacy, ownership, and universal access.
+
+## Our Business Model
+
+### 1. Open Core Foundation
+
+**Jan Desktop is free forever**
+- Download and run AI models locally
+- No subscriptions, no accounts required
+- Full privacy and ownership of your data
+- Community-driven development
+
+This isn't a loss leader—it's our foundation. A thriving open-source community creates:
+- Trust through transparency
+- Innovation through collaboration
+- Distribution without marketing spend
+- Feedback loops that improve the product
+
+### 2. Premium Features & Services
+
+We monetize through optional add-ons that enhance the Jan experience:
+
+**Jan Cloud (Coming Soon)**
+- Optional cloud sync across devices
+- Managed model hosting for teams
+- Advanced collaboration features
+- Pay only for what you use
+
+**Enterprise Support**
+- Priority support and SLAs
+- Custom model deployment
+- Air-gapped installations
+- Compliance and audit tools
+
+**Professional Tools**
+- Advanced model fine-tuning interfaces
+- Team management and permissions
+- Analytics and usage insights
+- API rate limit increases
+
+### 3. Model Marketplace
+
+**Curated Model Hub**
+- Premium access to specialized models
+- Early access to new model releases
+- Commercial licensing for enterprise models
+- Revenue sharing with model creators
+
+We use **Jan Exam** to ensure quality—whether models are free or paid, they must meet our standards.
+
+### 4. Hardware Integration
+
+**Optimized Hardware Solutions**
+- Pre-configured devices for AI inference
+- Plug-and-play AI workstations
+- Edge deployment solutions
+- Partnership with hardware manufacturers
+
+### 5. Training & Certification
+
+**Jan Certified Program**
+- Professional certification for developers
+- Enterprise training workshops
+- Educational institution partnerships
+- Community education initiatives
+
+## Revenue Philosophy
+
+### What We'll Never Do
+
+- **Sell your data**: Your conversations remain private
+- **Lock you in**: Always provide export and migration tools
+- **Paywall core features**: Local AI remains free
+- **Compromise on privacy**: No ads, no tracking, no surveillance
+
+### What We Always Do
+
+- **Align incentives**: We succeed when you succeed with AI
+- **Transparent pricing**: Clear, predictable costs
+- **Community first**: Free tier remains fully functional
+- **Open standards**: No proprietary lock-in
+
+## Unit Economics
+
+Our approach creates sustainable unit economics:
+
+1. **Zero marginal cost** for open-source users
+2. **High-margin** cloud and enterprise services
+3. **Network effects** as community grows
+4. **Reduced CAC** through open-source distribution
+
+## The Flywheel
+
+```
+Open Source Adoption → Community Growth → Better Models →
+More Users → Enterprise Adoption → Revenue → Investment in Open Source
+```
+
+Each turn of the wheel strengthens the next:
+- More users improve models through feedback
+- Better models attract enterprise customers
+- Enterprise revenue funds open development
+- Open development attracts more users
+
+## Long-term Vision
+
+We're building for a future where:
+
+- Every organization runs their own AI
+- Privacy is the default, not premium
+- Open models outperform closed ones
+- AI infrastructure is as common as web servers
+
+Our business model ensures we'll be here to see it through.
+
+## Current Status
+
+As of 2024:
+- ✅ Sustainable open-source project
+- ✅ Growing community adoption
+- 🚧 Enterprise features in development
+- 🚧 Cloud services in beta
+- 📅 Hardware partnerships planned
+
+## Join Our Mission
+
+Whether you're a:
+- **User**: Your adoption drives our mission
+- **Contributor**: Your code shapes our platform
+- **Customer**: Your success funds development
+- **Partner**: Your integration expands possibilities
+
+You're part of building the open superintelligence economy.
+
+---
+
+*"The best business model is one where doing the right thing is also the profitable thing. That's what we're building at Jan."*
diff --git a/website/src/content/docs/handbook/money/moat.mdx b/website/src/content/docs/handbook/money/moat.mdx
new file mode 100644
index 000000000..7faedf2db
--- /dev/null
+++ b/website/src/content/docs/handbook/money/moat.mdx
@@ -0,0 +1,169 @@
+---
+title: "What is our moat?"
+description: "Building defensibility through community, trust, and aligned incentives"
+---
+
+Traditional moats rely on keeping secrets, locking in users, or maintaining
+technical advantages. Jan's moat is different—it's built on openness, community,
+and aligned values that are impossible to replicate with a closed approach.
+
+## Community as Competitive Advantage
+
+Our strongest moat is our community. While others guard their code, we share everything:
+
+### Network Effects at Scale
+- Every user improves Jan for everyone else
+- Bug reports, feature requests, and contributions compound
+- Community support scales infinitely
+- Local communities solve local problems
+
+### Trust Through Transparency
+- Open source code earns trust closed systems can't buy
+- Users verify our privacy claims themselves
+- No hidden agendas or dark patterns
+- Mistakes are public, fixes are collaborative
+
+### Innovation Velocity
+- Thousands of contributors vs hundreds of employees
+- Global perspectives vs Silicon Valley groupthink
+- 24/7 development across time zones
+- Passionate users become passionate builders
+
+## Technical Moats
+
+### Local-First Architecture
+While others race to build bigger data centers, we're optimizing for edge computing:
+- Years of optimization for consumer hardware
+- Deep integration with local systems
+- Efficient model quantization expertise
+- Cross-platform compatibility knowledge
+
+### Privacy-Preserving Design
+Privacy isn't a feature we added—it's our foundation:
+- Architecture that makes surveillance impossible
+- No user data to monetize or leak
+- Local-first eliminates attack surfaces
+- Trust that compounds over time
+
+### Model Agnostic Platform
+We're not tied to any single model:
+- Support for all open models
+- Jan Exam ensures quality across providers
+- Users aren't locked to our models
+- Best-in-class always available
+
+## Business Model Alignment
+
+Our moat strengthens because our business model reinforces our values:
+
+### No Conflict of Interest
+- We don't profit from user data
+- No ads means no perverse incentives
+- Success comes from user success
+- Premium features enhance, not gatekeep
+
+### Sustainable Growth
+- Open source distribution = zero CAC
+- Community support = reduced support costs
+- User contributions = free R&D
+- Word of mouth = organic growth
+
+## Cultural Moats
+
+### Values Alignment
+Users choose Jan because they share our values:
+- Privacy is non-negotiable
+- Ownership matters
+- Local-first is the future
+- Community over corporation
+
+### Mission-Driven Team
+- We attract talent that believes in the mission
+- Lower salaries offset by meaningful work
+- Equity alignment with long-term vision
+- Cultural fit over pure technical skills
+
+## The Anti-Moat Strategy
+
+Paradoxically, our moat comes from not trying to build traditional moats:
+
+### No Lock-In Creates Loyalty
+- Easy to leave = users choose to stay
+- Export everything = trust in the platform
+- No switching costs = genuine preference
+- Freedom of choice = actual choice
+
+### Open Source as Defense
+- Can't be acquired and shut down
+- Can't be feature-frozen by new management
+- Community can fork if we lose our way
+- Immortal through decentralization
+
+## Compounding Advantages
+
+Our moats compound over time:
+
+```
+Year 1: Build trust through transparency
+Year 2: Community starts contributing significantly
+Year 3: Network effects accelerate development
+Year 4: Ecosystem becomes self-sustaining
+Year 5: Platform effect makes leaving costly (by choice)
+```
+
+## What Can't Be Copied
+
+Competitors can't replicate:
+
+### Time and Trust
+- Years of consistent privacy-first decisions
+- Track record of putting users first
+- Community relationships built over time
+- Reputation for reliability and openness
+
+### Community Culture
+- Shared values and mission
+- Collaborative problem-solving approach
+- Global perspective on AI needs
+- Bottom-up innovation mindset
+
+### Architectural Decisions
+- Local-first can't be bolted onto cloud-first
+- Privacy can't be added to surveillance systems
+- Community-driven can't be faked by corporations
+- Open source commitment can't be half-hearted
+
+## The Ultimate Moat
+
+Our ultimate moat is simple: **we're building what we'd want to use**.
+
+- We're users of our own platform
+- We feel the pain points personally
+- We can't betray users without betraying ourselves
+- Our incentives perfectly align with our community
+
+## Sustainable Defensibility
+
+Traditional moats erode:
+- Patents expire
+- Technical advantages get copied
+- Network effects can shift platforms
+- Regulations can break monopolies
+
+Jan's moats strengthen:
+- Trust compounds daily
+- Community grows stronger
+- Values attract like-minded people
+- Open source ensures immortality
+
+## Conclusion
+
+Our moat isn't about keeping others out—it's about bringing everyone in. By building in the open, aligning our incentives with users, and creating genuine value, we're building defensibility that transcends traditional business strategy.
+
+The question isn't "How do we prevent competition?" but rather "How do we make competition irrelevant by building something that can't exist any other way?"
+
+That's our moat: being the platform that puts users first, not because it's good business, but because it's the only way to build the AI future we want to live in.
+
+---
+
+*"The strongest moat is a community that would rebuild you if you disappeared."*
diff --git a/website/src/content/docs/handbook/philosophy/humanity.mdx b/website/src/content/docs/handbook/philosophy/humanity.mdx
new file mode 100644
index 000000000..c8a1e6807
--- /dev/null
+++ b/website/src/content/docs/handbook/philosophy/humanity.mdx
@@ -0,0 +1,62 @@
+---
+title: "Humanity-aligned"
+description: "Jan exists because technology lets humanity work as one."
+---
+
+
+
+
+> "The imagined community of the nation is always limited and sovereign, yet human connections and cooperation stretch far beyond borders." — Benedict Anderson
+
+## Why this matters
+
+- AI is geopolitical. Nation-states compete to control it
+- We are concerned about those using AI to stoke division, profit from lies, and spur conflict
+- Some of us come from places where war isn't distant—it's family history
+- Technology should unite, not divide
+
+## Who we are
+
+- Remote team across 7 countries
+- Diverse backgrounds, united by mission
+- Many of us are [third culture kids](https://en.wikipedia.org/wiki/Third_culture_kid)—growing up between worlds
+- We often find ourselves misunderstood: an Asian-majority team, English-speaking, influenced by global philosophies
+- Crossroad cultures shape us:
+ - Turkiye: Europe, Asia, Middle East converge
+ - Singapore: Trade hub bridging worlds
+- We respect each other's cultures and build on shared values
+
+## Jan's stance
+
+- Humanity first. We build for people, not governments or factions
+- AI should enable:
+ - Shared prosperity
+ - Universal education
+ - Peaceful collaboration
+- Technology **must** empower humanity to do more—together
+
+## The bigger picture
+
+- Human history is one of scaling cooperation:
+ - Small-scale [kin groups](https://www.sciencedirect.com/science/article/pii/S0960982219303343) → diverse political formations → [modern nation-states](https://en.wikipedia.org/wiki/Westphalian_system) → global networks
+ - Empires rose and fell. Nationalism united and divided. Globalization connected and excluded
+- History doesn't progress. It moves—messy, cyclical, and full of contradiction
+- Technology changes the terrain:
+ - Like ant colonies forming complex systems from simple interactions, humans have always built networks beyond central control
+ - Complexity emerges from countless small decisions—but unlike ants, we carry ideologies, ambitions, and fears
+- AI is another fork in the road. It can reinforce old hierarchies or dismantle them. It can be used to surveil or to liberate
+
+## Why we exist
+
+- 30 people, from different countries, met online to build together
+- The internet enables connections that were impossible a generation ago
+- Ideas cross borders: an anthropologist in Turkiye collaborates with a roboticist in Saigon
+- Jan exists because technology lets humanity work as one
+
+## Our vision
+
+- AI can accelerate global coordination and shared progress
+- Our goal: help humanity align, collaborate, and solve collective challenges
+- Success = contributing to humanity's long arc toward unity
+
+If our work helps the world coordinate better—even slightly—we've done something that matters
diff --git a/website/src/content/docs/handbook/philosophy/lateral.mdx b/website/src/content/docs/handbook/philosophy/lateral.mdx
new file mode 100644
index 000000000..b783a0880
--- /dev/null
+++ b/website/src/content/docs/handbook/philosophy/lateral.mdx
@@ -0,0 +1,147 @@
+---
+title: "Lateral Thinking with Proven Technology"
+description: "Building the future with battle-tested tools"
+---
+
+> "Innovation is not about using the newest technology, but about using technology in new ways." — Adapted from Nintendo's philosophy
+
+## Our Approach
+
+At Jan, we don't chase the bleeding edge for its own sake. We take proven, battle-tested technologies and combine them in innovative ways to solve real problems. This philosophy shapes everything we build.
+
+## Why Proven Technology Wins
+
+### Stability Over Novelty
+- **GGUF Format**: We didn't invent a new model format. We adopted the community standard that already works across platforms
+- **llama.cpp**: Years of optimization by brilliant minds—why reinvent when we can build upon?
+- **Tauri + Web Tech**: Proven UI stack that works everywhere, letting us focus on AI innovation
+- **SQLite**: The world's most deployed database for local data—perfect for privacy-first architecture
+
+### Accessibility Through Maturity
+When we use proven technology:
+- Hardware requirements are well understood
+- Optimization paths are clear
+- Community knowledge exists
+- Documentation is comprehensive
+- Edge cases are known
+
+This means Jan works on more devices, for more people, with fewer surprises.
+
+## Lateral Innovation Examples
+
+### Local AI That Actually Works
+**Traditional approach**: Build massive cloud infrastructure, require constant internet
+**Our approach**: Use proven quantization techniques to run powerful models on consumer hardware
+
+### Privacy Without Compromise
+**Traditional approach**: Complex encryption and privacy theater
+**Our approach**: Simply don't collect data. Use local SQLite. Problem solved.
+
+### Universal Model Support
+**Traditional approach**: Proprietary model formats and exclusive partnerships
+**Our approach**: Support the open GGUF standard that the community already uses
+
+### Cross-Platform Without Complexity
+**Traditional approach**: Native apps for each platform, massive development overhead
+**Our approach**: One codebase using Electron, works everywhere, ships faster
+
+## The Power of Constraints
+
+Working with proven technology creates healthy constraints:
+
+### Resource Efficiency
+- Can't throw infinite compute at problems
+- Must optimize for consumer hardware
+- Forces elegant solutions over brute force
+- Makes us accessible globally, not just in rich markets
+
+### Clear Boundaries
+- Known limitations guide design decisions
+- Realistic about what's possible today
+- Focus on solving real problems now
+- Build stepping stones to the future
+
+### Community Alignment
+- Use formats the community knows
+- Build on protocols people trust
+- Contribute improvements back upstream
+- Stand on the shoulders of giants
+
+## Innovation Through Integration
+
+Our innovation comes from how we combine proven pieces:
+
+```
+llama.cpp (proven inference)
+ + GGUF (standard format)
+ + Electron (cross-platform UI)
+ + SQLite (local data)
+ + MCP (extensibility)
+ = Jan (accessible AI for everyone)
+```
+
+Each component is battle-tested. The magic is in the combination.
+
+## Real Problems, Real Solutions
+
+This philosophy keeps us grounded:
+
+### What We Build
+- AI that runs on the laptop you already own
+- Privacy that doesn't require a PhD to understand
+- Tools that work offline in rural areas
+- Features users actually need today
+
+### What We Don't Build
+- Exotic architectures that need exotic hardware
+- Complex systems that break in simple ways
+- Features that demo well but fail in practice
+- Solutions looking for problems
+
+## The Compound Effect
+
+Using proven technology creates compounding benefits:
+
+1. **Faster Development**: Less time debugging novel approaches
+2. **Better Reliability**: Years of battle-testing by others
+3. **Easier Adoption**: Users already understand the patterns
+4. **Stronger Ecosystem**: Can leverage existing tools and knowledge
+5. **Lower Costs**: Commodity hardware and proven optimizations
+
+## Learning from History
+
+Great innovations often use "old" technology in new ways:
+- The iPhone used existing touchscreen tech
+- Tesla started with laptop batteries
+- SpaceX used proven rocket designs more efficiently
+- The web succeeded using simple protocols
+
+Jan follows this tradition: proven AI inference, standard formats, and simple principles—combined in a way that makes AI accessible to everyone.
+
+## Building for the Long Term
+
+Proven technology has staying power:
+- Standards that last decades
+- Protocols that survive company failures
+- Formats that outlive their creators
+- Communities that maintain forever
+
+By building on these foundations, Jan can focus on what matters: making AI useful, private, and accessible to everyone.
+
+## Our Commitment
+
+We commit to:
+- **Practical over Perfect**: Ship what works today
+- **Simple over Sophisticated**: Elegant solutions to real problems
+- **Proven over Promising**: Build on solid foundations
+- **Accessible over Advanced**: Reach everyone, not just enthusiasts
+
+## The Future is Already Here
+
+The technologies needed for private, local AI already exist. They're proven, optimized, and ready. Our job isn't to invent new technology—it's to combine what exists in ways that serve humanity.
+
+That's lateral thinking. That's how we build the future: not by reaching for tomorrow's technology, but by using today's technology in tomorrow's ways.
+
+---
+
+*"The best technology is invisible. It just works, everywhere, for everyone."*
diff --git a/website/src/content/docs/handbook/philosophy/open-superintelligence.mdx b/website/src/content/docs/handbook/philosophy/open-superintelligence.mdx
new file mode 100644
index 000000000..2fb1ecd35
--- /dev/null
+++ b/website/src/content/docs/handbook/philosophy/open-superintelligence.mdx
@@ -0,0 +1,213 @@
+---
+title: "Open Superintelligence Platform"
+description: "Building superintelligence as an ecosystem you can own and deploy anywhere"
+---
+
+
+> "Superintelligence isn't one massive model. It's an ecosystem of specialized models, tools, and applications working together - and you should own all of it." — Jan Philosophy
+
+## What is Open Superintelligence?
+
+Open superintelligence is AI that adapts to how you want to work, not the other way around. It's not about forcing you to choose between local or cloud, powerful or private, easy or extensible. It's about having it all, and owning it.
+
+### The Ecosystem Approach
+
+We're not trying to build GPT-5. We're building something better: specialized models that excel at specific tasks, tools that extend their capabilities, and applications that tie it all together.
+
+```
+Models (specialized for tasks)
+ + Tools (extend capabilities)
+ + Applications (work everywhere)
+ = Superintelligence you own
+```
+
+## Why Not One Giant Model?
+
+The "one model to rule them all" approach has fundamental flaws:
+
+- **Inefficient**: Using the same weights for poetry and mathematics
+- **Inflexible**: Can't optimize for specific use cases
+- **Expensive**: Massive compute for simple tasks
+- **Monolithic**: Single point of failure and control
+
+### Our Approach: Specialized Excellence
+
+- **Jan-Search**: Knows how to find and synthesize information
+- **Jan-Write**: Understands structure, tone, and creativity
+- **Jan-Analyze**: Excels at reasoning and data interpretation
+- **Jan-Code**: Optimized for programming tasks
+
+Each model does one thing brilliantly. Together, they form superintelligence.
+
+## The Three Pillars
+
+### 1. Models That Adapt
+
+Not just our models, but any model you need:
+- **Jan Models**: Purpose-built for specific tasks
+- **Community Models**: Any GGUF from Hugging Face
+- **Cloud Models**: Connect to OpenAI, Anthropic when needed
+- **Your Models**: Fine-tune and deploy your own
+
+### 2. Tools That Extend
+
+Models need capabilities beyond conversation:
+- **Search**: Get answers, not just links
+- **DeepResearch**: Autonomous multi-step investigation
+- **BrowserUse**: Let AI interact with the web
+- **MCP Protocol**: Connect any data source or API
+
+### 3. Applications That Scale
+
+Same experience, infinite deployment options:
+- **Jan Desktop**: Your local AI workstation
+- **Jan.ai**: Browser-based, no setup required
+- **Jan Mobile**: AI that follows you (coming soon)
+- **Jan Server**: Self-hosted for teams (coming soon)
+
+## Ownership Through Flexibility
+
+True ownership means having choices:
+
+### Choose Your Deployment
+- **Full Local**: Everything on your hardware, works offline
+- **Full Cloud**: We host it, you just use it
+- **Hybrid**: Local for sensitive work, cloud for heavy compute
+- **Self-Hosted**: Your servers, your control
+
+### Choose Your Models
+- Use our models when they're best
+- Use community models when they fit
+- Use cloud models when you need them
+- Train your own when you want to
+
+### Choose Your Scale
+- **Personal**: Run on your laptop
+- **Team**: Deploy on your server
+- **Enterprise**: Scale across infrastructure
+- **Global**: Distribute to the edge
+
+## Building in the Open
+
+We don't just open source our code. We open source our entire development process:
+
+### Watch Us Train
+Live at [train.jan.ai](https://train.jan.ai):
+- See models training in real-time
+- View our datasets and methods
+- Learn from our failures
+- Track actual progress
+
+### Test With Us
+Evaluate at [eval.jan.ai](https://eval.jan.ai):
+- Compare model outputs
+- Vote on what works
+- Suggest improvements
+- Access all evaluation data
+
+### Build With Us
+Everything on [GitHub](https://github.com/janhq):
+- Contribute features
+- Report issues
+- Fork and modify
+- Join discussions
+
+## The Network Effect of Openness
+
+Open systems create compound benefits:
+
+### For Users
+- More models to choose from
+- More tools to extend capabilities
+- More deployment options
+- More control over your AI
+
+### For Developers
+- Build on a stable platform
+- Access to growing ecosystem
+- No vendor lock-in fears
+- Community support
+
+### For Organizations
+- Deploy how you need
+- Customize for your use case
+- Control your costs
+- Own your infrastructure
+
+## Quality Without Compromise
+
+**Jan Exam** ensures excellence across the ecosystem:
+- Objective benchmarks
+- Real-world testing
+- Community validation
+- Transparent results
+
+Whether it's our model or someone else's, if it performs well, it belongs on Jan.
+
+## The Path Forward
+
+### Today (Available Now)
+- Jan Desktop with local and cloud models
+- Basic tools via MCP
+- Growing model ecosystem
+- Active community
+
+### Next 12 Months
+- Jan v1 specialized models
+- Advanced tools (search, research, browser)
+- Jan Server for teams
+- Mobile applications
+
+### The Vision
+- Models that understand your context
+- Tools that act autonomously
+- Applications that work everywhere
+- An ecosystem owned by its users
+
+## Why This Wins
+
+### Against Closed Platforms
+- No vendor lock-in vs. total dependence
+- Own forever vs. rent forever
+- Infinite flexibility vs. their way only
+- Community innovation vs. corporate roadmap
+
+### Against DIY Open Source
+- Complete ecosystem vs. fragmented tools
+- Works out of box vs. endless configuration
+- Unified experience vs. duct-tape integration
+- Professional polish vs. research prototypes
+
+## Join the Revolution
+
+This isn't just about building better AI. It's about ensuring AI serves humanity, not the other way around.
+
+When you use Jan, you're not just a user. You're part of a movement building:
+- AI that works anywhere
+- Intelligence you can own
+- Tools that adapt to you
+- A future without gatekeepers
+
+## Our Promise
+
+We promise to build superintelligence that:
+
+1. **Works Everywhere**: From laptop to data center to edge
+2. **Belongs to You**: Download it, modify it, own it forever
+3. **Stays Open**: Core will always be open source
+4. **Keeps Improving**: Community-driven development
+5. **Respects Choice**: Your deployment, your rules
+
+## The Bottom Line
+
+Superintelligence shouldn't be controlled by a few companies. It shouldn't force you to choose between power and ownership. It shouldn't lock you into one way of working.
+
+With Jan, it doesn't.
+
+You get superintelligence that adapts to how you work, runs where you need it, and belongs to you completely.
+
+That's open superintelligence. That's what we're building.
+
+---
+
+*"The best AI platform is one where choosing local or cloud is like choosing between laptop or desktop - same experience, different form factor, your choice."*
diff --git a/website/src/content/docs/handbook/philosophy/optimism.mdx b/website/src/content/docs/handbook/philosophy/optimism.mdx
new file mode 100644
index 000000000..a0221e026
--- /dev/null
+++ b/website/src/content/docs/handbook/philosophy/optimism.mdx
@@ -0,0 +1,38 @@
+---
+title: "Perpetual Optimism is a Force Multiplier"
+description: "We believe in perpetual optimism."
+---
+
+import YouTube from '@/components/YouTube.astro'
+
+
+
+> "Success consists of going from failure to failure without loss of enthusiasm." — Winston Churchill
+
+In 1903, [we barely left the ground](https://en.wikipedia.org/wiki/Wright_Flyer). By 1969, [we stood on the moon](https://en.wikipedia.org/wiki/Apollo_11). What once seemed impossible became reality through persistence and optimism.
+
+## Augmenting Humanity
+
+- We reject AI doomerism. Focus on possibilities, not fear
+- Children represent our hope. We build for their future
+- Humanity progresses faster than it feels
+- AI is a tool—like electricity or the telephone. It's not the first revolution, nor the last
+- History shows resilience. We adapt, mitigate risks, and move forward
+- Airplanes once terrified—yet they helped humanity reach the moon and connect the world
+
+## AI Safety
+
+- Safety is non-negotiable. Protecting people is the baseline
+ - AI safety == human safety. If we haven’t solved human safety, we haven’t solved AI safety.
+ - AI alignment == human alignment. Misaligned societies can’t build aligned systems.
+- AI safety requires human alignment first
+- Fear and progress must be balanced—panic stalls; awareness guides
+- Building for our kids ensures safety is built-in, purpose drives caution
+- Airplanes once terrified—yet they helped humanity reach the moon and connect the world
+
+## Why we believe in optimism
+
+- Optimism drives solutions; fear paralyzes
+- Hope fuels persistence. Failures aren't endpoints
+- Every breakthrough began as a dream. We build toward better because we believe in it
+- Perpetual optimism multiplies effort and impact
diff --git a/website/src/content/docs/handbook/sell/marketing.mdx b/website/src/content/docs/handbook/sell/marketing.mdx
new file mode 100644
index 000000000..71b4bd77b
--- /dev/null
+++ b/website/src/content/docs/handbook/sell/marketing.mdx
@@ -0,0 +1,94 @@
+---
+title: "Marketing"
+description: "How we structure marketing to drive revenue and growth"
+asIndexPage: true
+---
+
+import YouTube from '@/components/YouTube.astro';
+
+> "Marketing is about values." - Steve Jobs
+
+
+
+Without a strategy, marketing is:
+- throwing social media posts with reach anxiety
+- starting podcasts that no one listens
+- writing newsletter that even team members don't read
+
+Marketing is telling your own words in some channels for a purpose. Without a purpose, it's just noise - like how some do.
+
+Before starting we should align on some terms:
+
+- **Framework**: The blueprint that organizes our marketing efforts
+- **Purpose**: The fundamental reason behind our marketing activities
+- **Campaign**: Marketing actions
+- **Goal**: The specific, measurable result we aim to achieve through our marketing activities
+- **Brief**: The document outlining how we're executing a campaign
+
+## Framework(s)
+
+Marketings looks like art, must be managed like math. At Jan, we follow some frameworks for alignment.
+
+Our marketing efforts consist of 3 purposes and each marketing action must target at least one:
+
+1. Grow the market
+2. Increase the market share
+3. Capture market share in a more efficient way
+
+Each purpose requires campaigns with clear goals. Goal types:
+
+- KPI Goals
+- Project & Campaign Goals
+- Experiment Goals
+- Hiring Goals
+
+Campaign executions must leave no questions, so each marketing campaign requires a brief format:
+
+- **Goals**: KPIs, timeline, relevant OKRs
+- **Audience**: Who we're targeting
+- **Creatives**: Messaging & creative assets
+- **Channels**: Distribution
+- **Ownership**: Who is involved
+
+## Positioning
+
+Marketing starts with positioning - we always think thorough where to sit in the market before new launches.
+
+No one cares about product functions, it's all about what you provide. If your positioning requires explanation, it isn't working. We never talk about what our product does until we've established what problem it eliminates.
+
+We start with a positioning:
+
+- What is our product/service/platform?
+ - In customer language, what is it?
+ - What pain point do we eliminate?
+ - What we improve?
+- Who is this for?
+ - Who benefits most from this solution?
+ - What characteristics define this segment?
+- Why is it better?
+ - What are the other players?
+ - How do we outperform alternatives?
+ - What makes us uniquely valuable here?
+
+## Big no's on marketing
+
+We're playing our game, not theirs.
+
+- Throwing out marketing activities to see what sticks
+- Burning money at Ads
+- Random posts
+- Copying what others do
+- Actions without planning or goals
+- Prioritizing paid activities over organic
+- Jumping on hypes over strategy
+
+## Big yes's on marketing
+
+- Growing together with others
+- Playing our game in the highest level
+- Listening selectively - Absorb market feedback, but filter it through strategy
+- Adding value to what we're working on
+- Repurposing content
+- Being opinionated about the differentiation and why we're doing
+- Understanding the technical aspects at a level that explains a child
+- Being aware of the target audience and the messaging
diff --git a/website/src/content/docs/handbook/team/index.mdx b/website/src/content/docs/handbook/team/index.mdx
new file mode 100644
index 000000000..1465f6e86
--- /dev/null
+++ b/website/src/content/docs/handbook/team/index.mdx
@@ -0,0 +1,70 @@
+---
+title: "Team Structure"
+description: "How Jan's team is organized"
+---
+
+Jan is an open-source AI platform built by a distributed team focused on making AI
+accessible, private, and locally-run.
+
+Jan is currently ~8 people working across different areas to build the open
+superintelligence platform.
+
+## Jan Desktop & Platform
+
+Building the core Jan application and platform infrastructure.
+
+### Engineering
+- **Core Platform**: Building Jan Desktop and core infrastructure
+- **Dev Relations**: Ramon - Community engagement and developer experience
+- **QA & Release**: Ensuring quality and stability across releases
+
+### Model Team
+
+Focused on making the best models available to Jan users.
+
+| Focus Area | Team Members | Description |
+|------------|--------------|-------------|
+| **Model Hub** | Thien | Leading model quantization and evaluation for Jan's Local Model Hub - ensuring Jan supports the latest and best open source models |
+| **Jan Models** | Alan, Alex, Bách, Warren | Developing Jan's flagship model to be the default in Jan Desktop |
+| **Advanced Features** | Thinh, Norapat | Working with Ramon to spec and implement: • Jan Attachments • Jan DeepResearch • Jan Voice Mode |
+| **Engineering** | Akarshan, Louis, Dinh | Building the core of Jan |
+| **Design** | Faisal | Making Jan look gooood, function well, and be accessible to everyone |
+| **Product & DevRel** | Yuuki and Ramon | Thinking about the future of Jan and how to best communicate it to users |
+| **Management** | Yuuki | Keeping people in check by threatening everyone with corporate speak |
+| **Marketing** | Emre | Spreading the word about Jan and its capabilities |
+| **Infra** | Minh and Hien| They make things run smoothly |
+
+
+## Infrastructure & Operations
+
+### Technical Infrastructure
+- **Jan Cloud**: Cloud infrastructure for optional services
+- **Hardware**: Optimizing for local hardware performance
+- **CI/CD**: Continuous integration and deployment
+
+### Business Operations
+
+| Team | Focus | Members |
+|------|-------|---------|
+| **Marketing** | Brand, growth, and community | Emre |
+| **Product** | Product strategy and roadmap | Ramon |
+| **Finance & HR** | Operations and people | Nicole, finance team |
+
+## Working Groups
+
+Cross-functional teams that form around specific initiatives:
+
+- **Model Evaluation**: Jan Exam benchmarking and quality assurance
+- **Community Features**: Building based on user feedback
+- **Open Standards**: MCP and other protocol development
+
+## Our Approach
+
+- **Open Development**: Build in public with community input
+- **Quality First**: Jan Exam as our north star for model quality
+- **User Focused**: Every role ultimately serves our users' need for private, local AI
+- **Flexible Structure**: Teams adapt based on current priorities and user needs
+
+## Join Us
+
+Interested in joining our mission to build open superintelligence? Check out our [careers page](https://jan.ai/careers) or contribute to our [open source projects](https://github.com/janhq).
diff --git a/website/src/content/docs/handbook/who/curiosity.mdx b/website/src/content/docs/handbook/who/curiosity.mdx
new file mode 100644
index 000000000..9a0e8378d
--- /dev/null
+++ b/website/src/content/docs/handbook/who/curiosity.mdx
@@ -0,0 +1,139 @@
+---
+title: "The Fast and the Curious"
+description: "We hire people who move quickly and never stop learning"
+---
+
+> "In the AI age, the ability to learn faster than the rate of change is the only sustainable advantage." — Adapted from Eric Ries
+
+We hire people who are both fast and curious—those who ship today while learning for tomorrow.
+
+## Ship Fast, Ship Often
+
+Speed is our advantage in the rapidly evolving AI landscape:
+
+- **Weekly Releases**: We ship updates to Jan every week, not every quarter
+- **Rapid Experimentation**: Test ideas with real users, not focus groups
+- **Fail Fast, Learn Faster**: Kill features that don't work, double down on what does
+- **User Feedback Loop**: From idea to user's hands in days, not months
+
+### What Fast Means at Jan
+
+- **Bias for Action**: See a problem? Fix it. Don't wait for permission.
+- **MVP Mindset**: Launch at 80% perfect, iterate to 100%
+- **Quick Decisions**: Make reversible decisions quickly, deliberate only on irreversible ones
+- **Async by Default**: Don't let time zones slow us down
+
+We've shipped:
+- Major features in days that others debate for months
+- Model support hours after release, not weeks
+- Bug fixes while users are still typing the report
+
+## Stay Endlessly Curious
+
+In AI, yesterday's breakthrough is today's baseline:
+
+### Learning is Non-Negotiable
+
+- **New Models Weekly**: Understand and integrate the latest AI advances
+- **Cross-Domain Knowledge**: From quantization techniques to UI design
+- **Community Learning**: Our users teach us as much as we teach them
+- **Open Source Study**: Learn from the best codebases in the world
+
+### Curiosity in Practice
+
+- **Why Over What**: Don't just implement—understand the reasoning
+- **Question Everything**: "Why do we collect user data?" led to our privacy-first approach
+- **Learn in Public**: Share discoveries with the community
+- **Teach to Learn**: Explaining concepts deepens understanding
+
+## The Jan Learning Culture
+
+### Everyone is a Student
+
+- **No Experts**: In a field moving this fast, everyone is learning
+- **Share Knowledge**: Daily discoveries in our Discord channels
+- **Document Learning**: Today's experiment is tomorrow's documentation
+- **Celebrate Questions**: The "stupid" question often reveals the biggest insight
+
+### Everyone is a Teacher
+
+- **Onboarding**: New hires teach us fresh perspectives
+- **Community Education**: Blog posts, tutorials, and demos
+- **Code as Teaching**: Well-commented code educates future contributors
+- **Failure Stories**: Share what didn't work and why
+
+## What We Look For
+
+### Signs of Speed
+
+- **GitHub Velocity**: Frequent commits, quick iterations
+- **Project Completion**: Finished projects, not just started ones
+- **Response Time**: Quick to engage, quick to deliver
+- **Adaptation Speed**: How fast do you integrate feedback?
+
+### Signs of Curiosity
+
+- **Side Projects**: What do you build for fun?
+- **Learning Artifacts**: Blogs, notes, or projects showing learning
+- **Question Quality**: Do you ask insightful questions?
+- **Knowledge Breadth**: Interests beyond your specialty
+
+## Why This Matters for Jan
+
+### AI Moves Too Fast for Slow
+
+- Models improve monthly
+- User expectations evolve weekly
+- Competition ships daily
+- Standards change quarterly
+
+If we're not fast and curious, we're obsolete.
+
+### Local-First Demands Both
+
+- **Fast**: Users expect immediate responses, not cloud latency
+- **Curious**: Supporting every model requires understanding each one
+- **Fast**: Privacy bugs need instant fixes
+- **Curious**: New quantization methods need quick adoption
+
+## The Compound Effect
+
+Fast + Curious creates exponential growth:
+
+```
+Ship Fast → User Feedback → Learn →
+Ship Smarter → More Users → More Learning →
+Ship Even Faster → Compound Growth
+```
+
+Each cycle makes us:
+- Faster at shipping
+- Better at learning
+- More valuable to users
+- More attractive to talent
+
+## Join Us If...
+
+- You've shipped something this week (not this year)
+- You've learned something new today (not last month)
+- You see a Jan issue and think "I could fix that"
+- You read our codebase and think "I could improve that"
+- You use Jan and think "It could also do this"
+
+## The Promise
+
+If you join Jan as someone fast and curious, in a year you'll be:
+- **Faster**: Shipping features you can't imagine today
+- **Smarter**: Understanding AI at a level that surprises you
+- **Connected**: Part of a global community of builders
+- **Impactful**: Your code running on millions of devices
+
+## The Bottom Line
+
+We don't hire for what you know today. We hire for how fast you'll know what matters tomorrow.
+
+In the race to build open superintelligence, the fast and curious don't just keep up—they set the pace.
+
+---
+
+*"At Jan, we measure progress in iterations per week, not years of experience."*
diff --git a/website/src/content/docs/handbook/who/dydx.mdx b/website/src/content/docs/handbook/who/dydx.mdx
new file mode 100644
index 000000000..93c39d274
--- /dev/null
+++ b/website/src/content/docs/handbook/who/dydx.mdx
@@ -0,0 +1,28 @@
+---
+title: "dy/dx > y-intercept"
+description: "The rate of a function is more important than its starting point."
+---
+
+> "A little bit of slope makes up for a lot of Y-intercept." — John Ousterhout
+
+The rate of a function is more important than its starting point. We value growth
+rate over initial advantage.
+
+## Why slope matters
+
+- Knowledge fades, but the ability to learn compounds
+- A fast learner will outpace someone with more initial experience over time
+- Slow, steady improvement outperforms quick starts with no growth
+
+## What it means day-to-day
+
+- Don't hesitate to try something new—even if you start clueless. Learning speed matters
+- Mistakes are expected. Learning from them is required
+- We'll invest in helping you grow, but the drive has to come from you
+- Your trajectory is more important than where you begin
+
+## Why we believe in this
+
+- Building something lasting requires patience and commitment to improvement
+- We're not interested in shortcuts. We value the work that compounds quietly until it becomes obvious
+- If Jan's greatest impact is helping people and ideas grow steadily over time—that's the kind of success we stand for
diff --git a/website/src/content/docs/handbook/who/obsession.mdx b/website/src/content/docs/handbook/who/obsession.mdx
new file mode 100644
index 000000000..3b14c2da2
--- /dev/null
+++ b/website/src/content/docs/handbook/who/obsession.mdx
@@ -0,0 +1,35 @@
+---
+title: "Obsession"
+description: "We seek obsession."
+---
+
+> "If you want to build a ship, don't drum up the men to gather wood, divide the work, and give orders. Instead, teach them to yearn for the vast and endless sea." — Antoine de Saint-Exupéry
+
+We don't hire for skills alone. We hire for obsession.
+
+## Find the obsessed
+
+- Breakthroughs require long-term focus, not fleeting interest
+- Failure isn't the end—just another step in relentless experimentation
+- People who obsess push beyond limits others accept
+
+## We seek those already consumed
+
+- You're already working on the problem—because you can't not
+- We don't assign you a job; we support your life's work
+- Obsession isn't just personal—it becomes more powerful in a team
+- Together, a group of obsessives becomes greater than the sum of its parts
+
+## The best hires share common traits
+
+- Desperation: Solving the problem feels like a need, not a choice
+- Relentlessness: You try, fail, adapt—again and again
+- Defiance: Others call it crazy; we call it genius
+
+## We're looking for the dreamers
+
+- That deep, persistent song you've been trying to sing? We hear it
+- Maybe you've been mocked, dismissed, or misunderstood
+- We seek those lonely, weird dreamers who refuse to give up
+
+Wherever you are in the world, if this feels like you—apply here
diff --git a/website/src/content/docs/handbook/who/team.mdx b/website/src/content/docs/handbook/who/team.mdx
new file mode 100644
index 000000000..952780265
--- /dev/null
+++ b/website/src/content/docs/handbook/who/team.mdx
@@ -0,0 +1,26 @@
+---
+title: "Jan is a Sports Team"
+description: "Our approach is super simple: Retain. Build trust. Win together."
+---
+
+> "First, you need to learn how to play. You need 2-3 years together to learn how to play with each other." — [Nikola Jokić](https://www.reddit.com/r/NBASpurs/comments/1cdscti/wise_words_from_jokic_first_you_need_to_learn_how/)
+
+Jan is a sports team, not unlike a NBA team or a NFL team. We focus on putting the best players on the court, focusing on their 1% strengths, and building a team that can win together.
+
+## Team vs. Family
+
+We learn the best principles from [Netflix Culture Deck](https://www.slideshare.net/slideshow/culture-1798664/1798664):
+
+- We operate like a pro sports team
+- There are limited spots on every team — every roster spot counts
+- We hire, develop and cut smartly
+
+## Continuity
+
+However, we balance this with our belief in Continuity and Stability:
+
+- Teams with psychological safety take better, long-term risks
+- Teams need 2–3 years to truly gel
+- Continuity matters; balance change carefully
+- Cohesive teams outperform constant reshuffles
+- Time builds trust, speed, and better decisions
diff --git a/website/src/content/docs/handbook/who/underdogs.mdx b/website/src/content/docs/handbook/who/underdogs.mdx
new file mode 100644
index 000000000..f2db21c51
--- /dev/null
+++ b/website/src/content/docs/handbook/who/underdogs.mdx
@@ -0,0 +1,34 @@
+---
+title: "Underdogs"
+description: "We hire from unconventional backgrounds."
+---
+
+> "I am, somehow, less interested in the weight and convolutions of Einstein's brain than in the near certainty that people of equal talent have lived and died in cotton fields and sweatshops." — Stephen Jay Gould
+
+We hire from unconventional backgrounds.
+
+## Opportunity isn't equal
+
+- Where you're born shapes your chances—but it shouldn't define them
+- Talent is everywhere. Opportunity isn't
+- We believe the internet has changed how we work and live together. Our team—spread across 10 cities worldwide—couldn't have existed just five years ago
+- Fast connections (5G and beyond) mean geography no longer decides your future. We'll find you—wherever you are
+
+## We seek the overlooked
+
+- Lonely geniuses, relentless tinkerers, people with potential others missed
+- We don't care where you went to school—or if you went at all
+- We bet on people like startups: high risk, high potential
+- Skills can be taught; drive cannot. We'll teach you to communicate and collaborate
+- We trust you'll prove yourself through your work and integrity
+
+## Meaningful work, close to home
+
+- We create jobs that matter, in the places people call home
+- If Jan's greatest achievement is enabling people to build meaningful careers in their hometowns—that would be success worth celebrating
+
+## This is your shot
+
+- Underdogs, dreamers, the overlooked—this place is for you
+- We don't just welcome you. We're looking for you
+- Wherever you are in the world—apply here
diff --git a/website/src/content/docs/handbook/why/icp.mdx b/website/src/content/docs/handbook/why/icp.mdx
new file mode 100644
index 000000000..0393dfc41
--- /dev/null
+++ b/website/src/content/docs/handbook/why/icp.mdx
@@ -0,0 +1,112 @@
+---
+title: "Who we are building for"
+description: "Building for people who want to own their AI and choose how to run it"
+---
+
+Jan is built for people who want control over their AI experience. Our users span from individual developers to global enterprises, united by a desire to own their AI infrastructure and choose how they deploy it.
+
+## Primary Users
+
+### Developers Who Want Flexibility
+Engineers and creators who need:
+- The freedom to run models locally or in the cloud
+- Complete control over their AI stack
+- The ability to switch providers without losing their work
+- Integration options that fit their existing workflows
+
+### Individuals Who Value Ownership
+People who understand that AI should be like any other software:
+- Download it once, use it forever
+- Your conversations and data belong to you
+- No vendor lock-in or subscription traps
+- The freedom to modify and extend as needed
+
+### Teams That Need Options
+Organizations that require:
+- The flexibility to run on-premises or in the cloud
+- The ability to scale from laptop to server cluster
+- Control over costs and infrastructure
+- Options that adapt to changing requirements
+
+### Global Users Who Need Access
+Millions of potential AI users who face:
+- Expensive or unreliable cloud services
+- The need for AI that works offline
+- Hardware limitations that require optimization
+- Different deployment needs for different situations
+
+## Our North Star: Best Experience, Your Way
+
+We use **Jan Exam** as our benchmark to ensure quality across all deployment options. Whether you're running locally on a laptop or scaled across a server farm, you get:
+
+- The same powerful AI capabilities
+- Consistent performance for your hardware
+- The flexibility to change your setup anytime
+- No compromise between local and cloud
+
+## The Flexibility Spectrum
+
+Our users choose their own balance:
+
+### Full Local
+- Everything runs on your hardware
+- Complete offline capability
+- Total data ownership
+- Zero recurring costs
+
+### Hybrid Approach
+- Local for sensitive work
+- Cloud for heavy compute
+- Seamless switching between modes
+- Optimize cost vs performance
+
+### Full Cloud
+- Jan.ai for zero setup
+- Team collaboration features
+- Managed infrastructure
+- Scale without limits
+
+## What Sets Our Users Apart
+
+They understand that:
+- **Ownership matters**: Your AI tools should be assets, not rentals
+- **Flexibility is power**: Different tasks need different approaches
+- **Control creates value**: The ability to customize and extend is crucial
+- **Choice prevents lock-in**: Multiple options keep you free
+
+## Real User Scenarios
+
+### The Solo Developer
+Runs Jan locally during development, uses cloud for production deployments. Owns their entire stack.
+
+### The Research Team
+Uses local models for sensitive data, cloud models for general research. Switches seamlessly based on needs.
+
+### The Enterprise
+Deploys Jan Server on-premises for compliance, uses Jan.ai for non-critical workloads. One platform, multiple deployment options.
+
+### The Student
+Runs lightweight models locally on their laptop, accesses more powerful models via cloud when needed. Learns AI without breaking the bank.
+
+## Growing With Our Users
+
+We build for users at every stage:
+
+**Day 1**: Download Jan, start chatting locally
+**Week 1**: Connect to cloud providers for more power
+**Month 1**: Deploy your own server for team access
+**Year 1**: Scale across infrastructure you control
+
+## The Future User
+
+We're building for the next billion AI users who will expect:
+- Software they can own, not just rent
+- The freedom to run AI anywhere
+- Tools that adapt to their needs
+- No artificial limitations
+
+They'll start with Jan because it gives them choices, and they'll stay because those choices grow with them.
+
+---
+
+*"The best AI is the one that works where you need it, how you need it, when you need it."*
diff --git a/website/src/content/docs/handbook/why/ownership.mdx b/website/src/content/docs/handbook/why/ownership.mdx
new file mode 100644
index 000000000..67e90b34c
--- /dev/null
+++ b/website/src/content/docs/handbook/why/ownership.mdx
@@ -0,0 +1,180 @@
+---
+title: "Our Vision for Open Superintelligence"
+description: "Building superintelligence that you can own, modify, and deploy however you want"
+---
+
+Jan believes the path to superintelligence isn't through one massive model controlled by a few companies. It's through an ecosystem of specialized models, tools, and applications that anyone can own and deploy.
+
+## What is Open Superintelligence?
+
+Open superintelligence is AI that:
+
+- **Works Everywhere**: From your laptop to your data center to the edge
+- **Belongs to You**: Download it, modify it, deploy it - it's yours
+- **Scales Infinitely**: Same AI whether you're one person or ten thousand
+- **Evolves with Community**: Thousands of contributors, not hundreds of employees
+
+## The Ecosystem Approach
+
+Superintelligence isn't one thing - it's many things working together:
+
+### Specialized Models
+Instead of one model trying to do everything:
+- **Jan-Search** excels at finding and synthesizing information
+- **Jan-Write** understands creativity and structure
+- **Jan-Analyze** specializes in reasoning and data
+- **Jan-Code** focuses on programming tasks
+
+Each model does one thing brilliantly, and they hand off tasks to each other.
+
+### Extensible Tools
+Models alone aren't enough. Our tools make them useful:
+- Search the web and get answers, not links
+- Automate browser tasks naturally
+- Parse documents and understand context
+- Connect to your data sources via MCP
+
+### Flexible Deployment
+One codebase, infinite configurations:
+- **Local Mode**: Everything on your hardware
+- **Server Mode**: Self-host for your team
+- **Cloud Mode**: We host it for you
+- **Hybrid Mode**: Mix and match as needed
+
+## Why Ownership Matters
+
+When you own your AI infrastructure:
+
+### You Control the Experience
+- Customize models for your specific needs
+- Build workflows that match how you work
+- Integrate with your existing tools
+- Scale based on your requirements
+
+### You Keep What You Build
+- Your prompts and conversations
+- Your fine-tuned models
+- Your custom workflows
+- Your accumulated knowledge
+
+### You Choose How to Run
+- Offline when you need privacy
+- Cloud when you need power
+- Local when you need speed
+- Distributed when you need scale
+
+## Building in Public
+
+We're not building this in secret. Everything is open:
+
+### Watch Us Train
+See our models being trained in real-time at [train.jan.ai](https://train.jan.ai):
+- Live loss curves
+- Training datasets
+- Failed experiments
+- Actual progress
+
+### Test With Us
+Help evaluate models at [eval.jan.ai](https://eval.jan.ai):
+- Compare outputs side-by-side
+- Vote on what actually works
+- Suggest improvements
+- See all the data
+
+### Build With Us
+Everything on [GitHub](https://github.com/janhq):
+- Contribute code
+- Report issues
+- Suggest features
+- Fork and modify
+
+## The Path to Superintelligence
+
+### Today: Foundation (✅)
+- Jan Desktop works with local and cloud models
+- Basic tools via MCP
+- Community growing rapidly
+
+### Next 12 Months: Ecosystem
+- Jan v1 models optimized for specific tasks
+- Jan Server for self-hosted deployments
+- Advanced tools (browser automation, deep research)
+- Cross-device synchronization
+
+### Future: True Superintelligence
+- Models that understand your context
+- Tools that act on your behalf
+- Applications that work everywhere
+- An ecosystem that belongs to everyone
+
+## Our Commitment
+
+We commit to building superintelligence that:
+
+### Stays Open
+- Core will always be open source
+- Models will always be downloadable
+- Standards will always be public
+- Development will always be transparent
+
+### Stays Flexible
+- Run it anywhere you have compute
+- Mix local and cloud as needed
+- Scale up or down instantly
+- Switch providers without friction
+
+### Stays Yours
+- No vendor lock-in
+- No forced updates
+- No subscription traps
+- No data exploitation
+
+## Why This Approach Wins
+
+### Network Effects
+Every user makes Jan better:
+- Bug reports improve stability
+- Feature requests guide development
+- Model feedback enhances quality
+- Community support helps everyone
+
+### Compound Innovation
+Open development accelerates progress:
+- Best ideas come from anywhere
+- Solutions shared instantly
+- Problems solved collectively
+- Innovation happens 24/7
+
+### Aligned Incentives
+We succeed when you succeed:
+- Your productivity is our metric
+- Your ownership is our philosophy
+- Your freedom is our product
+- Your success is our business model
+
+## Join the Movement
+
+This isn't just about building better AI. It's about ensuring the most powerful technology in human history belongs to humanity, not corporations.
+
+Whether you:
+- Use Jan for daily work
+- Contribute code or ideas
+- Share it with others
+- Build on top of it
+- Deploy it in your organization
+
+You're part of building superintelligence that everyone can own.
+
+## The Choice is Yours
+
+In five years, AI will be everywhere. The question is:
+
+**Will you rent intelligence from a monopoly, or will you own your own superintelligence?**
+
+With Jan, you don't have to choose between powerful and private, between cloud and local, between easy and extensible.
+
+You get it all. And it's yours.
+
+---
+
+*"The best superintelligence is the one you can run anywhere, modify for anything, and own forever."*
diff --git a/website/src/content/docs/handbook/why/problem.mdx b/website/src/content/docs/handbook/why/problem.mdx
new file mode 100644
index 000000000..26a7b08d1
--- /dev/null
+++ b/website/src/content/docs/handbook/why/problem.mdx
@@ -0,0 +1,106 @@
+---
+title: "What problem are we solving?"
+description: "Building superintelligence that you can own and run anywhere"
+---
+
+Current AI platforms force unnecessary trade-offs:
+
+- **All or Nothing**: Either use their cloud or build everything yourself from scratch
+- **Vendor Lock-in**: Your prompts, workflows, and knowledge trapped in one platform
+- **One-Size-Fits-All**: Same solution whether you're a student or an enterprise
+- **Rent Forever**: Monthly subscriptions for software that should be yours
+- **Limited Deployment**: Can't run where you need it - offline, on-premises, or at the edge
+
+## Mission
+
+Jan is building a complete AI ecosystem that adapts to how you want to work. We believe AI should be:
+
+- **Truly Owned**: Like any software, download it and it's yours forever
+- **Infinitely Flexible**: Run locally, self-host, or use our cloud - same experience
+- **Professionally Complete**: Not just models, but the full stack of tools and applications
+- **Universally Deployable**: From your laptop to your data center to the edge
+
+## The Full Stack Approach
+
+We're not building another chatbot. We're building superintelligence you can own:
+
+### Models That Specialize
+- **Jan Models**: Purpose-built for specific tasks (search, analysis, writing)
+- **Community Models**: Any GGUF model from Hugging Face
+- **Cloud Models**: Connect to OpenAI, Anthropic, or others when needed
+- **Your Models**: Fine-tune and deploy your own
+
+### Tools That Extend
+- **Search**: Get answers, not just links
+- **DeepResearch**: Multi-step autonomous research
+- **BrowserUse**: Let AI interact with the web for you
+- **MCP Integration**: Connect to any data source or service
+
+### Applications That Scale
+- **Jan Desktop**: Your local AI workstation
+- **Jan.ai**: Web access with no setup
+- **Jan Mobile**: AI that follows you (coming soon)
+- **Jan Server**: Your own private AI cloud (coming soon)
+
+## Why This Matters
+
+In 5 years, AI will be embedded in everything you do. The question isn't whether you'll use AI, but how:
+
+**Option A**: Rent access from tech giants, accept their limitations, lose your data when you stop paying
+
+**Option B**: Own your AI infrastructure, deploy it anywhere, keep everything you build
+
+We're making Option B not just possible, but better than Option A.
+
+## The Problem With Status Quo
+
+### Cloud-Only Platforms
+- Can't work offline
+- Can't handle sensitive data
+- Can't customize deeply
+- Can't control costs
+
+### DIY Open Source
+- Requires deep technical knowledge
+- Fragmented tools and models
+- No unified experience
+- Massive integration overhead
+
+### Jan's Solution
+- Download and run in minutes
+- Complete ecosystem out of the box
+- Scale from laptop to cluster
+- Switch between local and cloud seamlessly
+
+## Real Problems We Solve Today
+
+### For Developers
+"I want to use AI in my app but don't want vendor lock-in"
+→ OpenAI-compatible API that runs anywhere
+
+### For Teams
+"We need AI but can't send data to the cloud"
+→ Self-host Jan Server on your infrastructure
+
+### For Individuals
+"I want ChatGPT capabilities without the subscription"
+→ Download Jan Desktop and own your AI
+
+### For Enterprises
+"We need to scale AI across different deployment scenarios"
+→ One platform that works from edge to cloud
+
+## Vision
+
+We envision a future where:
+
+- AI runs wherever you have compute - laptop, phone, server, edge device
+- You own your AI infrastructure like you own your other software
+- Models, tools, and applications work together seamlessly
+- The best AI isn't gated behind corporate APIs
+
+Jan is the platform that makes this future real, today.
+
+---
+
+*"The future of AI isn't about choosing between local or cloud. It's about having both, and everything in between, working perfectly together."*
diff --git a/website/src/content/docs/index.mdx b/website/src/content/docs/index.mdx
deleted file mode 100644
index c4caf13d0..000000000
--- a/website/src/content/docs/index.mdx
+++ /dev/null
@@ -1,171 +0,0 @@
----
-title: Jan
-description: Jan is an open-source ChatGPT-alternative and self-hosted AI platform - build and run AI on your own desktop or server.
-keywords:
- [
- Jan,
- Jan AI,
- ChatGPT alternative,
- OpenAI platform alternative,
- local API,
- local AI,
- private AI,
- conversational AI,
- no-subscription fee,
- large language model,
- LLM,
- ]
----
-
-import { Aside } from '@astrojs/starlight/components';
-
-
-
-
-Jan is a ChatGPT alternative that runs 100% offline on your desktop and (*soon*) on mobile. Our goal is to
-make it easy for anyone, with or without coding skills, to download and use AI models with full control and
-[privacy](https://www.reuters.com/legal/legalindustry/privacy-paradox-with-ai-2023-10-31/).
-
-Jan is powered by [Llama.cpp](https://github.com/ggerganov/llama.cpp), a local AI engine that provides an OpenAI-compatible
-API that can run in the background by default at `https://localhost:1337` (or your custom port). This enables you to power all sorts of
-applications with AI capabilities from your laptop/PC. For example, you can connect local tools like [Continue](https://jan.ai/docs/server-examples/continue-dev)
-and [Cline](https://cline.bot/) to Jan and power them using your favorite models.
-
-Jan doesn't limit you to locally hosted models, meaning, you can create an API key from your favorite model provider,
-add it to Jan via the configuration's page and start talking to your favorite models.
-
-### Features
-
-- Download popular open-source LLMs (Llama3, Gemma3, Qwen3, and more) from the HuggingFace [Model Hub](./jan/manage-models)
-or import any GGUF files (the model format used by llama.cpp) available locally
-- Connect to [cloud services](./jan/remote-models/openai) (OpenAI, Anthropic, Mistral, Groq, etc.)
-- [Chat](./jan/threads) with AI models & [customize their parameters](./jan/explanation/model-parameters) via our
-intuitive interface
-- Use our [local API server](https://jan.ai/api-reference) with an OpenAI-equivalent API to power other apps.
-
-### Philosophy
-
-Jan is built to be [user-owned](about#-user-owned), this means that Jan is:
-- Truly open source via the [Apache 2.0 license](https://github.com/menloresearch/jan/blob/dev/LICENSE)
-- [Data is stored locally, following one of the many local-first principles](https://www.inkandswitch.com/local-first)
-- Internet is optional, Jan can run 100% offline
-- Free choice of AI models, both local and cloud-based
-- We do not collect or sell user data. See our [Privacy Policy](./privacy).
-
-
-
-### Inspirations
-
-Jan is inspired by the concepts of [Calm Computing](https://en.wikipedia.org/wiki/Calm_technology), and the Disappearing Computer.
-
-## Acknowledgements
-
-Jan is built on the shoulders of many open-source projects like:
-
-- [Llama.cpp](https://github.com/ggerganov/llama.cpp/blob/master/LICENSE)
-- [Scalar](https://github.com/scalar/scalar)
-
-## FAQs
-
-
-What is Jan?
-
-Jan is a customizable AI assistant that can run offline on your computer - a privacy-focused alternative to tools like ChatGPT, Anthropic's Claude, and Google Gemini, with optional cloud AI support.
-
-
-
-
-How do I get started with Jan?
-
-Download Jan on your computer, download a model or add API key for a cloud-based one, and start chatting. For detailed setup instructions, see our installation guides.
-
-
-
-
-Is Jan compatible with my system?
-
-Jan supports all major operating systems:
-- [Mac](./jan/installation/mac#compatibility)
-- [Windows](./jan/installation/windows#compatibility)
-- [Linux](./jan/installation/linux)
-
-Hardware compatibility includes:
-- NVIDIA GPUs (CUDA)
-- AMD GPUs (Vulkan)
-- Intel Arc GPUs (Vulkan)
-- Any GPU with Vulkan support
-
-
-
-
-How does Jan protect my privacy?
-
-Jan prioritizes privacy by:
-- Running 100% offline with locally-stored data
-- Using open-source models that keep your conversations private
-- Storing all files and chat history on your device in the [Jan Data Folder](./jan/data-folder)
-- Never collecting or selling your data
-
-
-
-You can optionally share anonymous usage statistics to help improve Jan, but your conversations are never shared. See our complete [Privacy Policy](./jan/privacy).
-
-
-
-
-What models can I use with Jan?
-
-- Download optimized models from the [Jan Hub](./jan/manage-models)
-- Import GGUF models from Hugging Face or your local files
-- Connect to cloud providers like OpenAI, Anthropic, Mistral and Groq (requires your own API keys)
-
-
-
-
-Is Jan really free? What's the catch?
-
-Jan is completely free and open-source with no subscription fees for local models and features. When using cloud-based models (like GPT-4o or Claude Sonnet 3.7), you'll only pay the standard rates to those providers—we add no markup.
-
-
-
-
-Can I use Jan offline?
-
-Yes! Once you've downloaded a local model, Jan works completely offline with no internet connection needed.
-
-
-
-
-How can I contribute or get community help?
-
-- Join our [Discord community](https://discord.gg/qSwXFx6Krr) to connect with other users
-- Contribute through [GitHub](https://github.com/menloresearch/jan) (no permission needed!)
-- Get troubleshooting help in our [Discord](https://discord.com/invite/FTk2MvZwJH) channel [#🆘|jan-help](https://discord.com/channels/1107178041848909847/1192090449725358130)
-- Check our [Troubleshooting](./jan/troubleshooting) guide for common issues
-
-
-
-
-Can I self-host Jan?
-
-Yes! We fully support the self-hosted movement. Either download Jan directly or fork it on [GitHub repository](https://github.com/menloresearch/jan) and build it from source.
-
-
-
-
-What does Jan stand for?
-
-Jan stands for "Just a Name". We are, admittedly, bad at marketing 😂.
-
-
-
-
-Are you hiring?
-
-Yes! We love hiring from our community. Check out our open positions at [Careers](https://menlo.bamboohr.com/careers).
-
-
diff --git a/website/src/content/docs/jan/index.mdx b/website/src/content/docs/jan/index.mdx
new file mode 100644
index 000000000..fce7a8db0
--- /dev/null
+++ b/website/src/content/docs/jan/index.mdx
@@ -0,0 +1,172 @@
+---
+title: Jan
+description: Build, run, and own your AI. From laptop to superintelligence.
+keywords:
+ [
+ Jan,
+ self-hosted AI,
+ local AI,
+ open superintelligence,
+ AI ecosystem,
+ llama.cpp,
+ GGUF models,
+ MCP tools
+ ]
+---
+
+import { Aside } from '@astrojs/starlight/components';
+
+
+
+## Jan's Goal
+
+> Jan's goal is to build superintelligence that you can self-host and use locally.
+
+## What is Jan?
+
+Jan is an open-source AI ecosystem that runs on your hardware.
+
+- **Available Today**: Jan Desktop - run AI models on your computer with zero setup
+- **Coming Soon**: Complete ecosystem with specialized models, tools, and cross-device sync
+
+### The Stack
+
+**Models**: Specialized AI models trained for real tasks (search, analysis, writing)
+**Tools**: Browser automation, web search, memory systems via MCP
+**Applications**: Desktop (now), Web/Mobile/Server (coming 2025)
+
+Everything works together. Everything runs where you want it.
+
+
+
+## Core Features
+
+### Run Models Locally
+- Download any GGUF model from Hugging Face
+- Use our Jan models optimized for specific tasks
+- Automatic GPU acceleration (NVIDIA/AMD/Intel)
+- OpenAI-compatible API at `localhost:1337`
+
+### Connect to Cloud (Optional)
+- Your API keys for OpenAI, Anthropic, etc.
+- Jan.ai cloud models (coming soon)
+- Self-hosted Jan Server
+
+### Extend with Tools
+- Web search via MCP
+- Browser automation
+- File parsing and analysis
+- Custom tool development
+
+## Architecture
+
+Jan is built on:
+- [Llama.cpp](https://github.com/ggerganov/llama.cpp) for inference
+- [Model Context Protocol](https://modelcontextprotocol.io) for tools
+- Local-first data storage in `~/jan`
+
+## Why Jan?
+
+| Feature | Other AI Platforms | Jan |
+|:--------|:-------------------|:----|
+| **Deployment** | Their servers only | Your device, your servers, or our cloud |
+| **Models** | One-size-fits-all | Specialized models for specific tasks |
+| **Data** | Stored on their servers | Stays on your hardware |
+| **Cost** | Monthly subscription | Free locally, pay for cloud |
+| **Extensibility** | Limited APIs | Full ecosystem with MCP tools |
+
+## Development Philosophy
+
+1. **Local First**: Everything works offline. Cloud is optional.
+2. **User Owned**: Your data, your models, your compute.
+3. **Built in Public**: Watch our models train. See our code. Track our progress.
+
+
+
+## System Requirements
+
+**Minimum**: 8GB RAM, 10GB storage
+**Recommended**: 16GB RAM, NVIDIA GPU, 50GB storage
+**Supported**: Windows 10+, macOS 12+, Linux (Ubuntu 20.04+)
+
+## What's Next?
+
+
+When will mobile/web versions launch?
+
+- **Jan Web**: Beta Q1 2025
+- **Jan Mobile**: Q4 2025
+- **Jan Server**: Q3 2025
+
+All versions will sync seamlessly with your desktop.
+
+
+
+What models are available?
+
+**Now**:
+- Any GGUF model from Hugging Face
+- Cloud models via API keys (OpenAI, Anthropic, etc.)
+
+**Coming Q1 2025**:
+- Jan-Search: Optimized for web search and synthesis
+- Jan-Write: Creative and technical writing
+- Jan-Analyze: Data analysis and reasoning
+
+[Watch live training progress →](https://train.jan.ai)
+
+
+
+How does Jan make money?
+
+- **Local use**: Always free
+- **Cloud features**: Optional paid services (coming 2025)
+- **Enterprise**: Self-hosted deployment and support
+
+We don't sell your data. We sell software and services.
+
+
+
+Can I contribute?
+
+Yes. Everything is open:
+- [GitHub](https://github.com/janhq/jan) - Code contributions
+- [Model Training](https://jan.ai/docs/models) - See how we train
+- [Discord](https://discord.gg/FTk2MvZwJH) - Join discussions
+- [Model Testing](https://eval.jan.ai) - Help evaluate models
+
+
+
+Is this just another ChatGPT wrapper?
+
+No. We're building:
+- Our own models trained for specific tasks
+- Complete local AI infrastructure
+- Tools that extend model capabilities
+- An ecosystem that works offline
+
+ChatGPT is one model behind an API. Jan is a complete AI platform you own.
+
+
+
+What about privacy?
+
+**Local mode**: Your data never leaves your device. Period.
+**Cloud mode**: You choose when to use cloud features. Clear separation.
+
+See our [Privacy Policy](./privacy).
+
+
+## Get Started
+
+1. [Install Jan Desktop](./jan/installation) - Your AI workstation
+2. [Explore Models](./jan/models) - Download and configure
+3. [Learn the API](./api-reference) - Build with Jan
+
+---
+
+**Questions?** Join our [Discord](https://discord.gg/FTk2MvZwJH) or check [GitHub roadmap](https://github.com/janhq/jan/).
diff --git a/website/src/content/docs/jan/installation/windows.mdx b/website/src/content/docs/jan/installation/windows.mdx
index b50abf65f..45e80fad2 100644
--- a/website/src/content/docs/jan/installation/windows.mdx
+++ b/website/src/content/docs/jan/installation/windows.mdx
@@ -23,9 +23,6 @@ keywords:
import { Aside, Tabs, TabItem } from '@astrojs/starlight/components';
-
-# Windows Installation
-
## Compatibility
**System requirements:**
diff --git a/website/src/content/products/index.mdx b/website/src/content/products/index.mdx
index 3dab3e086..fdc2dfdaa 100644
--- a/website/src/content/products/index.mdx
+++ b/website/src/content/products/index.mdx
@@ -6,7 +6,6 @@ sidebar:
---
import { Aside, Card, CardGrid } from '@astrojs/starlight/components';
-**Jan's Goal is**
> **to build a superintelligence that you can self-host and use locally on your own devices.**
diff --git a/website/src/layouts/Layout.astro b/website/src/layouts/Layout.astro
index 591effc34..f13a41ebf 100644
--- a/website/src/layouts/Layout.astro
+++ b/website/src/layouts/Layout.astro
@@ -18,31 +18,84 @@ const { title } = Astro.props;
-
+
-
-
+
+
-
-
+ // Apply the correct theme immediately
+ if (theme === 'auto') {
+ const actualTheme = window.matchMedia('(prefers-color-scheme: light)').matches ? 'light' : 'dark';
+ document.documentElement.setAttribute('data-theme', actualTheme);
+ } else {
+ document.documentElement.setAttribute('data-theme', theme);
+ }
+
+ // Store non-auto preference
+ if (storedTheme && storedTheme !== 'auto') {
+ document.documentElement.setAttribute('data-theme', storedTheme);
+ }
+ })();
+
+
+
+