---
title: Quickstart
description: Cortex Quickstart.
keywords:
[
Jan,
Customizable Intelligence, LLM,
local AI,
privacy focus,
free and open source,
private and offline,
conversational AI,
no-subscription fee,
large language models,
Cortex,
Jan,
LLMs
]
---
import { Callout, Steps } from 'nextra/components'
import { Cards, Card } from 'nextra/components'
# Quickstart
🚧 Cortex is under construction.
To get started, confirm that your system meets the [hardware requirements](/cortex/hardware), and follow the steps below:
```bash
# 1. Install Cortex using NPM
npm i -g @janhq/cortex
# 2. Download a GGUF model
cortex models pull llama3
# 3. Run the model to start chatting
cortex models run llama3
# 4. (Optional) Run Cortex in OpenAI-compatible server mode
cortex serve
```
For more details regarding the Cortex server mode, please see here:
- [Server Endpoint](/cortex/server)
- [`cortex serve` command](/cortex/cli/serve)
## What's Next?
With Cortex now fully operational, you're ready to delve deeper:
- Explore how to [install Cortex](/cortex/installation) across various hardware environments.
- Familiarize yourself with the comprehensive set of [Cortex CLI commands](/cortex/cli) available for use.
- Gain insights into the system’s design by examining the [architecture](/cortex/architecture) of Cortex.