diff --git a/website/public/openapi/openapi.json b/website/public/openapi/openapi.json index cf4ecd136..976199abf 100644 --- a/website/public/openapi/openapi.json +++ b/website/public/openapi/openapi.json @@ -1,7 +1,7 @@ { "openapi": "3.1.0", "info": { - "title": "Jan API", + "title": "👋Jan API", "description": "OpenAI-compatible API for local inference with Jan. Run AI models locally with complete privacy using llama.cpp's high-performance inference engine. Supports GGUF models with CPU and GPU acceleration. No authentication required for local usage.", "version": "0.3.14", "contact": { @@ -49,9 +49,7 @@ "paths": { "/v1/completions": { "post": { - "tags": [ - "Completions" - ], + "tags": ["Completions"], "summary": "Create completion", "description": "Creates a completion for the provided prompt and parameters. This endpoint is compatible with OpenAI's completions API.", "operationId": "create_completion", @@ -92,11 +90,7 @@ "prompt": "# Python function to calculate fibonacci\ndef fibonacci(n):", "max_tokens": 200, "temperature": 0.3, - "stop": [ - "\n\n", - "def ", - "class " - ] + "stop": ["\n\n", "def ", "class "] } }, "streaming": { @@ -157,9 +151,7 @@ }, "/v1/chat/completions": { "post": { - "tags": [ - "Chat" - ], + "tags": ["Chat"], "summary": "Create chat completion", "description": "Creates a model response for the given chat conversation. This endpoint is compatible with OpenAI's chat completions API.", "operationId": "create_chat_completion", @@ -319,9 +311,7 @@ }, "/v1/models": { "get": { - "tags": [ - "Models" - ], + "tags": ["Models"], "summary": "List available models", "description": "Lists the currently available models and provides basic information about each one such as the owner and availability.", "operationId": "list_models", @@ -370,9 +360,7 @@ }, "/extras/tokenize": { "post": { - "tags": [ - "Extras" - ], + "tags": ["Extras"], "summary": "Tokenize text", "description": "Convert text input into tokens using the model's tokenizer.", "operationId": "tokenize", @@ -399,12 +387,7 @@ "$ref": "#/components/schemas/TokenizeResponse" }, "example": { - "tokens": [ - 15339, - 11, - 1917, - 0 - ] + "tokens": [15339, 11, 1917, 0] } } } @@ -414,9 +397,7 @@ }, "/extras/tokenize/count": { "post": { - "tags": [ - "Extras" - ], + "tags": ["Extras"], "summary": "Count tokens", "description": "Count the number of tokens in the provided text.", "operationId": "count_tokens", @@ -472,9 +453,7 @@ ] } }, - "required": [ - "input" - ] + "required": ["input"] }, "TokenizeResponse": { "type": "object", @@ -487,9 +466,7 @@ "description": "Array of token IDs" } }, - "required": [ - "tokens" - ] + "required": ["tokens"] }, "TokenCountResponse": { "type": "object", @@ -499,9 +476,7 @@ "description": "Number of tokens" } }, - "required": [ - "count" - ] + "required": ["count"] } }, "securitySchemes": { @@ -530,14 +505,11 @@ "no_telemetry": true, "offline_capable": true }, - "model_formats": [ - "GGUF", - "GGML" - ], + "model_formats": ["GGUF", "GGML"], "default_settings": { "context_length": 4096, "batch_size": 512, "threads": "auto" } } -} \ No newline at end of file +} diff --git a/website/src/components/ApiReferenceLayout.astro b/website/src/components/ApiReferenceLayout.astro index 35a75e265..ce9f35c18 100644 --- a/website/src/components/ApiReferenceLayout.astro +++ b/website/src/components/ApiReferenceLayout.astro @@ -152,7 +152,7 @@ const { title, description } = Astro.props; align-items: center; gap: 0.5rem; text-decoration: none; - color: var(--sl-color-white); + color: var(--sl-color-text); font-size: var(--sl-text-lg); font-weight: 600; }