separated scripts inside config file and fixed nav bar

This commit is contained in:
Ramon Perez 2025-09-05 21:39:33 +10:00
parent aea474bf57
commit afcaf531ed

View File

@ -1,7 +1,7 @@
{ {
"openapi": "3.1.0", "openapi": "3.1.0",
"info": { "info": {
"title": "Jan API", "title": "👋Jan API",
"description": "OpenAI-compatible API for local inference with Jan. Run AI models locally with complete privacy using llama.cpp's high-performance inference engine. Supports GGUF models with CPU and GPU acceleration. No authentication required for local usage.", "description": "OpenAI-compatible API for local inference with Jan. Run AI models locally with complete privacy using llama.cpp's high-performance inference engine. Supports GGUF models with CPU and GPU acceleration. No authentication required for local usage.",
"version": "0.3.14", "version": "0.3.14",
"contact": { "contact": {
@ -49,9 +49,7 @@
"paths": { "paths": {
"/v1/completions": { "/v1/completions": {
"post": { "post": {
"tags": [ "tags": ["Completions"],
"Completions"
],
"summary": "Create completion", "summary": "Create completion",
"description": "Creates a completion for the provided prompt and parameters. This endpoint is compatible with OpenAI's completions API.", "description": "Creates a completion for the provided prompt and parameters. This endpoint is compatible with OpenAI's completions API.",
"operationId": "create_completion", "operationId": "create_completion",
@ -92,11 +90,7 @@
"prompt": "# Python function to calculate fibonacci\ndef fibonacci(n):", "prompt": "# Python function to calculate fibonacci\ndef fibonacci(n):",
"max_tokens": 200, "max_tokens": 200,
"temperature": 0.3, "temperature": 0.3,
"stop": [ "stop": ["\n\n", "def ", "class "]
"\n\n",
"def ",
"class "
]
} }
}, },
"streaming": { "streaming": {
@ -157,9 +151,7 @@
}, },
"/v1/chat/completions": { "/v1/chat/completions": {
"post": { "post": {
"tags": [ "tags": ["Chat"],
"Chat"
],
"summary": "Create chat completion", "summary": "Create chat completion",
"description": "Creates a model response for the given chat conversation. This endpoint is compatible with OpenAI's chat completions API.", "description": "Creates a model response for the given chat conversation. This endpoint is compatible with OpenAI's chat completions API.",
"operationId": "create_chat_completion", "operationId": "create_chat_completion",
@ -319,9 +311,7 @@
}, },
"/v1/models": { "/v1/models": {
"get": { "get": {
"tags": [ "tags": ["Models"],
"Models"
],
"summary": "List available models", "summary": "List available models",
"description": "Lists the currently available models and provides basic information about each one such as the owner and availability.", "description": "Lists the currently available models and provides basic information about each one such as the owner and availability.",
"operationId": "list_models", "operationId": "list_models",
@ -370,9 +360,7 @@
}, },
"/extras/tokenize": { "/extras/tokenize": {
"post": { "post": {
"tags": [ "tags": ["Extras"],
"Extras"
],
"summary": "Tokenize text", "summary": "Tokenize text",
"description": "Convert text input into tokens using the model's tokenizer.", "description": "Convert text input into tokens using the model's tokenizer.",
"operationId": "tokenize", "operationId": "tokenize",
@ -399,12 +387,7 @@
"$ref": "#/components/schemas/TokenizeResponse" "$ref": "#/components/schemas/TokenizeResponse"
}, },
"example": { "example": {
"tokens": [ "tokens": [15339, 11, 1917, 0]
15339,
11,
1917,
0
]
} }
} }
} }
@ -414,9 +397,7 @@
}, },
"/extras/tokenize/count": { "/extras/tokenize/count": {
"post": { "post": {
"tags": [ "tags": ["Extras"],
"Extras"
],
"summary": "Count tokens", "summary": "Count tokens",
"description": "Count the number of tokens in the provided text.", "description": "Count the number of tokens in the provided text.",
"operationId": "count_tokens", "operationId": "count_tokens",
@ -472,9 +453,7 @@
] ]
} }
}, },
"required": [ "required": ["input"]
"input"
]
}, },
"TokenizeResponse": { "TokenizeResponse": {
"type": "object", "type": "object",
@ -487,9 +466,7 @@
"description": "Array of token IDs" "description": "Array of token IDs"
} }
}, },
"required": [ "required": ["tokens"]
"tokens"
]
}, },
"TokenCountResponse": { "TokenCountResponse": {
"type": "object", "type": "object",
@ -499,9 +476,7 @@
"description": "Number of tokens" "description": "Number of tokens"
} }
}, },
"required": [ "required": ["count"]
"count"
]
} }
}, },
"securitySchemes": { "securitySchemes": {
@ -530,10 +505,7 @@
"no_telemetry": true, "no_telemetry": true,
"offline_capable": true "offline_capable": true
}, },
"model_formats": [ "model_formats": ["GGUF", "GGML"],
"GGUF",
"GGML"
],
"default_settings": { "default_settings": {
"context_length": 4096, "context_length": 4096,
"batch_size": 512, "batch_size": 512,