diff --git a/extensions/conversational-extension/src/index.ts b/extensions/conversational-extension/src/index.ts index b3530e5d5..30766be45 100644 --- a/extensions/conversational-extension/src/index.ts +++ b/extensions/conversational-extension/src/index.ts @@ -40,7 +40,7 @@ export default class CortexConversationalExtension extends ConversationalExtensi async listThreads(): Promise { return this.queue.add(() => ky - .get(`${API_URL}/v1/threads`) + .get(`${API_URL}/v1/threads?limit=-1`) .json() .then((e) => e.data) ) as Promise @@ -133,7 +133,7 @@ export default class CortexConversationalExtension extends ConversationalExtensi async listMessages(threadId: string): Promise { return this.queue.add(() => ky - .get(`${API_URL}/v1/threads/${threadId}/messages?order=asc`) + .get(`${API_URL}/v1/threads/${threadId}/messages?order=asc&limit=-1`) .json() .then((e) => e.data) ) as Promise @@ -147,7 +147,9 @@ export default class CortexConversationalExtension extends ConversationalExtensi */ async getThreadAssistant(threadId: string): Promise { return this.queue.add(() => - ky.get(`${API_URL}/v1/assistants/${threadId}`).json() + ky + .get(`${API_URL}/v1/assistants/${threadId}?limit=-1`) + .json() ) as Promise } /** @@ -188,7 +190,7 @@ export default class CortexConversationalExtension extends ConversationalExtensi * Do health check on cortex.cpp * @returns */ - healthz(): Promise { + async healthz(): Promise { return ky .get(`${API_URL}/healthz`, { retry: { limit: 20, delay: () => 500, methods: ['get'] }, diff --git a/extensions/inference-cortex-extension/bin/version.txt b/extensions/inference-cortex-extension/bin/version.txt index cf13dd42b..af0b7ddbf 100644 --- a/extensions/inference-cortex-extension/bin/version.txt +++ b/extensions/inference-cortex-extension/bin/version.txt @@ -1 +1 @@ -1.0.5-rc2 +1.0.6 diff --git a/extensions/inference-openai-extension/resources/models.json b/extensions/inference-openai-extension/resources/models.json index fc6896882..0c822fde2 100644 --- a/extensions/inference-openai-extension/resources/models.json +++ b/extensions/inference-openai-extension/resources/models.json @@ -85,6 +85,63 @@ }, "engine": "openai" }, + { + "sources": [ + { + "url": "https://openai.com" + } + ], + "id": "gpt-4o-mini", + "object": "model", + "name": "OpenAI GPT 4o-mini", + "version": "1.1", + "description": "GPT-4o mini (“o” for “omni”) is a fast, affordable small model for focused tasks.", + "format": "api", + "settings": { + "vision_model": true + }, + "parameters": { + "max_tokens": 16384, + "temperature": 0.7, + "top_p": 0.95, + "stream": true, + "stop": [], + "frequency_penalty": 0, + "presence_penalty": 0 + }, + "metadata": { + "author": "OpenAI", + "tags": ["General"] + }, + "engine": "openai" + }, + { + "sources": [ + { + "url": "https://openai.com" + } + ], + "id": "o1", + "object": "model", + "name": "OpenAI o1", + "version": "1.0", + "description": "OpenAI o1 is a new model with complex reasoning", + "format": "api", + "settings": {}, + "parameters": { + "max_tokens": 100000, + "temperature": 1, + "top_p": 1, + "stream": true, + "frequency_penalty": 0, + "presence_penalty": 0 + }, + "metadata": { + "author": "OpenAI", + "tags": ["General"] + }, + "engine": "openai" + }, { "sources": [ { diff --git a/extensions/model-extension/src/cortex.ts b/extensions/model-extension/src/cortex.ts index 26316fbbc..ca5763962 100644 --- a/extensions/model-extension/src/cortex.ts +++ b/extensions/model-extension/src/cortex.ts @@ -53,7 +53,7 @@ export class CortexAPI implements ICortexAPI { */ getModels(): Promise { return this.queue - .add(() => ky.get(`${API_URL}/v1/models`).json()) + .add(() => ky.get(`${API_URL}/v1/models?limit=-1`).json()) .then((e) => typeof e === 'object' ? e.data.map((e) => this.transformModel(e)) : [] ) diff --git a/joi/src/core/TextArea/index.tsx b/joi/src/core/TextArea/index.tsx index 6807178ff..306e18d05 100644 --- a/joi/src/core/TextArea/index.tsx +++ b/joi/src/core/TextArea/index.tsx @@ -34,7 +34,11 @@ const TextArea = forwardRef( return (