diff --git a/electron/core/plugins/inference-plugin/index.js b/electron/core/plugins/inference-plugin/index.js index cbd48d682..6c001d9d2 100644 --- a/electron/core/plugins/inference-plugin/index.js +++ b/electron/core/plugins/inference-plugin/index.js @@ -17,9 +17,11 @@ const dispose = async () => .then((res) => resolve(res)); } }); +const inferenceUrl = () => "http://localhost:8080/llama/chat_completion"; // Register all the above functions and objects with the relevant extension points export function init({ register }) { register("initModel", "initModel", initModel); + register("inferenceUrl", "inferenceUrl", inferenceUrl); register("dispose", "dispose", dispose); } diff --git a/web/app/_hooks/useSendChatMessage.ts b/web/app/_hooks/useSendChatMessage.ts index 127c7dcec..f5bca3f96 100644 --- a/web/app/_hooks/useSendChatMessage.ts +++ b/web/app/_hooks/useSendChatMessage.ts @@ -11,7 +11,7 @@ import { import { useAtom, useAtomValue, useSetAtom } from "jotai"; import { selectAtom } from "jotai/utils"; -import { DataService } from "../../shared/coreService"; +import { DataService, InfereceService } from "../../shared/coreService"; import { MessageSenderType, RawMessage, @@ -52,9 +52,7 @@ export default function useSendChatMessage() { addNewMessage(newChatMessage); const recentMessages = [ - ...chatMessagesHistory.sort( - (a, b) => parseInt(a.id) - parseInt(b.id) - ), + ...chatMessagesHistory.sort((a, b) => parseInt(a.id) - parseInt(b.id)), newChatMessage, ] .slice(-10) @@ -67,23 +65,21 @@ export default function useSendChatMessage() { : "assistant", }; }); - const response = await fetch( - "http://localhost:8080/llama/chat_completion", - { - method: "POST", - headers: { - "Content-Type": "application/json", - Accept: "text/event-stream", - "Access-Control-Allow-Origi": "*", - }, - body: JSON.stringify({ - messages: recentMessages, - stream: true, - model: "gpt-3.5-turbo", - max_tokens: 500, - }), - } - ); + const url = await executeSerial(InfereceService.INFERENCE_URL); + const response = await fetch(url, { + method: "POST", + headers: { + "Content-Type": "application/json", + Accept: "text/event-stream", + "Access-Control-Allow-Origi": "*", + }, + body: JSON.stringify({ + messages: recentMessages, + stream: true, + model: "gpt-3.5-turbo", + max_tokens: 500, + }), + }); const stream = response.body; const decoder = new TextDecoder("utf-8"); diff --git a/web/shared/coreService.ts b/web/shared/coreService.ts index 35f7583a0..1de8f95ca 100644 --- a/web/shared/coreService.ts +++ b/web/shared/coreService.ts @@ -28,7 +28,7 @@ export enum ModelService { } export enum InfereceService { - PROMPT = "prompt", + INFERENCE_URL = "inferenceUrl", INIT_MODEL = "initModel", }