diff --git a/web/hooks/useSendChatMessage.ts b/web/hooks/useSendChatMessage.ts index 67501ab84..3e6131fac 100644 --- a/web/hooks/useSendChatMessage.ts +++ b/web/hooks/useSendChatMessage.ts @@ -24,11 +24,7 @@ import { ChatCompletionTool, ChatCompletionMessageToolCall, } from 'openai/resources/chat' -import { - CompletionResponse, - StreamCompletionResponse, - TokenJS, -} from 'token.js' +import { CompletionResponse, StreamCompletionResponse, TokenJS } from 'token.js' import { ulid } from 'ulidx' import { modelDropdownStateAtom } from '@/containers/ModelDropdown' @@ -225,7 +221,7 @@ export default function useSendChatMessage( }, activeThread, messages ?? currentMessages, - (tools && tools.length) ? tools : undefined, + tools && tools.length ? tools : undefined ).addSystemMessage(activeAssistant.instructions) requestBuilder.pushMessage(prompt, base64Blob, fileUpload) @@ -265,8 +261,9 @@ export default function useSendChatMessage( } // Start Model if not started - const isCortex = modelRequest.engine == InferenceEngine.cortex || - modelRequest.engine == InferenceEngine.cortex_llamacpp + const isCortex = + modelRequest.engine == InferenceEngine.cortex || + modelRequest.engine == InferenceEngine.cortex_llamacpp const modelId = selectedModel?.id ?? activeAssistantRef.current?.model.id if (base64Blob) { @@ -297,9 +294,10 @@ export default function useSendChatMessage( extendBuiltInEngineModels(tokenJS, provider, modelId) // llama.cpp currently does not support streaming when tools are used. - const useStream = (requestBuilder.tools && isCortex) ? - false : - modelRequest.parameters?.stream + const useStream = + requestBuilder.tools && isCortex + ? false + : modelRequest.parameters?.stream let parentMessageId: string | undefined while (!isDone) { @@ -577,4 +575,4 @@ export default function useSendChatMessage( sendChatMessage, resendChatMessage, } -} \ No newline at end of file +} diff --git a/web/screens/Hub/ModelPage/index.tsx b/web/screens/Hub/ModelPage/index.tsx index 904a73cda..dcd0c833b 100644 --- a/web/screens/Hub/ModelPage/index.tsx +++ b/web/screens/Hub/ModelPage/index.tsx @@ -144,10 +144,10 @@ const ModelPage = ({ model, onGoBack }: Props) => { {model.type !== 'cloud' && ( <> - + Format - + Size diff --git a/web/screens/Settings/Engines/RemoteEngineSettings.tsx b/web/screens/Settings/Engines/RemoteEngineSettings.tsx index fa3f7a668..e773b1957 100644 --- a/web/screens/Settings/Engines/RemoteEngineSettings.tsx +++ b/web/screens/Settings/Engines/RemoteEngineSettings.tsx @@ -168,7 +168,8 @@ const RemoteEngineSettings = ({

{!customEngineLogo ? ( - Enter your authentication key to activate this engine.{' '} + Enter your authentication key to activate this + engine.{' '} ) : (