jan/web/hooks/useSendMessage.ts
NamH 101268f6f3
feat: integrating cortex (#3001)
* feat: integrating cortex

* Temporary prevent crash

Signed-off-by: James <namnh0122@gmail.com>

* fix yarn lint

Signed-off-by: James <namnh0122@gmail.com>

* refactor: remove core node module - fs - extensions and so on (#3151)

* add migration script for threads, messages and models

Signed-off-by: James <namnh0122@gmail.com>

* remove freq_penalty and presence_penalty if model not supported

Signed-off-by: James <namnh0122@gmail.com>

* add back models in my models

Signed-off-by: James <namnh0122@gmail.com>

* fix api-url for setup API key popup

Signed-off-by: James <namnh0122@gmail.com>

* fix using model name for dropdown model

Signed-off-by: James <namnh0122@gmail.com>

* fix can't click to hotkey

Signed-off-by: James <namnh0122@gmail.com>

* fix: disable some UIs

Signed-off-by: James <namnh0122@gmail.com>

* fix build

Signed-off-by: James <namnh0122@gmail.com>

* reduce calling HF api

Signed-off-by: James <namnh0122@gmail.com>

* some ui update

Signed-off-by: James <namnh0122@gmail.com>

* feat: modal migration UI  (#3153)

* feat: handle popup migration

* chore: update loader

* chore: integrate script migration

* chore: cleanup import

* chore: moving out spinner loader

* chore: update check thread message success migrate

* chore: add handle script into retry button

* remove warning from joi

Signed-off-by: James <namnh0122@gmail.com>

* chore: fix duplicate children

* fix: path after migrating model

Signed-off-by: James <namnh0122@gmail.com>

* chore: apply mutation for config

* chore: prevent calling too many create assistant api

Signed-off-by: James <namnh0122@gmail.com>

* using cortexso

Signed-off-by: James <namnh0122@gmail.com>

* update download api

Signed-off-by: James <namnh0122@gmail.com>

* fix use on slider item

Signed-off-by: James <namnh0122@gmail.com>

* fix: ui no download model or simple onboarding (#3166)

* fix download huggingface model match with slider item

Signed-off-by: James <namnh0122@gmail.com>

* update owner_logo to logo and author

Signed-off-by: James <namnh0122@gmail.com>

* update new cortexso

Signed-off-by: James <namnh0122@gmail.com>

* Add install python step for macos

* add engine table

Signed-off-by: James <namnh0122@gmail.com>

* fix local icons

Signed-off-by: James <namnh0122@gmail.com>

* feat: add search feature for model hub

Signed-off-by: James <namnh0122@gmail.com>

* fix misalign switch

Signed-off-by: James <namnh0122@gmail.com>

* fix: delete thread not focus on other thread

Signed-off-by: James <namnh0122@gmail.com>

* add get model from hugging face

Signed-off-by: James <namnh0122@gmail.com>

* fix download from hugging face

Signed-off-by: James <namnh0122@gmail.com>

* small update

Signed-off-by: James <namnh0122@gmail.com>

* update

Signed-off-by: James <namnh0122@gmail.com>

* fix system monitor rounded only on the left

Signed-off-by: James <namnh0122@gmail.com>

* chore: update ui new hub screen (#3174)

* chore: update ui new hub screen

* chore: update layout centerpanel thread and hub screen

* chore: update detail model by group

* update cortexso 0.1.13

Signed-off-by: James <namnh0122@gmail.com>

* chore: add file size

Signed-off-by: James <namnh0122@gmail.com>

* chore: put engine to experimental feature

Signed-off-by: James <namnh0122@gmail.com>

* chore: open cortex folder

Signed-off-by: James <namnh0122@gmail.com>

* chore: add back user avatar

Signed-off-by: James <namnh0122@gmail.com>

* chore: minor UI hub (#3182)

* chore: add back right click thread list and update 3 dots are overlapping with the text

* chore: update position dropdown list my models

* chore: make on-device tab showing 6 items instead of 4

* chore: update style description modals detail model

* chore: update isGeneration loader and author name on modal

* feat: integrate cortex single executable

Signed-off-by: James <namnh0122@gmail.com>

* fix build

Signed-off-by: James <namnh0122@gmail.com>

* chore: added blank state

* chore: update ui component blank state

* bump cortex binary version

* fix: logic show modal migration (#3165)

* fix: logic show modal migration

* chore: fixed logic

* chore: read contain format gguf local models

* chore: change return hasLocalModel

* chore: intiial skipmigration state

* chore: filter embedding model

* fix: delete top thread not focus on any other thread

* chore: added UI no result component search models group (#3188)

* fix: remote model should show all when user config that engine

Signed-off-by: James <namnh0122@gmail.com>

* chore: set state thread and models migration using getOnInit (#3189)

* chore: set state thread and models migration using getOnInit

* chore: add state as dependecies hooks

* chore: system monitor panel show engine model (#3192)

* fix: remove config api, replace with engine

Signed-off-by: James <namnh0122@gmail.com>

* update

Signed-off-by: James <namnh0122@gmail.com>

* update reactquery

Signed-off-by: James <namnh0122@gmail.com>

* bump cortex 0.4.35

* feat: add waiting for cortex popup

Signed-off-by: James <namnh0122@gmail.com>

* chore: add loader detail model popup (#3195)

* chore: model start loader (#3197)

* chore: added model loader when user starting chat without model active

* chore: update copies loader

* fix: select min file size if recommended quant does not exist

Signed-off-by: James <namnh0122@gmail.com>

* chore: temporary hide gpu config

* fix: tensorrt not shown

Signed-off-by: James <namnh0122@gmail.com>

* fix lint

Signed-off-by: James <namnh0122@gmail.com>

* fix tests

Signed-off-by: James <namnh0122@gmail.com>

* fix e2e tests (wip)

Signed-off-by: James <namnh0122@gmail.com>

* update

Signed-off-by: James <namnh0122@gmail.com>

* fix: adding element and correct test to adapt new UI

* fix: temp skip unstable part

* fix: only show models which can be supported

Signed-off-by: James <namnh0122@gmail.com>

* Update version.txt

* update send message

Signed-off-by: James <namnh0122@gmail.com>

* fix: not allow user send message when is generating

Signed-off-by: James <namnh0122@gmail.com>

* chore: temp skip Playwright test due to env issue

* chore: temp skip Playwright test due to env issue

* update

Signed-off-by: James <namnh0122@gmail.com>

* chore: minor-ui-feedback (#3202)

---------

Signed-off-by: James <namnh0122@gmail.com>
Co-authored-by: Louis <louis@jan.ai>
Co-authored-by: Faisal Amir <urmauur@gmail.com>
Co-authored-by: Hien To <tominhhien97@gmail.com>
Co-authored-by: Van Pham <64197333+Van-QA@users.noreply.github.com>
Co-authored-by: Van-QA <van@jan.ai>
2024-07-26 17:52:43 +07:00

722 lines
21 KiB
TypeScript

import { useCallback, useRef } from 'react'
import {
ChatCompletionCreateParamsNonStreaming,
ChatCompletionMessageParam,
LocalEngines,
Message,
MessageContent,
RemoteEngines,
TextContentBlock,
Thread,
} from '@janhq/core'
import { useAtomValue, useSetAtom } from 'jotai'
import { currentPromptAtom, editPromptAtom } from '@/containers/Providers/Jotai'
import { toaster } from '@/containers/Toast'
import useCortex from './useCortex'
import useEngineInit from './useEngineInit'
import useEngineQuery from './useEngineQuery'
import useMessageCreateMutation from './useMessageCreateMutation'
import useMessageUpdateMutation from './useMessageUpdateMutation'
import useModelStart from './useModelStart'
import {
addNewMessageAtom,
getCurrentChatMessagesAtom,
updateMessageAtom,
} from '@/helpers/atoms/ChatMessage.atom'
import {
activeModelsAtom,
getSelectedModelAtom,
} from '@/helpers/atoms/Model.atom'
import {
activeThreadAtom,
addThreadIdShouldAnimateTitleAtom,
isGeneratingResponseAtom,
updateThreadTitleAtom,
} from '@/helpers/atoms/Thread.atom'
// TODO: NamH add this back
// const normalizeMessages = (messages: Message[]): Message[] => {
// const stack = new Stack<Message>()
// for (const message of messages) {
// if (stack.isEmpty()) {
// stack.push(message)
// continue
// }
// const topMessage = stack.peek()
// if (message.role === topMessage.role) {
// // add an empty message
// stack.push({
// role: topMessage.role === 'user' ? 'assistant' : 'user',
// content: '.', // some model requires not empty message
// })
// }
// stack.push(message)
// }
// return stack.reverseOutput()
// }
const useSendMessage = () => {
const createMessage = useMessageCreateMutation()
const updateMessage = useMessageUpdateMutation()
const initializeEngine = useEngineInit()
const addNewMessage = useSetAtom(addNewMessageAtom)
const { chatCompletionStreaming, chatCompletionNonStreaming, updateThread } =
useCortex()
const updateMessageState = useSetAtom(updateMessageAtom)
const setIsGeneratingResponse = useSetAtom(isGeneratingResponseAtom)
const setCurrentPrompt = useSetAtom(currentPromptAtom)
const setEditPrompt = useSetAtom(editPromptAtom)
const updateThreadTitle = useSetAtom(updateThreadTitleAtom)
const addThreadIdShouldAnimateTitle = useSetAtom(
addThreadIdShouldAnimateTitleAtom
)
const { data: engineData } = useEngineQuery()
const activeThread = useAtomValue(activeThreadAtom)
const activeModels = useAtomValue(activeModelsAtom)
const currentMessages = useAtomValue(getCurrentChatMessagesAtom)
const selectedModel = useAtomValue(getSelectedModelAtom)
const startModel = useModelStart()
const abortControllerRef = useRef<AbortController | undefined>(undefined)
const validatePrerequisite = useCallback(async (): Promise<boolean> => {
const errorTitle = 'Failed to send message'
if (!activeThread) {
toaster({
title: errorTitle,
description: 'No active thread! Please select a thread!',
type: 'error',
})
return false
}
if (!selectedModel) {
toaster({
title: errorTitle,
description: 'No model selected! Please select a model!',
type: 'error',
})
return false
}
if (!engineData) {
toaster({
title: errorTitle,
description:
'Jan failed to fetch available engine data! Please try restart the app!',
type: 'error',
})
return false
}
try {
if (selectedModel.model !== activeThread.assistants[0].model) {
activeThread.assistants[0].model = selectedModel.model
await updateThread(activeThread)
}
} catch (err) {
toaster({
title: errorTitle,
description: 'Please try select model for this thread again!',
type: 'error',
})
console.error(`Failed to update thread ${activeThread.id}, error: ${err}`)
return false
}
if (!selectedModel.engine) {
toaster({
title: errorTitle,
description: `Model ${selectedModel.model} does not have an engine`,
type: 'error',
})
console.error(`Model ${selectedModel.model} does not have an engine`)
return false
}
const engineStatus = engineData.find((e) => e.name === selectedModel.engine)
if (!engineStatus) {
toaster({
title: errorTitle,
description: `Engine ${selectedModel.engine} is not available`,
type: 'error',
})
console.error(`Engine ${selectedModel.engine} is not available`)
return false
}
if (
RemoteEngines.find((e) => e === selectedModel.engine) != null &&
engineStatus.status === 'missing_configuration'
) {
toaster({
title: errorTitle,
description: `Engine ${engineStatus.name} is missing configuration`,
type: 'error',
})
console.error(`Engine ${engineStatus.name} is missing configuration`)
return false
}
if (
LocalEngines.find((e) => e === selectedModel.engine) != null &&
engineStatus.status === 'not_initialized'
) {
toaster({
title: 'Please wait for engine to initialize',
description: `Please retry after engine ${engineStatus.name} is installed.`,
type: 'default',
})
initializeEngine.mutate(selectedModel.engine)
return false
}
if (engineStatus.status !== 'ready') {
toaster({
title: errorTitle,
description: `Engine ${engineStatus.name} is not ready`,
type: 'error',
})
console.error(`Engine ${engineStatus.name} is not ready`)
return false
}
return true
}, [activeThread, selectedModel, engineData, initializeEngine, updateThread])
const stopInference = useCallback(() => {
abortControllerRef.current?.abort()
}, [])
const summarizeThread = useCallback(
async (messages: string[], modelId: string, thread: Thread) => {
const maxWordForThreadTitle = 10
const summarizeMessages: ChatCompletionMessageParam[] = [
{
role: 'user',
content: `Summarize in a ${maxWordForThreadTitle}-word title the following conversation:\n\n${messages.join('\n')}`,
},
]
const summarizeParams: ChatCompletionCreateParamsNonStreaming = {
messages: summarizeMessages,
model: modelId,
max_tokens: 150,
temperature: 0.5,
}
const summarizeStream = await chatCompletionNonStreaming(summarizeParams)
const summarizedText = (
summarizeStream.choices[0].message.content ?? 'New Thread'
).replace(/"/g, '')
addThreadIdShouldAnimateTitle(thread.id)
updateThread({ ...thread, title: summarizedText })
updateThreadTitle(thread.id, summarizedText)
},
[
addThreadIdShouldAnimateTitle,
chatCompletionNonStreaming,
updateThreadTitle,
updateThread,
]
)
const resendMessage = useCallback(async () => {
const isValid = await validatePrerequisite()
if (!isValid) return
const modelId = activeThread!.assistants[0].model
try {
// start model if not yet started
if (LocalEngines.find((e) => e === selectedModel!.engine) != null) {
// start model if local and not started
if (!activeModels.map((model) => model.model).includes(modelId)) {
await startModel.mutateAsync(modelId)
}
}
} catch (err) {
console.error(`Failed to start model ${modelId}, error: ${err}`)
toaster({
title: 'Failed to start model',
description: `Failed to start model ${modelId}`,
type: 'error',
})
}
setIsGeneratingResponse(true)
// building messages
const systemMessage: ChatCompletionMessageParam = {
role: 'system',
content: activeThread!.assistants[0].instructions ?? '',
}
const messages: ChatCompletionMessageParam[] = currentMessages
.map((msg) => {
switch (msg.role) {
case 'user':
case 'assistant':
return {
role: msg.role,
content: (msg.content[0] as TextContentBlock).text.value,
}
// we will need to support other roles in the future
default:
break
}
})
.filter((msg) => msg != null) as ChatCompletionMessageParam[]
messages.unshift(systemMessage)
const modelOptions: Record<string, string | number> = {}
if (selectedModel!.frequency_penalty) {
modelOptions.frequency_penalty = selectedModel!.frequency_penalty
}
if (selectedModel!.presence_penalty) {
modelOptions.presence_penalty = selectedModel!.presence_penalty
}
try {
let assistantResponseMessage = ''
if (selectedModel!.stream === true) {
const stream = await chatCompletionStreaming({
messages,
model: selectedModel!.model,
stream: true,
max_tokens: selectedModel!.max_tokens,
stop: selectedModel!.stop,
temperature: selectedModel!.temperature ?? 1,
top_p: selectedModel!.top_p ?? 1,
...modelOptions,
})
abortControllerRef.current = stream.controller
const assistantMessage = await createMessage.mutateAsync({
threadId: activeThread!.id,
createMessageParams: {
role: 'assistant',
content: '',
},
})
const responseMessage: Message = {
id: assistantMessage.id,
thread_id: activeThread!.id,
assistant_id: activeThread!.id,
role: 'assistant',
content: [],
status: 'in_progress',
created_at: assistantMessage.created_at,
metadata: undefined,
attachments: null,
completed_at: Date.now(),
incomplete_at: null,
incomplete_details: null,
object: 'thread.message',
run_id: null,
}
addNewMessage(responseMessage)
for await (const chunk of stream) {
const content = chunk.choices[0]?.delta?.content || ''
assistantResponseMessage += content
const messageContent: MessageContent = {
type: 'text',
text: {
value: assistantResponseMessage,
annotations: [],
},
}
responseMessage.content = [messageContent]
updateMessageState(
responseMessage.id,
responseMessage.thread_id,
responseMessage.content,
responseMessage.status
)
}
abortControllerRef.current = undefined
responseMessage.status = 'completed'
updateMessageState(
responseMessage.id,
responseMessage.thread_id,
responseMessage.content,
responseMessage.status
)
updateMessage.mutateAsync({
threadId: activeThread!.id,
messageId: responseMessage.id,
data: {
content: responseMessage.content,
},
})
} else {
const abortController = new AbortController()
const response = await chatCompletionNonStreaming(
{
messages,
model: selectedModel!.model,
stream: false,
max_tokens: selectedModel!.max_tokens,
stop: selectedModel!.stop,
temperature: selectedModel!.temperature ?? 1,
top_p: selectedModel!.top_p ?? 1,
...modelOptions,
},
{
signal: abortController.signal,
}
)
assistantResponseMessage = response.choices[0].message.content ?? ''
const assistantMessage = await createMessage.mutateAsync({
threadId: activeThread!.id,
createMessageParams: {
role: 'assistant',
content: assistantResponseMessage,
},
})
const responseMessage: Message = {
id: assistantMessage.id,
thread_id: activeThread!.id,
assistant_id: activeThread!.id,
role: 'assistant',
content: [
{
type: 'text',
text: {
value: assistantResponseMessage,
annotations: [],
},
},
],
status: 'completed',
created_at: assistantMessage.created_at,
metadata: undefined,
attachments: null,
completed_at: Date.now(),
incomplete_at: null,
incomplete_details: null,
object: 'thread.message',
run_id: null,
}
updateMessage.mutate({
threadId: activeThread!.id,
messageId: responseMessage.id,
data: {
content: responseMessage.content,
},
})
addNewMessage(responseMessage)
}
} catch (err) {
console.error(err)
toaster({
title: 'Failed to generate response',
type: 'error',
})
}
setIsGeneratingResponse(false)
}, [
activeThread,
activeModels,
currentMessages,
selectedModel,
updateMessage,
createMessage,
validatePrerequisite,
startModel,
updateMessageState,
addNewMessage,
chatCompletionNonStreaming,
chatCompletionStreaming,
setIsGeneratingResponse,
])
const sendMessage = useCallback(
async (message: string) => {
const isValid = await validatePrerequisite()
if (!isValid) return
let shouldSummarize =
activeThread!.title === 'New Thread' ||
activeThread!.title.trim() === ''
const modelId = activeThread!.assistants[0].model
setCurrentPrompt('')
setEditPrompt('')
const userMessage = await createMessage.mutateAsync({
threadId: activeThread!.id,
createMessageParams: {
role: 'user',
content: message,
},
})
// Push to states
addNewMessage(userMessage)
try {
// start model if not yet started
if (LocalEngines.find((e) => e === selectedModel!.engine) != null) {
// start model if local and not started
if (!activeModels.map((model) => model.model).includes(modelId)) {
await startModel.mutateAsync(modelId)
}
}
} catch (err) {
console.error(`Failed to start model ${modelId}, error: ${err}`)
return
}
setIsGeneratingResponse(true)
// building messages
const systemMessage: ChatCompletionMessageParam = {
role: 'system',
content: activeThread!.assistants[0].instructions ?? '',
}
const messages: ChatCompletionMessageParam[] = currentMessages
.map((msg) => {
switch (msg.role) {
case 'user':
case 'assistant':
return {
role: msg.role,
content: (msg.content[0] as TextContentBlock).text.value,
}
// we will need to support other roles in the future
default:
break
}
})
.filter((msg) => msg != null) as ChatCompletionMessageParam[]
messages.push({
role: 'user',
content: message,
})
messages.unshift(systemMessage)
const modelOptions: Record<string, string | number> = {}
if (selectedModel!.frequency_penalty) {
modelOptions.frequency_penalty = selectedModel!.frequency_penalty
}
if (selectedModel!.presence_penalty) {
modelOptions.presence_penalty = selectedModel!.presence_penalty
}
let assistantResponseMessage = ''
try {
if (selectedModel!.stream === true) {
const stream = await chatCompletionStreaming({
messages,
model: selectedModel!.model,
stream: true,
max_tokens: selectedModel!.max_tokens,
stop: selectedModel!.stop,
temperature: selectedModel!.temperature ?? 1,
top_p: selectedModel!.top_p ?? 1,
...modelOptions,
})
abortControllerRef.current = stream.controller
const assistantMessage = await createMessage.mutateAsync({
threadId: activeThread!.id,
createMessageParams: {
role: 'assistant',
content: '',
},
})
const responseMessage: Message = {
id: assistantMessage.id,
thread_id: activeThread!.id,
assistant_id: activeThread!.id,
role: 'assistant',
content: [],
status: 'in_progress',
created_at: assistantMessage.created_at,
metadata: undefined,
attachments: null,
completed_at: Date.now(),
incomplete_at: null,
incomplete_details: null,
object: 'thread.message',
run_id: null,
}
if (responseMessage) {
setIsGeneratingResponse(false)
}
addNewMessage(responseMessage)
for await (const chunk of stream) {
const content = chunk.choices[0]?.delta?.content || ''
assistantResponseMessage += content
const messageContent: MessageContent = {
type: 'text',
text: {
value: assistantResponseMessage,
annotations: [],
},
}
responseMessage.content = [messageContent]
updateMessageState(
responseMessage.id,
responseMessage.thread_id,
responseMessage.content,
responseMessage.status
)
}
abortControllerRef.current = undefined
responseMessage.status = 'completed'
updateMessageState(
responseMessage.id,
responseMessage.thread_id,
responseMessage.content,
responseMessage.status
)
updateMessage.mutateAsync({
threadId: activeThread!.id,
messageId: responseMessage.id,
data: {
content: responseMessage.content,
},
})
} else {
const abortController = new AbortController()
const response = await chatCompletionNonStreaming(
{
messages,
model: selectedModel!.model,
stream: false,
max_tokens: selectedModel!.max_tokens,
stop: selectedModel!.stop,
temperature: selectedModel!.temperature ?? 1,
top_p: selectedModel!.top_p ?? 1,
...modelOptions,
},
{
signal: abortController.signal,
}
)
assistantResponseMessage = response.choices[0].message.content ?? ''
const assistantMessage = await createMessage.mutateAsync({
threadId: activeThread!.id,
createMessageParams: {
role: 'assistant',
content: assistantResponseMessage,
},
})
const responseMessage: Message = {
id: assistantMessage.id,
thread_id: activeThread!.id,
assistant_id: activeThread!.id,
role: 'assistant',
content: [
{
type: 'text',
text: {
value: assistantResponseMessage,
annotations: [],
},
},
],
status: 'completed',
created_at: assistantMessage.created_at,
metadata: undefined,
attachments: null,
completed_at: Date.now(),
incomplete_at: null,
incomplete_details: null,
object: 'thread.message',
run_id: null,
}
updateMessage.mutateAsync({
threadId: activeThread!.id,
messageId: responseMessage.id,
data: {
content: responseMessage.content,
},
})
if (responseMessage) {
setIsGeneratingResponse(false)
}
addNewMessage(responseMessage)
}
} catch (err) {
console.error(err)
setIsGeneratingResponse(false)
shouldSummarize = false
toaster({
title: 'Failed to generate response',
type: 'error',
})
}
try {
if (!shouldSummarize) return
// summarize if needed
const textMessages: string[] = messages
.map((msg) => {
if (typeof msg.content === 'string') return msg.content
})
.filter((msg) => msg != null) as string[]
textMessages.push(assistantResponseMessage)
summarizeThread(textMessages, modelId, activeThread!)
} catch (err) {
console.error(`Failed to summarize thread: ${err}`)
}
},
[
activeThread,
activeModels,
currentMessages,
selectedModel,
updateMessage,
createMessage,
validatePrerequisite,
setCurrentPrompt,
setEditPrompt,
setIsGeneratingResponse,
updateMessageState,
addNewMessage,
startModel,
chatCompletionNonStreaming,
chatCompletionStreaming,
summarizeThread,
]
)
return { resendMessage, sendMessage, stopInference }
}
export default useSendMessage