fix: Exposing PromptProgress to be passed as param
This commit is contained in:
parent
1c0e135077
commit
7127ff1244
@ -179,7 +179,6 @@ const ChatInput = ({
|
|||||||
const mcpExtension = extensionManager.get<MCPExtension>(ExtensionTypeEnum.MCP)
|
const mcpExtension = extensionManager.get<MCPExtension>(ExtensionTypeEnum.MCP)
|
||||||
const MCPToolComponent = mcpExtension?.getToolComponent?.()
|
const MCPToolComponent = mcpExtension?.getToolComponent?.()
|
||||||
|
|
||||||
|
|
||||||
const handleSendMesage = (prompt: string) => {
|
const handleSendMesage = (prompt: string) => {
|
||||||
if (!selectedModel) {
|
if (!selectedModel) {
|
||||||
setMessage('Please select a model to start chatting.')
|
setMessage('Please select a model to start chatting.')
|
||||||
|
|||||||
@ -4,7 +4,7 @@ import { MCPTool } from '@/types/completion'
|
|||||||
import { useAssistant } from './useAssistant'
|
import { useAssistant } from './useAssistant'
|
||||||
import { ChatCompletionMessageToolCall } from 'openai/resources'
|
import { ChatCompletionMessageToolCall } from 'openai/resources'
|
||||||
|
|
||||||
type PromptProgress = {
|
export type PromptProgress = {
|
||||||
cache: number
|
cache: number
|
||||||
processed: number
|
processed: number
|
||||||
time_ms: number
|
time_ms: number
|
||||||
|
|||||||
@ -3,7 +3,7 @@ import { flushSync } from 'react-dom'
|
|||||||
import { usePrompt } from './usePrompt'
|
import { usePrompt } from './usePrompt'
|
||||||
import { useModelProvider } from './useModelProvider'
|
import { useModelProvider } from './useModelProvider'
|
||||||
import { useThreads } from './useThreads'
|
import { useThreads } from './useThreads'
|
||||||
import { useAppState } from './useAppState'
|
import { useAppState, type PromptProgress } from './useAppState'
|
||||||
import { useMessages } from './useMessages'
|
import { useMessages } from './useMessages'
|
||||||
import { useRouter } from '@tanstack/react-router'
|
import { useRouter } from '@tanstack/react-router'
|
||||||
import { defaultModel } from '@/lib/models'
|
import { defaultModel } from '@/lib/models'
|
||||||
@ -23,7 +23,7 @@ import {
|
|||||||
ChatCompletionMessageToolCall,
|
ChatCompletionMessageToolCall,
|
||||||
CompletionUsage,
|
CompletionUsage,
|
||||||
} from 'openai/resources'
|
} from 'openai/resources'
|
||||||
import { MessageStatus, ContentType } from '@janhq/core'
|
import { MessageStatus, ContentType, ThreadMessage } from '@janhq/core'
|
||||||
|
|
||||||
import { useServiceHub } from '@/hooks/useServiceHub'
|
import { useServiceHub } from '@/hooks/useServiceHub'
|
||||||
import { useToolApproval } from '@/hooks/useToolApproval'
|
import { useToolApproval } from '@/hooks/useToolApproval'
|
||||||
@ -74,7 +74,7 @@ const finalizeMessage = (
|
|||||||
finalContent: ThreadMessage,
|
finalContent: ThreadMessage,
|
||||||
addMessage: (message: ThreadMessage) => void,
|
addMessage: (message: ThreadMessage) => void,
|
||||||
updateStreamingContent: (content: ThreadMessage | undefined) => void,
|
updateStreamingContent: (content: ThreadMessage | undefined) => void,
|
||||||
updatePromptProgress: (progress: unknown) => void,
|
updatePromptProgress: (progress: PromptProgress | undefined) => void,
|
||||||
updateThreadTimestamp: (threadId: string) => void,
|
updateThreadTimestamp: (threadId: string) => void,
|
||||||
updateMessage?: (message: ThreadMessage) => void,
|
updateMessage?: (message: ThreadMessage) => void,
|
||||||
continueFromMessageId?: string
|
continueFromMessageId?: string
|
||||||
@ -102,7 +102,7 @@ const processStreamingCompletion = async (
|
|||||||
updateStreamingContent: (content: ThreadMessage | undefined) => void,
|
updateStreamingContent: (content: ThreadMessage | undefined) => void,
|
||||||
updateTokenSpeed: (message: ThreadMessage, increment?: number) => void,
|
updateTokenSpeed: (message: ThreadMessage, increment?: number) => void,
|
||||||
setTokenSpeed: (message: ThreadMessage, tokensPerSecond: number, totalTokens: number) => void,
|
setTokenSpeed: (message: ThreadMessage, tokensPerSecond: number, totalTokens: number) => void,
|
||||||
updatePromptProgress: (progress: unknown) => void,
|
updatePromptProgress: (progress: PromptProgress | undefined) => void,
|
||||||
timeToFirstToken: number,
|
timeToFirstToken: number,
|
||||||
tokenUsageRef: { current: CompletionUsage | undefined },
|
tokenUsageRef: { current: CompletionUsage | undefined },
|
||||||
continueFromMessageId?: string,
|
continueFromMessageId?: string,
|
||||||
@ -475,7 +475,7 @@ export const useChat = () => {
|
|||||||
const accumulatedTextRef = {
|
const accumulatedTextRef = {
|
||||||
value: continueFromMessage?.content?.[0]?.text?.value || ''
|
value: continueFromMessage?.content?.[0]?.text?.value || ''
|
||||||
}
|
}
|
||||||
let currentAssistant: Assistant | undefined
|
let currentAssistant: Assistant | undefined | null
|
||||||
|
|
||||||
try {
|
try {
|
||||||
if (selectedModel?.id) {
|
if (selectedModel?.id) {
|
||||||
|
|||||||
@ -23,8 +23,9 @@ export const useTools = () => {
|
|||||||
updateTools(data)
|
updateTools(data)
|
||||||
|
|
||||||
// Initialize default disabled tools for new users (only once)
|
// Initialize default disabled tools for new users (only once)
|
||||||
if (!isDefaultsInitialized() && data.length > 0 && mcpExtension?.getDefaultDisabledTools) {
|
const mcpExt = mcpExtension as MCPExtension & { getDefaultDisabledTools?: () => Promise<string[]> }
|
||||||
const defaultDisabled = await mcpExtension.getDefaultDisabledTools()
|
if (!isDefaultsInitialized() && data.length > 0 && mcpExt?.getDefaultDisabledTools) {
|
||||||
|
const defaultDisabled = await mcpExt.getDefaultDisabledTools()
|
||||||
if (defaultDisabled.length > 0) {
|
if (defaultDisabled.length > 0) {
|
||||||
setDefaultDisabledTools(defaultDisabled)
|
setDefaultDisabledTools(defaultDisabled)
|
||||||
markDefaultsAsInitialized()
|
markDefaultsAsInitialized()
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user