fix: 4238 - fix default max_tokens set on remote models

This commit is contained in:
Louis 2024-12-12 13:20:39 +07:00
parent 09bfc0549e
commit aac2216aef
No known key found for this signature in database
GPG Key ID: 44FA9F4D33C37DE2
4 changed files with 16 additions and 7 deletions

View File

@ -99,10 +99,10 @@
"format": "api", "format": "api",
"settings": {}, "settings": {},
"parameters": { "parameters": {
"max_tokens": 32768,
"temperature": 1, "temperature": 1,
"top_p": 1, "top_p": 1,
"stream": true, "stream": true,
"max_tokens": 32768,
"frequency_penalty": 0, "frequency_penalty": 0,
"presence_penalty": 0 "presence_penalty": 0
}, },
@ -126,9 +126,9 @@
"format": "api", "format": "api",
"settings": {}, "settings": {},
"parameters": { "parameters": {
"max_tokens": 65536,
"temperature": 1, "temperature": 1,
"top_p": 1, "top_p": 1,
"max_tokens": 65536,
"stream": true, "stream": true,
"frequency_penalty": 0, "frequency_penalty": 0,
"presence_penalty": 0 "presence_penalty": 0

View File

@ -192,8 +192,12 @@ const ModelDropdown = ({
model?.settings.ctx_len ?? 8192 model?.settings.ctx_len ?? 8192
) )
const overriddenParameters = { const overriddenParameters = {
ctx_len: Math.min(8192, model?.settings.ctx_len ?? 8192), ctx_len: !isLocalEngine(model?.engine)
max_tokens: defaultContextLength, ? undefined
: defaultContextLength,
max_tokens: !isLocalEngine(model?.engine)
? (model?.parameters.max_tokens ?? 8192)
: defaultContextLength,
} }
const modelParams = { const modelParams = {

View File

@ -17,6 +17,7 @@ import { fileUploadAtom } from '@/containers/Providers/Jotai'
import { toaster } from '@/containers/Toast' import { toaster } from '@/containers/Toast'
import { isLocalEngine } from '@/utils/modelEngine'
import { generateThreadId } from '@/utils/thread' import { generateThreadId } from '@/utils/thread'
import { useActiveModel } from './useActiveModel' import { useActiveModel } from './useActiveModel'
@ -113,12 +114,14 @@ export const useCreateNewThread = () => {
) )
const overriddenSettings = { const overriddenSettings = {
ctx_len: defaultContextLength, ctx_len: !isLocalEngine(model?.engine) ? undefined : defaultContextLength,
} }
// Use ctx length by default // Use ctx length by default
const overriddenParameters = { const overriddenParameters = {
max_tokens: defaultContextLength, max_tokens: !isLocalEngine(model?.engine)
? (model?.parameters.token_limit ?? 8192)
: defaultContextLength,
} }
const createdAt = Date.now() const createdAt = Date.now()

View File

@ -38,7 +38,9 @@ export const getLogoEngine = (engine: InferenceEngine) => {
* @param engine * @param engine
* @returns * @returns
*/ */
export const isLocalEngine = (engine: string) => { export const isLocalEngine = (engine?: string) => {
if (!engine) return false
const engineObj = EngineManager.instance().get(engine) const engineObj = EngineManager.instance().get(engine)
if (!engineObj) return false if (!engineObj) return false
return ( return (