fix: wrong engine handling (#2363)
This commit is contained in:
parent
e40d0481b7
commit
58e12f35c9
@ -129,6 +129,8 @@ export default class TensorRTLLMExtension extends LocalOAIEngine {
|
|||||||
}
|
}
|
||||||
|
|
||||||
async onModelInit(model: Model): Promise<void> {
|
async onModelInit(model: Model): Promise<void> {
|
||||||
|
if (model.engine !== this.provider) return
|
||||||
|
|
||||||
if ((await this.installationState()) === 'Installed')
|
if ((await this.installationState()) === 'Installed')
|
||||||
return super.onModelInit(model)
|
return super.onModelInit(model)
|
||||||
else {
|
else {
|
||||||
@ -138,7 +140,6 @@ export default class TensorRTLLMExtension extends LocalOAIEngine {
|
|||||||
message: 'EXTENSION_IS_NOT_INSTALLED::TensorRT-LLM extension',
|
message: 'EXTENSION_IS_NOT_INSTALLED::TensorRT-LLM extension',
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
return
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -162,6 +163,7 @@ export default class TensorRTLLMExtension extends LocalOAIEngine {
|
|||||||
}
|
}
|
||||||
|
|
||||||
inference(data: MessageRequest): void {
|
inference(data: MessageRequest): void {
|
||||||
|
if (!this.isRunning) return
|
||||||
// TensorRT LLM Extension supports streaming only
|
// TensorRT LLM Extension supports streaming only
|
||||||
if (data.model) data.model.parameters.stream = true
|
if (data.model) data.model.parameters.stream = true
|
||||||
super.inference(data)
|
super.inference(data)
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user