* fix: turn off experimental settings should also turn off quick ask (#2411) * fix: app glitches 1s generating response before starting model (#2412) * fix: disable experimental feature should also disable vulkan (#2414) * fix: model load stuck on windows when can't get CPU core count (#2413) Signed-off-by: James <james@jan.ai> Co-authored-by: James <james@jan.ai> * feat: TensorRT-LLM engine update support (#2415) * fix: engine update * chore: add remove prepopulated models Signed-off-by: James <james@jan.ai> * update tinyjensen url Signed-off-by: James <james@jan.ai> * update llamacorn Signed-off-by: James <james@jan.ai> * update Mistral 7B Instruct v0.1 int4 Signed-off-by: James <james@jan.ai> * update tensorrt Signed-off-by: James <james@jan.ai> * update Signed-off-by: hiro <hiro@jan.ai> * update Signed-off-by: James <james@jan.ai> * prettier Signed-off-by: James <james@jan.ai> * update mistral config Signed-off-by: James <james@jan.ai> * fix some lint Signed-off-by: James <james@jan.ai> --------- Signed-off-by: James <james@jan.ai> Signed-off-by: hiro <hiro@jan.ai> Co-authored-by: James <james@jan.ai> Co-authored-by: hiro <hiro@jan.ai> * Tensorrt LLM disable turing support (#2418) Co-authored-by: Hien To <tominhhien97@gmail.com> * chore: add prompt template tensorrtllm (#2375) * chore: add prompt template tensorrtllm * Add Prompt template for mistral and correct model metadata --------- Co-authored-by: Hien To <tominhhien97@gmail.com> * fix: correct tensorrt mistral model.json (#2419) --------- Signed-off-by: James <james@jan.ai> Signed-off-by: hiro <hiro@jan.ai> Co-authored-by: Louis <louis@jan.ai> Co-authored-by: James <james@jan.ai> Co-authored-by: hiro <hiro@jan.ai> Co-authored-by: hiento09 <136591877+hiento09@users.noreply.github.com> Co-authored-by: Hien To <tominhhien97@gmail.com>
78 lines
2.9 KiB
TypeScript
78 lines
2.9 KiB
TypeScript
import resolve from '@rollup/plugin-node-resolve'
|
|
import commonjs from '@rollup/plugin-commonjs'
|
|
import sourceMaps from 'rollup-plugin-sourcemaps'
|
|
import typescript from 'rollup-plugin-typescript2'
|
|
import json from '@rollup/plugin-json'
|
|
import replace from '@rollup/plugin-replace'
|
|
const packageJson = require('./package.json')
|
|
|
|
export default [
|
|
{
|
|
input: `src/index.ts`,
|
|
output: [{ file: packageJson.main, format: 'es', sourcemap: true }],
|
|
watch: {
|
|
include: 'src/**',
|
|
},
|
|
plugins: [
|
|
replace({
|
|
EXTENSION_NAME: JSON.stringify(packageJson.name),
|
|
TENSORRT_VERSION: JSON.stringify(packageJson.tensorrtVersion),
|
|
PROVIDER: JSON.stringify(packageJson.provider),
|
|
DOWNLOAD_RUNNER_URL:
|
|
process.platform === 'win32'
|
|
? JSON.stringify(
|
|
'https://github.com/janhq/nitro-tensorrt-llm/releases/download/windows-v<version>-tensorrt-llm-v0.7.1/nitro-windows-v<version>-tensorrt-llm-v0.7.1-amd64-all-arch.tar.gz'
|
|
)
|
|
: JSON.stringify(
|
|
'https://github.com/janhq/nitro-tensorrt-llm/releases/download/linux-v<version>/nitro-linux-v<version>-amd64-tensorrt-llm-<gpuarch>.tar.gz'
|
|
),
|
|
NODE: JSON.stringify(`${packageJson.name}/${packageJson.node}`),
|
|
INFERENCE_URL: JSON.stringify(
|
|
process.env.INFERENCE_URL ||
|
|
`${packageJson.config?.protocol ?? 'http'}://${packageJson.config?.host}:${packageJson.config?.port}/v1/chat/completions`
|
|
),
|
|
COMPATIBILITY: JSON.stringify(packageJson.compatibility),
|
|
}),
|
|
json(),
|
|
typescript({ useTsconfigDeclarationDir: true }),
|
|
commonjs(),
|
|
resolve({
|
|
extensions: ['.js', '.ts', '.svelte'],
|
|
}),
|
|
sourceMaps(),
|
|
],
|
|
},
|
|
{
|
|
input: `src/node/index.ts`,
|
|
output: [
|
|
{ file: 'dist/node/index.cjs.js', format: 'cjs', sourcemap: true },
|
|
],
|
|
external: ['@janhq/core/node'],
|
|
watch: {
|
|
include: 'src/node/**',
|
|
},
|
|
plugins: [
|
|
replace({
|
|
EXTENSION_NAME: JSON.stringify(packageJson.name),
|
|
TENSORRT_VERSION: JSON.stringify(packageJson.tensorrtVersion),
|
|
PROVIDER: JSON.stringify(packageJson.provider),
|
|
LOAD_MODEL_URL: JSON.stringify(
|
|
`${packageJson.config?.protocol ?? 'http'}://${packageJson.config?.host}:${packageJson.config?.port}/inferences/tensorrtllm/loadmodel`
|
|
),
|
|
TERMINATE_ENGINE_URL: JSON.stringify(
|
|
`${packageJson.config?.protocol ?? 'http'}://${packageJson.config?.host}:${packageJson.config?.port}/processmanager/destroy`
|
|
),
|
|
ENGINE_HOST: JSON.stringify(packageJson.config?.host ?? '127.0.0.1'),
|
|
ENGINE_PORT: JSON.stringify(packageJson.config?.port ?? '3928'),
|
|
}),
|
|
json(),
|
|
typescript({ useTsconfigDeclarationDir: true }),
|
|
commonjs(),
|
|
resolve({
|
|
extensions: ['.ts', '.js', '.json'],
|
|
}),
|
|
sourceMaps(),
|
|
],
|
|
},
|
|
]
|