202 Commits

Author SHA1 Message Date
NamH
20657bb3da
fix: cannot cancel model which is progress at 0 percent (#2727)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-04-15 21:52:22 +07:00
Louis
9369ac3e8b
Merge branch 'dev' into main 2024-04-15 14:57:31 +07:00
Andreas Deininger
81e8889568
Fix typos (#2714) 2024-04-15 13:27:28 +07:00
Faisal Amir
1619478250
enhancement: update input component dynamic text alignment (#2712) 2024-04-15 09:48:57 +07:00
Louis
fa9d8ab9a5
fix: switch between models get stuck at generating (#2698) 2024-04-12 12:34:22 +07:00
Louis
065ed03099
fix: wrong monitoring system information type (#2679) 2024-04-11 11:07:31 +07:00
NamH
ddb73d8131
fix: can't read the setting at first time (#2677)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-04-11 10:56:47 +07:00
Louis
b19234ed71
chore: Extension should have product name in manifest (#2675)
* chore: Extension should have product name in manifest

* chore: typo
2024-04-11 09:50:58 +07:00
Louis
d93d74c86b
feat: nitro additional dependencies (#2674) 2024-04-11 09:13:02 +07:00
NamH
8917be5ef3
fix: add fallback as default endpoint for inference engine (#2669)
Co-authored-by: James <james@jan.ai>
2024-04-10 18:15:20 +07:00
NamH
69f73b86c4
fix: api server chat completion error for remote model (#2671)
* fix: api server chat completion error for remote model

Signed-off-by: James <james@jan.ai>

* fix: duplicate setting in local api server

Signed-off-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-04-10 18:13:32 +07:00
Inchoker
00049aac7d
fix error handling of core for mistral integration (#2672)
* fix error handling of core for mistral integration

* refactor logic using status code

* nitpicking fix

---------

Co-authored-by: Jack Tri Le <Jack>
2024-04-10 17:03:26 +07:00
Louis
3f23de6c28
feat: move log into monitoring extension (#2662) 2024-04-10 14:35:15 +07:00
hiento09
5fd6025175
Bump nitro version to 0.3.18 (#2652) 2024-04-09 12:31:10 +07:00
Louis
c31a6ef4d3
fix: missing api key error handling (#2645) 2024-04-09 11:33:12 +07:00
NamH
c0c038f4f7
chore: filter gguf files when importing model on windows (#2653)
Co-authored-by: Louis <louis@jan.ai>
2024-04-09 10:36:01 +07:00
Louis
9303e7aeb4
fix: object is destroyed block app exit (#2623) 2024-04-05 14:30:43 +07:00
NamH
e0d6049d66
chore: extension should register its own models (#2601)
* chore: extension should register its own models

Signed-off-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-04-05 14:18:58 +07:00
GenkaOk
089e311bfd
Add API prefix settings (#2528)
Co-authored-by: Van Pham <64197333+Van-QA@users.noreply.github.com>
2024-04-04 19:45:45 +07:00
Louis
a484315699
fix: wrong response header for non streamed responses (#2606) 2024-04-04 06:37:06 +07:00
Louis
fe89901a65
fix: groq frequently stops during long responses (#2584) 2024-04-02 22:19:34 +07:00
Louis
3fded8fd20
fix: regenerate message should not include the assistant's previous response (#2583) 2024-04-02 22:17:35 +07:00
Louis
f6d3b53ab5
Merge branch 'main' into dev
# Conflicts:
#	web/screens/Chat/ErrorMessage/index.tsx
2024-04-02 11:09:59 +07:00
NamH
fa35aa6e14
feat: dynamically register extension settings (#2494)
* feat: add extesion settings

Signed-off-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: Louis <louis@jan.ai>
2024-03-29 15:44:46 +07:00
NamH
132487acec
fix: show default error message from inference engine (#2533)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-03-29 11:53:34 +07:00
NamH
3ecdb81881
fix: file explore on windows show empty when importing model (#2484)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-03-29 01:24:53 +07:00
Louis
3b3eb119f0
fix: duplicate api definition (#2522) 2024-03-28 11:45:08 +07:00
Louis
75eea1fdb2
Merge branch 'dev'
# Conflicts:
#	core/src/browser/core.ts
#	core/src/browser/extensions/monitoring.ts
#	core/src/browser/fs.ts
#	core/src/extensions/ai-engines/LocalOAIEngine.ts
#	extensions/monitoring-extension/src/node/index.ts
#	extensions/tensorrt-llm-extension/src/index.ts
#	extensions/tensorrt-llm-extension/src/node/index.ts
#	web/hooks/useSendChatMessage.ts
2024-03-28 10:46:05 +07:00
Louis
8e8dfd4b37
refactor: introduce inference tools (#2493) 2024-03-25 23:26:05 +07:00
Louis
14a67463dc
chore: refactor core folder structure - module based 2024-03-25 16:20:06 +07:00
Louis
9551996e34
chore: load, unload model and inference synchronously 2024-03-25 12:25:30 +07:00
NamH
67e285fa96
chore: remove rmdirsync from core api since it is deprecated (#2459)
* chore: remove rmdirsync from core api since it is deprecated

Signed-off-by: James <james@jan.ai>

* chore: remove mkdirsync

Signed-off-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-03-22 17:57:16 +07:00
NamH
8303f74f58
fix: only allow 4 stop words for openai (#2457)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-03-22 13:46:12 +07:00
NamH
6dd26f6a4b
fix: post to /v1/threads not found (#2455)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-03-22 12:41:51 +07:00
Louis
acbec78dbf
fix: refactor inference engines to extends AIEngine (#2347)
* fix: refactor nitro to extends localoaiengine

* fix: refactor openai extension

* chore: refactor groq extension

* chore: refactor triton tensorrt extension

* chore: add tests

* chore: refactor engines
2024-03-22 09:35:14 +07:00
Louis
ff7ec39915
fix: incompatible browser dependency (#2439)
* fix: incompatible browser dependency

* fix: update model extension to use rollup

* fix: test timeout
2024-03-21 16:54:42 +07:00
NamH
b8d86df688
Fix/unable factory reset windows nitro running (#2422)
* fix: unable to factory reset when nitro is running on windows

---------

Signed-off-by: James <james@jan.ai>
2024-03-19 18:05:03 +07:00
Louis
489e8aab24
Sync release 0.4.9 to dev (#2407)
* fix: move tensorrt executable to engine (#2400)

* fix: move tensorrt executable to engine

Signed-off-by: James <james@jan.ai>

* some update

Signed-off-by: hiro <hiro@jan.ai>

* chore: bump tensorrt version

* fix: wrong destroy path

* fix: install extensions in parallel

* chore: update path for tensorrt engine (#2404)

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Signed-off-by: hiro <hiro@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: hiro <hiro@jan.ai>
Co-authored-by: Louis <louis@jan.ai>

* Release/v0.4.9 (#2421)

* fix: turn off experimental settings should also turn off quick ask (#2411)

* fix: app glitches 1s generating response before starting model (#2412)

* fix: disable experimental feature should also disable vulkan (#2414)

* fix: model load stuck on windows when can't get CPU core count (#2413)

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>

* feat: TensorRT-LLM engine update support (#2415)

* fix: engine update

* chore: add remove prepopulated models

Signed-off-by: James <james@jan.ai>

* update tinyjensen url

Signed-off-by: James <james@jan.ai>

* update llamacorn

Signed-off-by: James <james@jan.ai>

* update Mistral 7B Instruct v0.1 int4

Signed-off-by: James <james@jan.ai>

* update tensorrt

Signed-off-by: James <james@jan.ai>

* update

Signed-off-by: hiro <hiro@jan.ai>

* update

Signed-off-by: James <james@jan.ai>

* prettier

Signed-off-by: James <james@jan.ai>

* update mistral config

Signed-off-by: James <james@jan.ai>

* fix some lint

Signed-off-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Signed-off-by: hiro <hiro@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: hiro <hiro@jan.ai>

* Tensorrt LLM disable turing support (#2418)

Co-authored-by: Hien To <tominhhien97@gmail.com>

* chore: add prompt template tensorrtllm (#2375)

* chore: add prompt template tensorrtllm

* Add Prompt template for mistral and correct model metadata

---------

Co-authored-by: Hien To <tominhhien97@gmail.com>

* fix: correct tensorrt mistral model.json (#2419)

---------

Signed-off-by: James <james@jan.ai>
Signed-off-by: hiro <hiro@jan.ai>
Co-authored-by: Louis <louis@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: hiro <hiro@jan.ai>
Co-authored-by: hiento09 <136591877+hiento09@users.noreply.github.com>
Co-authored-by: Hien To <tominhhien97@gmail.com>

---------

Signed-off-by: James <james@jan.ai>
Signed-off-by: hiro <hiro@jan.ai>
Co-authored-by: NamH <NamNh0122@gmail.com>
Co-authored-by: James <james@jan.ai>
Co-authored-by: hiro <hiro@jan.ai>
Co-authored-by: hiento09 <136591877+hiento09@users.noreply.github.com>
Co-authored-by: Hien To <tominhhien97@gmail.com>
2024-03-19 12:20:09 +07:00
NamH
3a3bceb0c0
Release/v0.4.9 (#2421)
* fix: turn off experimental settings should also turn off quick ask (#2411)

* fix: app glitches 1s generating response before starting model (#2412)

* fix: disable experimental feature should also disable vulkan (#2414)

* fix: model load stuck on windows when can't get CPU core count (#2413)

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>

* feat: TensorRT-LLM engine update support (#2415)

* fix: engine update

* chore: add remove prepopulated models

Signed-off-by: James <james@jan.ai>

* update tinyjensen url

Signed-off-by: James <james@jan.ai>

* update llamacorn

Signed-off-by: James <james@jan.ai>

* update Mistral 7B Instruct v0.1 int4

Signed-off-by: James <james@jan.ai>

* update tensorrt

Signed-off-by: James <james@jan.ai>

* update

Signed-off-by: hiro <hiro@jan.ai>

* update

Signed-off-by: James <james@jan.ai>

* prettier

Signed-off-by: James <james@jan.ai>

* update mistral config

Signed-off-by: James <james@jan.ai>

* fix some lint

Signed-off-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Signed-off-by: hiro <hiro@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: hiro <hiro@jan.ai>

* Tensorrt LLM disable turing support (#2418)

Co-authored-by: Hien To <tominhhien97@gmail.com>

* chore: add prompt template tensorrtllm (#2375)

* chore: add prompt template tensorrtllm

* Add Prompt template for mistral and correct model metadata

---------

Co-authored-by: Hien To <tominhhien97@gmail.com>

* fix: correct tensorrt mistral model.json (#2419)

---------

Signed-off-by: James <james@jan.ai>
Signed-off-by: hiro <hiro@jan.ai>
Co-authored-by: Louis <louis@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: hiro <hiro@jan.ai>
Co-authored-by: hiento09 <136591877+hiento09@users.noreply.github.com>
Co-authored-by: Hien To <tominhhien97@gmail.com>
2024-03-19 10:06:47 +07:00
NamH
c81a33f382
fix: move tensorrt executable to engine (#2400)
* fix: move tensorrt executable to engine

Signed-off-by: James <james@jan.ai>

* some update

Signed-off-by: hiro <hiro@jan.ai>

* chore: bump tensorrt version

* fix: wrong destroy path

* fix: install extensions in parallel

* chore: update path for tensorrt engine (#2404)

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Signed-off-by: hiro <hiro@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: hiro <hiro@jan.ai>
Co-authored-by: Louis <louis@jan.ai>
2024-03-18 07:38:35 +07:00
Meta Spartan
0348aa3321
feat: Groq Inference Extension (#2263)
* feat: Groq Inference Extension

* Add Groq supported models

* Fix folder typo

* Add Groq options to interface and new API Key saving, tested working

* Fix linting
2024-03-18 06:40:20 +07:00
Louis
d85d02693b
feat: Nitro-Tensorrt-LLM Extension (#2280)
* feat: tensorrt-llm-extension

* fix: loading

* feat: add download tensorrt llm runner

Signed-off-by: James <james@jan.ai>

* feat: update to rollupjs instead of webpack for monitoring extension

Signed-off-by: James <james@jan.ai>

* feat: move update nvidia info to monitor extension

Signed-off-by: James <james@jan.ai>

* allow download tensorrt

Signed-off-by: James <james@jan.ai>

* update

Signed-off-by: James <james@jan.ai>

* allow download tensor rt based on gpu setting

Signed-off-by: James <james@jan.ai>

* update downloaded models

Signed-off-by: James <james@jan.ai>

* feat: add extension compatibility

* dynamic tensor rt engines

Signed-off-by: James <james@jan.ai>

* update models

Signed-off-by: James <james@jan.ai>

* chore: remove ts-ignore

* feat: getting installation state from extension

Signed-off-by: James <james@jan.ai>

* chore: adding type for decompress

Signed-off-by: James <james@jan.ai>

* feat: update according Louis's comment

Signed-off-by: James <james@jan.ai>

* feat: add progress for installing extension

Signed-off-by: James <james@jan.ai>

* chore: remove args from extension installation

* fix: model download does not work properly

* fix: do not allow user to stop tensorrtllm inference

* fix: extension installed style

* fix: download tensorrt does not update state

Signed-off-by: James <james@jan.ai>

* chore: replace int4 by fl16

* feat: modal for installing extension

Signed-off-by: James <james@jan.ai>

* fix: start download immediately after press install

Signed-off-by: James <james@jan.ai>

* fix: error switching between engines

* feat: rename inference provider to ai engine and refactor to core

* fix: missing ulid

* fix: core bundler

* feat: add cancel extension installing

Signed-off-by: James <james@jan.ai>

* remove mocking for mac

Signed-off-by: James <james@jan.ai>

* fix: show models only when extension is ready

* add tensorrt badge for model

Signed-off-by: James <james@jan.ai>

* fix: copy

* fix: add compatible check (#2342)

* fix: add compatible check

Signed-off-by: James <james@jan.ai>

* fix: copy

* fix: font

* fix: copy

* fix: broken monitoring extension

* chore: bump engine

* fix: copy

* fix: model copy

* fix: copy

* fix: model json

---------

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: Louis <louis@jan.ai>

* fix: vulkan support

* fix: installation button padding

* fix: empty script

* fix: remove hard code string

---------

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: NamH <NamNh0122@gmail.com>
2024-03-14 14:07:22 +07:00
Louis
af5bcea773
fix: gate quick ask with feature toggle (#2331) 2024-03-12 20:10:59 +07:00
NamH
f36d740b1e
feat: add quick ask (#2197)
* feat: add quick ask

Signed-off-by: James <james@jan.ai>

---------

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: Louis <louis@jan.ai>
2024-03-08 10:01:37 +07:00
NamH
510491253c
fix: error message being sent along with conversation when inference (#2242)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-03-05 22:09:38 +07:00
hiro
e6c10202e0
feat: Jan can see (#2069)
* feat: jan can see

feat: Add GPT-4 Vision model (Preview)

fix: Add visionModel as property in ModelInfo

fix: Fix condition to load local messages in useSetActiveThread hook

feat: Enable Image as input for chat

fix: Update model parameters in JSON files for remote GPT models

fix: Add thread as optional

fix: Add support for message as image

fix: Linter

fix: Update proxyModel to proxy_model and add textModel

chore: Change proxyModel to proxy_model

fix: Update settings with visionModel and textModel

fix: vision model passed through the retrieval tool

fix: linter

* fix: could not load image and request is not able to be sent

---------

Co-authored-by: Louis <louis@jan.ai>
2024-03-05 08:33:09 +07:00
NamH
5c185d2740
fix: download model error does not reset state in model hub (#2199)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-02-29 14:35:05 +07:00
NamH
d7070d8c4a
fix: some bugs for import model (#2181)
* fix: some bugs for import model

Signed-off-by: James <james@jan.ai>

Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-02-27 23:59:37 +07:00
NamH
773963a456
feat: add import model (#2104)
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
2024-02-26 16:15:10 +07:00
Helloyunho
e86cd7e661
feat: add a simple way to convert Hugging Face model to GGUF (#1972)
* chore: add react developer tools to electron

* feat: add small convert modal

* feat: separate modals and add hugging face extension

* feat: fully implement hugging face converter

* fix: forgot to uncomment this...

* fix: typo

* feat: try hf-to-gguf script first and then use convert.py

HF-to-GGUF has support for some unusual models
maybe using convert.py first would be better but we can change the usage order later

* fix: pre-install directory changed

* fix: sometimes exit code is undefined

* chore: download additional files for qwen

* fix: event handling changed

* chore: add one more necessary package

* feat: download gguf-py from llama.cpp

* fix: cannot interpret wildcards on GNU tar

Co-authored-by: hiento09 <136591877+hiento09@users.noreply.github.com>

---------

Co-authored-by: hiento09 <136591877+hiento09@users.noreply.github.com>
2024-02-26 10:57:53 +07:00