* feat: add extesion settings
Signed-off-by: James <james@jan.ai>
---------
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: Louis <louis@jan.ai>
* feat: tensorrt-llm-extension
* fix: loading
* feat: add download tensorrt llm runner
Signed-off-by: James <james@jan.ai>
* feat: update to rollupjs instead of webpack for monitoring extension
Signed-off-by: James <james@jan.ai>
* feat: move update nvidia info to monitor extension
Signed-off-by: James <james@jan.ai>
* allow download tensorrt
Signed-off-by: James <james@jan.ai>
* update
Signed-off-by: James <james@jan.ai>
* allow download tensor rt based on gpu setting
Signed-off-by: James <james@jan.ai>
* update downloaded models
Signed-off-by: James <james@jan.ai>
* feat: add extension compatibility
* dynamic tensor rt engines
Signed-off-by: James <james@jan.ai>
* update models
Signed-off-by: James <james@jan.ai>
* chore: remove ts-ignore
* feat: getting installation state from extension
Signed-off-by: James <james@jan.ai>
* chore: adding type for decompress
Signed-off-by: James <james@jan.ai>
* feat: update according Louis's comment
Signed-off-by: James <james@jan.ai>
* feat: add progress for installing extension
Signed-off-by: James <james@jan.ai>
* chore: remove args from extension installation
* fix: model download does not work properly
* fix: do not allow user to stop tensorrtllm inference
* fix: extension installed style
* fix: download tensorrt does not update state
Signed-off-by: James <james@jan.ai>
* chore: replace int4 by fl16
* feat: modal for installing extension
Signed-off-by: James <james@jan.ai>
* fix: start download immediately after press install
Signed-off-by: James <james@jan.ai>
* fix: error switching between engines
* feat: rename inference provider to ai engine and refactor to core
* fix: missing ulid
* fix: core bundler
* feat: add cancel extension installing
Signed-off-by: James <james@jan.ai>
* remove mocking for mac
Signed-off-by: James <james@jan.ai>
* fix: show models only when extension is ready
* add tensorrt badge for model
Signed-off-by: James <james@jan.ai>
* fix: copy
* fix: add compatible check (#2342)
* fix: add compatible check
Signed-off-by: James <james@jan.ai>
* fix: copy
* fix: font
* fix: copy
* fix: broken monitoring extension
* chore: bump engine
* fix: copy
* fix: model copy
* fix: copy
* fix: model json
---------
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: Louis <louis@jan.ai>
* fix: vulkan support
* fix: installation button padding
* fix: empty script
* fix: remove hard code string
---------
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: NamH <NamNh0122@gmail.com>
* feat: add quick ask
Signed-off-by: James <james@jan.ai>
---------
Signed-off-by: James <james@jan.ai>
Co-authored-by: James <james@jan.ai>
Co-authored-by: Louis <louis@jan.ai>
* feat: jan can see
feat: Add GPT-4 Vision model (Preview)
fix: Add visionModel as property in ModelInfo
fix: Fix condition to load local messages in useSetActiveThread hook
feat: Enable Image as input for chat
fix: Update model parameters in JSON files for remote GPT models
fix: Add thread as optional
fix: Add support for message as image
fix: Linter
fix: Update proxyModel to proxy_model and add textModel
chore: Change proxyModel to proxy_model
fix: Update settings with visionModel and textModel
fix: vision model passed through the retrieval tool
fix: linter
* fix: could not load image and request is not able to be sent
---------
Co-authored-by: Louis <louis@jan.ai>
* Revert "feat: temporary remove dark mode :( (#2168)"
This reverts commit 222b4ad897c275dab0eaec3c8a8472bf3df7afc4.
* fix: revert darkmode and fix darkmode for import model
* fix: prettier format import model
---------
Co-authored-by: Louis <louis@jan.ai>
* chore: add react developer tools to electron
* feat: add small convert modal
* feat: separate modals and add hugging face extension
* feat: fully implement hugging face converter
* fix: forgot to uncomment this...
* fix: typo
* feat: try hf-to-gguf script first and then use convert.py
HF-to-GGUF has support for some unusual models
maybe using convert.py first would be better but we can change the usage order later
* fix: pre-install directory changed
* fix: sometimes exit code is undefined
* chore: download additional files for qwen
* fix: event handling changed
* chore: add one more necessary package
* feat: download gguf-py from llama.cpp
* fix: cannot interpret wildcards on GNU tar
Co-authored-by: hiento09 <136591877+hiento09@users.noreply.github.com>
---------
Co-authored-by: hiento09 <136591877+hiento09@users.noreply.github.com>
* feat: add vulkan support on windows and linux
* fix: correct vulkan settings
* fix: gpu settings and enable Vulkan support
* fix: vulkan support 1 device at a time only
* inference-nitro-extension add download vulkaninfo
---------
Co-authored-by: Louis <louis@jan.ai>
Co-authored-by: Hien To <tominhhien97@gmail.com>