5666 Commits

Author SHA1 Message Date
Louis
83527a7533 fix: Jan hub repo detail and deep link 2025-08-06 12:24:21 +10:00
Faisal Amir
026b21f779 feat: jinja template customize per model instead provider level (#6053) 2025-08-06 12:24:21 +10:00
Akarshan Biswas
dcffa4fa0a Fix: Improve Llama.cpp model path handling and error handling (#6045)
* Improve Llama.cpp model path handling and validation

This commit refactors the load_llama_model function to improve how it handles and validates the model path.

Previously, the function extracted the model path but did not perform any validation. This change adds the following improvements:

It now checks for the presence of the -m flag.

It verifies that a path is provided after the -m flag.

It validates that the specified model path actually exists on the filesystem.

It ensures that the SessionInfo struct stores the canonical display path of the model, which is a more robust approach.

These changes make the model loading process more reliable and provide better error handling for invalid or missing model paths.

* Exp: Use short path on Windows

* Fix: Remove error channel and handling in llama.cpp server loading

The previous implementation used a channel to receive error messages from the llama.cpp server's stdout. However, this proved unreliable as the path names can contain 'errors strings' that we use to check even during normal operation. This commit removes the error channel and associated error handling logic.
The server readiness is still determined by checking for the "server is listening" message in stdout. Errors are now handled by relying on the process exit code and capturing the full stderr output if the process fails to start or exits unexpectedly. This approach provides a more robust and accurate error detection mechanism.

* Add else block in Windows path handling

* Add some path related tests

* Fix windows tests
2025-08-06 12:24:21 +10:00
Faisal Amir
318f6f504f feat: recommended label llamacpp setting (#6052)
* feat: recommended label llamacpp

* chore: remove log
2025-08-06 12:24:21 +10:00
Minh141120
8e4c696583 ci: disable autoqa on nightly build 2025-08-06 12:24:21 +10:00
Louis
7e52512d0e fix: should check for invalid backend to cover previous missing backend case 2025-08-06 12:24:21 +10:00
Louis
eb13189d07 fix: run dev should reinstall extensions 2025-08-06 12:24:21 +10:00
Louis
026383e92d test: add tests for new changes 2025-08-06 12:24:21 +10:00
Louis
4b6269a4f0 fix: add missing cuda backend support 2025-08-06 12:24:21 +10:00
Minh141120
3ffb30b544 chore: skip nightly build workflow for external contributor 2025-08-06 12:24:21 +10:00
Sherzod Mutalov
5f06a35f4e fix: use attributes to check the feature existence 2025-08-06 12:24:21 +10:00
Sherzod Mutalov
280ea1aa9f chore: extracted macos avx2 check code to the utility function 2025-08-06 12:23:18 +10:00
Sherzod Mutalov
ad9c4854a9 chore: added comments 2025-08-06 12:20:30 +10:00
Sherzod Mutalov
49c8334e40 chore: replaced with macros call to remove warning 2025-08-06 12:20:30 +10:00
Sherzod Mutalov
f1dd42de9e fix: use system npx on old mac's 2025-08-06 12:20:30 +10:00
Chaiyapruek Muangsiri
4e31e1d3a8 remove unnecessary try catch block 2025-08-06 12:20:30 +10:00
Chaiyapruek Muangsiri
00f686a733 fix connected servers status not in sync when edit mcp json 2025-08-06 12:20:30 +10:00
Ramon Perez
890a917dec removed nextra component in astro site 2025-08-06 12:20:30 +10:00
Akarshan Biswas
8d147c1774
fix: Add conditional Vulkan support check for better GPU compatibility (#6066)
Changes:
- Introduce conditional Vulkan support check for discrete GPUs with 6GB+ VRAM

fixes: #6009
2025-08-06 07:20:44 +05:30
Louis
c642076ec3
Merge pull request #6024 from menloresearch/fix/jan-hub-repo-data-and-deeplink
fix: Jan hub model detail and deep link
2025-08-06 08:46:07 +07:00
Louis
3b349a60f1 chore: add deep_link register_all 2025-08-05 22:32:27 +07:00
Ramon Perez
4ee6873ca5
Update docs/src/pages/docs/remote-models/huggingface.mdx
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-06 00:24:18 +10:00
Ramon Perez
fc4ecd3412
Update README.md
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-06 00:24:10 +10:00
Faisal Amir
5d001dfd5a
feat: jinja template customize per model instead provider level (#6053) 2025-08-05 21:21:41 +07:00
Ramon Perez
f95c6c4d3d updated readme 2025-08-05 23:11:05 +10:00
Ramon Perez
4c66b1f65b added huggingface page and updated readme 2025-08-05 22:57:49 +10:00
Akarshan Biswas
088b9d7f25
Fix: Improve Llama.cpp model path handling and error handling (#6045)
* Improve Llama.cpp model path handling and validation

This commit refactors the load_llama_model function to improve how it handles and validates the model path.

Previously, the function extracted the model path but did not perform any validation. This change adds the following improvements:

It now checks for the presence of the -m flag.

It verifies that a path is provided after the -m flag.

It validates that the specified model path actually exists on the filesystem.

It ensures that the SessionInfo struct stores the canonical display path of the model, which is a more robust approach.

These changes make the model loading process more reliable and provide better error handling for invalid or missing model paths.

* Exp: Use short path on Windows

* Fix: Remove error channel and handling in llama.cpp server loading

The previous implementation used a channel to receive error messages from the llama.cpp server's stdout. However, this proved unreliable as the path names can contain 'errors strings' that we use to check even during normal operation. This commit removes the error channel and associated error handling logic.
The server readiness is still determined by checking for the "server is listening" message in stdout. Errors are now handled by relying on the process exit code and capturing the full stderr output if the process fails to start or exits unexpectedly. This approach provides a more robust and accurate error detection mechanism.

* Add else block in Windows path handling

* Add some path related tests

* Fix windows tests
2025-08-05 14:17:19 +05:30
Faisal Amir
99567a1102
feat: recommended label llamacpp setting (#6052)
* feat: recommended label llamacpp

* chore: remove log
2025-08-05 13:55:33 +07:00
Louis
065a850a94 fix: test env 2025-08-05 13:44:40 +07:00
Louis
b8070f1871 chore: able to disable updater via env flag 2025-08-05 13:44:40 +07:00
Louis
90e46a2696 test: add tests 2025-08-05 13:44:40 +07:00
Louis
7f0c605651 fix: Jan hub repo detail and deep link 2025-08-05 13:44:40 +07:00
Nguyen Ngoc Minh
339a1957c8
Merge pull request #6051 from menloresearch/ci/disable-autoqa-on-nightly-build
ci: disable autoqa on nightly build
2025-08-05 12:47:31 +07:00
Louis
813c911487
Merge pull request #6046 from menloresearch/fix/support-missing-llamacpp-cuda-backends
fix: support missing llamacpp cuda backends
2025-08-05 12:37:31 +07:00
Minh141120
517e556552 ci: disable autoqa on nightly build 2025-08-05 12:14:16 +07:00
Louis
4a4bc35cce fix: should check for invalid backend to cover previous missing backend case 2025-08-05 11:41:02 +07:00
Nguyen Ngoc Minh
ac66b49eca
Merge pull request #6050 from menloresearch/chore/skip-nightly-build-for-external-contrib
chore: skip nightly build workflow for external contributor
2025-08-05 11:09:17 +07:00
Louis
80707c42e8
Merge pull request #5530 from shmutalov/feat/old-mac-support
feat: old mac support
2025-08-05 11:08:14 +07:00
Louis
48004024ee
Merge pull request #6020 from cmppoon/fix-mcp-servers-edit-json
fix connected servers status not in sync when edit mcp json
2025-08-05 11:06:05 +07:00
Minh141120
e3331dd508 chore: skip nightly build workflow for external contributor 2025-08-05 10:29:33 +07:00
Sherzod Mutalov
0a79cf79a6
Merge branch 'dev' into feat/old-mac-support 2025-08-05 08:17:55 +05:00
Ramon Perez
4c40236441 Merge branch 'rp/jan-docs-v2' into rp/docs-model-providers 2025-08-05 12:19:30 +10:00
Ramon Perez
71987a383a feat: add updated server documentation and MCP examples
- Update local-server docs with comprehensive configuration details
- Add new MCP examples: browserbase, canva, octagon
- Convert Nextra components to Starlight components
- Fix image paths to use correct relative paths
2025-08-05 12:18:42 +10:00
Faisal Amir
641df474fd
fix: Generate A Response button does not show context size error dialog (#6029)
* fix: Generate A Response button does not show context size error dialog

* chore: remove as a child button params
2025-08-05 08:34:06 +07:00
Chaiyapruek Muangsiri
da0cf10f91 remove unnecessary try catch block 2025-08-05 08:08:59 +07:00
Chaiyapruek Muangsiri
477651e5d5 fix connected servers status not in sync when edit mcp json 2025-08-05 08:08:59 +07:00
Louis
72de519430 fix: run dev should reinstall extensions 2025-08-05 00:09:12 +07:00
Louis
29ad6691d8
Merge pull request #6044 from cmppoon/error-toast-on-download-failed
fix: show error toast message on download error
2025-08-04 22:45:17 +07:00
Akarshan Biswas
5e533bdedc
feat: Improve llama.cpp argument handling and add device parsing tests (#6041)
* feat: Improve llama.cpp argument handling and add device parsing tests

This commit refactors how arguments are passed to llama.cpp,
specifically by only adding arguments when their values differ from
their defaults. This reduces the verbosity of the command and prevents
potential conflicts or errors when llama.cpp's default behavior aligns
with the desired setting.

Additionally, new tests have been added for parsing device output from
llama.cpp, ensuring the accurate extraction of GPU information (ID,
name, total memory, and free memory). This improves the robustness of
device detection.

The following changes were made:

* **Remove redundant `--ctx-size` argument:** The `--ctx-size`
    argument is now only explicitly added if `cfg.ctx_size` is greater
    than 0.
* **Conditional argument adding for default values:**
    * `--split-mode` is only added if `cfg.split_mode` is not empty
        and not 'layer'.
    * `--main-gpu` is only added if `cfg.main_gpu` is not undefined
        and not 0.
    * `--cache-type-k` is only added if `cfg.cache_type_k` is not 'f16'.
    * `--cache-type-v` is only added if `cfg.cache_type_v` is not 'f16'
        (when `flash_attn` is enabled) or not 'f32' (otherwise). This
        also corrects the `flash_attn` condition.
    * `--defrag-thold` is only added if `cfg.defrag_thold` is not 0.1.
    * `--rope-scaling` is only added if `cfg.rope_scaling` is not
        'none'.
    * `--rope-scale` is only added if `cfg.rope_scale` is not 1.
    * `--rope-freq-base` is only added if `cfg.rope_freq_base` is not 0.
    * `--rope-freq-scale` is only added if `cfg.rope_freq_scale` is
        not 1.
* **Add `parse_device_output` tests:** Comprehensive unit tests were
    added to `src-tauri/src/core/utils/extensions/inference_llamacpp_extension/server.rs`
    to validate the parsing of llama.cpp device output under various
    scenarios, including multiple devices, single devices, different
    backends (CUDA, Vulkan, SYCL), complex GPU names, and error
    conditions.

* fixup cache_type_v comparision
2025-08-04 19:47:04 +05:30
Chaiyapruek Muangsiri
38c5911460 fix: show error toast on download error 2025-08-04 20:40:17 +08:00