5530 Commits

Author SHA1 Message Date
Louis
3bdd5f00b6 chore: able to disable updater via env flag 2025-08-06 12:24:21 +10:00
Louis
de146f363a test: add tests 2025-08-06 12:24:21 +10:00
Louis
83527a7533 fix: Jan hub repo detail and deep link 2025-08-06 12:24:21 +10:00
Faisal Amir
026b21f779 feat: jinja template customize per model instead provider level (#6053) 2025-08-06 12:24:21 +10:00
Akarshan Biswas
dcffa4fa0a Fix: Improve Llama.cpp model path handling and error handling (#6045)
* Improve Llama.cpp model path handling and validation

This commit refactors the load_llama_model function to improve how it handles and validates the model path.

Previously, the function extracted the model path but did not perform any validation. This change adds the following improvements:

It now checks for the presence of the -m flag.

It verifies that a path is provided after the -m flag.

It validates that the specified model path actually exists on the filesystem.

It ensures that the SessionInfo struct stores the canonical display path of the model, which is a more robust approach.

These changes make the model loading process more reliable and provide better error handling for invalid or missing model paths.

* Exp: Use short path on Windows

* Fix: Remove error channel and handling in llama.cpp server loading

The previous implementation used a channel to receive error messages from the llama.cpp server's stdout. However, this proved unreliable as the path names can contain 'errors strings' that we use to check even during normal operation. This commit removes the error channel and associated error handling logic.
The server readiness is still determined by checking for the "server is listening" message in stdout. Errors are now handled by relying on the process exit code and capturing the full stderr output if the process fails to start or exits unexpectedly. This approach provides a more robust and accurate error detection mechanism.

* Add else block in Windows path handling

* Add some path related tests

* Fix windows tests
2025-08-06 12:24:21 +10:00
Faisal Amir
318f6f504f feat: recommended label llamacpp setting (#6052)
* feat: recommended label llamacpp

* chore: remove log
2025-08-06 12:24:21 +10:00
Minh141120
8e4c696583 ci: disable autoqa on nightly build 2025-08-06 12:24:21 +10:00
Louis
7e52512d0e fix: should check for invalid backend to cover previous missing backend case 2025-08-06 12:24:21 +10:00
Louis
eb13189d07 fix: run dev should reinstall extensions 2025-08-06 12:24:21 +10:00
Louis
026383e92d test: add tests for new changes 2025-08-06 12:24:21 +10:00
Louis
4b6269a4f0 fix: add missing cuda backend support 2025-08-06 12:24:21 +10:00
Minh141120
3ffb30b544 chore: skip nightly build workflow for external contributor 2025-08-06 12:24:21 +10:00
Sherzod Mutalov
5f06a35f4e fix: use attributes to check the feature existence 2025-08-06 12:24:21 +10:00
Sherzod Mutalov
280ea1aa9f chore: extracted macos avx2 check code to the utility function 2025-08-06 12:23:18 +10:00
Sherzod Mutalov
ad9c4854a9 chore: added comments 2025-08-06 12:20:30 +10:00
Sherzod Mutalov
49c8334e40 chore: replaced with macros call to remove warning 2025-08-06 12:20:30 +10:00
Sherzod Mutalov
f1dd42de9e fix: use system npx on old mac's 2025-08-06 12:20:30 +10:00
Chaiyapruek Muangsiri
4e31e1d3a8 remove unnecessary try catch block 2025-08-06 12:20:30 +10:00
Chaiyapruek Muangsiri
00f686a733 fix connected servers status not in sync when edit mcp json 2025-08-06 12:20:30 +10:00
Ramon Perez
890a917dec removed nextra component in astro site 2025-08-06 12:20:30 +10:00
Ramon Perez
4ee6873ca5
Update docs/src/pages/docs/remote-models/huggingface.mdx
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-06 00:24:18 +10:00
Ramon Perez
fc4ecd3412
Update README.md
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-06 00:24:10 +10:00
Ramon Perez
f95c6c4d3d updated readme 2025-08-05 23:11:05 +10:00
Ramon Perez
4c66b1f65b added huggingface page and updated readme 2025-08-05 22:57:49 +10:00
Ramon Perez
4c40236441 Merge branch 'rp/jan-docs-v2' into rp/docs-model-providers 2025-08-05 12:19:30 +10:00
Ramon Perez
71987a383a feat: add updated server documentation and MCP examples
- Update local-server docs with comprehensive configuration details
- Add new MCP examples: browserbase, canva, octagon
- Convert Nextra components to Starlight components
- Fix image paths to use correct relative paths
2025-08-05 12:18:42 +10:00
Faisal Amir
641df474fd
fix: Generate A Response button does not show context size error dialog (#6029)
* fix: Generate A Response button does not show context size error dialog

* chore: remove as a child button params
2025-08-05 08:34:06 +07:00
Louis
29ad6691d8
Merge pull request #6044 from cmppoon/error-toast-on-download-failed
fix: show error toast message on download error
2025-08-04 22:45:17 +07:00
Akarshan Biswas
5e533bdedc
feat: Improve llama.cpp argument handling and add device parsing tests (#6041)
* feat: Improve llama.cpp argument handling and add device parsing tests

This commit refactors how arguments are passed to llama.cpp,
specifically by only adding arguments when their values differ from
their defaults. This reduces the verbosity of the command and prevents
potential conflicts or errors when llama.cpp's default behavior aligns
with the desired setting.

Additionally, new tests have been added for parsing device output from
llama.cpp, ensuring the accurate extraction of GPU information (ID,
name, total memory, and free memory). This improves the robustness of
device detection.

The following changes were made:

* **Remove redundant `--ctx-size` argument:** The `--ctx-size`
    argument is now only explicitly added if `cfg.ctx_size` is greater
    than 0.
* **Conditional argument adding for default values:**
    * `--split-mode` is only added if `cfg.split_mode` is not empty
        and not 'layer'.
    * `--main-gpu` is only added if `cfg.main_gpu` is not undefined
        and not 0.
    * `--cache-type-k` is only added if `cfg.cache_type_k` is not 'f16'.
    * `--cache-type-v` is only added if `cfg.cache_type_v` is not 'f16'
        (when `flash_attn` is enabled) or not 'f32' (otherwise). This
        also corrects the `flash_attn` condition.
    * `--defrag-thold` is only added if `cfg.defrag_thold` is not 0.1.
    * `--rope-scaling` is only added if `cfg.rope_scaling` is not
        'none'.
    * `--rope-scale` is only added if `cfg.rope_scale` is not 1.
    * `--rope-freq-base` is only added if `cfg.rope_freq_base` is not 0.
    * `--rope-freq-scale` is only added if `cfg.rope_freq_scale` is
        not 1.
* **Add `parse_device_output` tests:** Comprehensive unit tests were
    added to `src-tauri/src/core/utils/extensions/inference_llamacpp_extension/server.rs`
    to validate the parsing of llama.cpp device output under various
    scenarios, including multiple devices, single devices, different
    backends (CUDA, Vulkan, SYCL), complex GPU names, and error
    conditions.

* fixup cache_type_v comparision
2025-08-04 19:47:04 +05:30
Chaiyapruek Muangsiri
38c5911460 fix: show error toast on download error 2025-08-04 20:40:17 +08:00
Ramon Perez
a8613e5d15
Merge pull request #5950 from menloresearch/rp/jan-docs-v2 2025-08-04 22:28:55 +10:00
Ramon Perez
3f71428c29 Merge dev branch changes into isolated branch 2025-08-04 10:32:42 +10:00
Akarshan Biswas
b1984a452e
Fix: Llama.cpp server hangs on model load (#6030)
* Fix: Llama.cpp server hangs on model load

Resolves an issue where the llama.cpp server would hang indefinitely when loading certain models, as described in the attached ticket. The server's readiness message was not being correctly detected, causing the application to stall.

The previous implementation used a line-buffered reader (BufReader::lines()) to process the stderr stream. This method proved to be unreliable for the specific output of the llama.cpp server.

This commit refactors the stderr handling logic to use a more robust, chunk-based approach (read_until(b'\n', ...)). This ensures that the output is processed as it arrives, reliably capturing critical status messages and preventing the application from hanging during model initialization.

Fixes: #6021

* Handle error gracefully with ServerError

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

* Revert "Handle error gracefully with ServerError"

This reverts commit 267a8a8a3262fbe36a445a30b8b3ba9a39697643.

* Revert "Fix: Llama.cpp server hangs on model load"

This reverts commit 44e5447f82f0ae32b6db7ffb213025f130d655c4.

* Add more guards, refactor and fix error sending to FE

---------

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-02 21:50:07 +05:30
Faisal Amir
787c4ee073
fix: wrong desc setting cont_batching (#6034) 2025-08-02 21:48:43 +07:00
Faisal Amir
3acb61b5ed
fix: react state loop from hooks useMediaQuery (#6031)
* fix: react state loop from hooks useMediaQuerry

* chore: update test cases hooks media query
2025-08-02 21:48:40 +07:00
Nguyen Ngoc Minh
6603460c39
Merge pull request #6014 from menloresearch/ci/add-trigger-pr-dev
ci: enable PR trigger for dev branch in tauri nightly workflow
2025-07-31 09:14:23 -07:00
Minh141120
b53d5f3638 ci: enable PR trigger for dev branch in tauri nightly workflow 2025-07-31 22:22:38 +07:00
Louis
9c0d09c487
refactor: clean up cortex (#6003)
* refactor: clean up cortex

* chore: clean up

* refactor: clean up
2025-07-31 21:58:12 +07:00
Louis
9573329d06
Merge pull request #6004 from menloresearch/release/v0.6.6
Sync release/v0.6.6 into dev
2025-07-31 21:34:52 +07:00
Louis
4bcfa84d75
Merge pull request #6008 from menloresearch/hotfix/regression-issue-with-colon-in-model-name
hotfix: regression issue with colon in model name
v0.6.6
2025-07-31 17:55:28 +07:00
Ramon Perez
76966135d3
Merge pull request #5958 from menloresearch/rp/nextra-product-section 2025-07-31 20:06:53 +10:00
Louis
8a7edbf3a7
Merge pull request #6005 from menloresearch/fix/save_my_life
Add RunEvent::Exit event to tauri to handle macos context menu exit
2025-07-31 16:05:22 +07:00
Ramon Perez
a93d4f380b updated changelog 2025-07-31 18:52:01 +10:00
Ramon Perez
23ba41fa9d docs: add MCP integration tutorials and update llama.cpp settings for v0.6.6 2025-07-31 18:52:01 +10:00
Ramon Perez
183c028e46 tutorials started 2025-07-31 18:52:01 +10:00
Ramon Perez
aa117cf917 added products section, and reorganized and improved the docs 2025-07-31 18:52:01 +10:00
Ramon Perez
b5d54e48d1 fixed docs workflow issues 2025-07-31 18:52:01 +10:00
Ramon Perez
8511e35df8 feat(docs): Migrate to dual Nextra/Astro deployment & recreate products section
This commit introduces a significant restructuring of the documentation deployment and content strategy to support a gradual migration from Nextra to Astro.

- **New Astro Workflow (`jan-astro-docs.yml`)**: Implemented a new, separate GitHub Actions workflow to build and deploy the Astro site from the `/website` directory to a new subdomain (`v2.jan.ai`). This isolates the new site from the existing one, allowing for independent development and testing.
- **Removed Combined Workflow**: Deleted the previous, more complex combined workflow (`jan-combined-docs.yml`) and its associated test scripts to simplify the deployment process and eliminate routing conflicts.
- **Astro Config Update**: Simplified the Astro configuration (`astro.config.mjs`) by removing the conditional `base` path. The Astro site is now configured to deploy to the root of its own subdomain.

- **Mirrored Content**: Recreated the entire `/products` section from the Astro site within the Nextra site at `/docs/src/pages/products`. This provides content parity and a consistent user experience on both platforms during the transition period.
- **File Structure**: Established a clear, organized structure for platforms, models, and tools within the Nextra `products` directory.
- **Nextra Sidebar Fix**: Implemented the correct `_meta.json` structure for the new products section. Created nested meta files to build a collapsible sidebar, fixing the UI bug that caused duplicated navigation items.

- **"Coming Soon" Pages**: Added clear, concise "Coming Soon" and "In Development" banners and content for upcoming products like Jan V1, Mobile, Server, and native Tools, ensuring consistent messaging across both sites.
- **.gitignore**: Updated the root `.gitignore` to properly exclude build artifacts, caches, and environment files for both the Nextra (`/docs`) and Astro (`/website`) projects.
- **Repository Cleanup**: Removed temporary and unused files related to the previous combined deployment attempt.

This new architecture provides a stable, predictable, and low-risk path for migrating our documentation to Astro while ensuring the current production site remains unaffected.
2025-07-31 18:52:00 +10:00
Ramon Perez
6c189ea961 website layout complete, next, polish wording 2025-07-31 18:52:00 +10:00
Ramon Perez
df5fac06b4 interim product page layout 2025-07-31 18:52:00 +10:00