5371 Commits

Author SHA1 Message Date
Nguyen Ngoc Minh
af892428a5
chore: sync make build with dev (#5847)
* chore: sync up make build with dev

* ci: update macOS self-hosted runner
2025-07-22 11:12:14 +07:00
Nguyen Ngoc Minh
9ea081576b
fix: custom tauri nsis template CheckIfAppIsRunning macro (#5840)
* fix: update CheckIfAppIsRunning macro to include args
2025-07-21 20:54:06 +07:00
Nguyen Ngoc Minh
275cab7538
Merge pull request #5839 from menloresearch/fix/appimage-url-with-latest-tauri-cli
fix: update @taur-apps/cli to newest verison to fix appimage download
2025-07-21 03:41:27 -07:00
Minh141120
db962b2ba6 fix: update @taur-apps/cli to newest verison to fix appimage download issue 2025-07-21 16:32:27 +07:00
Akarshan Biswas
08de0fa42d
fix: prevent terminal window from opening on model load on WindowsOS (#5837)
On Windows, spawning the llamacpp server was causing an unwanted terminal window
to appear. This is now fixed by combining `CREATE_NO_WINDOW` with
`CREATE_NEW_PROCESS_GROUP` using `.creation_flags(...)`, ensuring that the
process runs in the background without a console window.

This change only applies to 64-bit Windows builds.
2025-07-21 13:24:31 +05:30
Louis
05b9d4e9fd
feat: add claude-4 (#5829)
* feat: add claude-4

* fix: sorting order
2025-07-21 12:30:56 +07:00
Akarshan Biswas
81d6ed3785
feat: support per-model overrides in llama.cpp load() (#5820)
* feat: support per-model overrides in llama.cpp load()

Extend the `load()` method in the llama.cpp extension to accept optional
`overrideSettings`, allowing fine-grained per-model configuration.

This enables users to override provider-level settings such as `ctx_size`,
`chat_template`, `n_gpu_layers`, etc., when loading a specific model.

Fixes: #5818 (Feature Request - Jan v0.6.6)

Use cases enabled:
- Different context sizes per model (e.g., 4K vs 32K)
- Model-specific chat templates (ChatML, Alpaca, etc.)
- Performance tuning (threads, GPU layers)
- Better memory management per deployment

Maintains full backward compatibility with existing provider config.

* swap overrideSettings and isEmbedding argument
2025-07-21 08:59:50 +05:30
Louis
bc4fe52f8d
fix: llama.cpp integration model load and chat experience (#5823)
* fix: stop generating should not stop running models

* fix: ensure backend ready before loading model

* fix: backend setting should not block onLoad
2025-07-21 09:29:26 +07:00
Louis
5696e951f2
fix: Legacy threads show on top of new threads (#5696) (#5810)
* fix: #5696 - legacy threads show on top of new threads

* fix: tests
2025-07-20 16:58:22 +07:00
Louis
19cb1c96e0
fix: llama.cpp backend download on windows (#5813)
* fix: llama.cpp backend download on windows

* test: add missing cases

* clean: linter

* fix: build
2025-07-20 16:58:09 +07:00
hiento09
b7b3eb9d19
fix: autoqa requirements.txt (#5812) 2025-07-19 22:47:34 +07:00
Louis
c550f6cf0d
Merge pull request #5809 from menloresearch/refactor/simplify-proxy-settings
refactor: simplify proxy settings by removing unused SSL verification options
2025-07-19 16:34:37 +07:00
Akarshan Biswas
8f1a36c8e3
fix: Improve stream error handling and parsing (#5807)
* fix: Enhance stream error handling and parsing

This commit improves the robustness of stream processing in the llamacpp-extension.

- Adds explicit handling for 'error:' prefixed lines in the stream, parsing the contained JSON error and throwing an appropriate JavaScript Error.
- Centralizes JSON parsing of 'data:' and 'error:' lines, ensuring consistent error propagation by re-throwing parsing exceptions.
- Ensures the async iterator terminates correctly upon encountering stream errors or malformed JSON.

* Address bot comments and cleanup
2025-07-18 18:36:33 +05:30
Akarshan
59ad2eb784
Merge branch 'dev' into release/v0.6.6 2025-07-18 18:29:20 +05:30
hiento09
4d44f4324d
feat: add autoqa (#5779)
* feat: add autoqa

* chore: add auto start computer_server

* chore: add ci autoqa windows

* chore: add ci support for both windows and linux

* chore: add ci support for macos

* chore: refactor auto qa

* chore: refactor autoqa workflow

* chore: fix upload turn
2025-07-18 15:22:31 +07:00
Louis
a56e58f69b
Merge pull request #5782 from ethanova/fix/no-more-code-line-number-selection
set line number userSelect to none so that code can be copied without line number
2025-07-18 10:08:46 +07:00
Louis
8d84c3b884
feat: add model load error handling to improve UX (#5802)
* feat: model load error handling

* chore: clean up

* test: add tests

* fix: provider name
2025-07-18 08:25:54 +05:30
Akarshan Biswas
bcb60378c0
fix: Add --reasoning-format none to support rendering of reasoning content (#5803) 2025-07-18 08:22:37 +05:30
Louis
8ca507c01c
feat: proxy support for the new downloader (#5795)
* feat: proxy support for the new downloader

* test: remove outdated test

* ci: clean up
2025-07-17 23:10:21 +07:00
Louis
32966f9259
Merge pull request #5801 from menloresearch/release/v0.6.5
Sync release/0.6.5 into dev to start new development cycle
2025-07-17 22:22:32 +07:00
Akarshan Biswas
92703bceb2
refactor: move thinking toggle to runtime settings for dynamic control (#5800)
* refactor: move thinking toggle to runtime settings for per-message control

Replaces the static `reasoning_budget` config with a dynamic `enable_thinking` flag under `chat_template_kwargs`, allowing models like Jan-nano and Qwen3 to enable/disable thinking behavior at runtime, even mid-conversation.
Requires UI update

* remove engine argument
2025-07-17 20:18:24 +05:30
Louis
4699b07ca6
Merge pull request #5797 from menloresearch/test/deprecate-webdriver-test
test: deprecate webdriver test in favor of auto qa using CUA
2025-07-17 21:28:12 +07:00
Ramon Perez
333cd3a694
Merge pull request #5799 from menloresearch/ramon/docs-v0.6.5 2025-07-17 20:26:04 +10:00
Ramon Perez
f606ba6906
Update settings.mdx
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-07-17 20:24:58 +10:00
Ramon Perez
5d87dc8d32
Update 2025-07-17-responsive-ui.mdx
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-07-17 20:24:48 +10:00
Ramon Perez
2f77c6ba9c Addedd changelog, reorganized docs, added MCP examples, and updated different pages. 2025-07-17 19:49:00 +10:00
Louis
c2839794cf
fix: bump token.js to fix gemini provider base url issue v0.6.5 2025-07-17 14:27:49 +07:00
Akarshan Biswas
b736d09168
fix: Prevent spamming /health endpoint and improve startup and resolve compiler warnings (#5784)
* fix: Prevent spamming /health endpoint and improve startup and resolve compiler warnings

This commit introduces a delay and improved logic to the /health endpoint checks in the llamacpp extension, preventing excessive requests during model loading.

Additionally, it addresses several Rust compiler warnings by:
- Commenting out an unused `handle_app_quit` function in `src/core/mcp.rs`.
- Explicitly declaring `target_port`, `session_api_key`, and `buffered_body` as mutable in `src/core/server.rs`.
- Commenting out unused `tokio` imports in `src/core/setup.rs`.
- Enhancing the `load_llama_model` function in `src/core/utils/extensions/inference_llamacpp_extension/server.rs` to better monitor stdout/stderr for readiness and errors, and handle timeouts.
- Commenting out an unused `std::path::Prefix` import and adjusting `normalize_path` in `src/core/utils/mod.rs`.
- Updating the application version to 0.6.904 in `tauri.conf.json`.

* fix grammar!

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

* fix grammar 2

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

* reimport prefix but only on Windows

* remove instead of commenting

* remove redundant check

* sync app version in cargo.toml with tauri.conf

---------

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-07-16 18:18:11 +05:30
Louis
b17c6f4eb6
Merge pull request #5791 from menloresearch/fix/fetch-models-from-custom-endpoint-crashes-app
fix: fetch models from custom provider causes app to crash
2025-07-16 16:01:19 +07:00
Louis
3eaa3424e1
fix: fetch models from custom provider causes app to crash 2025-07-16 15:36:45 +07:00
Ramon Perez
44082846d5 new tutorials, reorganization, and pages 2025-07-16 16:04:30 +10:00
Louis
646f40d664
chore: token-js version bump 2025-07-16 11:37:40 +07:00
Ethan Garber
a3b95f01de add comment to justify style block 2025-07-15 21:14:24 -04:00
Ethan Garber
b0e66993fe set line number userSelect to none so that code can be copied without line number 2025-07-15 21:10:22 -04:00
Louis
bd3b8bff35
Merge pull request #5781 from menloresearch/test/add-missing-tests
test: add missing unit tests
2025-07-15 22:56:22 +07:00
Louis
9872a6e82a test: add missing unit tests 2025-07-15 22:29:28 +07:00
Louis
f083fafcfd
Merge pull request #5776 from menloresearch/fix/download-icon
🐛fix: download icon when left panel close
2025-07-15 11:44:29 +07:00
Louis
08fe2c27fd
fix: translations 2025-07-15 11:11:00 +07:00
Nguyen Ngoc Minh
f674c786ba
Merge pull request #5778 from menloresearch/fix/revert-windows-installation-mode
fix: revert installationmode in nsis template
2025-07-15 11:02:42 +07:00
Faisal Amir
02c049653e
🐛fix: revert back stat hover for three dots (#5777) 2025-07-15 10:34:02 +07:00
Minh141120
65bc24530f fix: change installationmode in nsis template 2025-07-15 10:23:22 +07:00
Louis
8e85c2fd06
fix: bump llama.cpp b5857 on windows 2025-07-15 10:15:03 +07:00
Faisal Amir
55b68df956 🐛fix: download icon when left panel close 2025-07-15 09:39:51 +07:00
Sam Hoang Van
9a76c94e22
update rmcp to fix issues (#5290) 2025-07-14 16:49:27 +07:00
Akarshan Biswas
dee98f41d1
Feat: Improved llamacpp Server Stability and Diagnostics (#5761)
* feat: Improve llamacpp server error reporting and model load stability

This commit introduces significant improvements to how the llamacpp server
process is managed and how its errors are reported.

Key changes:
- **Enhanced Error Reporting:** The llamacpp server's stdout and stderr
  are now piped and captured. If the llamacpp process exits prematurely
  or fails to start, its stderr output is captured and returned as a
  `LlamacppError`. This provides much more specific and actionable
  diagnostic information for users and developers.
- **Increased Model Load Timeout:** The `waitForModelLoad` timeout has
  been increased from 30 seconds to 240 seconds (4 minutes). This
  addresses issues where larger models or slower systems would
  prematurely time out during the model loading phase.
- **API Secret Update:** The internal API secret for the llamacpp
  extension has been updated from 'Jan' to 'JustAskNow'.
- **Version Bump:** The application version in `tauri.conf.json` has
  been incremented to `0.6.901`.

* fix: should not spam load requests

* test: add test to cover the fix

* refactor: clean up

* test: add more test case

---------

Co-authored-by: Louis <louis@jan.ai>
2025-07-14 11:55:44 +05:30
Akarshan Biswas
96ba42e411
feat: Add missing ctx-shift toggle (#5765)
* feat: Add missing ctx_shift

* fix typo

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

* refine description

---------

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-07-14 11:51:34 +05:30
Louis
eaf4b1b954
Merge pull request #5757 from menloresearch/test/add-tests
test: add missing unit tests
2025-07-14 09:55:53 +07:00
Louis
03bcd02002
test: add missing unit tests 2025-07-12 22:46:27 +07:00
Louis
c2790d9181
test: remove route tests 2025-07-12 21:35:49 +07:00
Louis
864ad50880
test: add missing tests 2025-07-12 21:29:51 +07:00