5457 Commits

Author SHA1 Message Date
hiento09
4d44f4324d
feat: add autoqa (#5779)
* feat: add autoqa

* chore: add auto start computer_server

* chore: add ci autoqa windows

* chore: add ci support for both windows and linux

* chore: add ci support for macos

* chore: refactor auto qa

* chore: refactor autoqa workflow

* chore: fix upload turn
2025-07-18 15:22:31 +07:00
Louis
a56e58f69b
Merge pull request #5782 from ethanova/fix/no-more-code-line-number-selection
set line number userSelect to none so that code can be copied without line number
2025-07-18 10:08:46 +07:00
Louis
8d84c3b884
feat: add model load error handling to improve UX (#5802)
* feat: model load error handling

* chore: clean up

* test: add tests

* fix: provider name
2025-07-18 08:25:54 +05:30
Akarshan Biswas
bcb60378c0
fix: Add --reasoning-format none to support rendering of reasoning content (#5803) 2025-07-18 08:22:37 +05:30
Louis
8ca507c01c
feat: proxy support for the new downloader (#5795)
* feat: proxy support for the new downloader

* test: remove outdated test

* ci: clean up
2025-07-17 23:10:21 +07:00
Louis
32966f9259
Merge pull request #5801 from menloresearch/release/v0.6.5
Sync release/0.6.5 into dev to start new development cycle
2025-07-17 22:22:32 +07:00
Akarshan Biswas
92703bceb2
refactor: move thinking toggle to runtime settings for dynamic control (#5800)
* refactor: move thinking toggle to runtime settings for per-message control

Replaces the static `reasoning_budget` config with a dynamic `enable_thinking` flag under `chat_template_kwargs`, allowing models like Jan-nano and Qwen3 to enable/disable thinking behavior at runtime, even mid-conversation.
Requires UI update

* remove engine argument
2025-07-17 20:18:24 +05:30
Louis
4699b07ca6
Merge pull request #5797 from menloresearch/test/deprecate-webdriver-test
test: deprecate webdriver test in favor of auto qa using CUA
2025-07-17 21:28:12 +07:00
Ramon Perez
333cd3a694
Merge pull request #5799 from menloresearch/ramon/docs-v0.6.5 2025-07-17 20:26:04 +10:00
Ramon Perez
f606ba6906
Update settings.mdx
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-07-17 20:24:58 +10:00
Ramon Perez
5d87dc8d32
Update 2025-07-17-responsive-ui.mdx
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-07-17 20:24:48 +10:00
Ramon Perez
2f77c6ba9c Addedd changelog, reorganized docs, added MCP examples, and updated different pages. 2025-07-17 19:49:00 +10:00
Louis
c2839794cf
fix: bump token.js to fix gemini provider base url issue v0.6.5 2025-07-17 14:27:49 +07:00
Akarshan Biswas
b736d09168
fix: Prevent spamming /health endpoint and improve startup and resolve compiler warnings (#5784)
* fix: Prevent spamming /health endpoint and improve startup and resolve compiler warnings

This commit introduces a delay and improved logic to the /health endpoint checks in the llamacpp extension, preventing excessive requests during model loading.

Additionally, it addresses several Rust compiler warnings by:
- Commenting out an unused `handle_app_quit` function in `src/core/mcp.rs`.
- Explicitly declaring `target_port`, `session_api_key`, and `buffered_body` as mutable in `src/core/server.rs`.
- Commenting out unused `tokio` imports in `src/core/setup.rs`.
- Enhancing the `load_llama_model` function in `src/core/utils/extensions/inference_llamacpp_extension/server.rs` to better monitor stdout/stderr for readiness and errors, and handle timeouts.
- Commenting out an unused `std::path::Prefix` import and adjusting `normalize_path` in `src/core/utils/mod.rs`.
- Updating the application version to 0.6.904 in `tauri.conf.json`.

* fix grammar!

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

* fix grammar 2

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

* reimport prefix but only on Windows

* remove instead of commenting

* remove redundant check

* sync app version in cargo.toml with tauri.conf

---------

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-07-16 18:18:11 +05:30
Louis
b17c6f4eb6
Merge pull request #5791 from menloresearch/fix/fetch-models-from-custom-endpoint-crashes-app
fix: fetch models from custom provider causes app to crash
2025-07-16 16:01:19 +07:00
Louis
3eaa3424e1
fix: fetch models from custom provider causes app to crash 2025-07-16 15:36:45 +07:00
Ramon Perez
44082846d5 new tutorials, reorganization, and pages 2025-07-16 16:04:30 +10:00
Louis
646f40d664
chore: token-js version bump 2025-07-16 11:37:40 +07:00
Ethan Garber
a3b95f01de add comment to justify style block 2025-07-15 21:14:24 -04:00
Ethan Garber
b0e66993fe set line number userSelect to none so that code can be copied without line number 2025-07-15 21:10:22 -04:00
Louis
bd3b8bff35
Merge pull request #5781 from menloresearch/test/add-missing-tests
test: add missing unit tests
2025-07-15 22:56:22 +07:00
Louis
9872a6e82a test: add missing unit tests 2025-07-15 22:29:28 +07:00
Louis
f083fafcfd
Merge pull request #5776 from menloresearch/fix/download-icon
🐛fix: download icon when left panel close
2025-07-15 11:44:29 +07:00
Louis
08fe2c27fd
fix: translations 2025-07-15 11:11:00 +07:00
Nguyen Ngoc Minh
f674c786ba
Merge pull request #5778 from menloresearch/fix/revert-windows-installation-mode
fix: revert installationmode in nsis template
2025-07-15 11:02:42 +07:00
Faisal Amir
02c049653e
🐛fix: revert back stat hover for three dots (#5777) 2025-07-15 10:34:02 +07:00
Minh141120
65bc24530f fix: change installationmode in nsis template 2025-07-15 10:23:22 +07:00
Louis
8e85c2fd06
fix: bump llama.cpp b5857 on windows 2025-07-15 10:15:03 +07:00
Faisal Amir
55b68df956 🐛fix: download icon when left panel close 2025-07-15 09:39:51 +07:00
Sam Hoang Van
9a76c94e22
update rmcp to fix issues (#5290) 2025-07-14 16:49:27 +07:00
Akarshan Biswas
dee98f41d1
Feat: Improved llamacpp Server Stability and Diagnostics (#5761)
* feat: Improve llamacpp server error reporting and model load stability

This commit introduces significant improvements to how the llamacpp server
process is managed and how its errors are reported.

Key changes:
- **Enhanced Error Reporting:** The llamacpp server's stdout and stderr
  are now piped and captured. If the llamacpp process exits prematurely
  or fails to start, its stderr output is captured and returned as a
  `LlamacppError`. This provides much more specific and actionable
  diagnostic information for users and developers.
- **Increased Model Load Timeout:** The `waitForModelLoad` timeout has
  been increased from 30 seconds to 240 seconds (4 minutes). This
  addresses issues where larger models or slower systems would
  prematurely time out during the model loading phase.
- **API Secret Update:** The internal API secret for the llamacpp
  extension has been updated from 'Jan' to 'JustAskNow'.
- **Version Bump:** The application version in `tauri.conf.json` has
  been incremented to `0.6.901`.

* fix: should not spam load requests

* test: add test to cover the fix

* refactor: clean up

* test: add more test case

---------

Co-authored-by: Louis <louis@jan.ai>
2025-07-14 11:55:44 +05:30
Akarshan Biswas
96ba42e411
feat: Add missing ctx-shift toggle (#5765)
* feat: Add missing ctx_shift

* fix typo

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

* refine description

---------

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-07-14 11:51:34 +05:30
Louis
eaf4b1b954
Merge pull request #5757 from menloresearch/test/add-tests
test: add missing unit tests
2025-07-14 09:55:53 +07:00
Louis
03bcd02002
test: add missing unit tests 2025-07-12 22:46:27 +07:00
Louis
c2790d9181
test: remove route tests 2025-07-12 21:35:49 +07:00
Louis
864ad50880
test: add missing tests 2025-07-12 21:29:51 +07:00
Louis
c5fd964bf2 test: add missing tests 2025-07-12 20:15:45 +07:00
Louis
191537884e
Merge pull request #5759 from menloresearch/fix/test-coverage-upload-lcov
fix: vitest config - coverage lcov
2025-07-12 20:14:42 +07:00
Louis
09a45baa5d
fix: vitest config - coverage lcov 2025-07-12 20:09:31 +07:00
Louis
c23c34583d
Merge pull request #4965 from menloresearch/feat/inference-llamacpp-extension
feat: llama.cpp provider extension to replace cortex.cpp
2025-07-11 09:25:47 +07:00
Louis
b2ce138ea0
test: add tests 2025-07-11 09:21:11 +07:00
Louis
b8259e7794 feat: add HF token setting 2025-07-11 00:05:52 +07:00
Louis
9cea579c8e
fix: build issue 2025-07-10 22:16:31 +07:00
Louis
114f2b9092
ci: attempt to upload cov 2025-07-10 21:51:13 +07:00
Louis
963ad448f5
fix: build 2025-07-10 21:23:04 +07:00
Nguyen Ngoc Minh
0f7f2a7b38
Merge pull request #5745 from menloresearch/refactor/tauri-workflow
refactor: tauri workflow
2025-07-10 21:18:26 +07:00
Louis
a770e08013
test: migrate jest to vitest 2025-07-10 21:14:21 +07:00
Louis
1c7a20be44
fix: linux build 2025-07-10 21:14:20 +07:00
Louis
bc0ea343cc
fix: remove legacy build step 2025-07-10 20:24:11 +07:00
Louis
37718d1e71
fix: build issue with legacy libs 2025-07-10 20:17:20 +07:00