5593 Commits

Author SHA1 Message Date
Louis
010f0091b9
refactor: remove Cargo workspace 2025-08-07 21:22:35 +07:00
Louis
cafd8741ca
fix: test step 2025-08-07 21:01:40 +07:00
Louis
b8f5fd510a
test: fix failed tests 2025-08-07 20:54:00 +07:00
Louis
25699995fc
refactor: add cargo test 2025-08-07 20:32:55 +07:00
Louis
ab44faeda3
test: fix test 2025-08-07 20:09:07 +07:00
Louis
c1668a4e4a
refactor: clean up unused hardware apis 2025-08-07 20:04:23 +07:00
Akarshan Biswas
469d787888
refactor: Use more precise terminology in API server logs (#6085)
* refactor: Use more precise terminology in API server logs and error messages

This commit refactors several log and error messages to use more accurate and consistent terminology.

-   Replaced "backend servers" and "backend model servers" with "models" or "sessions" to better reflect the service's internal structure.
-   Changed "Proxy server" to "Jan API server" to more accurately describe the server's function.
-   Removed a redundant debug log message.

These changes are cosmetic and improve the readability and consistency of the logging output.

* Update src-tauri/src/core/server.rs

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

---------

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-07 17:48:33 +05:30
Faisal Amir
d8e1fef3f0
🐛fix/onboarding-loop (#6054) 2025-08-07 18:11:22 +07:00
Akarshan Biswas
6a699d8004
refactor: move session management & port allocation to backend (#6083)
* refactor: move session management & port allocation to backend

- Remove the in‑process `activeSessions` map and its cleanup logic from the TypeScript side.
- Introduce new Tauri commands in Rust:
  - `get_random_port` – picks an unused port using a seeded RNG and checks availability.
  - `find_session_by_model` – returns the `SessionInfo` for a given model ID.
  - `get_loaded_models` – returns a list of currently loaded model IDs.
- Update the extension’s TypeScript code to use these commands via `invoke`:
  - `findSessionByModel`, `load`, `unload`, `chat`, `getLoadedModels`, and `embed` now operate asynchronously and query the backend.
  - Remove the old `is_port_available` command and the custom port‑checking loop.
  - Simplify `onUnload` – session termination is now handled by the backend.
- Drop unused helpers (`sleep`, `waitForModelLoad`) and related port‑availability code.
- Add missing Rust imports (`rand::{StdRng,Rng,SeedableRng}`, `HashSet`) and improve error handling.
- Register the new commands in `src-tauri/src/lib.rs` (replace `is_port_available` with the three new commands).

This refactor centralises session state and port allocation in the Rust backend, eliminates duplicated logic, and resolves race conditions around model loading and session cleanup.

* Use String(e) for error

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>

---------

Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-07 13:06:21 +05:30
Akarshan Biswas
1f1605bdf9
feat: Add support for overriding tensor buffer type (#6062)
* feat: Add support for overriding tensor buffer type

This commit introduces a new configuration option, `override_tensor_buffer_t`, which allows users to specify a regex for matching tensor names to override their buffer type. This is an advanced setting primarily useful for optimizing the performance of large models, particularly Mixture of Experts (MoE) models.

By overriding the tensor buffer type, users can keep critical parts of the model, like the attention layers, on the GPU while offloading other parts, such as the expert feed-forward networks, to the CPU. This can lead to significant speed improvements for massive models.

Additionally, this change refines the error message to be more specific when a model fails to load. The previous message "Failed to load llama-server" has been updated to "Failed to load model" to be more accurate.

* chore: update FE to suppoer override-tensor

---------

Co-authored-by: Faisal Amir <urmauur@gmail.com>
2025-08-07 10:31:34 +05:30
Emre Can Kartal
c1cdc434a8
Add gpt-oss local installation blog post (#6075)
- Complete beginner guide for running OpenAI's gpt-oss locally
- Step-by-step instructions using Jan AI
- Alternative installation methods (llama.cpp, Ollama, LM Studio)
- Performance benchmarks and troubleshooting guide
- SEO-optimized with FAQ section and comparison tables
- 4 supporting screenshots showing the installation process
2025-08-07 09:48:05 +07:00
Nguyen Ngoc Minh
06941b932d
Merge pull request #6078 from menloresearch/ci/deprecate-jan-docs-new-release
ci: deprecate jan docs new release workflow in favor of jan-docs
2025-08-07 00:22:22 +07:00
Minh141120
c3cca93850 ci: deprecate jan docs new release workflow in favor of jan-docs 2025-08-07 00:04:21 +07:00
Nguyen Ngoc Minh
d55a5e695f
Merge pull request #6073 from menloresearch/chore/update-workflow-name
chore: update workflow name
2025-08-06 23:46:25 +07:00
Nguyen Ngoc Minh
397f71db6e
chore: update workflow name 2025-08-06 17:36:03 +07:00
Louis
b0785e9db0
Merge pull request #6072 from menloresearch/fix/should-not-include-reasoning-content-in-completion-request
fix: should not include reasoning text in the chat completion request
2025-08-06 17:34:16 +07:00
Louis
0b1b84dbf4
test: add tests for new change 2025-08-06 17:13:22 +07:00
Louis
fc815dc98e
fix: should not include reasoning text in the chat completion request 2025-08-06 17:07:32 +07:00
Faisal Amir
ffdb6829e1
fix: gpt-oss thinking block (#6071) 2025-08-06 16:10:24 +07:00
Ramon Perez
1739958664
Added new model provider and updated main repo readme 2025-08-06 13:14:28 +10:00
Ramon Perez
683fb34709 fixed components in troubleshooting tab 2025-08-06 12:49:01 +10:00
Ramon Perez
2306da0e84 added troubleshooting server instructions to config 2025-08-06 12:38:55 +10:00
Akarshan Biswas
fec4cce560 fix: Add conditional Vulkan support check for better GPU compatibility (#6066)
Changes:
- Introduce conditional Vulkan support check for discrete GPUs with 6GB+ VRAM

fixes: #6009
2025-08-06 12:24:21 +10:00
Louis
e74601443f chore: add deep_link register_all 2025-08-06 12:24:21 +10:00
Louis
f41a04b1a2 fix: test env 2025-08-06 12:24:21 +10:00
Louis
3bdd5f00b6 chore: able to disable updater via env flag 2025-08-06 12:24:21 +10:00
Louis
de146f363a test: add tests 2025-08-06 12:24:21 +10:00
Louis
83527a7533 fix: Jan hub repo detail and deep link 2025-08-06 12:24:21 +10:00
Faisal Amir
026b21f779 feat: jinja template customize per model instead provider level (#6053) 2025-08-06 12:24:21 +10:00
Akarshan Biswas
dcffa4fa0a Fix: Improve Llama.cpp model path handling and error handling (#6045)
* Improve Llama.cpp model path handling and validation

This commit refactors the load_llama_model function to improve how it handles and validates the model path.

Previously, the function extracted the model path but did not perform any validation. This change adds the following improvements:

It now checks for the presence of the -m flag.

It verifies that a path is provided after the -m flag.

It validates that the specified model path actually exists on the filesystem.

It ensures that the SessionInfo struct stores the canonical display path of the model, which is a more robust approach.

These changes make the model loading process more reliable and provide better error handling for invalid or missing model paths.

* Exp: Use short path on Windows

* Fix: Remove error channel and handling in llama.cpp server loading

The previous implementation used a channel to receive error messages from the llama.cpp server's stdout. However, this proved unreliable as the path names can contain 'errors strings' that we use to check even during normal operation. This commit removes the error channel and associated error handling logic.
The server readiness is still determined by checking for the "server is listening" message in stdout. Errors are now handled by relying on the process exit code and capturing the full stderr output if the process fails to start or exits unexpectedly. This approach provides a more robust and accurate error detection mechanism.

* Add else block in Windows path handling

* Add some path related tests

* Fix windows tests
2025-08-06 12:24:21 +10:00
Faisal Amir
318f6f504f feat: recommended label llamacpp setting (#6052)
* feat: recommended label llamacpp

* chore: remove log
2025-08-06 12:24:21 +10:00
Minh141120
8e4c696583 ci: disable autoqa on nightly build 2025-08-06 12:24:21 +10:00
Louis
7e52512d0e fix: should check for invalid backend to cover previous missing backend case 2025-08-06 12:24:21 +10:00
Louis
eb13189d07 fix: run dev should reinstall extensions 2025-08-06 12:24:21 +10:00
Louis
026383e92d test: add tests for new changes 2025-08-06 12:24:21 +10:00
Louis
4b6269a4f0 fix: add missing cuda backend support 2025-08-06 12:24:21 +10:00
Minh141120
3ffb30b544 chore: skip nightly build workflow for external contributor 2025-08-06 12:24:21 +10:00
Sherzod Mutalov
5f06a35f4e fix: use attributes to check the feature existence 2025-08-06 12:24:21 +10:00
Sherzod Mutalov
280ea1aa9f chore: extracted macos avx2 check code to the utility function 2025-08-06 12:23:18 +10:00
Sherzod Mutalov
ad9c4854a9 chore: added comments 2025-08-06 12:20:30 +10:00
Sherzod Mutalov
49c8334e40 chore: replaced with macros call to remove warning 2025-08-06 12:20:30 +10:00
Sherzod Mutalov
f1dd42de9e fix: use system npx on old mac's 2025-08-06 12:20:30 +10:00
Chaiyapruek Muangsiri
4e31e1d3a8 remove unnecessary try catch block 2025-08-06 12:20:30 +10:00
Chaiyapruek Muangsiri
00f686a733 fix connected servers status not in sync when edit mcp json 2025-08-06 12:20:30 +10:00
Ramon Perez
890a917dec removed nextra component in astro site 2025-08-06 12:20:30 +10:00
Akarshan Biswas
8d147c1774
fix: Add conditional Vulkan support check for better GPU compatibility (#6066)
Changes:
- Introduce conditional Vulkan support check for discrete GPUs with 6GB+ VRAM

fixes: #6009
2025-08-06 07:20:44 +05:30
Louis
c642076ec3
Merge pull request #6024 from menloresearch/fix/jan-hub-repo-data-and-deeplink
fix: Jan hub model detail and deep link
2025-08-06 08:46:07 +07:00
Louis
3b349a60f1 chore: add deep_link register_all 2025-08-05 22:32:27 +07:00
Ramon Perez
4ee6873ca5
Update docs/src/pages/docs/remote-models/huggingface.mdx
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-06 00:24:18 +10:00
Ramon Perez
fc4ecd3412
Update README.md
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-08-06 00:24:10 +10:00