6054 Commits

Author SHA1 Message Date
Faisal Amir
272ef9f8b8 fix/revalidate-model-gguf 2025-09-16 15:59:59 +07:00
Faisal Amir
ea354ce621 fix/validate-mmproj-from-general-basename 2025-09-16 15:44:58 +07:00
Faisal Amir
fd052149fe
Merge pull request #6474 from menloresearch/fix/model-selection
fix: immediate update value model selection
2025-09-16 13:54:37 +07:00
Faisal Amir
0e972646e8
Merge pull request #6465 from menloresearch/fix/attachment-edit-message
fix: attachment edit message
2025-09-16 11:17:17 +07:00
Faisal Amir
4ef64ce18c fix: linter 2025-09-16 11:16:53 +07:00
Faisal Amir
9fe517dbf8 fix: immediate dropdown value update 2025-09-16 11:16:35 +07:00
Faisal Amir
da2f790f25
Merge pull request #6464 from menloresearch/fix/pathname-install-be
fix: pathname file install BE
2025-09-16 10:14:02 +07:00
Faisal Amir
52fe8e8c0e fix: type imageurl 2025-09-15 21:58:24 +07:00
Faisal Amir
7b9b9666cb fix: imporove edit message with attachment image 2025-09-15 21:48:19 +07:00
Faisal Amir
3b22f0b7c0 fix: imporove edit message with attachment image 2025-09-15 21:48:01 +07:00
Faisal Amir
e02be47aae fix: remove log 2025-09-15 21:09:08 +07:00
Faisal Amir
5736d7b110 fix: auto update should not block popup 2025-09-15 20:51:27 +07:00
Akarshan Biswas
9e3a77a559
fix: set default memory mode and clean up unused import (#6463)
Use fallback value 'high' for memory_util config and remove unused GgufMetadata import.
2025-09-15 19:00:46 +05:30
Faisal Amir
18114c0a15 fix: pathname file install BE 2025-09-15 18:05:11 +07:00
Akarshan Biswas
e80a865def
fix: detect allocation failures as out-of-memory errors (#6459)
The Llama.cpp backend can emit the phrase “failed to allocate” when it runs out of memory.
Adding this check ensures such messages are correctly classified as out‑of‑memory errors,
providing more accurate error handling CPU backends.
2025-09-15 12:35:24 +05:30
Nguyen Ngoc Minh
55edc7129e
Merge pull request #6457 from menloresearch/chore/makefile-rust-targets
chore: makefile rust targets macos
2025-09-15 12:03:17 +07:00
Akarshan Biswas
489c5a3d9c
fix: KVCache size calculation and refactor (#6438)
- Removed the unused `getKVCachePerToken` helper and replaced it with a unified `estimateKVCache` that returns both total size and per‑token size.
- Fixed the KV cache size calculation to account for all layers, correcting previous under‑estimation.
- Added proper clamping of user‑requested context lengths to the model’s maximum.
- Refactored VRAM budgeting: introduced explicit reserves, fixed engine overhead, and separate multipliers for VRAM and system RAM based on memory mode.
- Implemented a more robust planning flow with clear GPU, Hybrid, and CPU pathways, including fallback configurations when resources are insufficient.
- Updated default context length handling and safety buffers to prevent OOM situations.
- Adjusted usable memory percentage to 90 % and refined logging for easier debugging.
2025-09-15 10:16:13 +05:30
Minh141120
1db67ea9a2 chore: simplify macos workflow 2025-09-15 11:24:11 +07:00
Faisal Amir
7a2782e6fd
Merge pull request #6456 from menloresearch/enhancement/copy-mcp-permission
enhancement: copy MCP permission
2025-09-15 11:01:09 +07:00
Faisal Amir
a4483b7eb7
Update web-app/src/locales/en/tool-approval.json
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-09-15 10:38:36 +07:00
Faisal Amir
a26445e557 chore: make action mutton capitalize 2025-09-15 10:34:27 +07:00
Faisal Amir
44893bc3c3 enhancement: copy MCP permission 2025-09-15 10:33:05 +07:00
Minh141120
4fa78fa892 fix: make install-rust-targets a dependency 2025-09-15 09:44:21 +07:00
Minh141120
6959329fd6 chore: add install-rust-targets step for macOS universal builds 2025-09-15 09:32:51 +07:00
Akarshan Biswas
654e566dcb
fix: correct context shift flag handling in LlamaCPP extension (#6404) (#6431)
* fix: correct context shift flag handling in LlamaCPP extension

The previous implementation added the `--no-context-shift` flag when `cfg.ctx_shift` was disabled, which conflicted with the llama.cpp CLI where the presence of `--context-shift` enables the feature.
The logic is updated to push `--context-shift` only when `cfg.ctx_shift` is true, ensuring the extension passes the correct argument and behaves as expected.

* feat: detect model out of context during generation

---------

Co-authored-by: Dinh Long Nguyen <dinhlongviolin1@gmail.com>
2025-09-12 13:43:31 +05:30
Faisal Amir
ad428f587b
Merge pull request #6426 from menloresearch/fix/error-validate-nested-dom
fix: avoid error validate nested DOM
2025-09-12 12:28:46 +07:00
Faisal Amir
4293fe7edc fix: avoid error validate nested dom 2025-09-12 10:58:34 +07:00
Dinh Long Nguyen
ea72c1ae0f
exclude jan extension web from desktop build (#6419) 2025-09-11 19:51:49 +07:00
Dinh Long Nguyen
db52057030
fix ollama error (#6418) 2025-09-11 18:38:06 +07:00
Faisal Amir
e709d200aa
Merge pull request #6416 from menloresearch/enhancement/experimental-label
enhancement: add label experimental for optimize setting
2025-09-11 16:12:35 +07:00
Dinh Long Nguyen
4856cfbfc4
bug: Deleted model file from imported models blocking model loading (#6317) (#6417) 2025-09-11 15:56:19 +07:00
Faisal Amir
19aa15ffcd chore: update return value 2025-09-11 15:51:21 +07:00
Akarshan
7c41408a1a
feat: add relative path support for model loading
Implemented `isAbsolutePath` helper to correctly identify POSIX, Windows drive‑letter, and UNC absolute paths. Updated `planModelLoad` to automatically resolve relative model and mmproj paths against the Jan data folder, enhancing usability for users supplying non‑absolute paths. Also refined minor formatting for readability.
2025-09-11 13:45:29 +05:30
Akarshan
8f67f29317
feat: add support for mmproj offload setting
Expose the new `mmproj_offload` option in the model settings UI and include it in the `ModelPlan` type. The component now collects the offload flag (`result.offloadMmproj`) and queues it with other setting updates to ensure a single atomic change, preventing race conditions when toggling this feature. This enables users to control MMProj offloading directly from the app.
2025-09-11 13:08:01 +05:30
Faisal Amir
14c7fc0450 chore: update argument 2025-09-11 14:23:56 +07:00
Akarshan
abd0cbe599
refactor: rename noOffloadMmproj flag to offloadMmproj and reorder args
The flag `noOffloadMmproj` was misleading – it actually indicates when the mmproj file **is** offloaded to VRAM. Renaming it to `offloadMmproj` clarifies its purpose and aligns the naming with the surrounding code.

Additionally, the `planModelLoad` signature has been reordered to place `mmprojPath` before `requestedCtx`, improving readability and making the optional parameters more intuitive. All related logic, calculations, and log messages have been updated to use the new flag name.
2025-09-11 12:29:53 +05:30
Faisal Amir
198955285e
Merge pull request #6412 from menloresearch/fix/render-new-line
fix: render new line for user message
2025-09-11 13:29:18 +07:00
Faisal Amir
bc29046c06 enhancement: send params mmptoj_path for optimize setting 2025-09-11 13:23:25 +07:00
Louis
7fea6e1ab0
fix: clean up unused packages (#6414) 2025-09-11 13:16:26 +07:00
Faisal Amir
791563e6ba enhancement: add label experimental for optimize setting 2025-09-11 13:11:37 +07:00
Akarshan Biswas
5ff7935d91
fix: include lm_head and embedding layers in totalLayers count (#6415)
The original calculation used only the `block_count` from the model metadata, which excludes the final LM head and the embedding layer. This caused an underestimation of the total number of layers and consequently an incorrect `layerSize` value. Adding `+2` accounts for these two missing layers, ensuring accurate model size metrics.
2025-09-11 11:40:39 +05:30
Nguyen Ngoc Minh
d856651380
Merge pull request #6413 from menloresearch/ci/add-nightly-external-contrib
ci: add nightly build for external contributors
2025-09-11 13:03:30 +07:00
Minh141120
65a515a9db chore: add upload artifact steps for 3 platforms 2025-09-11 12:21:56 +07:00
Minh141120
773b252555 ci: add nightly build for external contributors 2025-09-11 11:30:43 +07:00
Akarshan Biswas
7a94e74d6b
Merge pull request #6360 from menloresearch/feat/llamacpp_backend
feat: enhance llamacpp backend management and installation
2025-09-11 09:57:16 +05:30
Akarshan
13806a3f06
Fixup sorting in determineBestBackend 2025-09-11 09:56:46 +05:30
Akarshan Biswas
3cd099ee87
Update extensions/llamacpp-extension/src/index.ts
Co-authored-by: ellipsis-dev[bot] <65095814+ellipsis-dev[bot]@users.noreply.github.com>
2025-09-11 09:55:57 +05:30
Akarshan
42411b5f33
feat: prioritize Vulkan backend only when GPU has ≥6 GB VRAM
Added a GPU memory check using `getSystemInfo` to ensure Vulkan is selected only on systems with at least 6 GB of VRAM.
* Made `determineBestBackend` asynchronous and updated all callers to `await` it.
* Adjusted backend priority list to include or demote Vulkan based on the memory check.
* Updated Vulkan support detection in `backend.ts` to rely solely on API version (memory check moved to selection logic).
* Imported `getSystemInfo` and refined file‑existence validation.

These changes prevent sub‑optimal Vulkan usage on low‑memory GPUs and improve backend selection reliability.
2025-09-11 09:55:55 +05:30
Akarshan
84874c6039
fix file condition 2025-09-11 09:55:08 +05:30
Akarshan
0eff1bfaa9
Throw error when invalid file 2025-09-11 09:55:08 +05:30