* feat: improve local provider connectivity with CORS bypass
- Add @tauri-apps/plugin-http dependency
- Implement dual fetch strategy for local vs remote providers
- Auto-detect local providers (localhost, Ollama:11434, LM Studio:1234)
- Make API key optional for local providers
- Add comprehensive test coverage for provider fetching
refactor: simplify fetchModelsFromProvider by removing preflight check logic
* feat: extend config options to include custom fetch function for CORS handling
* feat: conditionally use Tauri's fetch for openai-compatible providers to handle CORS
* Feat: auto restart mcp (#5226)
* feat: implement retry mechanism for MCP server activation with exponential backoff
feat: enhance MCP server activation with configurable retry attempts
feat: implement MCP server restart monitoring and cleanup functionality
feat: enhance MCP server restart logic with improved monitoring and configuration handling
feat: add manual deactivation for MCP servers to prevent automatic restarts
* feat: enhance MCP server startup with initial attempt tracking and health monitoring
* 🐛fix: prevent render error when additional information missing from hardware (#5413)
---------
Co-authored-by: Sam Hoang Van <samhv.ict@gmail.com>
Co-authored-by: Faisal Amir <urmauur@gmail.com>
* feat: implement retry mechanism for MCP server activation with exponential backoff
feat: enhance MCP server activation with configurable retry attempts
feat: implement MCP server restart monitoring and cleanup functionality
feat: enhance MCP server restart logic with improved monitoring and configuration handling
feat: add manual deactivation for MCP servers to prevent automatic restarts
* feat: enhance MCP server startup with initial attempt tracking and health monitoring
* 🐛fix: setting provider hide model capabilities
* 🐛fix: hide tools icon on dropdown model providers
* fix: stop server on app close or reload
* ✨enhancement: reset heading class
---------
Co-authored-by: Louis <louis@jan.ai>
- Add filtering logic to proxy server for GET /models requests
- Keep only models with status "downloaded" in response
- Remove Content-Length header to prevent mismatch after filtering
- Support both ListModelsResponseDto and direct array formats
- Add comprehensive tests for filtering functionality
- Fix Content-Length header conflict causing empty responses
Fixes issue where all models were returned regardless of download status.
* 🐛fix: llama.cpp default NGL setting does not offload all layers to GPU
* chore: cover more cases
* chore: clean up
* fix: should not show GPU section on Mac