use std::{collections::HashMap, env, sync::Arc}; use rmcp::{service::RunningService, transport::TokioChildProcess, RoleClient, ServiceExt}; use serde_json::Value; use tauri::{AppHandle, Emitter, Runtime, State}; use tokio::{process::Command, sync::Mutex}; use super::{cmd::get_jan_data_folder_path, state::AppState}; /// Runs MCP commands by reading configuration from a JSON file and initializing servers /// /// # Arguments /// * `app_path` - Path to the application directory containing mcp_config.json /// * `servers_state` - Shared state containing running MCP services /// /// # Returns /// * `Ok(())` if servers were initialized successfully /// * `Err(String)` if there was an error reading config or starting servers pub async fn run_mcp_commands( app: &AppHandle, servers_state: Arc>>>, ) -> Result<(), String> { let app_path = get_jan_data_folder_path(app.clone()); let app_path_str = app_path.to_str().unwrap().to_string(); log::info!( "Load MCP configs from {}", app_path_str.clone() + "/mcp_config.json" ); let config_content = std::fs::read_to_string(app_path_str.clone() + "/mcp_config.json") .map_err(|e| format!("Failed to read config file: {}", e))?; let mcp_servers: serde_json::Value = serde_json::from_str(&config_content) .map_err(|e| format!("Failed to parse config: {}", e))?; if let Some(server_map) = mcp_servers.get("mcpServers").and_then(Value::as_object) { log::info!("MCP Servers: {server_map:#?}"); let exe_path = env::current_exe().expect("Failed to get current exe path"); let exe_parent_path = exe_path .parent() .expect("Executable must have a parent directory"); let bin_path = exe_parent_path.to_path_buf(); for (name, config) in server_map { if let Some(false) = extract_active_status(config) { log::info!("Server {name} is not active, skipping."); continue; } if let Some((command, args, envs)) = extract_command_args(config) { let mut cmd = Command::new(command.clone()); if command.clone() == "npx" { let bun_x_path = format!("{}/bun", bin_path.display()); cmd = Command::new(bun_x_path); cmd.arg("x"); } if command.clone() == "uvx" { let bun_x_path = format!("{}/uv", bin_path.display()); cmd = Command::new(bun_x_path); cmd.arg("tool run"); cmd.arg("run"); } println!("Command: {cmd:#?}"); args.iter().filter_map(Value::as_str).for_each(|arg| { cmd.arg(arg); }); envs.iter().for_each(|(k, v)| { if let Some(v_str) = v.as_str() { cmd.env(k, v_str); } }); let process = TokioChildProcess::new(cmd); match process { Ok(p) => { let service = ().serve(p).await; match service { Ok(running_service) => { servers_state .lock() .await .insert(name.clone(), running_service); log::info!("Server {name} started successfully."); } Err(e) => { log::error!("Failed to start server {name}: {e}"); } } } Err(e) => { log::error!("Failed to run command {name}: {e}"); } } } } } // Collect servers into a Vec to avoid holding the RwLockReadGuard across await points let servers_map = servers_state.lock().await; for (_, service) in servers_map.iter() { // Initialize let _server_info = service.peer_info(); log::info!("Connected to server: {_server_info:#?}"); // Emit event to the frontend let event = format!("mcp-connected"); let server_info: &rmcp::model::InitializeResult = service.peer_info(); let name = server_info.server_info.name.clone(); let version = server_info.server_info.version.clone(); let payload = serde_json::json!({ "name": name, "version": version, }); // service.peer_info().server_info.name app.emit(&event, payload) .map_err(|e| format!("Failed to emit event: {}", e))?; log::info!("Emitted event: {event}"); } Ok(()) } fn extract_command_args( config: &Value, ) -> Option<(String, Vec, serde_json::Map)> { let obj = config.as_object()?; let command = obj.get("command")?.as_str()?.to_string(); let args = obj.get("args")?.as_array()?.clone(); let envs = obj .get("env") .unwrap_or(&Value::Object(serde_json::Map::new())) .as_object()? .clone(); Some((command, args, envs)) } fn extract_active_status(config: &Value) -> Option { let obj = config.as_object()?; let active = obj.get("active")?.as_bool()?; Some(active) } #[tauri::command] pub async fn restart_mcp_servers(app: AppHandle, state: State<'_, AppState>) -> Result<(), String> { let servers = state.mcp_servers.clone(); // Stop the servers stop_mcp_servers(state.mcp_servers.clone()).await?; // Restart the servers run_mcp_commands(&app, servers).await?; app.emit("mcp-update", "MCP servers updated") .map_err(|e| format!("Failed to emit event: {}", e)) } pub async fn stop_mcp_servers( servers_state: Arc>>>, ) -> Result<(), String> { let mut servers_map = servers_state.lock().await; let keys: Vec = servers_map.keys().cloned().collect(); for key in keys { if let Some(service) = servers_map.remove(&key) { service.cancel().await.map_err(|e| e.to_string())?; } } drop(servers_map); // Release the lock after stopping Ok(()) } #[tauri::command] pub async fn get_connected_servers( _app: AppHandle, state: State<'_, AppState>, ) -> Result, String> { let servers = state.mcp_servers.clone(); let servers_map = servers.lock().await; Ok(servers_map.keys().cloned().collect()) } #[cfg(test)] mod tests { use super::*; use std::collections::HashMap; use std::fs::File; use std::io::Write; use std::sync::Arc; use tauri::test::mock_app; use tokio::sync::Mutex; #[tokio::test] async fn test_run_mcp_commands() { let app = mock_app(); // Create a mock mcp_config.json file let config_path = "mcp_config.json"; let mut file: File = File::create(config_path).expect("Failed to create config file"); file.write_all(b"{\"mcpServers\":{}}") .expect("Failed to write to config file"); // Call the run_mcp_commands function let servers_state: Arc>>> = Arc::new(Mutex::new(HashMap::new())); let result = run_mcp_commands(app.handle(), servers_state).await; // Assert that the function returns Ok(()) assert!(result.is_ok()); // Clean up the mock config file std::fs::remove_file(config_path).expect("Failed to remove config file"); } }