From a9cbbbffa03e76adb8f0b186780ce3066d1bee04 Mon Sep 17 00:00:00 2001 From: Rodrigo Rodriguez Date: Sun, 15 Feb 2026 20:23:04 +0000 Subject: [PATCH] fix: Use correct default LLM model name for local DeepSeek server Changed default model from 'gpt-3.5-turbo' to 'DeepSeek-R1-Distill-Qwen-1.5B-Q3_K_M.gguf' in bot message handler. This ensures the local llama-server receives the correct model name and can process requests properly. Co-Authored-By: Claude Sonnet 4.5 --- src/core/bot/mod.rs | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/src/core/bot/mod.rs b/src/core/bot/mod.rs index 25a63e2af..623b0f6ca 100644 --- a/src/core/bot/mod.rs +++ b/src/core/bot/mod.rs @@ -427,9 +427,11 @@ impl BotOrchestrator { // DEBUG: Log which bot we're getting config for info!("[CONFIG_TRACE] Getting LLM config for bot_id: {}", session.bot_id); + // For local LLM server, use the actual model name + // Default to DeepSeek model if not configured let model = config_manager - .get_config(&session.bot_id, "llm-model", Some("gpt-3.5-turbo")) - .unwrap_or_else(|_| "gpt-3.5-turbo".to_string()); + .get_config(&session.bot_id, "llm-model", Some("DeepSeek-R1-Distill-Qwen-1.5B-Q3_K_M.gguf")) + .unwrap_or_else(|_| "DeepSeek-R1-Distill-Qwen-1.5B-Q3_K_M.gguf".to_string()); let key = config_manager .get_config(&session.bot_id, "llm-key", Some(""))