From 58adf7c4ae211aba8c5ed5039f46c4caf82d3bad Mon Sep 17 00:00:00 2001 From: Rodrigo Rodriguez Date: Sun, 15 Feb 2026 13:11:14 +0000 Subject: [PATCH] fix: Set default llm_server_path and correct model file paths When no default.gbai/config.csv exists, the system now: - Sets default llm_server_path to ./botserver-stack/bin/llm/build/bin - Uses correct relative paths to model files: ../../../../data/llm/ - Uses actual model filenames from 3rdparty.toml This fixes the issue where LLM/embedding servers couldn't find model files because the paths were constructed incorrectly. Model filenames: - LLM: DeepSeek-R1-Distill-Qwen-1.5B-Q3_K_M.gguf - Embedding: bge-small-en-v1.5-f32.gguf Co-Authored-By: Claude Sonnet 4.5 --- src/llm/local.rs | 14 ++++++++++---- 1 file changed, 10 insertions(+), 4 deletions(-) diff --git a/src/llm/local.rs b/src/llm/local.rs index 498fe2d38..f3f01213f 100644 --- a/src/llm/local.rs +++ b/src/llm/local.rs @@ -74,16 +74,22 @@ pub async fn ensure_llama_servers_running( let llm_server_enabled = llm_server_enabled.to_lowercase() == "true"; // Use default models when config is empty (no default.gbai/config.csv) + let llm_server_path = if llm_server_path.is_empty() { + "./botserver-stack/bin/llm/build/bin".to_string() + } else { + llm_server_path + }; + let llm_model = if llm_model.is_empty() { - info!("No LLM model configured, using default: deepseek-small"); - "deepseek-small".to_string() + info!("No LLM model configured, using default: ../../../../data/llm/DeepSeek-R1-Distill-Qwen-1.5B-Q3_K_M.gguf"); + "../../../../data/llm/DeepSeek-R1-Distill-Qwen-1.5B-Q3_K_M.gguf".to_string() } else { llm_model }; let embedding_model = if embedding_model.is_empty() { - info!("No embedding model configured, using default: bge-small-en-v1.5"); - "bge-small-en-v1.5".to_string() + info!("No embedding model configured, using default: ../../../../data/llm/bge-small-en-v1.5-f32.gguf"); + "../../../../data/llm/bge-small-en-v1.5-f32.gguf".to_string() } else { embedding_model };