diff --git a/src/automation/compact_prompt.rs b/src/automation/compact_prompt.rs index ed779515..08887828 100644 --- a/src/automation/compact_prompt.rs +++ b/src/automation/compact_prompt.rs @@ -71,15 +71,9 @@ async fn compact_prompt_for_bots( } if !has_new_messages && last_summary_index.is_some() { - trace!( - "Skipping session {} - no new messages since last summary", - session.id - ); continue; } if messages_since_summary < compact_threshold as usize { - trace!("Skipping compaction for session {} - only {} new messages since last summary (threshold: {})", - session.id, messages_since_summary, compact_threshold); continue; } @@ -123,7 +117,6 @@ async fn compact_prompt_for_bots( })]; let llm_provider = state.llm_provider.clone(); - trace!("Starting summarization for session {}", session.id); let mut filtered = String::new(); let config_manager = crate::config::ConfigManager::new(state.conn.clone()); let model = config_manager diff --git a/src/llm/local.rs b/src/llm/local.rs index 8992e923..75ea41ea 100644 --- a/src/llm/local.rs +++ b/src/llm/local.rs @@ -229,7 +229,7 @@ pub async fn start_llm_server( .get_config(&default_bot_id, "llm-server-n-predict", None) .unwrap_or("50".to_string()); let mut args = format!( - "-m {} --host 0.0.0.0 --port {} --reasoning-format deepseek --top_p 0.95 --temp 0.6 --repeat-penalty 1.2 --n-gpu-layers {}", + "-m {} --host 0.0.0.0 --port {} --top_p 0.95 --temp 0.6 --repeat-penalty 1.2 --n-gpu-layers {}", model_path, port, gpu_layers ); if !reasoning_format.is_empty() { diff --git a/web/html/index.html b/web/html/index.html index b2cdee07..ef466798 100644 --- a/web/html/index.html +++ b/web/html/index.html @@ -772,7 +772,7 @@
-