Some checks failed
BotServer CI / build (push) Failing after 1m34s
Split 20+ files over 1000 lines into focused subdirectories for better maintainability and code organization. All changes maintain backward compatibility through re-export wrappers. Major splits: - attendance/llm_assist.rs (2074→7 modules) - basic/keywords/face_api.rs → face_api/ (7 modules) - basic/keywords/file_operations.rs → file_ops/ (8 modules) - basic/keywords/hear_talk.rs → hearing/ (6 modules) - channels/wechat.rs → wechat/ (10 modules) - channels/youtube.rs → youtube/ (5 modules) - contacts/mod.rs → contacts_api/ (6 modules) - core/bootstrap/mod.rs → bootstrap/ (5 modules) - core/shared/admin.rs → admin_*.rs (5 modules) - designer/canvas.rs → canvas_api/ (6 modules) - designer/mod.rs → designer_api/ (6 modules) - docs/handlers.rs → handlers_api/ (11 modules) - drive/mod.rs → drive_handlers.rs, drive_types.rs - learn/mod.rs → types.rs - main.rs → main_module/ (7 modules) - meet/webinar.rs → webinar_api/ (8 modules) - paper/mod.rs → (10 modules) - security/auth.rs → auth_api/ (7 modules) - security/passkey.rs → (4 modules) - sources/mod.rs → sources_api/ (5 modules) - tasks/mod.rs → task_api/ (5 modules) Stats: 38,040 deletions, 1,315 additions across 318 files Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
44 lines
1.2 KiB
Rust
44 lines
1.2 KiB
Rust
use std::sync::Arc;
|
|
|
|
use crate::core::shared::state::AppState;
|
|
|
|
#[cfg(feature = "llm")]
|
|
use crate::llm::OpenAIClient;
|
|
|
|
pub async fn call_llm(
|
|
state: &Arc<AppState>,
|
|
system_prompt: &str,
|
|
user_content: &str,
|
|
) -> Result<String, String> {
|
|
#[cfg(feature = "llm")]
|
|
{
|
|
let llm = &state.llm_provider;
|
|
|
|
let messages = OpenAIClient::build_messages(
|
|
system_prompt,
|
|
"",
|
|
&[("user".to_string(), user_content.to_string())],
|
|
);
|
|
|
|
let config_manager = crate::core::config::ConfigManager::new(state.conn.clone());
|
|
let model = config_manager
|
|
.get_config(&uuid::Uuid::nil(), "llm-model", None)
|
|
.unwrap_or_else(|_| "gpt-3.5-turbo".to_string());
|
|
let key = config_manager
|
|
.get_config(&uuid::Uuid::nil(), "llm-key", None)
|
|
.unwrap_or_else(|_| String::new());
|
|
|
|
llm.generate(user_content, &messages, &model, &key)
|
|
.await
|
|
.map_err(|e| format!("LLM error: {}", e))
|
|
}
|
|
|
|
#[cfg(not(feature = "llm"))]
|
|
{
|
|
let _ = (state, system_prompt);
|
|
Ok(format!(
|
|
"[LLM not available] Processing: {}...",
|
|
&user_content[..50.min(user_content.len())]
|
|
))
|
|
}
|
|
}
|