fix: Increase default n_predict to 512 for DeepSeek R1 reasoning
All checks were successful
BotServer CI / build (push) Successful in 9m26s
All checks were successful
BotServer CI / build (push) Successful in 9m26s
DeepSeek R1 model outputs reasoning_content first, then content. With n_predict=50, responses were truncated during reasoning phase. Increased to 512 to allow full reasoning + response. Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
This commit is contained in:
parent
a9cbbbffa0
commit
0a1bd25869
1 changed files with 2 additions and 2 deletions
|
|
@ -386,8 +386,8 @@ pub fn start_llm_server(
|
||||||
|
|
||||||
let n_predict = config_manager
|
let n_predict = config_manager
|
||||||
.get_config(&default_bot_id, "llm-server-n-predict", None)
|
.get_config(&default_bot_id, "llm-server-n-predict", None)
|
||||||
.unwrap_or_else(|_| "50".to_string());
|
.unwrap_or_else(|_| "512".to_string()); // Increased default for DeepSeek R1 reasoning
|
||||||
let n_predict = if n_predict.is_empty() { "50".to_string() } else { n_predict };
|
let n_predict = if n_predict.is_empty() { "512".to_string() } else { n_predict };
|
||||||
|
|
||||||
let n_ctx_size = config_manager
|
let n_ctx_size = config_manager
|
||||||
.get_config(&default_bot_id, "llm-server-ctx-size", None)
|
.get_config(&default_bot_id, "llm-server-ctx-size", None)
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue