Commit graph

4 commits

Author SHA1 Message Date
3befc141e5 Fix token limits for local llama.cpp server
- Add token-aware text truncation utility in core/shared/utils.rs
- Fix embedding generators to use 600 token limit (safe under 768)
- Fix LLM context limit detection for local models (768 vs 4096)
- Prevent 'exceed context size' errors for both embeddings and chat
2026-02-02 11:56:13 -03:00
43b168a4b5 Update workspace configuration and submodules 2026-01-30 23:25:02 -03:00
68f3ada343 Enable LLM feature and fix compilation issues
- Update botserver submodule with LLM feature enabled by default
- Fix duplicate module declarations and move errors
- Remove unused imports and variables
- Improve code formatting

This enables the LLM feature by default so it's always available
when building the project.
2026-01-28 16:58:25 -03:00
c4544468f1 Add workspace README and configuration files 2025-12-25 11:03:05 -03:00