gb/.gitignore
Rodrigo Rodriguez (Pragmatismo) 3befc141e5 Fix token limits for local llama.cpp server
- Add token-aware text truncation utility in core/shared/utils.rs
- Fix embedding generators to use 600 token limit (safe under 768)
- Fix LLM context limit detection for local models (768 vs 4096)
- Prevent 'exceed context size' errors for both embeddings and chat
2026-02-02 11:56:13 -03:00

38 lines
383 B
Text

# Build artifacts
target/
*.out
bin/
# Logs
*.log
*logfile*
*-log*
# Temporary files
.tmp*
.tmp/*
# Environment files
.env
*.env
# Work directories
work/
# Redis
*.rdb
# Documentation build
docs/book
# Installers (keep gitkeep)
botserver-installers/*
!botserver-installers/.gitkeep
botserver-stack
TODO*
work
# Lock file (regenerated from Cargo.toml)
Cargo.lock
.kiro
config