8.2 KiB
Configuration Parameters
Complete reference of all available parameters in config.csv.
Server Parameters
Web Server
| Parameter | Description | Default | Type |
|---|---|---|---|
server-host |
Server bind address | 0.0.0.0 |
IP address |
server-port |
Server listen port | 8080 |
Number (1-65535) |
sites-root |
Generated sites directory | /tmp |
Path |
MCP Server
| Parameter | Description | Default | Type |
|---|---|---|---|
mcp-server |
Enable MCP protocol server | false |
Boolean |
LLM Parameters
Core LLM Settings
| Parameter | Description | Default | Type |
|---|---|---|---|
llm-key |
API key for LLM service | none |
String |
llm-url |
LLM service endpoint | http://localhost:8081 |
URL |
llm-model |
Model path or identifier | Required | Path/String |
LLM Cache
| Parameter | Description | Default | Type |
|---|---|---|---|
llm-cache |
Enable response caching | false |
Boolean |
llm-cache-ttl |
Cache time-to-live | 3600 |
Seconds |
llm-cache-semantic |
Semantic similarity cache | true |
Boolean |
llm-cache-threshold |
Similarity threshold | 0.95 |
Float (0-1) |
Embedded LLM Server
| Parameter | Description | Default | Type |
|---|---|---|---|
llm-server |
Run embedded server | false |
Boolean |
llm-server-path |
Server binary path | botserver-stack/bin/llm/build/bin |
Path |
llm-server-host |
Server bind address | 0.0.0.0 |
IP address |
llm-server-port |
Server port | 8081 |
Number |
llm-server-gpu-layers |
GPU offload layers | 0 |
Number |
llm-server-n-moe |
MoE experts count | 0 |
Number |
llm-server-ctx-size |
Context size | 4096 |
Tokens |
llm-server-n-predict |
Max predictions | 1024 |
Tokens |
llm-server-parallel |
Parallel requests | 6 |
Number |
llm-server-cont-batching |
Continuous batching | true |
Boolean |
llm-server-mlock |
Lock in memory | false |
Boolean |
llm-server-no-mmap |
Disable mmap | false |
Boolean |
Hardware-Specific LLM Tuning
For RTX 3090 (24GB VRAM)
You can run impressive models with proper configuration:
- DeepSeek-R1-Distill-Qwen-7B: Set
llm-server-gpu-layersto 35-40 - Qwen2.5-32B-Instruct (Q4_K_M): Fits with
llm-server-gpu-layersto 40-45 - DeepSeek-V3 (with MoE): Set
llm-server-n-moeto 2-4 to run even 120B models! MoE only loads active experts - Optimization: Use
llm-server-ctx-sizeof 8192 for longer contexts
For RTX 4070/4070Ti (12-16GB VRAM)
Mid-range cards work great with quantized models:
- Qwen2.5-14B (Q4_K_M): Set
llm-server-gpu-layersto 25-30 - DeepSeek-R1-Distill-Llama-8B: Fully fits with layers at 32
- Tips: Keep
llm-server-ctx-sizeat 4096 to save VRAM
For CPU-Only (No GPU)
Modern CPUs can still run capable models:
- DeepSeek-R1-Distill-Qwen-1.5B: Fast on CPU, great for testing
- Phi-3-mini (3.8B): Excellent CPU performance
- Settings: Set
llm-server-mlocktotrueto prevent swapping - Parallel: Increase
llm-server-parallelto CPU cores -2
Recommended Models (GGUF Format)
- Best Overall: DeepSeek-R1-Distill series (1.5B to 70B)
- Best Small: Qwen2.5-3B-Instruct-Q5_K_M
- Best Medium: DeepSeek-R1-Distill-Qwen-14B-Q4_K_M
- Best Large: DeepSeek-V3, Qwen2.5-32B, or GPT2-120B-GGUF (with MoE enabled)
Pro Tip: The llm-server-n-moe parameter is magic for large models - it enables Mixture of Experts, letting you run 120B+ models on consumer hardware by only loading the experts needed for each token!
Local vs Cloud: A Practical Note
General Bots excels at local deployment - you own your hardware, your data stays private, and there are no recurring costs. However, if you need cloud inference:
Groq is the speed champion - They use custom LPU (Language Processing Unit) chips instead of GPUs, delivering 10x faster inference than traditional cloud providers. Their hardware is purpose-built for transformers, avoiding the general-purpose overhead of NVIDIA GPUs.
This isn't about market competition - it's about architecture. NVIDIA GPUs are designed for many tasks, while Groq's chips do one thing incredibly well: transformer inference. If speed matters and you're using cloud, Groq is currently the fastest option available.
For local deployment, stick with General Bots and the configurations above. For cloud bursts or when you need extreme speed, consider Groq's API with these settings:
llm-url,https://api.groq.com/openai/v1
llm-key,your-groq-api-key
llm-model,mixtral-8x7b-32768
Embedding Parameters
| Parameter | Description | Default | Type |
|---|---|---|---|
embedding-url |
Embedding service endpoint | http://localhost:8082 |
URL |
embedding-model |
Embedding model path | Required for KB | Path |
Prompt Parameters
| Parameter | Description | Default | Type |
|---|---|---|---|
prompt-compact |
Context compaction level | 4 |
Number |
prompt-history |
Messages in history | Not set | Number |
Email Parameters
| Parameter | Description | Default | Type |
|---|---|---|---|
email-from |
Sender address | Required for email | |
email-server |
SMTP hostname | Required for email | Hostname |
email-port |
SMTP port | 587 |
Number |
email-user |
SMTP username | Required for email | String |
email-pass |
SMTP password | Required for email | String |
Theme Parameters
| Parameter | Description | Default | Type |
|---|---|---|---|
theme-color1 |
Primary color | Not set | Hex color |
theme-color2 |
Secondary color | Not set | Hex color |
theme-logo |
Logo URL | Not set | URL |
theme-title |
Bot display title | Not set | String |
Custom Database Parameters
| Parameter | Description | Default | Type |
|---|---|---|---|
custom-server |
Database server | localhost |
Hostname |
custom-port |
Database port | 5432 |
Number |
custom-database |
Database name | Not set | String |
custom-username |
Database user | Not set | String |
custom-password |
Database password | Not set | String |
Parameter Types
Boolean
Values: true or false (case-sensitive)
Number
Integer values, must be within valid ranges:
- Ports: 1-65535
- Tokens: Positive integers
- Percentages: 0-100
Float
Decimal values:
- Thresholds: 0.0 to 1.0
Path
File system paths:
- Relative:
../../../../data/model.gguf - Absolute:
/opt/models/model.gguf
URL
Valid URLs:
- HTTP:
http://localhost:8081 - HTTPS:
https://api.example.com
String
Any text value (no quotes needed in CSV)
Valid email format: user@domain.com
Hex Color
HTML color codes: #RRGGBB format
Required vs Optional
Always Required
- None - all parameters have defaults or are optional
Required for Features
- LLM:
llm-modelmust be set - Email:
email-from,email-server,email-user - Embeddings:
embedding-modelfor knowledge base - Custom DB:
custom-databaseif using external database
Configuration Precedence
- Built-in defaults (hardcoded)
- config.csv values (override defaults)
- Environment variables (if implemented, override config)
Special Values
none- Explicitly no value (forllm-key)- Empty string - Unset/use default
false- Feature disabledtrue- Feature enabled
Performance Tuning
For Local Models
llm-server-ctx-size,8192
llm-server-n-predict,2048
llm-server-parallel,4
llm-cache,true
llm-cache-ttl,7200
For Production
llm-server-cont-batching,true
llm-cache-semantic,true
llm-cache-threshold,0.90
llm-server-parallel,8
For Low Memory
llm-server-ctx-size,2048
llm-server-n-predict,512
llm-server-mlock,false
llm-server-no-mmap,false
llm-cache,false
Validation Rules
- Paths: Model files must exist
- URLs: Must be valid format
- Ports: Must be 1-65535
- Emails: Must contain @ and domain
- Colors: Must be valid hex format
- Booleans: Exactly
trueorfalse