Instead of downloading pre-built binaries (which may require AVX2), compile llama.cpp from source during installation. This ensures: - Works on older CPUs (Sandy Bridge, Haswell, etc.) - Uses GGML_NATIVE=ON to optimize for the current CPU - Binary path updated to build/bin/llama-server Reverts the AVX2 detection that was incorrectly disabling LLM. |
||
|---|---|---|
| .. | ||
| context | ||
| llm_models | ||
| prompt_manager | ||
| cache.rs | ||
| cache_test.rs | ||
| episodic_memory.rs | ||
| local.rs | ||
| mod.rs | ||
| observability.rs | ||