A lightweight chat terminal-interface for llama.cpp server written in C++ with many features and windows/linux support.
-
Updated
May 31, 2025 - C++
A lightweight chat terminal-interface for llama.cpp server written in C++ with many features and windows/linux support.
This is a Bash script to automatically launch llama-server, detects available .gguf models, and selects GPU layers based on your free VRAM.
Add a description, image, and links to the llama-server topic page so that developers can more easily learn about it.
To associate your repository with the llama-server topic, visit your repo's landing page and select "manage topics."