LLama.cpp Server
DB-GPT supports native llama.cpp server, which supports concurrent requests and continuous batching inference.
Install dependencies
pip install -e ".[llama_cpp_server]"
DB-GPT supports native llama.cpp server, which supports concurrent requests and continuous batching inference.
pip install -e ".[llama_cpp_server]"