Llama.cpp
A C++ inference engine that enables running large language models locally with minimal setup and state-of-the-art performance across CPU and GPU hardware.
Related Technologies
Competition
Integration
Integrates With
80% confident
text generation webui
Official integration support
Integrates With
80% confident
gpt4all
Official integration support
Integrates With
80% confident
lm studio
Official integration support
Integrates With
70% confident
jan
Official integration support
Integrates With
80% confident
open webui
Official integration support
Compatible With
90% confident
cuda
Works well together
Compatible With
90% confident
metal
Works well together
Compatible With
80% confident
vulkan
Works well together
Dependencies
Built On
100% confident
ggml
Foundation technology
Powers
100% confident
llama cpp python
Enables other tools
Powers
100% confident
node llama cpp
Enables other tools
Powers
100% confident
go llama cpp
Enables other tools
Powers
100% confident
llama cpp rs
Enables other tools
Depends On
100% confident
cmake
Requires for operation
Powers
90% confident
ollama
Enables other tools
Built On
90% confident
koboldcpp
Foundation technology