Skip to main content

C/C++ implementation for running quantized LLMs on consumer hardware (CPU, GPU, Apple Silicon).extracted from the official website or Wikipedia.

Learn llama.cpp

Recommended resources to get started

Let's Connect

Interested in this technology?

Feel free to reach out if you would like to discuss this technology or explore how it can be applied to your projects.