looking through the llama.cpp source code, to learn about language models. Disclaimer: it's january 2026, so things will likely change, at some point.
- llama.cpp repo link
- build instructions
- contains instructions on how to build the debug version
- default build creates devices: 'libggml-base.so' and 'libggml-cpu.so' - probably need to set up additional requirements for the other backends.
- if you need additional backends: documentation on each backend includes build instructions
- looking at main.cpp for llama-simple - a CLI program to continue with a given prompt, specified on the command line. command line:
./llama-simple -m model.gguf [-n n_predict] [-ngl n_gpu_layers] {prompt-text}