Once installed, you'll need a model to work with. Head to the Obtaining and quantizing models section to learn more. The main goal of llama.cpp is to enable LLM inference with minimal setup and ...