Skip to main content


The llama provider is compatible with the HTTP server bundled with llama.cpp. This allows you to leverage the power of llama.cpp models within Promptfoo.


To use the llama provider, specify llama as the provider in your promptfooconfig.yaml file.

Supported environment variables:

  • LLAMA_BASE_URL - Scheme, hostname, and port (defaults to http://localhost:8080)

For a detailed example of how to use Promptfoo with llama.cpp, including configuration and setup, refer to the example on GitHub.