Llama.cpp
The llama
provider is compatible with the HTTP server bundled with llama.cpp. This allows you to leverage the power of llama.cpp
models within Promptfoo.
Configuration
To use the llama
provider, specify llama
as the provider in your promptfooconfig.yaml
file.
Supported environment variables:
LLAMA_BASE_URL
- Scheme, hostname, and port (defaults tohttp://localhost:8080
)
For a detailed example of how to use Promptfoo with llama.cpp
, including configuration and setup, refer to the example on GitHub.