Ollama requires Ollama to be installed and running on your machine before you can use it with evalflow. Pull the model you intend to use with
ollama pull <model> before running evals.Configure evalflow.yaml
Add anollama block under providers. Because Ollama runs locally, you still set api_key_env in the config — but you do not need to export a real key. The field must be present for evalflow to validate your config:
Set the placeholder environment variable
Start the Ollama server
evalflow communicates with Ollama over its local HTTP server. Start it before running any evals:Verify the connection
Runevalflow doctor with the --check-providers flag to confirm evalflow can reach the local Ollama server:
Run evals
eval.default_provider is already set to ollama in your evalflow.yaml, you can omit the --provider flag:
Provider notes
- Default model:
llama3.2. You can setdefault_modelto any model you have pulled withollama pull. - Offline support: Once a model is pulled, evals run fully offline. No network connection is required.
- No billing: Ollama is free and open source. There are no API costs or rate limits.
- Performance: Inference speed depends on your hardware. Running large models on CPU will be significantly slower than GPU-accelerated hardware.
- Judge model: By default, evalflow uses Groq as the LLM judge, which requires network access. To keep everything local, configure Ollama as the judge provider too: