Prompt & EvalOpen Source✦ Free Tier

PromptFoo

CLI/library for prompt testing and red-teaming

5,000 stars● Health 80ActiveApp Infrastructure

About

Test and compare prompts across models. Built-in red-teaming, regression testing, and side-by-side model comparison.

Choose PromptFoo when…

  • You want CLI-first, config-driven LLM evals
  • Running eval suites in CI/CD pipelines is a goal
  • You need red-teaming and safety testing built in

Builder Slot

How do you know it's working?Optional for most stacks

Tests, evals, and experiment tracking to measure and improve your AI output quality

Dev Tools
Not applicable
App Infra
Recommended
Hybrid
Optional

Other tools in this slot:

Stack Genome Detection

AIchitect's Genome scanner detects PromptFoo in your project via these signals:

npm packages
promptfoo
config files
promptfooconfig.yamlpromptfooconfig.yml

Integrates with (2)

LangfuseLLM Infrastructure

Langfuse production traces can be exported as eval datasets that Promptfoo uses for regression testing in CI.

Close the eval loop: real failures captured in Langfuse become the regression test cases Promptfoo runs on every deploy.

Compare →
OpenAI APILLM Infrastructure

Promptfoo calls OpenAI's API directly to run prompts through configured test cases and compare outputs against assertions.

Automated prompt regression testing against GPT-4o — catch output quality changes before they reach production.

Compare →

Often paired with (1)

Alternatives to consider (3)

Pricing

✦ Free tier available

In 4 stacks

Badge

Add to your GitHub README

PromptFoo on AIchitect[![PromptFoo](https://aichitect.dev/badge/tool/promptfoo)](https://aichitect.dev/tool/promptfoo)

Explore the full AI landscape

See how PromptFoo fits into the bigger picture — browse all 207 tools and their relationships.

Explore graph →