These tools integrates with
DeepEvalvsLangfuse
LLM evaluation framework — 14+ metrics versus OSS LLM engineering platform
Compare interactively in Explore →Choose DeepEval when…
- •You want a pytest-style framework for LLM testing
- •Unit-test-like evals for LLM outputs fit your workflow
- •You need RAG-specific metrics like faithfulness and relevancy
Choose Langfuse when…
- •You want open-source LLM observability
- •Self-hosting your tracing stack is important
- •You need cost tracking across models and users
Side-by-side comparison
Field
DeepEval
Langfuse
Category
Prompt & Eval
LLM Infrastructure
Type
Open Source
Open Source
Free Tier
✓ Yes
✓ Yes
Pricing Plans
—
Cloud: $59/mo
GitHub Stars
⭐ 5,500
⭐ 7,000
Health
●80 — Active
●80 — Active
DeepEval
Open-source evaluation framework with 14+ metrics including faithfulness, relevancy, and hallucination detection. Integrates with CI/CD.
Shared Connections3 tools both integrate with
Only DeepEval (4)
LangfuseTruLensInspectGalileo
Only Langfuse (27)
Claude CodeCrewAILangGraphLangChainCursorLlamaIndexOpenHandsDifyMastraAutoGen
Explore the full AI landscape
See how DeepEval and Langfuse fit into the bigger picture — 207 tools, 452 relationships, all mapped.