Keywords AI
DeepEval is an open-source LLM evaluation framework built for unit testing AI outputs. It provides 14+ evaluation metrics including hallucination detection, answer relevancy, and contextual recall. Integrates with pytest, supports custom metrics, and works with any LLM provider for automated quality assurance in CI/CD pipelines.
Top companies in Observability, Prompts & Evals you can use instead of DeepEval.
Companies from adjacent layers in the AI stack that work well with DeepEval.