> ## Documentation Index
> Fetch the complete documentation index at: https://langwatch.ai/docs/llms.txt
> Use this file to discover all available pages before exploring further.

# How to evaluate an LLM when you don't have defined answers

> Measure LLM performance using LLM-as-a-Judge when no ground-truth answers exist to support scalable AI agent evaluations.

For some AI applications, it's not really possible to define a golden answer, this happens for example in creative tasks, where it's hard to define a single correct answer.

On the video below, we show how to use the LangWatch Experiments via UI to evaluate a Business Coaching Agent, where we don't have defined answers, but we can use an LLM-as-a-judge to evaluate the quality of the answers:

<iframe width="720" height="420" src="https://www.youtube.com/embed/PQPGOJqYR2Q" title="YouTube video player" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture" allowFullScreen />
