Back to Opik

Summarization coherence

apps/opik-documentation/documentation/fern/docs-v2/evaluation/metrics/summarization_coherence.mdx

2.0.22-6605-merge-20651.4 KB
Original Source

Summarization Coherence Judge

SummarizationCoherenceJudge evaluates the writing quality of a summary: structure, clarity, and logical flow. It complements SummarizationConsistencyJudge by focusing on how the summary reads rather than whether it is factual, returning a 0.0–1.0 score derived from a raw 0–10 judgement.

python
from opik.evaluation.metrics import SummarizationCoherenceJudge

metric = SummarizationCoherenceJudge()

score = metric.score(
    output="""SUMMARY: First, the product launched. Revenue grew. Margins fell. Next steps TBD.""",
)

print(score.value)   # 0.0–1.0 after normalisation
print(score.reason)

Inputs

ArgumentTypeRequiredDescription
outputstrYesSummary text to evaluate.
inputstrOptionalOriginal document/talk track for additional context (not required).

Configuration

ParameterDefaultNotes
modelgpt-5-nanoUpgrade when assessing long-form or domain-specific summaries.
temperature0.0Raise slightly (≤0.3) to expose diverse stylistic critiques.
trackTrueToggle off to skip logging.
project_nameNoneOverride when tracking across projects.

Pair this judge with SummarizationConsistencyJudge to ensure summaries are both factual and easy to skim. The evaluator returns a 0–10 integer that Opik normalises to 0–1.