Skip to content

BLEU Score

A metric for evaluating machine-generated text by comparing n-gram overlap with reference texts. Originally designed for machine translation, BLEU is widely used but has known limitations in capturing semantic meaning and fluency.

Related terms

BenchmarkMachine Translation
← Back to glossary