BLEU Score
A metric for evaluating machine-generated text by comparing n-gram overlap with reference texts. Originally designed for machine translation, BLEU is widely used but has known limitations in capturing semantic meaning and fluency.
A metric for evaluating machine-generated text by comparing n-gram overlap with reference texts. Originally designed for machine translation, BLEU is widely used but has known limitations in capturing semantic meaning and fluency.