Definition
A metric for evaluating machine-translated text against reference translations.
Detailed Explanation
Bilingual Evaluation Understudy (BLEU) score measures the quality of machine translation by comparing it to human reference translations. It considers the overlap of n-grams between the candidate and reference translations, with penalties for length mismatches. Scores range from 0 to 1.
Use Cases
Essential in machine translation systems, language generation evaluation, and automated content creation assessment.