BLEU Score
Bilingual Evaluation Understudy - a metric for evaluating machine translation quality by comparing generated text to reference translations. Measures n-gram overlap. Scores range from 0 to 1 (or 0-100). Higher is better, but BLEU has limitations for creative or diverse outputs.