Perplexity

A metric measuring how well a language model predicts text. Lower perplexity indicates better prediction. Calculated as the exponential of the average negative log-likelihood. While useful for comparing models, it doesn't always correlate with human-perceived quality.