by Michele Laurelli
Measurement of how well a probability model predicts a sample, for evaluating language models.
Perplexity = exp(cross_entropy). Lower perplexity indicates better prediction. Commonly used for LM evaluation.
Language model evaluation
GPT performance metric
Text generation quality