Learn Before
n-grams
Perplexity
Perplexity is a probability-based metric for evaluating language models. It is the weighted average of the number of possible next words that can follow any word, a.k.a. the weighted average branching factor.
Given a mini-language of 10 words "zero, one ... ten", each word's occurrence probability is 1/10 (unigram), the perplexity is the inverse is 10:
0
1
4 years ago
Tags
Natural language processing
Data Science
Related
Markov used in NLP
MLE & Normalizing
Perplexity
General equation of n-gram model