Visualize how n-gram coverage drops exponentially as sequence length increases
Exponential growth of possible combinations vs. linear growth of training data
Percentage of possible n-grams actually seen in training
As n-gram size increases, the number of possible combinations grows exponentially, but training data grows linearly. This creates massive "holes" in our knowledge.