Posts
Comments
Comment by
almath123 on
Forecasting progress in language models ·
2021-10-29T18:21:25.405Z ·
LW ·
GW
Perplexity depends on the vocabulary and is sensitive to preprocessing which could skew the results presented here. This is a common problem. See the following reference:
Unigram-Normalized Perplexity as a Language Model Performance Measure with Different Vocabulary Sizes Jihyeon Roha, Sang-Hoon Ohb, Soo-Young Lee, 2020