Process Tokens
Original data 237, 091
Tokenization 34, 770
Token normalization 15, 051
Stop words 14, 847
Stemming 10, 878