Tokenization and Lemmatization: Core Components of Natural Language Processing.