Words and Morphology¶ Tokenization: the process that converts running text (i.e., a sequence of characters) into a sequence of tokens.