The process of converting text into tokens that can be fed into a Large Language Model (LLM).
For example, tokenizing the sentence ‘The quick brown fox jumps over the lazy dog’ results in the following tokens: [‘The’, ‘quick’, ‘brown’, ‘fox’, ‘jumps’, ‘over’, ‘lazy’, ‘dog’].