A token is the smallest unit of text that a language model processes. Rather than reading character by character or word by word, LLMs break text into tokens - which are typically common words, word pieces, or individual characters for rare words.
Understanding tokens is crucial because LLM pricing, context limits, and processing all work in terms of tokens rather than words or characters.