Introduction
Large Language Models (LLMs) like GPT-4 don't understand text the way humans do. Instead, they break it down into smaller pieces called tokens — which can be words, parts of words, punctuation, or even spaces. Each token is assigned a unique token ID (integer), which the model uses internally to understand and generate language.