When talking about artificial intelligence and language models like GPT-4, it's essential to understand what tokens are. To explain it in a simple way, let's take a well-known example from the book The Little Prince by Antoine de Saint-Exupéry.
Let's say you're reading this famous passage:
“Please... tame me! he said. “I want to,” replied the little prince, “but I don't have much time. I have friends to discover and lots of things to know.”
When we read this text, our brain processes every word, every space, every punctuation mark. For an AI model like GPT-4, it's exactly the same, except that it cuts the text into Tokens.
One Token is a unit of text that can be:
Each time you interact with an AI model, it breaks down your request into tokens.
These two elements consume tokens. For example, if you ask a 100 token question and the AI generates a 200 token response, the total request is 300 tokens.
AI models like GPT-4 have a token limit per request. For example, GPT-4 can process up to 8192 tokens per interaction. The longer your message is, the fewer tokens are left available for a detailed response.
In API, this concept is particularly important because each token consumed has a cost. Optimizing your messages therefore allows you to reduce your costs while obtaining more relevant answers.
Understanding what a token is is essential to effectively interact with an artificial intelligence model. Whether it's to structure your requests or reduce your API costs, this concept helps you better control your exchanges with AI.