LLMs don't think, reason, or understand. They predict the next token.
What's a token?
What the model does:
Given: "The capital of France is"
Model predicts: "Paris" (most likely next token based on training)
That's it. Every single thing an LLM does is just repeating this process over and over.
Marketing says: "The model thinks through the problem step by step"
Reality: The model generates more tokens that look like thinking before giving the final answer.
Question: "What is 547 + 832?"
Without "thinking" (direct prediction):
Input: "What is 547 + 832?"
↓
Model predicts: "1379"