.png)
AI & UXR, TOKEN, LLM
Why AI Sometimes Can’t Count to 3 – And What That Has to Do With Tokens
3
MIN
Sep 25, 2025
When AI miscounts
‘How many rs are there in strawberry?’ – a simple question, right?
Not for language models. For a long time, the common AI answer was ‘2’. Anyone who counts along quickly realises that this is wrong. Strawberry contains three rs – clearly.
It's a mistake that seems so absurdly simple that it makes you wonder: how can a highly developed language model like ChatGPT fail at this?
The answer takes us to the heart of language model architecture – more specifically, to the world of tokenisation. And that's more fascinating than it seems at first glance.
What is actually happening here?
Language models such as ChatGPT do not really count. Nor do they analyse letters, at least not in the way we would.
Instead, they break down text into so-called tokens – smaller units with which the model was trained. Tokens can be an entire word, part of a word or even just a syllable.
And here's the crux of the matter: ‘r’ is not a token.
The model does not “see” the letter individually, but embedded in larger text segments. It only recognises the ‘r’ if explicitly prompted to do so – and even then, it may be wrong depending on the prompt and context.
This is because language models do not work deterministically, but probabilistically: they guess what is probably meant – and not necessarily what would be mathematically correct.
Tokenisation using the example of ‘strawberry’
The word strawberry, for example, is broken down into exactly two tokens by GPT-4o:
[“straw”, ‘berry’]
This means that the model recognises strawberry as two typical word components. And the ‘r’? It is contained in both tokens – but never in isolation. The language model does not count letters, but probability-based clusters of meaning.
So anyone who asks how many rs there are in strawberry is asking an accounting question of a semantic probability model. No wonder it often got it wrong in the past.
Even more exciting: German words
German language, difficult tokens: our beloved compound words are a real stress test for tokenisers. But surprisingly, GPT-4o doesn't do too badly here:
Word | Token | Number |
Herausforderung (Challenge) | ["Hera", "us", "ford", "er", "ung"] | 5 |
Krankenhausaufenthalt (Hospital stay) | ["Kranken", "haus", "auf", "ent", "halt"] | 5 |
Datenschutzgrundverordnung (General Data Protection Regulation) | ["Datenschutz", "grund", "ver", "ord", "nung"] | 5 |
Arbeitszeiterfassungspflicht (Obligation to record working hours) | ["Arbeits", "zeit", "er", "fass", "ungs", "pflicht"] | 6 |
Selbstverständlichkeit (Matter of course) | ["Selbst", "ver", "ständ", "lich", "keit"] | 5 |
This shows that the tokeniser recognises many meaningful units – such as ‘-ity’, ‘-ity’, “self”, ‘evident’ – and splits compounds in a semantically intelligent way.
But here, too, the same applies: no model counts letters. It recognises, processes and combines tokens. Only if it is additionally trained or prepared for this with an example can it count correctly.
Why this is important (also for UX & prompting)
These small counting errors tell a big story – about the nature of language models.
Language models do not calculate; they probabilistically rank language. They can be brilliant at leaps in meaning and nuances, but they can also be completely off the mark when it comes to simple structural questions – such as counting letters, alphabetical sorting or mathematical sequences.
When it comes to prompting, this means:
If precision is important (e.g. for counting, formatting, extraction) → formulate tasks very clearly
If hallucinations are to be avoided → provide examples
If UX researchers work with AI → keep the behaviour of the tokeniser in mind
Because many supposed ‘errors’ are actually consequences of the architecture. And once you understand that, you can write much better prompts – and get better results.
Conclusion: Tokens are the new semantics – or the new tripwire
The error with the “r” in strawberry is not a trivial bug – it is an invitation to better understand language models.
Anyone who works with AI should be aware that:
AI does not understand letters – it understands tokens.
AI does not count – it estimates probabilities.
AI is not stupid – it is just trained differently.
Those who know this are less likely to stumble over simple tasks – and get more out of complex prompts.
Bonus: Try it yourself
🔧 Tool tip
If you want to try for yourself how words are broken down into tokens, you can use this OpenAI tool, for example:
👉 https://platform.openai.com/tokenizer
🧠 Prompt tip for counting
‘Please count exactly how many times the letter “r” appears in the following word: strawberry. Just give me the number.’
📣 Participatory question
How many s are there in ‘Mississippi’?
💌 Not enough? Then read on – in our newsletter. It comes four times a year. Sticks in your mind longer. To subscribe: https://www.uintent.com/newsletter
RELATED ARTICLES YOU MIGHT ENJOY
AUTHOR
Tara Bosenick
Tara has been active as a UX specialist since 1999 and has helped to establish and shape the industry in Germany on the agency side. She specialises in the development of new UX methods, the quantification of UX and the introduction of UX in companies.
At the same time, she has always been interested in developing a corporate culture in her companies that is as ‘cool’ as possible, in which fun, performance, team spirit and customer success are interlinked. She has therefore been supporting managers and companies on the path to more New Work / agility and a better employee experience for several years.
She is one of the leading voices in the UX, CX and Employee Experience industry.




















