AIM Banners_978 x 90

It is Stupid to Ask How Many R’s ‘Strawberry’ Has

LLMs can’t count letters directly because they process text in chunks called “tokens”.
Whenever a new LLM is released, users tend to quiz it first with basic questions like: “How many R’s does 'Strawberry' have?” or “Which one is bigger – 9.9 or 9.11?”.  Most models, like GPT-3.5, Claude, and Llama, get the answer wrong. The problem starts when users try to benchmark the reasoning capabilities of a model based on these questions.  Steve Wilson, the CEO of Exabeam, explained that when an LLM processes a word, it doesn’t see it as individual letters but rather as tokens. These tokens may represent entire words or subword units, depending on the model’s design.  “For example, “strawberry” might be broken into one or more tokens that don’t directly correspond to individual letters. This can lead to errors when trying to analyse or
Subscribe or log in to Continue Reading

Uncompromising innovation. Timeless influence. Your support powers the future of independent tech journalism.

Already have an account? Sign In.

📣 Want to advertise in AIM? Book here

Picture of Sagar Sharma
Sagar Sharma
A software engineer who loves to experiment with new-gen AI. He also happens to love testing hardware and sometimes they crash. While reviving his crashed system, you can find him reading literature, manga, or watering plants.
Related Posts
AIM Print and TV
Don’t Miss the Next Big Shift in AI.
Get one year subscription for ₹5999
Download the easiest way to
stay informed