this post was submitted on 12 Aug 2025
18 points (90.9% liked)

Hacker News

2472 readers
368 users here now

Posts from the RSS Feed of HackerNews.

The feed sometimes contains ads and posts that have been removed by the mod team at HN.

founded 11 months ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] theunknownmuncher@lemmy.world -2 points 2 weeks ago* (last edited 2 weeks ago) (3 children)

As if a stochastic parrot could ever be able to count or calculate…

https://www.anthropic.com/research/tracing-thoughts-language-model

well that's provably false as the neural network can be traced to understand the exact process and calculations that it follows when doing math. See the above research under heading "Mental Math".

The reason LLMs struggle to count letters is because of tokenization. The text is first converted to tokens, numeric vectors which represents whole words or parts of words, before being given to the LLM, and the LLM has no concept of or access to the underlying letters that make up the words. The LLM outputs only tokens, which are converted back into text.

EDIT: you can downvote me, but facts are facts lol.

[–] jacksilver@lemmy.world 5 points 2 weeks ago (1 children)

There is definitely more going on with LLMs than just direct parroting.

However, there is also an upper limit to what an LLM can logic/calculate. Since LLMs basically boil down to a series of linear operations, there is an upper bound on all of them as to how accurately they can calculate anything.

Most chat systems use python behind the scene for any actual math, but if you run a raw LLM you can see the errors grow faster as you look at higher orders of growth (addition, multiplication, powers, etc.).

[–] theunknownmuncher@lemmy.world -1 points 2 weeks ago* (last edited 2 weeks ago)

Yes, exactly. It can do basic math and also doesn't really matter because it is really good at interfacing with tools/functions for calculation anyway

However, there is also an upper limit to what an LLM can logic/calculate. Since LLMs basically boil down to a series of linear operations, there is an upper bound on all of them as to how accurately they can calculate anything.

Also this is only true when LLMs are not using Chain of Thought.

[–] Onomatopoeia@lemmy.cafe 2 points 2 weeks ago (1 children)

So what you're saying is it can't count.

Got it.

[–] theunknownmuncher@lemmy.world 2 points 2 weeks ago* (last edited 2 weeks ago)

Reading comprehension. The only thing it cannot count are specifically letters of tokenized words. If you separate the letters out into separate tokens ("B L U E B E R R Y", etc) it will get it correct 100% of the time.