Anyone even believing that a generic word auto completer would beat classic algorithms wherever possible probably belongs into a psychiatry.
Technology
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
There are a lot of people out there that think LLM's are somehow reasoning. Even reasoning models aren't really doing it. It important to do demonstrations like this in the hopes that the general public will understand the limitations of this tech.
It is important to do demonstrations like this in the hopes that the general public will understand the limitations of this tech.
THIS is the thing. The general public's perception of ChatGPT is basically whatever OpenAI's marketing department tells them to believe, plus their single memory of that one time they tested out ChatGPT and it was pretty impressive. Right now, OpenAI is telling everyone that they are a few years away from Artificial General Intelligence. Tests like this one demonstrate how wrong OpenAI is in that assertion.
It's almost as bad as the opposition's comparison of it to Skynet. People are never going to understand technology without applying some fucking nuance.
Stop hyping new technology... in either direction.
I think the problem is that, while the model isn't actually reasoning, it's very good at convincing people it actually is.
I see current LLMs kinda like an RPG character build with all ability points put into Charisma. It's actually not that good at most tasks, but it's so good at convincing people that they start to think it's actually doing a great job.
I think I remember some doge goon asking online about using an LLM to parse JSON. Many people don't understand things.
That’s too much critical thinking for most people
In a quite unexpected turn of events, it is claimed that OpenAI’s ChatGPT “got absolutely wrecked on the beginner level” while playing Atari Chess.
Who the hell thought this was "unexpected"?
What's next? ChatGPT vs. Microwave to see which can make instant oatmeal the fastest? 😂
Considering how much heat the servers probably generate, ChatGPT might have a decent chance in that competition 😁
Air-fried oatmeal, FTW!
A simple calculator will also beat it at math.
Atari game programmed to know chess moves: knight to B4
Chat-GPT: many Redditors have credited Chesster A. Pawnington with inventing the game when he chased the queen across the palace before crushing the king with a castle tower. Then he became the king and created his own queen by playing "The Twist" and "Let's Twist Again" at the same time.
This article makes ChatGPT sound like a deranged blowhard, blaming everything but its own ineptitude for its failure.
So yeah, that tracks.
Isn’t this kind of like ridiculing that same Atari for not being able to form coherent sentences? It’s not all that surprising that a system not designed to play chess loses to a system designed specifically for that purpose.
Pretty much, but the marketers are still trying to tell people it can totally do logic anyway. Hopefully the apple paper opens some eyes
For anyone wondering what "the" apple paper is: https://machinelearning.apple.com/research/illusion-of-thinking
A PE teacher got absolutely wrecked by a former Olympic sprinter at a sprint competition.
Change "PE teacher" to "stack of health magazines" and it's a more accurate equivalence.
Well... yeah. That's not what LLMs do. That's like saying "A leafblower got absolutely wrecked by 1998 Dodge Viper in beginner's drag race". It's only impressive if you don't understand what a leafblower is.
People write code with LLMs. Programming language is just a language specialised at precise logic. That’s what „AI” is advertised to be good at. How can you do that an not the other?
It's not very good at it though, if you've ever used it to code. It automates and eases a lot of mundane tasks, but still requires a LOT of supervision and domain knowledge to not have it go off the rails or hallucinate code that's either full of bugs or will never work. It's not a "prompt and forget" thing, not by a long shot. It's just an easier way to steal code it picked up from Stackoverflow and GitHub.
Me as a human will know to check how much data is going into a fixed size buffer somewhere and break out of the code if it exceeds it. The LLM will have no qualms about putting buffer overflow vulnerabilities all over your shit because it doesn't care, it only wants to fulfill the prompt and get something to work.
I’m not saying it’s good at coding, I’m saying it’s specifically advertised as being very good at it.
"Precise logic" is specifically what AI is not any good at whatsoever.
AI might be able to write a program that beats an A2600 in chess, but it should not be expected to win at chess itself.
I shall await the moment when AI pretends to be as confident about communicating not being able to do something as it is with the opposite because it looks like it’s my job somehow.
Yeah, LLMs seem pretty unlikely to do that, though if they figure it out that would be great. That's just not their wheelhouse. You have to know enough about what you're attempting to ask the right questions and recognize bad answers. The thing you're trying to do needs be within your reach without AI or you are unlikely to be successful.
I think the problem is more the over-promising what AI can do (or people who don't understand it at all making assumptions because it sounds human-like).