this post was submitted on 29 Aug 2025
491 points (98.4% liked)
Technology
74585 readers
3951 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Its just an API.
There's a few ways they could go about it. They could have part of the prompt be something like "when the customer is done taking their order, create a JSON file with the order contents" and set up a dumb register essentially that looks for those files and adds that order like a standard POS would.
They could spell out a tutorial in the prompt, "to order a number 6 meal, type "system.order.meal(6)" calling the same functions that a POS system would, and have that output right to a terminal.
They could have their POS system be open on an internal screen, and have a model that can process images, and have it specify a coordinate pair, to simulate a touch screen, and make it manually enter an order that way as an employee would.
There's lots of ways to hook up the AI, and it's not actually that different from hooking up a normal POS system in the first place, although just because one method does allow an AI to interact doesn't mean it'll go about it correctly.
LLMs, with a little coaxing, perform well at returning well formed JSON.
They do, my concern is more about if that JSON is correct, not just well-formed.
Also, 18000 waters might be correct JSON, but makes an AI a bad cashier.
There is a lot more that goes into it than just being correct. 18000 waters may have been the actual order, because somebody decided to screw with the machine. A human who isn't terminally autistic would reliably interpret that as a joke and would simply refuse to punch that in. The LLM will likely do what a human tells it to do, since it has no contextual awareness, it only has the system prompt and whatever interaction with the user it had so far.
Thats part of correctness to me, delivering an order that taco bell actually would make is important.
Semantics aside, though, we agree. That's very important.
So they just trim the instructions so it doesn't take joke orders, so it can make more reasonable decisions, like:
"May I take your order?"
"Two double whoppers with extra mayo and a chocolate cherry banana sundae"
"Oh you've GOTTA be joking!"
It's trivial to get LLMs to act against the instructions