this post was submitted on 19 Nov 2025
6 points (87.5% liked)

AI News

66 readers
2 users here now

This community is for posting articles covering AI.

https://lemmy.world/c/AIGenerated to post any content generated using AI.

founded 1 month ago
MODERATORS
all 6 comments
sorted by: hot top controversial new old
[–] T00l_shed@lemmy.world 5 points 1 month ago (1 children)

Is it easier to not use? That's my most important benchmark when it comes to llms

[–] jaykrown@lemmy.world -1 points 4 weeks ago (2 children)

That's what the benchmarks attempt to find out, I think it depends on what you're trying to do as well. Easier to use for what?

[–] T00l_shed@lemmy.world 4 points 4 weeks ago

Which one is the easiest not to use. Llm are catastrophic for the environment

[–] TonyTonyChopper@mander.xyz 1 points 4 weeks ago

If all goes well: nothing

[–] brucethemoose@lemmy.world 1 points 4 weeks ago* (last edited 4 weeks ago)

It seems to have regressed vs Gemini 2.5 in some long context comprehension, like asking stuff about papers or stories... Which is basically the only thing I use Gemini for, since open/local models are so good at shorter contexts now.

This isn't suprising. For that stuff, Gemini's peak was somewhere in the 2.0/2.5 previews, but then they deep-fried it to benchmaxx coding and lm-arena.