this post was submitted on 24 Jan 2024
246 points (90.2% liked)
Open Source
31114 readers
1 users here now
All about open source! Feel free to ask questions, and share news, and interesting stuff!
Useful Links
- Open Source Initiative
- Free Software Foundation
- Electronic Frontier Foundation
- Software Freedom Conservancy
- It's FOSS
- Android FOSS Apps Megathread
Rules
- Posts must be relevant to the open source ideology
- No NSFW content
- No hate speech, bigotry, etc
Related Communities
- !libre_culture@lemmy.ml
- !libre_software@lemmy.ml
- !libre_hardware@lemmy.ml
- !linux@lemmy.ml
- !technology@lemmy.ml
Community icon from opensource.org, but we are not affiliated with them.
founded 6 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It's an interesting debate isn't it? Does AI transform something free into something that's not? Or does it simply study the code?
There's no debate. LLMs are plagiarism with extra steps. They take data (usually illegally) wholesale and then launder it.
A lot of people have been doing research into the ethics of these systems and that's more or less what they found. The reason why they're black boxes is precisely the reason we all suspected; they were made that way because if they weren't we'd all see them for what they are.
Can you link it please? I'd like to inform myself.
I doubt they have a factual basis for their opinion, considering
Is just plain wrong. Researchers would love to have a non black box AI (i.e. a white box AI), but it's unfortunately impossible with the current architecture.
Their use of language also feels more emotional and if anything it makes me more skeptical.