this post was submitted on 11 Dec 2025
547 points (96.6% liked)

Technology

77631 readers
1584 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] killea@lemmy.world 45 points 2 days ago (2 children)

So in a just world, google would be heavily penalized for not only allowing csam on their servers, but also for violating their own tos with a customer?

[–] shalafi@lemmy.world 19 points 2 days ago (3 children)

We really don't want that first part to be law.

Section 230 was enacted as part of the Communications Decency Act of 1996 and is a crucial piece of legislation that protects online service providers and users from being held liable for content created by third parties. It is often cited as a foundational law that has allowed the internet to flourish by enabling platforms to host user-generated content without the fear of legal repercussions for that content.

Though I'm not sure if that applies to scraping other server's content. But I wouldn't say it's fair for the scraper to review everything. If we don't like that take, then we should illegalize scraping altogether, but I'm betting there are unwanted side effects to that.

While I agree with Section 230 in theory, it is often only used in practice to protect megacorps. For example, many Lemmy instances started getting spammed by CSAM after the Reddit API migration. It was very clearly some angry redditors who were trying to shut down instances, to try and keep people on Reddit.

But individual server owners were legitimately concerned that they could be held liable for the CSAM existing on their servers, even if they were not the ones who uploaded it. The concern was that Section 230 would be thrown out the window if the instance owners were just lone devs and not massive megacorps.

Especially since federation caused content to be cached whenever a user scrolled past another instance’s posts. So even if they moderated their own server’s content heavily (which wasn’t even possible with the mod tools that existed at the time), then there was still the risk that they’d end up cacheing CSAM from other instances. It led to a lot of instances moving from federation blacklists to whitelists instead. Basically, default to not federating with an instance, unless that instance owner takes the time to jump through some hoops and promises to moderate their own shit.

[–] killea@lemmy.world 1 points 1 day ago

Oh my, yes, you are correct. That was sort of knee jerk, as opposed to it being the reporting party's burden somehow. I simply cannot understand the legal gymnastics needed to punish your customers for this sort of thing; I'm tired but I feel like this is not exactly an uncommon occurrence. Anyways let us all learn from my mistake and do not be rash and curtail your own freedoms.

[–] vimmiewimmie@slrpnk.net 2 points 2 days ago

Not to create an argument, which isn't my intent, as certainty there may be a thought such as, "scraping as it stands is good because of the simplification and 'benefit'". Which, sure, it's easiest to wide net and absorb, to simply the concept, at least as I'm also understanding it.

Yet, maybe it is the process of scraping, and also absorbing into databases including AI, which is a worthwhile point of conversation. Maybe how we've been doing something isn't the continued 'best course' for a situation.

Undeniably, more minutely monitoring what is scraped and stored creates large quantities, and large in scope, of questions and obstacles, but, maybe having that conversation is where things should go.

Thoughts?

[–] abbiistabbii@lemmy.blahaj.zone 5 points 2 days ago (4 children)

This, literally the only reason I could guess is that it is to teach AI to recognise childporn, but if that is the case, why is google going it instead of like, the FBI?

[–] forkDestroyer 1 points 1 day ago

Google isn't the only service checking for csam. Microsoft (and other file hosting services, likely) also have methods to do this. This doesn't mean they also host csam to detect it. I believe their checks use hash values to determine if a picture is already clocked as being in that category.

This has existed since 2009 and provides good insight on the topic, used for detecting all sorts of bad category images:

https://technologycoalition.org/news/the-tech-coalition-empowers-industry-to-combat-online-child-sexual-abuse-with-expanded-photodna-licensing/

[–] gustofwind@lemmy.world 6 points 2 days ago (1 children)

Who do you think the FBI would contract to do the work anyway 😬

Maybe not Google but it would sure be some private company. Our government doesn’t do stuff itself almost ever. It hires the private sector

guess i gotta get into the private sector, lmao

[–] alias_qr_rainmaker@lemmy.world 2 points 2 days ago* (last edited 2 days ago) (1 children)

i know it's really fucked up, but the FBI needs to train an AI on CSAM if it is to be able to identify it.

i'm trying to help, i have a script that takes control of your computer and opens the folder where all your fucked up shit is downloaded it's basically a pedo destroyer. they all just save everything to the downloads folder of their tor browser, so the script just takes control of their computer, opens tor, and pressed cmd+j to open up downloads and then it copies the files names and all that.

will it work? dude, how the fuck am i supposed to know, i don't even do this shit for a living

i'm trying to use steganography to embed the applescript in a png

[–] vimmiewimmie@slrpnk.net 1 points 2 days ago (1 children)
[–] alias_qr_rainmaker@lemmy.world 2 points 2 days ago (1 children)

the applescript opens tor from spotlight search and presses the shortcut to open downloads

i dunno how much y'all know about applescript. it's used to automate apps on your mac. i know y'all hate mac shit but dude, whatever, if you get osascript -e aliased to o you can run applescript easily from your terminal

just pass in a heredoc

[–] frongt@lemmy.zip 1 points 2 days ago (1 children)

Google wants to be able to recognize and remove it. They don't want the FBI all up in their business.

[–] Allero@lemmy.today 1 points 1 day ago* (last edited 1 day ago)

So, Google could be allowed to have the tools to collect, store, and process CSAM all over the Web without oversight?

Pretty much everyone else would get straight to jail for attempting that.