FaceDeer

joined 2 years ago
[–] FaceDeer@fedia.io 4 points 1 year ago

They invented Bacta about 20 minutes after Qui-Gon got stabbed. Bad luck.

[–] FaceDeer@fedia.io -5 points 1 year ago (11 children)

You can turn off Recall with a simple toggle in the settings.

There's no need to switch operating systems, just turn it off.

[–] FaceDeer@fedia.io 0 points 1 year ago (3 children)
  • Stable
  • Lots of features
  • Very widespread support
[–] FaceDeer@fedia.io 15 points 1 year ago (1 children)

The problem appears to be that the magazine deceptively portrayed it as an interview with the actual Michael Schumacher, rather than explaining that it was fictional. The lawsuit would probably be the same if the magazine had had a human writer come up with it all instead.

[–] FaceDeer@fedia.io 13 points 1 year ago

Yeah, it's unclear whether copyright is even relevant when it comes to training AI. It feels a lot like people who feel very strongly about intellectual property but have clearly confused trademarks, patents, copyright, and maybe even regular old property law - they've got an idea of what they think is "right" and "wrong" but it's not closely attached to any actual legal theory.

[–] FaceDeer@fedia.io 2 points 1 year ago

Well, I haven't gone to any of my image AIs and actually asked them to generate naked pictures of young people. So unless you want to go there this will necessarily involve some degree of theoretical elements.

However, according to the article it's possible to generate this stuff with Stable Diffusion models, and Stable Diffusion models have a negligible amount of CSAM in the training set. So short of actually doing the experiment that would seem to settle it.

I think a lot of people don't appreciate just how surprisingly sophisticated the "world model" that these image AIs have learned is. There was a paper a while back where some researchers were trying to analyze how image generators were working internally, and they discovered that if you were to for example ask one to make a picture of a bicycle it will first come up with a depth map of the image before it starts doing anything to the visual output. That shows that the AI has figured out what the three-dimensional form of a bicycle is based entirely on a pile of two-dimensional training images, with no other clues telling it that the third dimension even exists in the first place.

[–] FaceDeer@fedia.io 0 points 1 year ago

We aren't disagreeing because that's not what I was addressing in the first place. The comment I'm responding to, from Dave, reads:

In that case probably the strongest argument is that if it were legal, many people would get off charges of real CSAM because the prosecuter can't prove that it wasn't AI generated.

Emphasis added. The premise of the scenario is that possession of such images (ie, AI-generated CSAM) is not illegal. Given that, for purposes of argument, it follows that this would indeed be a valid defense. You'd need to prove in court that the CSAM pictures that you're charging someone with possessing are not AI-generated, in that scenario.

If you want to have a wider discussion of whether AI-generated CSAM images should be illegal, that's a separate matter.

[–] FaceDeer@fedia.io 3 points 1 year ago* (last edited 1 year ago) (2 children)

It's not the specific thing being made illegal, it's the underlying philosophy of "Better a dozen innocent men go to prison than one guilty man go free" I'm arguing against here. Most western justice systems operate under a principle of requiring guilt to be proven beyond a reasonable doubt, and if there is doubt then guilt cannot be considered proven and the person is not convicted.

The comment I'm responding to is proposing a situation where non-AI-generated images are illegal but AI-generated ones aren't, and that there's no way to tell the difference just by looking at the image itself. In that situation you couldn't convict someone merely based on the existence of the image because it could have been AI-generated. That's fundamental to the "innocent until proven guilty beyond all reasonable doubt" philosophy I'm talking about, to do otherwise would mean that innocent people could very easily be convicted of crimes they didn't do.

[–] FaceDeer@fedia.io 1 points 1 year ago

The comment I'm responding to is proposing a situation in which it isn't illegal.

[–] FaceDeer@fedia.io 6 points 1 year ago (2 children)

But it doesn't fully understand young and "naked young person" isn't just a scaled down "naked adult".

Do you actually know that, or are you just assuming it?

Personally, I'm basing my assertions off of experience with related situations, where I've asked image AIs to generate images of things that I'm quite sure weren't in its training set and that require conceptual understanding to create "hybrids." It's done a decent job of those so I'm assuming that it can figure out this specific situation as well, since most of these models have a lot of examples of naked people and young people in their training sets. But I haven't actually asked any AIs to generate images of naked young people to test this one specific case.

view more: ‹ prev next ›