this post was submitted on 01 Jul 2025
2102 points (98.3% liked)

Microblog Memes

9953 readers
1902 users here now

A place to share screenshots of Microblog posts, whether from Mastodon, tumblr, ~~Twitter~~ X, KBin, Threads or elsewhere.

Created as an evolution of White People Twitter and other tweet-capture subreddits.

Rules:

  1. Please put at least one word relevant to the post in the post title.
  2. Be nice.
  3. No advertising, brand promotion or guerilla marketing.
  4. Posters are encouraged to link to the toot or tweet etc in the description of posts.

Related communities:

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] brucethemoose@lemmy.world 47 points 5 months ago* (last edited 5 months ago) (4 children)

Let’s do the math.

Let’s take an SDXl porn model, with no 4-step speed augmentations, no hand written quantization/optimization schemes like svdquant, or anything, just an early, raw inefficient implementation:

https://www.baseten.co/blog/40-faster-stable-diffusion-xl-inference-with-nvidia-tensorrt/#sdxl-with-tensorrt-in-production

So 2.5 seconds on an A100 for a single image. Let’s batch it (because that’s what’s done in production), and run it on the now popular H100 instead, and very conservatively assume 1.5 seconds per single image (though it’s likely much faster).

That’s on a 700W SXM Nvidia H100. Usually in a server box with 7 others, so let’s say 1000W including its share of the CPU and everything else. Let’s say 1400W for networking, idle time, whatever else is going on.

That’s 2 kJ, or 0.6 watt hours.

…Or about the energy of browsing Lemmy for 30-60 seconds. And again, this is an high estimate, but also a fraction of a second of usage for a home AC system.


…So yeah, booby pictures take very little energy, and the usage is going down dramatically.

Training light, open models like Deepseek or Qwen or SDXL takes very little energy, as does running them. The GPU farms they use are tiny, and dwarfed by something like an aluminum plant.

What slurps energy is AI Bros like Musk or Altman trying to brute force their way to a decent model by scaling out instead of increasing efficiency, and mostly they’re blowing that out of proportion to try the hype the market and convince them AI will be expensive and grow infinitely (so people will give them money).

That isn’t going to work very long. Small on-device models are going to be too cheap to compete.

https://escholarship.org/uc/item/2kc978dg

So this is shit, they should be turning off AI farms too, but your porn images are a drop in the bucket compared to AC costs.


TL;DR: There are a bazillion things to flame AI Bros about, but inference for small models (like porn models) is objectively not one of them.

The problem is billionaires.

[–] SmoothLiquidation@lemmy.world 33 points 5 months ago (3 children)

I don’t disagree with you but most of the energy that people complain about AI using is used to train the models, not use them. Once they are trained it is fast to get what you need out of it, but making the next version takes a long time.

[–] brucethemoose@lemmy.world 13 points 5 months ago* (last edited 5 months ago) (1 children)

Only because of brute force over efficient approaches.

Again, look up Deepseek's FP8/multi GPU training paper, and some of the code they published. They used a microscopic fraction of what OpenAI or X AI are using.

And models like SDXL or Flux are not that expensive to train.

It doesn’t have to be this way, but they can get away with it because being rich covers up internal dysfunction/isolation/whatever. Chinese trainers, and other GPU constrained ones, are forced to be thrifty.

[–] HK65@sopuli.xyz 5 points 5 months ago (1 children)

And I guess they need it to be inefficient and expensive, so that it remains exclusive to them. That's why they were throwing a tantrum at Deepseek, because they proved it doesn't have to be.

[–] brucethemoose@lemmy.world 5 points 5 months ago* (last edited 5 months ago)

Bingo.

Altman et al want to kill open source AI for a monopoly.

This is what the entire AI research space already knew even before deepseek hit, and why they (largely) think so little of Sam Altman.

The real battle in the space is not AI vs no AI, but exclusive use by AI Bros vs. open models that bankrupt them. Which is what I keep trying to tell /c/fuck_ai, as the "no AI" stance plays right into the AI Bro's hands.

[–] drmoose@lemmy.world 6 points 5 months ago

people complain about AI using is used to train the models, not use them

that's absolutely not true. In fact, most people who complain don't even know the difference.

[–] Knock_Knock_Lemmy_In@lemmy.world 1 points 5 months ago

This is a specious argument.

Once a model has been trained once they don't just stop training. They refine and/or start training new models. Showing demand for these models is what has encouraged construction on 100s of new datacenters.

[–] illusionist@lemmy.zip 4 points 5 months ago

Thx for doing the math

[–] Track_Shovel@slrpnk.net 2 points 5 months ago (2 children)

I'm really OOTL when it comes to AI GHG impact. How is it any worse than crypto farms, or streaming services?

How do their outputs stack up to traditional emitters like Ag and industry? I need a measuring stick

[–] brucethemoose@lemmy.world 2 points 5 months ago* (last edited 5 months ago)

The UC paper above touches on that. I will link a better one if I find it.

But specifically:

streaming services

Almost all the power from this is from internet infrastructure and the end device. Encoding videos (for them to be played thousands/millions of times) is basically free since its only done once, with the exception being YouTube (which is still very efficient). Storage servers can handle tons of clients (hence they're dirt cheap), and (last I heard) Netflix even uses local cache boxes to shorten the distance.

TBH it must be less per capita than CRTs. Old TVs burned power like crazy.

[–] brucethemoose@lemmy.world 1 points 5 months ago

Also, one other thing is that Nvidia clocks their GPUs (aka the world's AI accelerators) very inefficiently, because they have a pseudo monopoly, and they can.

It doesn't have to be this way, and likely wont in the future.