Could someone please help me save some power and just post the image with the 5tits so I don’t need to have it regenerated de novo?
Meanwhile I’m down town I’m my city cleaning windows in office buildings that are 75% empty but the heat or ac is blasting on completely empty floors and most of the lights are on.
The HVAC does serve a purpose, it reduces the moisture in the building, which would otherwise ruin the building
I had my energy company remove their LVTC smart meter this week after they started using it to shut off our condenser unit during our 100 degree days
The fact that it exists at all is bad enough, but they were doing this at a time when our AC was already malfunctioning due to low refrigerant. On the day they first shut it off, our house reached 94 degrees.
The program that the previous owner signed up for that enabled them to do this gave them a fucking two dollar a month discount.
I use a smart thermostat to optimize my home conditioning - having a second meter fucking with my schedule ends up making us all miserable. Energy providers need to stop fucking around and just build out their infrastructure to handle worst case peak loads, and enable customers to install solar to reduce peak loading to begin with.
The other thing that kills me about this is that our provider administers our city’s solar electric subsidy program themselves. When i had them come out to give us a quote, they inflated their price by more than 100% because they knew what our electricity bill was. All they did was take our average monthly bill and multiplied it by the repayment period. I could have been providing them more energy to the grid at their peak load if they hadn’t tried scamming me.
FUCK private energy providers.
How tf can a meter shut of an applience? Did you also have smart breakers from them?
Anyway absolutely ridiculous
It’s separate from the main meter and connected directly at the condenser unit.
It monitors power draw and acts as a relay when the provider sends a shutoff signal. The thermostat thinks the system is still going, and the fans still push air through the vents, but the coils aren’t being cooled anymore so the air gets hot and musty.
Peak load of households is not during peak solar power generation. Households installing pv isn’t a solution to what you described.
Today, you could also use a battery to buy power during mid day and use it in the evening when you need it the most.
In moderate climates in the US, peak loads are typically the hottest and sunniest hours of the day since condenser units are the most energy-hungry appliance in most homes. Clouds notwithstanding, peak solar generation would typically align (or closely align) with peak load time.
Batteries would also help a lot - they should definitely be subsidizing the installation of those as well but unfortunately they aren’t yet (at least not in my state).
This is incorrect. Look up the “duck curve” or if you prefer real-world examples look at the California electricity market (CAISO) where they have an excellent “net demand curve” that illustrates the problem.
I watch big state and national grid loads (for fun) and I see two distinct peaks: 7-8AM when everyone goes to work, and then around 5-7 PM when people commute home and heat up dinner.
Otherwise it’s a linear diagonal curve coinciding with temperatures.
I personally try to keep my own energy usage a completely flat line so I can benefit from baseline load generator plants like nuclear (located not that far away).
I personally try to keep my own energy usage a completely flat line so I can benefit from baseline load generator plants like nuclear (located not that far away).
If you consume energy during peak hours you are a peak load consumer. Consuming in other hours doesn’t change this fact.
Your personal energy use pattern does not determine where you are drawing your power from, wtf logic is this?
This curve has changed somewhat since this study in 2016. More efficient home insulation, remote working, and energy-efficient cooling systems have large impact in this pattern. But assuming you have a well-insulated home, setting your thermostat to maintain a consistent temperature throughout the day will shift this peak earlier and lower the peak load at sunset, when many people are returning home. More efficient heat pumps with variable pressure capabilities also helps this a lot, too.
Given just how many variables are involved, it’s better to assume peak cooling load to be mid-day and work toward equalizing that curve, rather than reacting to transient patterns that are subject to changes in customer behavior. Solar installations are just one aspect of this mitigation strategy, along with energy storage, energy-efficient cooling systems, and more efficient insulation and solar heat gain mitigation strategies.
If we’re discussing infrastructure improvements we might as well discuss home efficiency improvements as well.
Do you have a source for the cooling off effect of the duck curve?
Following is a 2 year old article hinting an increase in the effect https://www.powermag.com/epri-head-duck-curve-now-looks-like-a-canyon/ afaik it hasn’t changed much but I’m open to news
I’m not really saying that the curve itself is changing (sorry, I was really not clear), only that those other variables reduce actual energy demand later in the day because of the efficiency gains and thermal banking that happens during the peak energy production. The overproduction during max solar hours is still a problem. Even if the utility doesn’t have a way of banking the extra supply, individual customers can do it themselves at a smaller scale, even if just by over-cooling their homes to reduce their demand after sundown.
Overall, the problem of the duck curve isn’t as much about maxing out the grid, it’s about the utility not having instantaneous power availability when the sun suddenly goes down. For people like me who work from home and have the flexibility to keep my home cool enough to need less cooling in the evening, having solar power means I can take advantage of that free energy and bank it to reduce my demand in the evening.
I get what you were saying now, but having solar would absolutely reduce my demand during peak hours.
It’s a neat idea to over-cool in order to reduce consumption later on!
Study or no study, you can see this problem in the real world https://www.caiso.com/todays-outlook#section-net-demand-trend
Ok now go just one step further and ask yourself what variables factor into this.
There’s a reason that pattern exists, and it isn’t because solar and cooling hours don’t align.
the difference between demand and net demand in that graph is purely solar/wind generation, isn’t it?
Why do you want a subsidy for batteries? Installing batteries at a large scale at homes is incredibly expensive compared to an off site battery. Especially with regards to the move towards hydrogen.
For the same reason we want to subsidize solar production in residential construction even though it’s more efficient and cost-productive to do it at-scale. Having energy production and storage at the point of use reduces strain on power infrastructure and helps alleviate the types of load surging ayyy is talking about.
It’s not a replacement for modernizing our power grids, too - it simply helps to make them more resilient.
That’s understandable but do we need it now? Neither pv nor batteries last forever. I’m just not sure if we need them now (or short-medium term future). But I’m not in the position to decide upon it
I don’t know how to answer that question.
Smart meters with this ability are great, when done well. Without them they have the ability to turn off all of your power if they need to. If they can’t keep up with demand, they have to turn things off. It’s better for them to have the ability to shut off a few appliances or decrease your AC usage rather than shut people down entirely.
People always complain that they don’t want to give the energy company power over their electricity, but they already do. However, without this their power is total, and only total. With it they can moderate it. It’s better if everyone has a smart meter instead of only people who care about others, and greedy people only look out for themselves.
I agree though, fuck private providers.
I agree though, fuck private providers
If it was a public utility i’d feel very differently.
our city’s solar electric subsidy program
It sounds like there’s two different things there. There’s a solar installation (hardware, etc.), and there’s likely some kind of net metering program (where they pay you or give you credit for electricity you generate). That paragraph sounds like the first, but the phrase sounds like the second.
You shouldn’t have to go through them for the solar installation, if your conditions accommodate it. Granted, the conditions don’t apply to everyone. You’ll want to have a suitable roof that ideally faces south-ish, own your home, and plan to stay there for at least 10 years. In the US, you also kind of need to get it done within this calendar year, which is a rough ask, before the federal 30% tax credit goes away. But maybe you can find an installer that isn’t trying to scam you quite as much.
(It’s early and cloudy today.)
Sorry, maybe I wasn’t being clear.
My area has solar incentive programs that are run through the energy utility - meaning the state makes available zero-interest loans for the purposes of solar installation, but those loans are only available through an entity partnered with our utility. They limit the number of homes in each area that are eligible through this program so that solar generation never exceeds demand. Our home was eligible through the program, so I had them come out to give us a quote. Our utility is also transitioning to surge pricing and smart metering, so there’s a pretty high demand for solar installation in my area and they know that they’d lose out on a lot of revenue if everyone installed their own solar systems.
A part of that process was them asking for the last year of energy bills, along with taking measurements and doing daylighting analysis on our roof area. At the end, they gave us a quote for a 15 year loan for the equipment and installation, and it just so happened that the monthly payment was the same as our average energy bill. I work in AEC and I know what solar panels cost, and they had inflated their price by more than double what it would cost at market rate.
Of course I could install my own panels, but it would be out-of-pocket and I would have to seek out and apply for out-of-state incentive programs myself, but I can’t afford the up-front costs and the loan terms don’t make sense for how long we’ll be in this house. Id love nothing more than to do it myself, even at a loss if that’s what it took, but I have a spouse that is less spiteful than I am.
more than double what it would cost at market rate
I definitely paid more for labor than for materials. My payoff time is about 13 years with a Tesla Powerwall 3, maybe a bit less now that I have an EV. I had a team of 4 guys plus an electrician here for about five days.
I did go with a slightly more reputable company that charged slightly more, but I would have gone elsewhere if it was a huge difference.
Maybe I should get around to making a post in !Solarpunk@slrpnk.net or something, even though it isn’t very punk.
I’m factoring in labor. It was an extremely bad deal - they were praying on the fact most home owners do not have familiarity with solar installation pricing.
Like I said, I would love to still do it on my own, but it just doesn’t make sense for our household.
Your HA dashboard derailed this conversation for me. lol.
I would love to know more about the equipment you are using to push this info into your HA.
I’ll write up a post now in !solarpunk@slrpnk.net
Worse is Google that insists on shoving a terrible AI-based result in your face every time you do a search, with no way to turn it off.
I’m not telling these systems to generate images of cow-like girls, but I’m getting AI shoved in my face all the time whether I want it or not. (I don’t).
Then I guess it’s time to stop using Google!
And including the word “fuck” in your query no longer stops it.
And when it did it also altered the results, making them worse, because it was trying to satisfy “fuck” as part of your search.
If you can’t search for “fuck,” you can’t search for “fuck google.”
With apologies to Lenny Bruce.
Well fuck…
I am trying to understand what Google’s motivation for this even is. Surely it is not profitable to be replacing their existing, highly lucrative product with an inferior alternative that eats up way more power?
Their motivation is always ads. The ai response is longer and takes time to read so more time looking at their ads. If the answer is sufficient, you might not even click away to the search result.
AI is a potential huge bonanza to search sites, letting them suck up the ad revenue that used to goto the search results
They don’t want to direct you to the thing you’re searching for anymore because that means you’re off their site quickly. Instead they want to provide themselves whatever it is you were searching for, so you will stay on their site and generate ad money. They don’t care if their results are bad, because that just means you’ll stick around longer, looking for an answer.
To make search more lucrative, they’ve enshitified it and went too far, but for a short time there were great quarterly resukts. Now they’re slowly losing users. So they try AI to fix it up.
It’s also a signal to the shareholders that they’re implementing the latest buzzword, plus they’re all worried AI will take off and they’ve missed that train.
Piling on to the google alternatives heap: https://searx.space/
You can pick a public instance of searxng and choose which engines it queries by going to the setting cog, then Engines. A few of these public instances I’ve checked out have only google enabled, though, so you really do need to check the settings.
If you want to add a searxng instance as your default engine and your browser doesn’t automatically do it, the URL for that is: https://<searxng_url>/search?q=%s
I have to add this manually for things like ironfox/firefox mobile.
Someone posted here a while ago - if you use the URL https://www.google.com/search?q=%25s&udm=14 it doesn’t include the AI search. I’ve updated my Google search links to use that instead of the base Google URL.
You can also use alternatives like startpage and ecosia which use google results, I believe.
Both of which are probably training their own AI as middle men or stealing your search terms to tell Walmart what type of peanut butter you’re most likely to buy if they could lock it up on a plastic covered shelve.
Probably, but neither automatically opt into AI replies. Ecosia has an AI chat, but it doesn’t run until you go to it. Startpage has no AI option that I can see.
Ecosia has the upside of planting trees depending on user search rate. Not sure how true that is, though. I prefer startpage either way. Startpage claims to be privacy first, and I’ve never received tailored results or ads.
That doesn’t mean they don’t sell info. We can’t know that for sure, but it sure as hell beats using Google and it’s automatic AI searching.
Firefox has a plugin that blocks the AI results. It works pretty well most of the time, but it occasionally has hiccups when Google updates stuff or something.
I’ll have to look for this, thanks!
I have to look stuff up for medical school (usually trying to find studies and whatnot) so the gemini results are just obnoxious garbage to me.
There is a way to “turn it off” with some search parameters. However there is no guarantee that the AI is not consuming resources at the backend.
Also the search parameters are undocumented internal things that can change or be disabled at any time.
When I’m told there’s power issues and to conserve power I drop my AC to 60 and leave all my lights on. Only way for them to fix the grid is to break it.
Literally rolling coal to own the cons
wow based
Yeah, that thing that nobody wanted? Everybody has to have it. Fuck corporations and capitalism.
Just like screens in cars, and MASSIVE trucks. We don’t want this. Well, some dumbass Americans do, but intelligent people don’t need a 32 ton 6 wheel drive pickup to haul jr to soccer.
Massive trucks? They need those trucks for truck stuff, like this giant dilhole parking with his wife to go to Aldi today. Not even a flag on the end of that ladder, it filled a whole spot by itself.
My couch wouldn’t fit in that bed, and every giant truck I see is sparkling shiny and looks like it hasn’t done a day of hard labor, much like the drivers.
You underestimate the number of people you wouldn’t class as intelligent. If no one wanted massive trucks, they would have disappeared off the market within a couple of years because they wouldn’t sell. They’re ridiculous, inefficient hulks that basically no one really needs but they sell, so they continue being made.
It’s actually because small trucks were regulated out of the US market. Smaller vehicles have more stringent mileage standards that trucks aren’t able to meet. That forces companies to make all their trucks bigger, because bigger vehicles are held to a different standard.
So the people who want or need a truck are pushed to buy a larger one.
They can meet them. But the profit margin is slimmer than if they use the giant frame.
Do you have any data to support this is actually the case? I see this all the time but absolutely zero evidence but a 2015 Axios survey with no methodology or dataset. Nearly every article cites this one industry group with 3 questions that clearly aren’t exclusive categorical and could be picked apart by a high school student.
I ask this question nearly every time I see this comment and in 5 years I have not found a single person who can actually cite where this came from or a complete explanation of even hope they got to that conclusion.
The truck owners I know, myself included, use them all the time for towing and like the added utility having the bed as as secondary feature.
The truck owners I know, myself included, use them all the time for towing and like the added utility having the bed as as secondary feature.
Then you put it beside a truck from 30 years ago that’s a quarter the overall size but has the same bed capacity and towing power along with much better visibility instead of not being able to see the child you’re about to run over. And then you understand what people mean when they say massive trucks - giant ridiculously unnecessary things that are all about being a status symbol and dodging regulations rather than practicality.
Absolutely 100% incorrect on towing. The 95 top f150 towed about 7700 compared to 13500 today. That’s an f350 in 95. It’ll also fit a family of 4 comparable to a full size sedan eliminating any need of a secondary vehicle. The old f150/1500s were miserable in the back.
As for the safety I find the argument disingenuous not based on reality. Roughly 160 kids were killed in 23 with the EU27. It was 220 in the US. Much of that could be correlated to traffic density as well.
Country / Region Est. Fatalities/Year Child Pop. (0–14) Fatalities per Million
United States ~225 ~61 million ~3.7 United Kingdom ~22 ~11.5 million ~1.9 Canada ~12 ~6 million ~2.0 Australia ~11 ~4.8 million ~2.3 Germany ~20 ~11 million ~1.8 France ~18 ~11 million ~1.6 Japan ~18 ~15 million ~1.2 India ~3,000 (est.) ~360 million ~8.3 Brazil ~450 ~50 million ~9.0 European Union (EU-27) ~140–160 ~72 million ~2.0–2.2
I think we should offer incentives for manufacturers to start reducing size and weight, but things you are saying here aren’t really based off of any data nor was it what I was asking.
I just wish I could find one person to show me what they are referencing when they repeat that seemingly false fact.
Let me express it to you with some numbers… The US is ~3.81 million square miles in size.
The F150 has sold 8.810 million units in the US in the last 10 years.
There are ~ 2.3 F150s fewer than 10 years old for every square mile in this country.
There is no way the majority of those trucks are going to job sites, or hauling junk, or pulling a trailer, just look around. That’s not even all trucks. Thats just one model, from one brand, for a single 10 yr period.
These trucks are primarily sold as a vanity vehicle, and a minivan alternative, and that’s what I think when I see one.
Nonsense. This is some Trump style math.
No, trump style math would be saying that The number of Trucks Towing has gone DOWN 400% PERCENT after the EVIL AMERICA HATING COMMUNIST Dems elected a soon-to-be-illegal Migrant Gang member as Mayor of New York NYC.
Oh, and you don’t want it and want the stupid model? You can still buy it for 3x the price.
Do the new models even have non-“smart” fittings? I thought all the electronic chip plants closed during covid.
Also they can build nuclear power generators for the data centers but never for the residential power grid.
There’s no money in selling residential energy.
Let’s do the math.
Let’s take an SDXl porn model, with no 4-step speed augmentations, no hand written quantization/optimization schemes like svdquant, or anything, just an early, raw inefficient implementation:
So 2.5 seconds on an A100 for a single image. Let’s batch it (because that’s what’s done in production), and run it on the now popular H100 instead, and very conservatively assume 1.5 seconds per single image (though it’s likely much faster).
That’s on a 700W SXM Nvidia H100. Usually in a server box with 7 others, so let’s say 1000W including its share of the CPU and everything else. Let’s say 1400W for networking, idle time, whatever else is going on.
That’s 2 kJ, or 0.6 watt hours.
…Or about the energy of browsing Lemmy for 30-60 seconds. And again, this is an high estimate, but also a fraction of a second of usage for a home AC system.
…So yeah, booby pictures take very little energy, and the usage is going down dramatically.
Training light, open models like Deepseek or Qwen or SDXL takes very little energy, as does running them. The GPU farms they use are tiny, and dwarfed by something like an aluminum plant.
What slurps energy is AI Bros like Musk or Altman trying to brute force their way to a decent model by scaling out instead of increasing efficiency, and mostly they’re blowing that out of proportion to try the hype the market and convince them AI will be expensive and grow infinitely (so people will give them money).
That isn’t going to work very long. Small on-device models are going to be too cheap to compete.
https://escholarship.org/uc/item/2kc978dg
So this is shit, they should be turning off AI farms too, but your porn images are a drop in the bucket compared to AC costs.
TL;DR: There are a bazillion things to flame AI Bros about, but inference for small models (like porn models) is objectively not one of them.
The problem is billionaires.
I don’t disagree with you but most of the energy that people complain about AI using is used to train the models, not use them. Once they are trained it is fast to get what you need out of it, but making the next version takes a long time.
Only because of brute force over efficient approaches.
Again, look up Deepseek’s FP8/multi GPU training paper, and some of the code they published. They used a microscopic fraction of what OpenAI or X AI are using.
And models like SDXL or Flux are not that expensive to train.
It doesn’t have to be this way, but they can get away with it because being rich covers up internal dysfunction/isolation/whatever. Chinese trainers, and other GPU constrained ones, are forced to be thrifty.
And I guess they need it to be inefficient and expensive, so that it remains exclusive to them. That’s why they were throwing a tantrum at Deepseek, because they proved it doesn’t have to be.
Bingo.
Altman et al want to kill open source AI for a monopoly.
This is what the entire AI research space already knew even before deepseek hit, and why they (largely) think so little of Sam Altman.
The real battle in the space is not AI vs no AI, but exclusive use by AI Bros vs. open models that bankrupt them. Which is what I keep trying to tell /c/fuck_ai, as the “no AI” stance plays right into the AI Bro’s hands.
people complain about AI using is used to train the models, not use them
that’s absolutely not true. In fact, most people who complain don’t even know the difference.
This is a specious argument.
Once a model has been trained once they don’t just stop training. They refine and/or start training new models. Showing demand for these models is what has encouraged construction on 100s of new datacenters.
Thx for doing the math
Not only are they cheaper than AC, but doing the math shows that they are more energy efficient than a human doing the same work, since humans operate at around 80-100W, 24 hours a day. (Assuming that the output is worth anything, of course.)
let’s not use the term “efficiency” with humans making art, please. you’re not helping anyone with that argument, you’re just annoying both sides.
Well if humans could run on coal it would be a valid argument…
Humans essentially do run on fossil fuels. Modern agriculture is very energy intensive.
Humans at least run on renewable energy.
The computer you draw your art on, not so much. Reject modern art, embrace traditional carvings and cave paintings!
I think that’s going a bit far. ML models are tools to augment people, mostly.
Oh for sure. But if (for example) an artist can save time by tracing over an SDXL reference image, that is energy-efficient as well as time-efficient, despite most people claiming the contrary.
I’m really OOTL when it comes to AI GHG impact. How is it any worse than crypto farms, or streaming services?
How do their outputs stack up to traditional emitters like Ag and industry? I need a measuring stick
How is it any worse than crypto farms, or streaming services?
These two things are so different.
Streaming services are extremely efficient; they tend to be encode-once and decode-on-user’s-device. Video was for a long time considered a tough thing to serve, so engineers put tons of effort into making it efficient.
Crypto currency is literally designed to be as wasteful as possible while still being feasible. “Proof-of-work” (how Bitcoin and many other currencies operate) literally means that crypto mining algorithms must waste as much computation as they can get away with doing pointless operations just to say they tried. It’s an abomination.
I legit don’t know much about tech, and it ts showing. I didn’t know streaming was so efficient.
What I. Trying to get at (I still have to read that article in the parent comment) is that how is AI any worse than crypto? As far as I can tell crypto impact, while bad, was relatively minor and it drastically decreased in popularity; it’s kind of logical AI does the same, unless it’s impact is way higher.
Meanwhile we have cargo ships burning bunker crude .
If you are expecting AI to not have much impact and turn out to be a bubble, then I guess there isn’t much reason to believe it it will have much environmental impact. If you expect AI to not be a fad, then yeah it could have big environmental consequences if we can’t find renewable power and coolant. If AI is all it is hyped up to be, then it would dwarf the rest of humanity’s power consumption down to a footnote. So it really depends on how bullish you are about AI, or at least how bullish you expect the market to be going forward.
Regarding proof-of-work crypto, well, bitcoin is currently at its all-time high in terms of value, exceeding USD$100k/BTC. So I’m not sure I exactly buy the idea that it’s less popular, though perhaps people aren’t reporting on it as much. If the power consumption of crypto has levelled off, which I don’t know if it has, then it might be because it’s expensive to build a mining rig and the yield goes down over time as more bitcoin is mined. (It’s presumably true of other proof-of-work crypto, too, but as more BTC is mined, the marginal yield of mining more BTC decreases.)
Honestly, all of this is really interesting. It’s a whole side of humanity that I very much do NOT think about or follow. I previously spent the last decade much, much, too busy stomping through the forest, so I really didn’t follow anything during that time. A new game or phone came out? sure, cool, I might look that up. When I finally emerged from the fens, sodden and fly-bitten, I was very much out of the loop, despite the algorithm trying to cram articles about NFTs, crypto etc., down my throat. I actually tend to avoid tech stuff because it’s too much of a learning curve at this point. I get the fundamentals, but beyond that I don’t dig in.
I agree with you on the bubble - it depends on the size. I guess my original take is how could it actually get bigger than it is? I just don’t see how it can scale beyond begin in phones or used in basic data analysis/like another google. The AIs can definitely get more advanced, sure, but with that should come some sort of efficiency. We’re also seemingly on the cusp of quantum computing, which I imagine would reduce power requirements.
Meanwhile (and not to detract from the environmental concerns AI could pose) we have very, very real and very, very large environmental concerns that need addressing. Millions of cubic metres of sulphur are sitting in stockpiles in northern Alberta, and threatening the Athabasca river. That’s not even close to the top of the list of things we need to focus on before we can get out in front of the damage AI can cause.
We’re in a real mess.
The AIs can definitely get more advanced, sure, but with that should come some sort of efficiency.
This is what AI researchers/pundits believed until roughly 2020, when it was discovered you could brute force your way to have more advanced AIs (so-called “scaling laws”) just by massively scaling up existing algorithms. That’s essentially what tech companies have been doing ever since. Nobody knows what the limit on this is going to be, but as far as I know nobody has any good evidence to suggest that we’re near the limit of what’s going to be possible with scaling.
We’re also seemingly on the cusp of quantum computing, which I imagine would reduce power requirements.
Quantum computing is not faster than regular computers. Quantum computing has efficiency advantages for some particular algorithms, such as breaking certain types of encryption. As far as I’m aware, nobody is really looking to replace computers with quantum computers in general. Even if they did, I don’t think anyone has thought of a way to accelerate AI using quantum computing. Even if there were a way to, it would presumably require quantum computers like, 15 orders of magnitude more powerful than the ones we have today.
We have very, very real and very, very large environmental concerns that need addressing.
Yeah. I don’t think AI is really at the highest level of concern for environmental impact, especially since it is looking plausible it will lead to investing in nuclear power, which would be a net positive IMO. (Coolant could still be an issue though.)
How do they brute force their way to a better algorithm? Just trial and error? How do they check outcomes to determine that their new model is good?
I don’t expect you to answer those musings - you’ve been more than patient with me.
Honestly, I’m a tree hugger, and the fact that we aren’t going for nuclear simply because of smear campaigns and changes in public opinion is insanity. We already treat some mining wastes in perpetuity, or plan to have them entombed for the rest of time - how is nuclear waste any different?
The UC paper above touches on that. I will link a better one if I find it.
But specifically:
streaming services
Almost all the power from this is from internet infrastructure and the end device. Encoding videos (for them to be played thousands/millions of times) is basically free since its only done once, with the exception being YouTube (which is still very efficient). Storage servers can handle tons of clients (hence they’re dirt cheap), and (last I heard) Netflix even uses local cache boxes to shorten the distance.
TBH it must be less per capita than CRTs. Old TVs burned power like crazy.
Also, one other thing is that Nvidia clocks their GPUs (aka the world’s AI accelerators) very inefficiently, because they have a pseudo monopoly, and they can.
It doesn’t have to be this way, and likely wont in the future.
We’re going away folks, and nothing of any true value will be lost, except all the species that did live in homeostasis with the Earth that we’re taking with us in our species’ avarice induced murder-suicide
Carlin had some good material, but this is an absolutely stupid mindset. We can cause an extreme level of ecological damage. Will the planet eventually recover? Quite possibly. But that’s not a certainty, and in the mean time we’re triggering a mass extinction precisely because irresponsible humans figure there’s no way we can hurt the Earth and it’s self-important hubris to think that we can.
But the time we’re living through and the time we’re heading into are all the proof we should need that it’s actually hubris to assume our actions have no meaningful impact.
Immediate to Short-Term (Days to Centuries)
- Hours to weeks: Power grids fail; nuclear reactors melt down without maintenance[11].
- Months to decades: Urban areas flood as drainage systems fail; buildings decay from weather and plant growth[6][11].
- 100–300 years: Steel structures collapse; concrete buildings crumble[5][7]. Most cities become overgrown forests[6].
Medium-Term (Thousands of Years)
- 1,000 years: Visible surface structures (e.g., roads, monuments) are buried or eroded. Plastics fragment but persist chemically[5][7].
- 10,000–250,000 years: Nuclear isotopes (e.g., plutonium-239) remain detectable in sediments and ice cores[7]. Mining tunnels fill with sediment but leave identifiable “industrial fossils”[7].
- 500,000 years: Microplastics and polymer layers in ocean sediments endure[5][10].
Long-Term (Millions of Years)
- 1–7 million years: Fossils of humans and domesticated animals persist. Geological strata show elevated carbon levels and mass extinction markers[4][8]. Deep mines and landfills remain as distinct layers[7][10].
- 50–100 million years: Continental drift subducts surface evidence; satellites decay or drift into space[3][10]. Only deep geological traces (e.g., mine shafts, isotope ratios) might endure[3][10].
- 250 million years: Next predicted mass extinction eradicates all mammals, including any remaining human traces[9].
Near-Permanent Traces
- Space artifacts: Lunar landers, Mars rovers, and Voyager probes persist for billions of years[3][10].
- Radio signals: Human broadcasts travel through space indefinitely at light speed[5].
Key Factors
- Detection likelihood: Aliens or future species could find traces for 100+ million years via deep geological analysis or space exploration[5][10].
- Total erasure: Requires Earth’s destruction (e.g., solar expansion in 5 billion years)[10].
Citations: [1] Human extinction https://en.wikipedia.org/wiki/Human_extinction [2] What If Humans Suddenly Went Extinct? https://www.youtube.com/watch?v=yuOKTZISXhc [3] How long would it take for all traces of humans to be gone? https://www.reddit.com/r/answers/comments/1azu120/how_long_would_it_take_for_all_traces_of_humans/ [4] What would happen to Earth if humans went extinct? https://www.livescience.com/earth-without-people.html [5] How long before all human traces are wiped out? https://www.newscientist.com/lastword/2215950-how-long-before-all-human-traces-are-wiped-out/ [6] Vanishing Act: What Earth Will Look Like 100 Years After Humans Disappear - Brilliantio https://brilliantio.com/if-people-dissapeared-what-will-happen-to-earth-in-100-years/ [7] If humans became extinct, how long would it take for all … https://www.sciencefocus.com/science/if-humans-became-extinct-how-long-would-it-take-for-all-traces-of-us-to-vanish [8] Nature will need up to five million years to fill the gaps caused by man-made mass extinctions, study finds https://www.independent.co.uk/climate-change/news/mass-extinctions-five-million-years-nature-mammals-crisis-animal-plants-pnas-aarhus-a8585066.html [9] Humans Will Go Extinct on Earth in 250 Million Years; Mass Extinction Will Occur Sooner if Burning Fossil Fuels Continues [Study] https://www.sciencetimes.com/articles/49951/20240430/humans-will-go-extinct-earth-250-million-years-mass-extinction.htm [10] How long would it take for all evidence of humanity to be … https://worldbuilding.stackexchange.com/questions/153618/how-long-would-it-take-for-all-evidence-of-humanity-to-be-erased-from-earth [11] What Would Happen If Every Human On Earth Just Disappeared? https://www.scienceabc.com/humans/life-like-humans-suddenly-disappeared.html
Stephen Baxter over here. I’m going back to bed, it won’t change anything…
We do have an impact but the earth will 100% be ok when we are dead and gone eventually. A million years ain’t shit to the earth.
We humans are a virus…a parasite and the earth will be better off once we are extinct.
I’ve been trying to write this comment the more concise possible, I’m trying my best. “We’re going away”, yes, that’s true. No matter what we are leaving this place, but, that doesn’t mean that the last days of humanity have to be surrounded by pollution and trash. All I can get of that quote in the image is that we should let big companies shit on us till we die.
“let”
The sociopath fascist capitalists won. They have multilayered protection from us from propaganda dividing us and turning us on one another without end to government capture and having the exclusive use of state violence to protect the capital markets and literally having state sanctioned murder for for private profit. This isnt a war, this is a well oiled Orwellian occupation. The people surrendered half a century ago without terms and received the delusion that they’ll be the rich ones one day herp derp.
We can’t do anything about the misery they spread, and that sucks. We don’t have to add to our misery by pretending there’s hope and we can turn any of it around. They’re going to do what we’re going to do and short of being a lone “terrorist” that takes a pot shot at them like Luigi there’s nothing to be done, because half of us are too cowardly and social opiate addicted(fast food, social media, literal opiates, etc) or too deluded and actually on the robber baron’s side out of pick me mentality to take it as a rallying cry.
Only the planet itself, our shared habitat, can stop them. And it will, regardless of all the studies they kill, ecological treaties they betray, and all the propaganda they spread. The capitalists reign of terror will end when enough of their suckers are starving, not before.
Can’t be soon enough, really. I’d rather we take fewer other species along with us.
Let’s not forget billionaires in this consideration.
Laughs in total recall
I know she’s exaggerating but this post yet again underscores how nobody understands that it is training AI which is computationally expensive. Deployment of an AI model is a comparable power draw to running a high-end videogame. How can people hope to fight back against things they don’t understand?
She’s not exaggerating, if anything she’s undercounting the number of tits.
Well you asked for six tits but you’re getting five. Why? Because the AI is intelligent and can count, obviously.
I mean, continued use of AI encourages the training of new models. If nobody used the image generators, they wouldn’t keep trying to make better ones.
you are correct, and also not in any way disagreeing with me.
I try lol.
TBH most people still use old SDXL finetunes for porn, even with the availability of newer ones.
It’s closer to running 8 high-end video games at once. Sure, from a scale perspective it’s further removed from training, but it’s still fairly expensive.
really depends. You can locally host an LLM on a typical gaming computer.
You can, but that’s not the kind of LLM the meme is talking about. It’s about the big LLMs hosted by large companies.
True, and that’s how everyone who is able should use AI, but OpenAI’s models are in the trillion parameter range. That’s 2-3 orders of magnitude more than what you can reasonably run yourself
This is still orders of magnitude less than what it takes to run an EV, which are an eco-friendly form of carbrained transportation. Especially if you live in an area where the power source is renewable. On that note, it looks to me like AI is finally going to be the impetus to get the U.S. to invest in and switch to nuclear power – isn’t that altogether a good thing for the environment?
lol just purely wrong on that one. Hilarious.
Well that’s sort of half right. Yes you can run the smaller models locally, but usually it’s the bigger models that we want to use. It would also be very slow on a typical gaming computer and even a high end gaming computer. To make it go faster not only is the hardware used in datacenters more optimised for the task, it’s also a lot faster. This is both a speed increase per unit as well as more units being used than you would normally find in a gaming PC.
Now these things aren’t magic, the basic technology is the same, so where does the speed come from? The answer is raw power, these things run insane amounts of power through them, with specialised cooling systems to keep them cool. This comes at the cost of efficiency.
So whilst running a model is much cheaper compared to training a model, it is far from free. And whilst you can run a smaller model on your home PC, it isn’t directly comparable to how it’s used in the datacenter. So the use of AI is still very power hungry, even when not counting the training.
Yeh but those local models are usually pretty underpowered compared to the ones that run via online services, and are still more demanding than any game.
Not at all. Not even close.
Image generation is usually batched and takes seconds, so 700W (a single H100 SXM) for a few seconds for a batch of a few images to multiple users. Maybe more for the absolute biggest (but SFW, no porn) models.
LLM generation takes more VRAM, but is MUCH more compute-light. Typically one has banks of 8 GPUs in multiple servers serving many, many users at once. Even my lowly RTX 3090 can serve 8+ users in parallel with TabbyAPI (and modestly sized model) before becoming more compute bound.
So in a nutshell, imagegen (on an 80GB H100) is probably more like 1/4-1/8 of a video game at once (not 8 at once), and only for a few seconds.
Text generation is similarly efficient, if not more. Responses take longer (many seconds, except on special hardware like Cerebras CS-2s), but it parallelized over dozens of users per GPU.
This is excluding more specialized hardware like Google’s TPUs, Huawei NPUs, Cerebras CS-2s and so on. These are clocked far more efficiently than Nvidia/AMD GPUs.
…The worst are probably video generation models. These are extremely compute intense and take a long time (at the moment), so you are burning like a few minutes of gaming time per output.
ollama/sd-web-ui are terrible analogs for all this because they are single user, and relatively unoptimized.
How exactly did you come across this “fact”?
I compared the TDP of an average high-end graphics card with the GPUs required to run big LLMs. Do you disagree?
I do, because they’re not at full load the entire time it’s in use
They are, it’d be uneconomical not to use them fully the whole time. Look up how batching works.
I mean I literally run a local LLM, while the model sits in memory it’s really not using up a crazy amount of resources, I should hook up something to actually measure exactly how much it’s pulling vs just looking at htop/atop and guesstimating based on load TBF.
Vs when I play a game and the fans start blaring and it heats up and you can clearly see the usage increasing across various metrics
He isn’t talking about locally, he is talking about what it takes for the AI providers to provide the AI.
To say “it takes more energy during training” entirely depends on the load put on the inference servers, and the size of the inference server farm.
One user vs a public service is apples to oranges and it’s actually hilarious you’re so willing to compare them.
My guy, we’re not talking about just leaving a model loaded, we’re talking about actual usage in a cloud setting with far more GPUs and users involved.
nice name btw
Right, but that’s kind of like saying “I don’t kill babies” while you use a product made from murdered baby souls. Yes you weren’t the one who did it, but your continued use of it caused the babies too be killed.
There is no ethical consumption under capitalism and all that, but I feel like here is a line were crossing. This fruit is hanging so low it’s brushing the grass.
“The plane is flying, anyway.”
Are you interpreting my statement as being in favour of training AIs?
I’m interpreting your statement as “the damage is done so we might as well use it”
And I’m saying that using it causes them to train more AIs, which causes more damage.I agree with your second statement. You have misunderstood me. I am not saying the damage is done so we might as well use it. I am saying people don’t understand that it is the training of AIs which is directly power-draining.
I don’t understand why you think that my observation people are ignorant about how AIs work is somehow an endorsement that we should use AIs.
I guess.
It still smells like an apologist argument to be like “yeah but using it doesn’t actually use a lot of power”.
I’m actually not really sure I believe that argument either, through. I’m pretty sure that inference is hella expensive. When people talk about training, they don’t talk about the cost to train on a single input, they talk about the cost for the entire training. So why are we talking about the cost to infer on a single input?
What’s the cost of running training, per hour? What’s the cost of inference, per hour, on a similarly sized inference farm, running at maximum capacity?Maybe you should stop smelling text and try reading it instead. :P
Running an LLM in deployment can be done locally on one’s machine, on a single GPU, and in this case is like playing a video game for under a minute. OpenAI models are larger than by a factor of 10 or more, so it’s maybe like playing a video game for 15 minutes (obviously varies based on the response to the query.)
It makes sense to measure deployment usage marginally based on its queries for the same reason it makes sense to measure the environmental impact of a car in terms of hours or miles driven. There’s no natural way to do this for training though. You could divide training by the number of queries, to amortize it across its actual usage, which would make it seem significantly cheaper, but it comes with the unintuitive property that this amortization weight goes down as more queries are made, so it’s unclear exactly how much of the cost of training should be assigned to a given query. It might make more sense to talk in terms of expected number of total queries during the lifetime deployment of a model.
You’re way overcomplicating how it could be done. The argument is that training takes more energy:
Typically if you have a single cost associated with a service, then you amortize that cost over the life of the service: so you take the total energy consumption of training and divide it by the total number of user-hours spent doing inference, and compare that to the cost of a single user running inference for an hour (which they can estimate by the number of user-hours in an hour divided by their global inference energy consumption for that hour).
If these are “apples to orange” comparisons, then why do people defending AI usage (and you) keep making the comparison?
But even if it was true that training is significantly more expensive that inference, or that they’re inherently incomparable, that doesn’t actually change the underlying observation that inference is still quite energy intensive, and the implicit value statement that the energy spent isn’t worth the affect on society
How about, fuck AI, end story.
how about, fuck capitalism? Have you lost sight of the goal?
What tools do you think capitalism is going to use to fuck us harder and faster than ever before?
All of them at their disposal, we should get rid of all tools
Running a concept to its extreme just to try and dismiss it to sound smart is an entire damn logical fallacy. Why are you insisting on using fallacies that brainless morons use?
Have you never heard of a straw man fallacy? That’s you. That’s what you’re doing.
So mad at a joke
Still dodging any actual point. Have fun being too stupid to actually engage with a discussion. Genuinely pitiful of you.
I did, as a matter of fact, fuck AI.
You thought blind anger came from well informed opinions?
But then the rage machine couldn’t rage
there is so much rage today. why don’t we uh, destroy them with facts and logic
Hahaha at this point even facts and logic is a rage inducing argument. “My facts” vs “Your facts”
Crossposted to fuck ai community
Classic neo-liberalism - privatize the benefits, socialize the costs.
Corporations : “We should get to gobble all power with our projects… and you should have the personal responsibility to reduce power usage even though it would - at best - only improve things at the very edges of the margins… and then we can get away with whatever we want.”
Just like with paper straws. You get crappy straws and they hope you feel like you’re helping the environment (even though the plastic straws account for like 0.00002% of plastic waste generated) … meanwhile 80% of the actual pollution and waste being generated by like 12 corporations gets to continue.
I feel like i’ve read a very similar argument somewhere recently, but i have difficulty remembering it precisely. It went something like this:
- If a company kills 5 people, it was either an accident, an unfortunate mishap, a necessity of war (in case of the weapons industry) or some other bullshit excuse.
- If the people threaten to kill 5 billionaires, they’re charged with “terrorism” (see Luigi Mangione’s case).
1 prompt is avg 1Wh of electricity -> typical AC runs avg 1,500 W = 2.4 seconds of AC per prompt.
Energy capacity is really not a problem first world countries should face. We have this solved and you’re just taking the bait of blaming normal dudes using miniscule amounts of power while billionaires fly private jets for afternoon getaways.
They are blaming the billionaires (or their companies), for making the thing nobody wanted so they can make money off of it. The guy making a five-breasted woman is a side effect.
And sure, that one image only uses a moderate amount of power. But there still exists giant data centers for only this purpose, gobbling up tons of power and evaporating tons of water for power and cooling. And all this before considering the training of the models (which you better believe they’re doing continuously to try to come up with better ones).
nobody wanted according to whom? It’s literally the most used product of this century stop deluding yourself.
All datacenters in the world combined use like 5% of our energy now and the value we get from computing far outweighs any spending we have here. You’re better off not buying more trash from Temu rather than complain about software using electricity. This is ridiculous.
Nah dawg im pretty sure the most used product this century is food.
the most used product of this century is actually your momma
Where are you getting your false information. Its certainly not the most used. And, the reason it’s used at all is from advertising and ownership of the media by the billionaire class to shove the gibbity in our faces at every waking moment so people use it. They’re losing money like never before on ai.
This level of collective delusion is crazy. I don’t think any amount of stats will change your mind so you’re clearly argueing in bad faith but sure:
https://explodingtopics.com/blog/chatgpt-users says 5.2B monthly visits compared to Facebook 12.7 and Instagram’s 7.5. Chatgpt is literally bigger than X.com already. Thats just one tool and LLM’s have direct integrations in phones and other apps.
I really don’t understand what’s the point of purposefully lying here? We can all hate billionaires together without the need for this weird anti-intellectual bullshit.
I don’t agree that it’s the most used invention this century. Also, typing “hi there gibbity” hardly counts as actually using the tool. If thats what you mean though, sure.
People hate AI so much (for many good reasons!) that they can’t see or accept the truth: many many people want to use it, not just “billionaires”
A lot of things are solved, but capitalism means that we need a profit motive to act. World hunger is another good example. We know how to make fertilizer and how to genetically alter crops to ensure we never have a crop failure. We have trains and refrigeration to take food anywhere we want. Pretty much any box that we need to check to solve this problem has been. The places that have food problems largely have to do with poverty, which at this point is a polite way to say “I won’t make money, so I am okay with them starving”
Im not sure what’s the point here? If we dont like LLMs and data centers using power then we use existing strategies that work like taxing their power use and subsidizing household power use which btw we’re already doing almost everywhere around the world in some form or another.
The data centers are actually easier to negotiate and work with than something like factories or households where energy margins are much more brittle. Datacenter employs like 5 people and you can squeeze with policy to match social expectations - you can’t do that with factories or households. So datacenter energy problem is not that difficult relatively speaking.
I am agreeing with you that the solutions exist, but the will to implement them is going to be the hard part. A big dampener is simply going to be the profit motive. There is more money in siding with the data center than a the households. Are households okay with an increasing in price? Data center is likely to manage that better, or even just pay a bribe to someone. I used food as another example of a problem that is solved. We can grow food without fail and build the rail to get it to where it needs. We just don’t because need does not match profit expectation. There are talks of building nuclear power for some data centers, but such talk would not happen for normal households.
People definitely underestimate how cooperative big tech is relative to every other business mostly because big tech has a lot of money and very few expenses so friction is relatively a bigger bottle neck than almost any other industry. So I still think that pressuring openAi into green energy is easier than pressuring Volvo (or any manufacturer) which already is really brittle and has huge negotiation leverage in the form of jobs it provides.
Take a look at any other business niche and no one’s comitting to green energy other than big tech. As you said yourself no other niche want to build their own nuclear reactors to satisfy their own green energy need.
I think its OK to hate on big tech because they’re billionaires but people really lose sight here and complain about wrong things that distract from real much bigger issues and paints the entire movement as idiots.
A 12000 BTUs inverter split system at peak capacity requires less than 1500 W to run. After it reaches equilibrium it drops the power requirement significantly.
ok so 5 seconds of AC then? my point still stands.