• 3 Posts
  • 851 Comments
Joined 2 years ago
cake
Cake day: July 4th, 2023

help-circle

  • I hope you dont use any of the other standard quality of life features day to day that consume substantially more power per day then.

    There’s plenty of stuff you likely take for granted every day that you use, that burn way more fossil fuels than training GPT took.

    GPT did cost a lot of power, but if you put it beside other fairly standard day-to-day things people tend to take for granted, it’s a drop in the bucket.

    • Air conditioning, both at your home if you use it, your work if you have it, stores you visit, etc etc
    • Public transit
    • Your stove
    • Your microwave
    • Your water kettle
    • Your heating systems everywhere you go
    • Your computer
    • Your phone
    • The internet
    • Emergency response systems
    • Your clothes washer and dryer

    The list goes on and on. ESPECIALLY your clothes dryer, that thing uses a massive amount of power

    People seriously underestimate how much power the internet uses overall. GPT’s training provides a concrete, discrete, measured amount of power one specific thing used.

    Whereas the internet, as a whole, over one day, uses way more power than all of GPT’s training took total. The issue is “the internet” has its power consumption broadly distributed across the entire globe, in a manner that makes it basically impossible to actually measure how much “total” power you are burning just browsing the web.

    But it’s non trivial. Every switch between you and your destination is burning in the range of 150 watts, easily, every router is burning easily 80 watts, etc etc.

    And theres dozens of those between you and 1 given destination. The process of routing your packets from your machine all the way across countries at the speed of light, and then a response back, takes a non trivial amount of power. Theres often around 8 to 15 hops between you and the destination, and every single hop tends to have multiple machines involved in that one single packet.

    Its easy to handwave that enormous power consumption away because, well, you can’t see it. You aren’t privvy to how much power your ISP burns every day, how much power the nameservers use, etc etc.

    GPT is a non trivial chunk of power… but its not THAT much compared to all the other shit going on in the web, its genuinely just a tiny drop in the bucket.

    You are extremely naive if you think using GPT makes any kind of notable shift in your total carbon footprint, it doesnt even move the dial at all.

    If you actually wanna pick something as a real target for reducing your carbon footprint, the two biggest contenders are:

    1. Use public transit, or better yet, bike/walk/run to work. If you work from home, good!
    2. Dry your clothes via hangers instead of a dryer

  • Thats irrelevant to the discussion at hand.

    That’s like arguing needles were a bad invention because many people use them for heroin.

    People using the tool wrong to hurt themselves doesn’t mean the tool is bad, it just means better regulations and education needs to be put in place.


  • pixxelkick@lemmy.worldtoFuck AI@lemmy.worldOn Exceptions
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    4 days ago

    The training costs effectively enter a “divide by infinity” argument given enough time.

    While they continue to train models at this time, eventually you hit a point where a given model can be used in perpetuity.

    Costs to train go down, whereas the usability of that model stretches on to effectively infinity.

    So you hit a point where you have a one time energy cost to make the model, and an infinite timescale to use it on.


  • Personally I don’t really trust the LLMs to synthesize disparate sources.

    The #1 best use case for LLMs is using them as extremely powerful fuzzy searchers on very large datasets, so stuff like hunting down published papers on topics.

    Dont actually use their output as the basis for reasoning, but use it to find the original articles.

    For example, as a software dev, I use them often to search for the specific documentation for what I need. I then go look at the actual documentation, but the LLM is exceptionally fast at locating the document itself for me.

    Basically, using them as a powerful resource to look up and find resources is key, and was why I was able to find documentation on the symptoms of my pet so fast. It would have taken me ages to find those esoteric published papers on my own, there’s so much to sift through, especially when many papers cover huge amounts of info and what Im looking for is one small piece of info in that one paper.

    But with an LLM I can trim down the search space instantly to a way way smaller set, and then go through that by hand. Thousands of papers turn into a couple in a matter of seconds.


  • pixxelkick@lemmy.worldtoFuck AI@lemmy.worldOn Exceptions
    link
    fedilink
    arrow-up
    1
    arrow-down
    1
    ·
    4 days ago

    The power server side for 5 minutes of chatgpt, vs the power burned browsing the internet to find the info on my own (which would take hours to manually sift through)

    Thats the comparison.

    Even though server side power consumption to run GPT is very high, its not so high that its more than hours and hours of a laptop usage


  • pixxelkick@lemmy.worldtoFuck AI@lemmy.worldOn Exceptions
    link
    fedilink
    arrow-up
    3
    arrow-down
    2
    ·
    edit-2
    4 days ago

    Not at tremendously less of a power cost anyways. My laptop draws 35W

    5 minutes of GPT is genuinely less power consumption than several hours of my laptop being actively used to do the searching manually. Laptops burn non trivial amounts of power when in use. Anyone who has held a laptop on their lap can attest to the fact they aren’t exactly running cold.

    Hell even a whole day of using your mobile phone is non trivial in power consumption, they also use 8~10W or so.

    Using GPT for dumb shit is arguably unethical, but only in the sense that baking cookies in the oven is. You gonna go and start yelling at people for making cookies? Cooking up one batch of cookies burns WAAAY more energy than fucking around with GPT. And yet I don’t see people going around bashing people for using their ovens to cook things as a hobby.

    There’s no good argument against what I did, by all metrics it genuinely was the ethical choice.


  • …no that’s not the summarization.

    The summarization is:

    if you reinforce your model via user feedback, via “likes” or “dislikes” or etc, such that you condition the model towards getting positive user feedback, it will start to lean towards just telling users whatever they want to hear in order to get those precious likes, cuz obviously you trained it to do that

    They demo’d in the same paper other examples.

    Basically, if you train it on likes, the model becomes duper sycophantic, laying it on super thick…

    Which should sound familiar to you.


  • pixxelkick@lemmy.worldtoFuck AI@lemmy.worldOn Exceptions
    link
    fedilink
    arrow-up
    8
    arrow-down
    12
    ·
    edit-2
    5 days ago

    AI saved my pets life. You won’t convince me it’s 100% all bad and there’s no “right” way to use it.

    The way it is trained isnt intellectual theft imo.

    It only becomes intellectual theft if it is used to generate something that then competes with and takes away profits from the original creators.

    Thus the intellectual theft only kicks in at generation time, but the onus is still on the AI owners for not preventing it

    However if I use AI to generate anything that doesn’t “compete” with anyone, then “intellectual theft” doesn’t matter.

    For example, when I used it to assist with diagnosing a serious issue my pet was having 2 months ago that was stumping even our vet and it got the answer right, which surprised our vet when we asked them to check a very esoteric possibility (which they dubious checked and then they were shocked to find something there.

    They asked us how on earth we managed to guess to check that place of all things, how could we have known. As a result we caught the issue very early when it was easy to treat and saved our pets life

    It was a gallbladder infection, and her symptoms had like 20 other more likely causes individually.

    But when I punched all her symptoms into GPT, everytime, it asserted if was likely the gallbladder. It had found some papers on other animals and mammals and how gallbladder infections cause that specific combo of symptoms rarely, and encouraged us to check it out.

    If you think “intellectual theft” still applies here, despite it being used to save an animals life, then you are the asshole. No one “lost” profit or business to this, no one’s intellectual property was infringed, and consuming the same amount of power it takes to cook 1 pizza in my oven to save my pets life is a pretty damn good trade, in my opinion.

    So, yes. I think I used AI ethically there. Fight me.


  • Exceedingly false representation of the actual experiment.

    They took Llama 3 and then trained it further on specific conditions (reinforcing it on “likes” / "thumbs up"s based on positive feedback from a simulated userbase)

    And then after that the scientists found the new model (which you can’t really call Llama 3 anymore, it’s been trained further and it’s behavior fundamentally altered) behaved like this when prior informed that the user was easily influenced by the model specifically

    What is important to gather though, is the fact that when a model gets trained on the metrics of “likes”, it starts to behave in a manner like this, telling the user whatever they want to hear… Which makes sense, the model is effectively getting trained to min/max positive feedback from users, rather than being trained on being right / correct

    But to try and represent this as a “real” chatbot’s behavior is definitely false, this was a model trained by scientists explicitly to test if this behavior happens under extreme conditioning.



  • Eh, Im not conservative but I do think this is a valid bipartisan criticism.

    If they pulled this stunt with a liberal leaning byelection Id consider it a greasy trick.

    I have no issue with putting a cap on ballot names via some kind of reasonable system (IE if you get more than x candidates, then the top y candidates can only go on the ballot based on who has the most vouches or whatever)

    That way you can avoid dozens and dozens of randos flooding the ballot to confuse people. It’s a valid hole to patch in our system that both sides should be on board with fixing in a reasonable way.



  • pixxelkick@lemmy.worldtoFuck AI@lemmy.worldEfficency
    link
    fedilink
    arrow-up
    3
    arrow-down
    1
    ·
    27 days ago

    If they are mandated, that’s just as bad I agree.

    At my company we have tonnes of in house Lunch and Learns (on paid time, non mandatory) that are effectively “I found this super useful thing and want others to know about it”

    And I’ll join these things, and see (person), who is on my team, in it too. Later I’ll hat with them about it, or at least try, and they’ll have zero clue wtf I’m talking about.

    And it becomes obvious they just joined the meeting to give the illusion of caring, they prolly were afk the whole time. And I suspect this cuz they often do the same for our “in team” mandatory important meetings discussing critical stuff on the project.


  • pixxelkick@lemmy.worldtoFuck AI@lemmy.worldEfficency
    link
    fedilink
    arrow-up
    17
    arrow-down
    5
    ·
    27 days ago

    Sorta just sounds like you can probably fire a few employees who don’t give a fuck.

    From experience, a lot of companies tend to be propped up by like 10% of their developers doing 90% of the work, maybe 50% of developers doing the last 10%, and then like 40% of developers either doing fuck all or actively harming the codebases and making more work for the other 60%.

    And more often than not, these people are the ones sending stuff like “AI Note Takers” merely to give the illusion of existing.

    In reality you have like three devs who actually care and do most of the work having a discussion and like 10 to 30 AFK participants with their cameras off who probably arent even listening or care.

    And the thing is, it shows later. The same devs have zero clue wtf is going on, they have zero clue how to do their job, and they constantly get their code sent back or they plague up some codebase that doesnt have reviewers to catch them.

    The AI note takers are just the new version of people showing up to meetings with their camera off and never speaking a word.

    Except now they burn orders of magitude more power to do it, which is awful.