• givesomefucks@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    7
    ·
    13 hours ago

    Stop burning the planet down to generate social media comments

    I mean, I thought it would be obvious my issue was with using AI to do so…

    Even if it had been a serious question.

    But, to be fair I was thinking of what a normal.person would be able to parse, and not people who’s critical thinking had already atrophied from offloading to AI.

    They probably don’t have any idea what I meant and would need it explicitly spelled out.

    • thefactremains@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      2
      ·
      edit-2
      6 hours ago

      If It makes you feel better (or at least more educated)……the entire three-prompt interaction to calculate dogpower consumed roughly the same amount of energy as making three Google searches.

      A single Google search uses about 0.3 watt-hours (Wh) of energy. A typical AI chat query with a modern model uses a similar amount, roughly 0.2 to 0.34 Wh. Therefore, my dogpower curiosity discussion used approximately 0.9 Wh in total.

      For context, this is less energy than an LED lightbulb consumes in a few minutes. While older AI models were significantly more energy-intensive (sometimes using 10 times more power than a search) the latest versions have become nearly as efficient for common tasks.

      For even more context, It would take approximately 9 Lemmy comments to equal the energy consumed by my 3-prompt dogpower calculation discussion.

      • verdi@feddit.org
        link
        fedilink
        English
        arrow-up
        1
        ·
        3 hours ago

        This is not correct and can easily be disproven, even if one assumes less than 480g/Kwh.

        And that is ignoring the infrastructure necessary to perform a search vs AI query.

        • thefactremains@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          49 minutes ago

          You’re absolutely right! According to the research you cited, the energy use is actually much LOWER than I stated in my comment.

          Your source shows that an efficient AI model (Qwen 7B) used only 0.058 watt-hours (Wh) per query.

          Based on that, my entire 3-prompt chat only used about 0.17 Wh. That’s actually less energy than a single Google search (~0.3 Wh). Thanks for sharing the source and correcting me.

    • Nima@leminal.space
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      1
      ·
      13 hours ago

      I didn’t realize it even was ai generated. but even if it is, that’s still a fairly off-putting way to respond.

      • givesomefucks@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        6
        ·
        13 hours ago

        but even if it is, that’s still a fairly off-putting way to respond.

        No you’re right…

        It’s not like it’s literally burning our planet down and the people profiting off it aren’t tech bro fascists…