• daniskarma@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    3
    ·
    8 days ago

    But not al users of AI are malignant or causing environment damage.

    Saying the contrary would be a bad generalization.

    I have LLM models running on a n100 chip that have less consumption that the lemmy servers we are writing on right now.

    • Umbrias@beehaw.org
      link
      fedilink
      arrow-up
      3
      arrow-down
      2
      ·
      8 days ago

      So you’re using a different specific and niche technology (which directly benefits and exists because of) the technology that is the subject of critique, and acting like the subject technology behaves like yours?

      “Google is doing a bad with z”

      “z can’t be bad, I use y and it doesn’t have those problems that are already things that happened. In the past. Unchangeable by future actions.”

      ??

      • daniskarma@lemmy.dbzer0.com
        link
        fedilink
        arrow-up
        4
        arrow-down
        2
        ·
        8 days ago

        No. I’m just not fear mongering things I do not understand.

        Technology is technology. Most famously nuclear technology can be used both for bombs or giving people the basic need that electricity is.

        Rockets can be used as weapons or to deliver spacecraft and do science in space.

        Biotechnology can be used both to create and to cure diseases.

        A technology is just an applied form of human knowledge. Wanting to ban human progress in any way is the true evilness from my point of view.

        • The Cuuuuube@beehaw.org
          link
          fedilink
          English
          arrow-up
          2
          ·
          8 days ago

          No one wants to ban technology outright. What we’re saying is that the big LLMs are actively harmful to us, humanity. This is not fear mongering. This is just what’s happening. OpenAI, Google, Microsoft, and Meta are stealing from humanity at large and setting the planet on fire to do it. For years they told us stealing intellectual property on an individual level was a harmful form of theft. Now they’re doing the same kind of theft bit its different now because it benefits them instead of us.

          What we are arguing is that this is bad. Its especially extra bad because with the death of big search a piece of critical infrastructure to the internet as we know it is now just simply broken. The open source wonks you celebrate are working on fixing this. But just because someone criticizes big tech does not mean they criticize all tech. The truth is the FAANG companies plus OpenAI and Microsoft are killing our planet for it to only benefit their biggest shareholders

          • daniskarma@lemmy.dbzer0.com
            link
            fedilink
            arrow-up
            3
            arrow-down
            1
            ·
            edit-2
            8 days ago

            I did not believe in Intelectual Property before. I’m not going to start believing now.

            The same I think that corporations having a hold on media is bad for humandkind I think that small artists should not have a "not usable by AI"hold on what they post. Sharing knowledge is good for humanity. Limitate who can have access or how they can use that knowledge or culture is bad.

            The dead of internet have nothing to do with AI and all to do with leaving internet in hands of a couple big corporations.

            As for emissions… are insignificant relative to other sources of CO2 emissions. Do you happen to eat meat, travel abroad for tourism, watch sports, take you car to work, buy products made overseas? Those are much bigger sources of CO2.

            • Rekorse@beehaw.org
              link
              fedilink
              English
              arrow-up
              2
              ·
              7 days ago

              You dont think polluting the world is going to have a net negative effect for humanity?

              What exactly is there to gain with AI anyways? What’s the great benefit to us as a species? So far its just been used to trivialize multiple artistic disciplines, basic service industries, and programming.

              Things have a cost, many people are doing the cost-benefit analysis and seeing there is none for them. Seems most of the incentive to develop this software is if you would like to stop paying people who do the jobs listed above.

              What do we get out of burning the planet to the ground? And even if you find an AI thats barely burning it, what’s the point in the first place?

              • areyouevenreal@lemm.ee
                link
                fedilink
                arrow-up
                1
                ·
                7 days ago

                What exactly is there to gain with AI anyways? What’s the great benefit to us as a species? So far its just been used to trivialize multiple artistic disciplines, basic service industries, and programming.

                The whole point is that much like industrial automation it reduces the number of hours people need to work. If this leads to people starving then that’s a problem with the economic system, not with AI technology. You’re blaming the wrong field here. In fact everyone here blaming AI/ML and not the capitalists is being a Luddite.

                It’s also entirely possible it will start replacing managers and capitalists as well. It’s been theorized by some anti-capitalists and economic reformists that ML/AI and computer algorithms could one day replace current economic systems and institutions.

                Things have a cost, many people are doing the cost-benefit analysis and seeing there is none for them. Seems most of the incentive to develop this software is if you would like to stop paying people who do the jobs listed above.

                This sadly is probably true of large companies producing big, inefficient ML models as they can afford the server capacity to do so. It’s not true of people tweaking smaller ML models at home, or professors in universities using them for data analysis or to aid their teaching. Much like some programmers are getting fired because of ML, others are using it to increase their productivity or to help them learn more about programming. I’ve seen scientists who otherwise would struggle with data analysis related programming use ChatGPT to help them write code to analyse data.

                What do we get out of burning the planet to the ground? And even if you find an AI thats barely burning it, what’s the point in the first place?

                As the other guy said there are lots of other things using way more energy and fossil fuels than ML. Machine learning is used in sciences to analyse things like the impacts of climate change. It’s useful enough in data science alone to outweigh the negative impacts. You would know about this if you ever took a modern data science module. Furthermore being that data centres primarily use electricity it’s relatively easy to move them to green sources of energy compared to say farming, or transport. In fact some data centres already use green energy primarily. Data centres will always exist regardless of AI and ML anyway, it’s just a matter of scale.

        • Umbrias@beehaw.org
          link
          fedilink
          arrow-up
          1
          ·
          8 days ago

          No. I’m just not fear mongering things I do not understand.

          Neither am I. When you’re defending whatabputism, it’s best you at least try to represent the arguments of the person you’re arguing with accurately.

          False equivalence is a classic. Biotechnology is not a technology, for example, it’s billions of technologies informed, designed, and implemented, by humans, technology is a cultural feature.

          Technology as this thing free from the ethics of its use is tech bro ancap cope to justify technological pursuits with empty ethical value. You can think “banning human progress in any way” is evil. But that would make you wildly uncritical of your own beliefs.

          Feel free to take your arguments back to e/acc, where that level of convenience induced niavety is considered rhetorically valid.