The majority of U.S. adults don’t believe the benefits of artificial intelligence outweigh the risks, according to a new Mitre-Harris Poll released Tuesday.

  • Uncle_Iroh@lemmy.world
    link
    fedilink
    English
    arrow-up
    94
    arrow-down
    37
    ·
    1 year ago

    Most of the U.S. adults also don’t understand what AI is in the slightest. What do the opinions of people who are not in the slightest educated on the matter affect lol.

    • Mac@mander.xyz
      link
      fedilink
      English
      arrow-up
      61
      arrow-down
      1
      ·
      1 year ago

      “What do the opinions of people who are not in the slightest educated on the matter affect”

      Judging by the elected leaders of the USA: quite a lot, in fact.

      • Armen12@lemm.ee
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        12
        ·
        1 year ago

        So you’d rather only the 1% get the right to vote? How about only white land owners? How about only men get to vote in this wonderful utopia of yours

        • 4am@lemm.ee
          link
          fedilink
          English
          arrow-up
          16
          arrow-down
          1
          ·
          1 year ago

          Stop stop there isn’t any straw left!

          • Armen12@lemm.ee
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            5
            ·
            edit-2
            1 year ago

            Making a mockery of the workforce who rely on jobs to not be homeless is not appropriate in this conversation, nor is it even an argument to begin with, it’s just a snobbish incel who probably lives in a gated community mocking poor people

      • Wolf_359@lemmy.world
        link
        fedilink
        English
        arrow-up
        45
        arrow-down
        19
        ·
        1 year ago

        Prime example. Atomic bombs are dangerous and they seem like a bad thing. But then you realize that, counter to our intuition, nuclear weapons have created peace and security in the world.

        No country with nukes has been invaded. No world wars have happened since the invention of nukes. Countries with nukes don’t fight each other directly.

        Ukraine had nukes, gave them up, promptly invaded by Russia.

        Things that seem dangerous aren’t always dangerous. Things that seem safe aren’t always safe. More often though, technology has good sides and bad sides. AI does and will continue to have pros and cons.

        • Hexagon@feddit.it
          link
          fedilink
          English
          arrow-up
          44
          arrow-down
          2
          ·
          1 year ago

          Atomic bomb are also dangerous because if someone end up launching one by mistake, all hell is gonna break loose. This has almost happened multiple times:

          https://en.wikipedia.org/wiki/List_of_nuclear_close_calls

          We’ve just been lucky so far.

          And then there are questionable state leaders who may even use them willingly. Like Putin, or Kim, maybe even Trump.

          • gravitas_deficiency@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            6
            ·
            1 year ago

            …and the development and use of nuclear power has been one of the most important developments in civil infrastructure in the last century.

            Nuclear isn’t categorically free from the potential to harm, but it can also do a whole hell of a lot for humanity if used the right way. We understand it enough to know how to use it carefully and safely in civil applications.

            We’ll probably get to the same place with ML… eventually. Right now, everyone’s just throwing tons of random problems at it to see what sticks, which is not what one could call responsible use - particularly when outputs are used in a widespread sense in production environments.

        • cheery_coffee@lemmy.ca
          link
          fedilink
          English
          arrow-up
          33
          arrow-down
          3
          ·
          1 year ago

          Alright, when the AI takes my job and I can’t feed my family while the billionaires add another digit to their net worth I’ll consider the pros.

          There’s about 0% chance we reform society for AI, it will just funnel more wealth to the rich. People claim it will open new jobs but I don’t see it.

          • Jerkface@lemmy.world
            link
            fedilink
            English
            arrow-up
            10
            arrow-down
            9
            ·
            edit-2
            1 year ago

            People have had the same concerns about automation since basically forever. Automation isn’t the problem. The people who use automation to perpetuate the systems that work against us will continue to find creative ways to exploit us with or without AI. Those people and those systems-- they are the problem. And believe it or not, that problem is imminently solvable.

            • cheery_coffee@lemmy.ca
              link
              fedilink
              English
              arrow-up
              10
              arrow-down
              1
              ·
              1 year ago

              It’s fair to compare but you can’t dismiss concerns based on that.

              Past automation often removed duplicate or superfluous work type things, AI removes thought work. It’s a fundamentally different kind of automation than we’ve seen before.

              It will make many things cheaper to do and easier to start some businesses, but it will also decimate workers. It’s also not something that’s generally available to lower classes to wield yet.

              It’s here but I don’t have to be optimistic.

              • Jerkface@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                2
                ·
                1 year ago

                I fully agree with everything you said. My point is more that if we look at AI as the culprit, we’re missing the point. If I may examine the language you are using a bit-

                AI removes thought work.

                Employers are the agents. They remove thought work.

                it will also decimate workers.

                Employers will decimate workers.

                It would be smart to enact legislation that will mitigate the damage employers enabled by AI will do to wokers, but they will continue to exploit us regardless.

                Using language that makes AI the antagonist helps tyrants deflect their overwhelming share of the blame. The responsible parties are people, who can and should be held accountable.

                • cheery_coffee@lemmy.ca
                  link
                  fedilink
                  English
                  arrow-up
                  6
                  ·
                  1 year ago

                  I don’t think you’re wrong either, but at the same time it’s not feasible for everyone to be their own agent and it’s not feasible to say employers can’t use AI.

                  I don’t know what the solution is, but I’m prepping for a sudden career change in the next few years.

                  • Jerkface@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    1 year ago

                    In general, progressive taxation can do quite a lot to ease the widening wealth gap. One such strategy is the robot tax. There exist other, perhaps better, legislative solutions, but more broadly we need to restore voting rights and diminish the influence the wealthy have on our political system so that smart, progressive legislation doesn’t have to fight tooth and nail against lobbying and other mechanisms that tie wealth to political influence.

              • Jerkface@lemmy.world
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                2
                ·
                edit-2
                1 year ago

                I want to avoid using the term solution, not least of all because implementation has its own set of challenges, but some of us used to dream that automation would do that work for us. Perhaps naively, some of us assumed that people just wouldn’t have to work as much. And perhaps I continue to be naive in thinking that that should still be our end goal. If automation reduces the required work hours by 20% with no reduction in profit, full time workers should have a 32 hour week with no reduction in income.

                But since employers will always pocket that money if given the option, we need more unionization, we need unions to fight for better contracts, we need legislation that will protect and facilitate them, and we need progressive taxation that will decouple workers most essential needs from their employers so they have more of a say in where and how they work, be that universal public services, minimum income guarantee, or what have you.

                We’re quite far behind in this fight but there has been some recent progress about which I am pretty optimistic.

                Edit: for clarification

                • Franzia@lemmy.blahaj.zone
                  link
                  fedilink
                  English
                  arrow-up
                  3
                  ·
                  edit-2
                  1 year ago

                  This was so very thoughtful, and after reading it, I feel optimistic too. Fuck yeah.

                  Edit: thank you.

          • PsychedSy@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            6
            ·
            1 year ago

            Technology tends to drive costs down and create more jobs, but in different areas. It’s not like there hasn’t been capture by the super rich in the past 150 years, but somehow we still enjoy better lives decade by decade.

        • richieadler@lemmy.myserv.one
          link
          fedilink
          English
          arrow-up
          21
          ·
          edit-2
          1 year ago

          If you’re from one of the countries with nukes, of course you’ll see it as positive. For the victims of the nuke-wielding countries, not so much.

        • walrusintraining@lemmy.world
          link
          fedilink
          English
          arrow-up
          12
          arrow-down
          1
          ·
          1 year ago

          That’s a good point, however just because the bad thing hasn’t happened yet, doesn’t mean it wont. Everything has pros and cons, it’s a matter of whether or not the pros outweigh the cons.

        • bogdugg@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          4
          ·
          1 year ago

          I don’t disagree with your overall point, but as they say, anything that can happen, will happen. I don’t know when it will happen; tomorrow, 50 years, 1000 years… eventually nuclear weapons will be used in warfare again, and it will be a dark time.

        • Techmaster@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          1 year ago

          No world wars have happened since the invention of nukes

          Except the current world war.

      • GigglyBobble@kbin.social
        link
        fedilink
        arrow-up
        7
        arrow-down
        5
        ·
        edit-2
        1 year ago

        You need to understand to correctly classify the danger though.

        Otherwise you make stupid decisions such as quiting nuclear energy in favor of coal because of an incident like Fukushima even though that incident just had a single casualty due to radiation.

      • StereoTrespasser@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        arrow-down
        13
        ·
        1 year ago

        I’m over here asking chatGPT for help with a pandas dataframe and loving every minute of it. At what point am I going to feel the effects of nuclear warfare?

        • walrusintraining@lemmy.world
          link
          fedilink
          English
          arrow-up
          24
          arrow-down
          5
          ·
          1 year ago

          I’m confused how this is relevant. Just pointing out this is a bad take, not saying nukes are the same as AI. chatGPT isn’t the only AI out there btw. For example NYC just allowed the police to use AI to profile potential criminals… you think that’s a good thing?

              • Jerkface@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                2
                ·
                1 year ago

                The take is “let’s not forget to hold people accountable for the shitty things they do.” AI is not a killing machine. Guns aren’t particularly productive.

      • Uncle_Iroh@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        4
        ·
        1 year ago

        You chose an analogy with the most limited scope possible but sure I’ll go with it. To understand how dangerous an atomic bomb is exactly without just looking up a hiroshima you need to have atleast some knowledge on the subject, you’d also have to understand all the nuances etc. The thing about AI is that most people haven’t a clue what it is, how it works, what it can do. They just listen to the shit their telegram loving uncle spewed at the family gathering. A lot of people think AI is fucking sentient lmao.

        • walrusintraining@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          1
          ·
          1 year ago

          I don’t think most people think ai is sentient. In my experience the people that think that are the ones who think they’re the most educated saying stuff like “neural networks are basically the same as a human brain.”

          • Uncle_Iroh@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            2
            ·
            1 year ago

            You don’t think, yet a software engineer from google, Blake Lemoine, thought LaMDA was sentient. He took a lot of idiots down with him when he went public with said claims. Not to mention the movies that were made with the premise of sentient AI.

            Your anecdotal experience and your feelings don’t in the slightest affect the reality that there is tons of people who think AI is sentient and will somehow start some fucking robo revolution.

      • WhyIDie@kbin.social
        link
        fedilink
        arrow-up
        5
        arrow-down
        9
        ·
        edit-2
        1 year ago

        you also don’t have to understand how 5g works to know it spreads covid /s

        point is, I don’t see how your analogy works beyond the limited scope of only things that result in an immediate loss of life

        • walrusintraining@lemmy.world
          link
          fedilink
          English
          arrow-up
          10
          arrow-down
          1
          ·
          1 year ago

          I don’t need to know the ins and outs of how the nazi regime operated to know it was bad for humanity. I don’t need to know how a vaccine works to know it’s probably good for me to get. I don’t need to know the ins and outs of personal data collection and exploitation to know it’s probably not good for society. There are lots of examples.

          • WhyIDie@kbin.social
            link
            fedilink
            arrow-up
            3
            ·
            edit-2
            1 year ago

            okay, I’ll concede, my scope also was pretty limited. I still stand by not trusting the public with deciding what’s the best use of AI, when most people think what we have now is anything more than statistics supercharged in its implementation.

          • linearchaos@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            I can certainly give that “you” don’t need to know but there are a lot of differing opinions on even the things you’re talking about inside of the people that are in this very community.

            I would say that the Royal we need to know because there are a lot of opinions on facts that don’t line up with actual facts for a lot of people. Sure, not you, not me but a hell of a lot of people.

            • walrusintraining@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              1
              ·
              1 year ago

              I don’t disagree that people are stupid, but the majority of people got/supported the vaccine. Majority is sometimes a good indicator, that’s how democracy works. Again, it’s not perfect, but it’s not useless either.

    • kitonthenet@kbin.social
      link
      fedilink
      arrow-up
      7
      arrow-down
      2
      ·
      1 year ago

      Because they live in the same society as you, and they get to decide who goes to jail as much as you do

    • Franzia@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      3
      ·
      1 year ago

      Well and being a snob about it doesn’t help. If all the average joe knows about AI is what google or openAI pushed to corporate media, that shouldn’t be where the conversation ends.

      • Uncle_Iroh@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        4
        ·
        1 year ago

        The average joe can have their thoughts on it all they want, but their opinions on the matter aren’t really valid or of any importance. AI is best left to the people who have a deep knowledge of the subject, just as nuclear fusion is best left to scientists studying the field. I’m not going to tell average Joe the mechanic that I think the engine he just revised might just blow up, because I have no fucking clue about it. Sure I have some very basic knowledge of it, that’s pretty much where it end too though.

    • gravitas_deficiency@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      1 year ago

      You can not know the nuanced details of something and still be (rightly) sketched out by it.

      I know a decent amount about the technical implementation details, and that makes me trust its use in (what I perceive as) inappropriate contexts way less than the average layperson.

    • Armen12@lemm.ee
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      8
      ·
      1 year ago

      What a terrible thing to say, they’re human beings so I hope they matter to you

          • Uncle_Iroh@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            7
            ·
            1 year ago

            I am a terrible person simply because they don’t matter to me? Do you cry for every death victim your military caused? Do you cry for every couple with a stillborn baby? No, you don’t. You think it’s shitty, because it is. But you don’t really care, they don’t truly matter to you. The way you throw those words around makes their meaning less.

            • Armen12@lemm.ee
              link
              fedilink
              English
              arrow-up
              3
              ·
              1 year ago

              Lot of words to just say you’re a terrible person, we got it already, you don’t need to explain why you’re terrible