As soon as Apple announced its plans to inject generative AI into the iPhone, it was as good as official: The technology is now all but unavoidable. Large language models will soon lurk on most of the world’s smartphones, generating images and text in messaging and email apps. AI has already colonized web search, appearing in Google and Bing. OpenAI, the $80 billion start-up that has partnered with Apple and Microsoft, feels ubiquitous; the auto-generated products of its ChatGPTs and DALL-Es are everywhere. And for a growing number of consumers, that’s a problem.

Rarely has a technology risen—or been forced—into prominence amid such controversy and consumer anxiety. Certainly, some Americans are excited about AI, though a majority said in a recent survey, for instance, that they are concerned AI will increase unemployment; in another, three out of four said they believe it will be abused to interfere with the upcoming presidential election. And many AI products have failed to impress. The launch of Google’s “AI Overview” was a disaster; the search giant’s new bot cheerfully told users to add glue to pizza and that potentially poisonous mushrooms were safe to eat. Meanwhile, OpenAI has been mired in scandal, incensing former employees with a controversial nondisclosure agreement and allegedly ripping off one of the world’s most famous actors for a voice-assistant product. Thus far, much of the resistance to the spread of AI has come from watchdog groups, concerned citizens, and creators worried about their livelihood. Now a consumer backlash to the technology has begun to unfold as well—so much so that a market has sprung up to capitalize on it.


Obligatory “fuck 99.9999% of all AI use-cases, the people who make them, and the techbros that push them.”

  • Quokka@quokk.au
    link
    fedilink
    arrow-up
    22
    arrow-down
    2
    ·
    5 months ago

    Good thing about this is it’s self selecting, all the luddites who refuse to use AI will find themselves at a disadvantage just the same as refusing to use a computer isn’t doing anyone any favours.

    • SkyNTP@lemmy.ml
      link
      fedilink
      arrow-up
      39
      ·
      5 months ago

      The benefit of AI is overblown for a majority of product tiers. Remember how everything was supposed to be block chain? And metaverse? And web 3.0? And dot.com? This is just the next tech trend for dumb VCs to throw money at.

      • CanadaPlus@lemmy.sdf.org
        link
        fedilink
        arrow-up
        22
        ·
        5 months ago

        Yes, it’s very hyped and being overused. Eventually the bullshit artists will move on to the next buzzword, though, and then there’s plenty of tasks it is very good at where it will continue to grow.

      • Kedly@lemm.ee
        link
        fedilink
        arrow-up
        7
        ·
        5 months ago

        Yeah, but the dot com bubble didnt kill the internet entirely, and the video game bubble that prompted nintendo to create its own quality seal of approval didnt kill video games entirely. This fad, when it dies, already has useful applications and when the bubble pops, those applications will survive

      • Zaktor@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        7
        ·
        5 months ago

        Except those things didn’t really solve any problems. Well, dotcom did, but that actually changed our society.

        AI isn’t vaporware. A lot of it is premature (so maybe overblown right now) or just lies, but ChatGPT is 18 months old and look where it is. The core goal of AI is replacing human effort, which IS a problem wealthy people would very much like to solve and has a real monetary benefit whenever they can. It’s not going to just go away.

        • BurningRiver@beehaw.org
          link
          fedilink
          arrow-up
          13
          ·
          edit-2
          5 months ago

          Can you trust whatever AI you use, implicitly? I already know the answer, but I really want to hear people say it. These AI hype men are seriously promising us capabilities that may appear down the road, without actually demonstrating use cases that are relevant today. “Some day it may do this, or that”. Enough already, it’s bullshit.

          • Zaktor@sopuli.xyz
            link
            fedilink
            English
            arrow-up
            3
            ·
            edit-2
            5 months ago

            Yes? AI is a lot of things, and most have well-defined accuracy metrics that regularly exceed human performance. You’re likely already experiencing it as a mundane tool you don’t really think about.

            If you’re referring specifically to generative AI, that’s still premature, but as I pointed out, the interactive chat form most people worry about is 18 months old and making shocking levels of performance gains. That’s not the perpetual “10 years away” it’s been for the last 50 years, that’s something that’s actually happening in the near term. Jobs are already being lost.

            People are scared about AI taking over because they recognize it (rightfully) as a threat. That’s not because they’re worthless. If that were the case you’d have nothing to fear.

        • PeteBauxigeg@lemm.ee
          link
          fedilink
          arrow-up
          7
          ·
          5 months ago

          ChatGPT didn’t begin 18 months ago, the research that it originates from has been ongoing for years, how old is alexnet?

          • Zaktor@sopuli.xyz
            link
            fedilink
            English
            arrow-up
            2
            ·
            5 months ago

            I’m referencing ChatGPT’s initial benchmarks to its capabilities to today. Observable improvements have been made in less than two years. Even if you just want to track time from the development of modern LLM transformers (All You Need is Attention/BERT), it’s still a short history with major gains (alexnet isn’t really meaningfully related). These haven’t been incremental changes on a slow and steady march to AI sometime in the scifi scale future.

              • Zaktor@sopuli.xyz
                link
                fedilink
                English
                arrow-up
                1
                ·
                5 months ago

                No, not even remotely. And that’s kind of like citing “the first program to run on a CPU” as the start of development for any new algorithm.

                • PeteBauxigeg@lemm.ee
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  5 months ago

                  As far as I can find out, there was only one use of GPUs prior to alexnet for CNN, and it certainty didn’t have the impact alexnet had. Besides, running this stuff on GPUs not CPUs is a relevant technological breakthrough, imagine how slow chayGPT would be running on a CPU. And it’s not at all as obvious as it seems, most weather forecasts still run on CPU clusters despite them being obvious targets for GPUs.

                  • Zaktor@sopuli.xyz
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    5 months ago

                    What? Alexnet wasn’t a breakthrough in that it used GPUs, it was a breakthrough for its depth and performance on image recognition benchmarks.

                    We knew GPUs could speed up neural networks in 2004. And I’m not sure that was even the first.

      • jarfil@beehaw.org
        link
        fedilink
        arrow-up
        2
        ·
        5 months ago

        Blockchain is used in more places than you’d expect… not the P2P version, or the “cryptocurrency” version, just the “signature based chained list” one. For example, all signed Git commits, form a blockchain.

        The Metaverse has been bubbling on and off for the last 30 years or so, each iteration it gets slightly better… but it keeps failing at the same points (I think I wrote about it 20+ years ago, with points which are still valid).

        Web 3.0, not to be confused with Web3, is the Semantic Web, in the works for the last 20+ years. Web3 is a cool idea for a post-scarcity world, pretty useless right now.

        Dot.com was the original Web bubble… and here we are, on the Web, post-bubble.

    • mayooooo@beehaw.org
      link
      fedilink
      arrow-up
      37
      ·
      5 months ago

      Luddites were not idiots, they were people who understood the only use of tech at their time was to fuck them. Like this complete garbage shit is going to be used to fuck people. Nobody is opposed to having tools, we just don’t like Musk fanboys blowing spit bubbles while trying to get peepee hard

    • Rozaŭtuno@lemmy.blahaj.zone
      link
      fedilink
      arrow-up
      12
      ·
      5 months ago

      Good thing about this is it’s self selecting, all the technobros who obsess over AI will find themselves bankrupted like when the blockchain bubble bursted.

      • Echo Dot@feddit.uk
        link
        fedilink
        arrow-up
        9
        ·
        5 months ago

        The blockchain bubble burst because everyone with a brain could see from the start that it wasn’t really a useful technology. AI actually does have some advantages so they won’t go completely bust as long as they don’t go completely mad and start declaring that it can do things it can’t do.

        • Rozaŭtuno@lemmy.blahaj.zone
          link
          fedilink
          arrow-up
          11
          ·
          5 months ago

          they won’t go completely bust as long as they don’t go completely mad and start declaring that it can do things it can’t do.

          Which is exactly what’s happening.

          • Echo Dot@feddit.uk
            link
            fedilink
            arrow-up
            6
            ·
            5 months ago

            The fact that it is useful technology though means they’ll always have a fullback. It’s not going to go way like bitcoin I guarantee it.

            • technocrit@lemmy.dbzer0.com
              link
              fedilink
              arrow-up
              7
              ·
              edit-2
              5 months ago

              Bitcoin went away? It’s at like $67k today. Personally I prefer sustainable cryptos but unfortunately Bitcoin is far from dead.

              And sure, there’s lots of data processing and statistics that’s extremely useful. That’s been the case for a long time. But anybody talking about “intelligence” is a con.

              • Zaktor@sopuli.xyz
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                5 months ago

                GameStop also went up. It doesn’t mean GameStop is a good company that’s valuable to own, it just means that dumb people will buy things without value if they think they can eventually pass the bag to someone else. If someone purchased every share of Amazon they’d own a massive asset that would continually produce value for them. If someone bought every outstanding Bitcoin, it both wouldn’t produce ongoing value, but the value would actually go to zero.

        • Sonori@beehaw.org
          link
          fedilink
          arrow-up
          5
          ·
          edit-2
          5 months ago

          Like say, treating a program that shows you the next most likely word to follow the previous one on the internet like it is capable of understanding a sentence beyond this is the most likely string of words to follow the given input on the internet. Boy it sure is a good thing no one would ever do something so brainless as that in the current wave of hype.

          It’s also definitely becuse autocompletes have made massive progress recently, and not just because we’ve fed simpler and simpler transformers more and more data to the point we’ve run out of new text on the internet to feed them. We definitely shouldn’t expect that the field as a whole should be valued what it was say back in 2018, when there were about the same number of practical uses and the foucus was on better programs instead of just throwing more training data at it and calling that progress that will continue to grow rapidly even though the amount of said data is very much finite.

      • Kedly@lemm.ee
        link
        fedilink
        arrow-up
        3
        ·
        5 months ago

        How does using free software to play dress up with anime characters bankrupt me financially?