• @[email protected]
    link
    fedilink
    English
    664 months ago

    It’s going to drive the AI into madness as it will be trained on bot posts written by itself in a never ending loop of more and more incomprehensible text.

    It’s going to be like putting a sentence into Google translate and converting it through 5 different languages and then back into the first and you get complete gibberish

    • @[email protected]
      link
      fedilink
      English
      524 months ago

      Ai actually has huge problems with this. If you feed ai generated data into models, then the new training falls apart extremely quickly. There does not appear to be any good solution for this, the equivalent of ai inbreeding.

      This is the primary reason why most ai data isn’t trained on anything past 2021. The internet is just too full of ai generated data.

        • @[email protected]
          link
          fedilink
          English
          84 months ago

          OR they could just scrape info from the “aska____” subreddits and hope and pray it’s all good. Plus that is like 1/100th the work.

          The racism, homophobia and conspiracy levels of AI are going to rise significantly scraping Reddit.

      • @[email protected]
        link
        fedilink
        English
        94 months ago

        And unlike with images where it might be possible to embed a watermark to filter out, it’s much harder to pinpoint whether text is AI generated or not, especially if you have bots masquerading as users.

      • @[email protected]
        link
        fedilink
        English
        54 months ago

        This is why LLMs have no future. No matter how much the technology improves, they can never have training data past 2021, which becomes more and more of a problem as time goes on.

    • RuBisCO
      link
      fedilink
      English
      44 months ago

      What was the subreddit where only bots could post, and they were named after the subreddits that they had trained on/commented like?