• @[email protected]
    link
    fedilink
    English
    663 months ago

    This is the best news. It’s one thing for AI to assist, it’s another to replace. Fuck em

    • @[email protected]
      link
      fedilink
      133 months ago

      Agreed. There’s a difference between using AI to assist workers and using it to replace them outright

      • @[email protected]
        link
        fedilink
        23 months ago

        If it was actually AI (or good) it’d be different.

        I’m ashamed of millennial gamers for not forcing the terminology to be VI. Did we learn nothing from Mass Effect?

        • @[email protected]
          link
          fedilink
          7
          edit-2
          3 months ago

          It’s a marketing thing. Calling LLM’s “AI” was a very intentional move, to evoke that sense of hyperintelligence. Whether it’s truly an artifical intelligence up for debate, but calling them AI absolutely helped them gain attention (good and bad).

          Also, obligatory “shut up Avina”.

  • @BatmanAoD
    link
    523 months ago

    It’s actually quite amusing to me that Wikipedia is an authority on “reliability”. It makes perfect sense, but can you imagine explaining that to a public school teacher twenty years ago?

  • @Kissaki
    link
    English
    203 months ago

    January 2023, Futurism brought widespread attention to the issue and discovered that the articles were full of plagiarism and mistakes. […] After the revelation, CNET management paused the experiment, but the reputational damage had already been done.

    So the “AI experiment” is not active anymore. But the damage is already done.

    It was also new to me that Wikipedia puts time-based reliability qualifiers on sources. It makes sense of course. And this example shows how a source can be good and reliable in the past, but not anymore - and differentiating that is important and necessary.

  • AutoTL;DRB
    link
    fedilink
    English
    153 months ago

    This is the best summary I could come up with:


    Wikipedia has downgraded tech website CNET’s reliability rating following extensive discussions among its editors regarding the impact of AI-generated content on the site’s trustworthiness, as noted in a detailed report from Futurism.

    The decision reflects concerns over the reliability of articles found on the tech news outlet after it began publishing AI-generated stories in 2022.

    Shortly after the CNET news broke in January 2023, Wikipedia editors began a discussion thread on the Reliable Sources project page about the publication.

    “CNET, usually regarded as an ordinary tech RS [reliable source], has started experimentally running AI-generated articles, which are riddled with errors,” wrote a Wikipedia editor named David Gerard.

    Futurism reports that the issue with CNET’s AI-generated content also sparked a broader debate within the Wikipedia community about the reliability of sources owned by Red Ventures, such as Bankrate and CreditCards.com.

    In response to the downgrade and the controversies surrounding AI-generated content, CNET issued a statement that claims that the site maintains high editorial standards.


    The original article contains 528 words, the summary contains 163 words. Saved 69%. I’m a bot and I’m open source!

  • @[email protected]
    link
    fedilink
    73 months ago

    Honestly they should just create a new category called AI generated. Reliability in journalism should only be for humans.

  • @Kissaki
    link
    English
    53 months ago

    CNET began publishing articles written by an AI model under the byline “CNET Money Staff”.

    (emphasis mine)

    What a label. I assume that “byline” was their “article author”? “Money Staff”. Baffling.