• SubArcticTundra@lemmy.ml
      link
      fedilink
      arrow-up
      14
      arrow-down
      1
      ·
      8 months ago

      Agreed. There’s a difference between using AI to assist workers and using it to replace them outright

      • DragonTypeWyvern@literature.cafe
        link
        fedilink
        arrow-up
        6
        arrow-down
        4
        ·
        8 months ago

        If it was actually AI (or good) it’d be different.

        I’m ashamed of millennial gamers for not forcing the terminology to be VI. Did we learn nothing from Mass Effect?

        • funnystuff97@lemmy.world
          link
          fedilink
          arrow-up
          8
          arrow-down
          1
          ·
          edit-2
          8 months ago

          It’s a marketing thing. Calling LLM’s “AI” was a very intentional move, to evoke that sense of hyperintelligence. Whether it’s truly an artifical intelligence up for debate, but calling them AI absolutely helped them gain attention (good and bad).

          Also, obligatory “shut up Avina”.

  • BatmanAoD@programming.dev
    link
    fedilink
    arrow-up
    53
    arrow-down
    1
    ·
    8 months ago

    It’s actually quite amusing to me that Wikipedia is an authority on “reliability”. It makes perfect sense, but can you imagine explaining that to a public school teacher twenty years ago?

    • Rodeo@lemmy.ca
      link
      fedilink
      arrow-up
      5
      arrow-down
      4
      ·
      8 months ago

      Try explaining anything to a public school teacher lol. They always think they know better.

  • Kissaki@programming.dev
    link
    fedilink
    English
    arrow-up
    21
    ·
    8 months ago

    January 2023, Futurism brought widespread attention to the issue and discovered that the articles were full of plagiarism and mistakes. […] After the revelation, CNET management paused the experiment, but the reputational damage had already been done.

    So the “AI experiment” is not active anymore. But the damage is already done.

    It was also new to me that Wikipedia puts time-based reliability qualifiers on sources. It makes sense of course. And this example shows how a source can be good and reliable in the past, but not anymore - and differentiating that is important and necessary.

  • AutoTL;DR@lemmings.worldB
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    1
    ·
    8 months ago

    This is the best summary I could come up with:


    Wikipedia has downgraded tech website CNET’s reliability rating following extensive discussions among its editors regarding the impact of AI-generated content on the site’s trustworthiness, as noted in a detailed report from Futurism.

    The decision reflects concerns over the reliability of articles found on the tech news outlet after it began publishing AI-generated stories in 2022.

    Shortly after the CNET news broke in January 2023, Wikipedia editors began a discussion thread on the Reliable Sources project page about the publication.

    “CNET, usually regarded as an ordinary tech RS [reliable source], has started experimentally running AI-generated articles, which are riddled with errors,” wrote a Wikipedia editor named David Gerard.

    Futurism reports that the issue with CNET’s AI-generated content also sparked a broader debate within the Wikipedia community about the reliability of sources owned by Red Ventures, such as Bankrate and CreditCards.com.

    In response to the downgrade and the controversies surrounding AI-generated content, CNET issued a statement that claims that the site maintains high editorial standards.


    The original article contains 528 words, the summary contains 163 words. Saved 69%. I’m a bot and I’m open source!

  • Kissaki@programming.dev
    link
    fedilink
    English
    arrow-up
    5
    ·
    8 months ago

    CNET began publishing articles written by an AI model under the byline “CNET Money Staff”.

    (emphasis mine)

    What a label. I assume that “byline” was their “article author”? “Money Staff”. Baffling.