“The chatbot gave wildly different answers to the same math problem, with one version of ChatGPT even refusing to show how it came to its conclusion.”

It’s getting worse. And because it’s a black box model they don’t know why. The computer science professor here likens it to how human students make mistakes… but human students make mistakes because they don’t have perfect recall, mishear things being told to them, are tired and/or not paying attention… A bunch of reason that basically relate to having a human body that needs food, rest and water. A thing a computer does not have.

The only reason ChatGPT should be getting math wrong is that it’s getting inputs that are wrong, but without view into it they can’t figure out where it’s getting it wrong and who told it the wrong info.

      • BraveSirZaphod@kbin.social
        link
        fedilink
        arrow-up
        6
        ·
        1 year ago

        I mean, they’ve gotta to be blowing absurd amounts of money at it. It’s not remotely cheap to build a massively complicated web service at that scale, and eventually the numbers need to start adding up. I’m sure they have several good monetization plans, but not every instance of a business attempting to stop hemorraghing money is a conspiracy. You’d be doing the exact same thing in their shoes.

        • Ragnell@kbin.socialOP
          link
          fedilink
          arrow-up
          6
          ·
          edit-2
          1 year ago

          Enshittification is not a conspiracy because a conspiracy requires communication and planning. Enshittification is just how idiots act when trying to make money.

          • Swedneck
            link
            fedilink
            arrow-up
            8
            ·
            1 year ago

            i wouldn’t even say that, it’s just the logical end point of capitalism

      • Ragnell@kbin.socialOP
        link
        fedilink
        arrow-up
        3
        ·
        1 year ago

        Yes.
        And extra infuriating they want to roll this stuff out after making it LESS reliable.

    • CarrieForle@kbin.social
      link
      fedilink
      arrow-up
      3
      ·
      1 year ago

      And there are more and more offline GPT AIs available for free. Now everyone with an above average computer can have their own chatGPT.

      • Hellsadvocate@kbin.social
        link
        fedilink
        arrow-up
        3
        ·
        1 year ago

        I mean an “average” computer would require a pretty beefy set of hardware. I think most of the average local llama’s would run fairly decently on a MacBook without issue nowadays (that m3 is going to be a pretty awesome beast). But the quality is pretty reduced even compared to something like 3.5 which most people thought wasn’t all that great.

        But really, I’m excited about researchers have access to more computer for smaller amounts (see this https://www.chatgptguide.ai/2023/07/20/worlds-largest-supercomputer-for-ai-training-is-out/) currently we have 1T models that are good, but we could pretty soon have 100T models from the open source community. Let’s see whether we can scale the hardware needs with the parameter growth so we don’t need A100s to run a decent model.

      • BarbecueCowboy@kbin.social
        link
        fedilink
        arrow-up
        2
        ·
        edit-2
        1 year ago

        It’s still pretty rough to selfhost an LLM. You can get one that’s kind of okay on an average computer, but to get a really competitive one running locally at a good speed, you need a huge amount of RAM that is still beyond most average users (VRAM for GPU based projects).

        I’ve been trying to get Vicuna going and the RAM usage is rough, 60gb is suggested, and I’ve got 64 and I think I need a lot more honestly.

  • Frog-Brawler@kbin.social
    link
    fedilink
    arrow-up
    17
    ·
    1 year ago

    Huh… so after months of being exposed to people that aren’t quite as smart as world class computer scientists and engineers, it gets dumber. Maybe it’s more human that I previously thought.

    • paper_clip@kbin.social
      link
      fedilink
      arrow-up
      5
      ·
      edit-2
      1 year ago

      it gets dumber

      In six months, ChatGPT will be talking up Brawndo, because it’s got the electrolytes that plants crave.

    • C4RP3_N0CT3M@kbin.social
      link
      fedilink
      arrow-up
      5
      ·
      1 year ago

      I wonder if it is in fact learning from people’s prompts; I didn’t think that was part of the operation. That’s a huge design flaw if so.

        • palordrolap@kbin.social
          link
          fedilink
          arrow-up
          1
          ·
          1 year ago

          A single word can be a full sentence, unless answers to either/or questions are not sentences.

          Or is this one of those logic things where a train is only a train when the railway engine is connected to something?

          • effingjoe@kbin.social
            link
            fedilink
            arrow-up
            2
            ·
            1 year ago

            A sentence needs a subject and a verb, if I remember grade school. Fun fact: “I’m.” is a sentence. There can be an implied “You” in there. Like “[You] Stop!” or “[You] Go!”

            • palordrolap@kbin.social
              link
              fedilink
              arrow-up
              1
              ·
              edit-2
              1 year ago

              The verb can be implied too. “Would you like mashed potatoes or fries?”

              “[I would like] Fries.”

              There’s also the joke sentence(?): “This sentence(,) no verb.”

              • effingjoe@kbin.social
                link
                fedilink
                arrow-up
                1
                ·
                1 year ago

                I’m pretty far away from an expert, but I’m pretty sure your example isn’t a “real” sentence. It implies the subject and the verb. (I, like).

  • stravanasu@lemmy.ca
    link
    fedilink
    English
    arrow-up
    11
    ·
    edit-2
    1 year ago

    For me it’s like using a coffee machine to measure a time lapse, and then complaining that it doesn’t always yield the same time lapse.

  • LineNoise@kbin.social
    link
    fedilink
    arrow-up
    9
    ·
    1 year ago

    For any question the number of incorrect answers is larger than the number of correct answers.

    This is a fundamental problem, constrained by energy costs, and one that will only be exacerbated as training datasets becomes more and more tainted by generated content.

  • Hazdaz@lemmy.world
    link
    fedilink
    arrow-up
    4
    ·
    1 year ago

    …but, but, but that guy and his prediction that all programmers will be unemployed within 2 years because of this technology?!?

    We couldn’t possibly have overhyped this technology when it was first announced, could we have?! Noooo! The internet, corporations, and our useless media would never do that!1!! /s

    • MrZigZag@kbin.social
      link
      fedilink
      arrow-up
      4
      ·
      1 year ago

      I think it’s the same hype train that said ten years ago that by now every vehicle would be self driving and all the truckers would be out of work. Or back when that first Avater movie came out that in a short while every movie and TV show would be in 3D.

      • Hazdaz@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        Oh absolutely. It is one hypetrain after another. VR was going to be the Next Big Thing… and then seemed to have fizzled faster than anything else. Companies would dump billions into these projects and then 6, 12, 18 months later cancel them when the public didn’t see interested. I am sure AI is not going totally away, but with the way it was being described just a few short weeks ago, it was the second coming of Jebus.

  • Talaraine@kbin.social
    link
    fedilink
    arrow-up
    4
    ·
    1 year ago

    I dunno, one of the very first tests I gave ChatGPT was one of those order of operations equations, and this was in like week 2. It gave me wrong answers every time, even when I explained the correct one. It was very polite about its mistakes, but this has always been here.

  • theyresocool@lemmy.world
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    I am writing a screenplay about a ship AI and problems I imagined and sleep is one of them. Just because it’s technology doesn’t mean it doesn’t need rest.