Abacus.ai:

We recently released Smaug-72B-v0.1 which has taken first place on the Open LLM Leaderboard by HuggingFace. It is the first open-source model to have an average score more than 80.

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      4
      arrow-down
      4
      ·
      5 months ago

      And at 72 billion parameters it’s something you can run on a beefy but not special-purpose graphics card.

      • glimse@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        2
        ·
        5 months ago

        Based on the other comments, it seems like this needs 4x as much ram than any consumer card has

        • FaceDeer@kbin.social
          link
          fedilink
          arrow-up
          4
          ·
          5 months ago

          It hasn’t been quantized, then. I’ve run 70B models on my consumer graphics card at a reasonably good tokens-per-second rate.

        • DarkThoughts@fedia.io
          link
          fedilink
          arrow-up
          2
          ·
          5 months ago

          I’m curious how local generation goes with potentially dedicated AI extensions using stuff like tensor cores and their own memory instead of hijacking parts of consumer GPUs for this.