• No_Ones_Slick_Like_Gaston@lemmy.world
    link
    fedilink
    arrow-up
    16
    ·
    3 months ago

    There’s a lot of explaining to do for Meta, OpenAI, Claude and Google gemini to justify overpaying for their models now that there’s l a literal open source model that can do the basics.

    • suoko@feddit.itOP
      link
      fedilink
      arrow-up
      4
      ·
      3 months ago

      I’m testing right now vscode+continue+ollama+gwen2.5-coder. With a simple GPU it’s already OK.

    • suoko@feddit.itOP
      link
      fedilink
      arrow-up
      4
      ·
      3 months ago

      You still need an expensive hardware to run it. Unless myceliumwebserver project will start

    • normalexit@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      3 months ago

      The cost is a function of running an LLM at scale. You can run small models on consumer hardware, but the real contenders are using massive amounts of memory and compute on GPU arrays (plus electricity and water for cooling).

      ChatGPT is reportedly losing money on their $200/mo pro subscription plan.

    • howrar@lemmy.ca
      link
      fedilink
      arrow-up
      1
      ·
      3 months ago

      The same could be said for when Meta “open sourced” their models. Someone has to do the training, or else these models wouldn’t exist in the first place.