• DavidGarcia@feddit.nl
    link
    fedilink
    arrow-up
    20
    arrow-down
    1
    ·
    2 months ago

    It is kind of interesting how open machine learning already is without much explicit advocacy for it.

    It’s the only field I can think of where the open version is just a few months behind SOTA in all of IT.

    Open training pipelines and open data are the only aspects that could still use improvements in ML, but there are plenty of projects that are near-SOTA and fully open.

    ML is extremely open compared to consumer mobile or desktop apps that are always ~10 years behind SOTA

    • underscores@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      9
      ·
      2 months ago

      I feel like it’s really far from being open. Besides the training data not being open, the more popular ones like llama and stable diffusion have these weird source available licenses with anti-competitive clauses, user count limits, or arbitrary morality clauses.

  • WalnutLum@lemmy.ml
    link
    fedilink
    arrow-up
    14
    ·
    2 months ago

    Ironically thanks in no small part to Facebook releasing Llama and kind of salting the earth for similar companies trying to create proprietary equivalents.

    Nowadays you either have gigantic LLMs with hundreds of billions of parameters like Claude and ChatGPT or you have open Models that are sub-200B.

  • thann@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    2 months ago

    We make a new image license where if it it gets trained in an ai model, they have to open source it, then we create a meme with that license and get the corpos to train on the image, then we sue their pants off to open source the model!