• hessenjunge
    link
    fedilink
    English
    arrow-up
    14
    ·
    3 months ago

    The LLMs they train on their code will only be accessible internally. They won’t leak their own intellectual property.

      • Echo Dot@feddit.uk
        link
        fedilink
        English
        arrow-up
        5
        ·
        3 months ago

        Yeah which is why this is a dumb statement from Amazon. But then again I don’t expect C-suite managers to really understand the intricacies of their own companies.

      • Hawk@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        3
        ·
        3 months ago

        Depends on the use case. Training local llms is a lot cheaper after Galore and there are ways to get useful local models with only a moderate amount of effort, see e.g. augmentoolkit.

        This may or may not be practical in many use cases.

        24 months is pretty generous but no doubt there will be significantly less demand for junior developers in the near future.

      • hessenjunge
        link
        fedilink
        English
        arrow-up
        2
        ·
        3 months ago

        Possibly. It’s hard to know without seeing the numbers and assessing output quality and volume.

        Also it’s not unheard of that some bigwig wastes millions of company €€ for some project they fancy. (Billions if they happen to be Elon)

    • prole@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      3 months ago

      If only we had an overarching structure that everyone in society has agreed exists for the purposes of enforcing laws and regulating things. Something that governs people living in a region… Maybe then they could be compelled to show exactly what they’re using, and what those models are being trained with.

      Oh well.