• rumba@lemmy.zip
    link
    fedilink
    English
    arrow-up
    3
    ·
    1 day ago

    I am SUPER happy to have a model that performs as well, if not better than OpenAI, that I can run myself. Completely overjoyed. But I’m not certain that they actually pulled it off for that price. They have everything to gain from lying about it. Knowing the hit the US companies would take from it, it could have been state-sponsored. I’m not saying that it was, but if it were, we’d likely never know.

    • BrikoX@lemmy.zipM
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 day ago

      It was a nice sleight of hand on their part. There is a lot of misleading information about all of it since they only release pre-training details on DeepSeek-V3 model, but not DeepSeek-R1. But the media reported on it as it was one and the same without any distinction.

      Based on reports, the parent company had access to more GPUs than reported amount used. Hard to tell if they were utilized though.

      • rumba@lemmy.zip
        link
        fedilink
        English
        arrow-up
        5
        ·
        1 day ago

        Yeah, whatever the case, They were all trained on data from the public. The very least they can do is make the models available to the public.