While I am quite excited about the Walton Goggins-infused Amazon Fallout series, the show debuted some promo art for the project ahead of official stills or footage and…it appears to be AI generated.

  • FaceDeer@kbin.social
    link
    fedilink
    arrow-up
    8
    arrow-down
    11
    ·
    1 year ago

    but it has already stolen work already done by artists and further stealing upcoming work by mashing together older works.

    You keep using that word “stolen”, I do not think it means what you think it means.

    Also, AIs do not “mash together” works from their training sets. This is a very common and very incorrect conception of how they work. They are not collage generators or copy-and-paste machines. They learn concepts from the images they train on, they don’t actually remember fragments of those images to later regurgitate in some sort of patched-together Frankenstein’s Monster.

    • Send_me_nude_girls@feddit.de
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      3
      ·
      1 year ago

      You’re correct but it’s still too early and most people haven’t spend enough time with AI to fully understand. Maybe they never will.

      • FaceDeer@kbin.social
        link
        fedilink
        arrow-up
        5
        arrow-down
        3
        ·
        1 year ago

        Like the classic quote says, it is difficult to get a man to understand something when his salary depends upon his not understanding it.

    • Pandemanium@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      1 year ago

      I just asked Wombo Dream to make the Mona Lisa and it did. Sure, you can tell it’s not exactly the real thing, but I don’t know how you can say it didn’t copy any of the actual Mona Lisa original.

      • FaceDeer@kbin.social
        link
        fedilink
        arrow-up
        3
        ·
        edit-2
        1 year ago

        I considered including mention of overfitting in my earlier comment, but since it’s such an edge case I felt it would just be an irrelevant digression.

        When a particular image has a great many duplicates in the training set - hundreds or even thousands of copies are necessary - then you get the phenomenon of overfitting. In that case you do get this sort of “memorization” of a particular image, because during training you are hitting the neural net over and over with the exact same inputs and really drilling it into them. This is universally considered undesirable, because there’s no point to it - why spend thousands of dollars to do something that a copy/paste command could do so much better and more easily? So when image generators are trained the training data goes through a “de-duplication” step intended to try to prevent this sort of thing from happening. Images like the Mona Lisa are so incredibly common that they still slip through the cracks, though.

        There’s a paper from some months back that commonly comes up when people want to go “aha, generative AI copies its training data!” But in reality this paper shows just how difficult it is to arrange for overfitting to happen. The researchers used an older version of Stable Diffusion whose training set was not well curated and is no longer used due to its poor quality, and even then it took them hundreds of millions of attempts to find just a handful of images from the training set that they could dredge back out of it in recognizable form.

      • emeralddawn45
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        2
        ·
        1 year ago

        People have also copied art for as long as art has existed. You can buy a copy of the Mona Lisa in the gift shop, or print your own. That’s why the market for art has always been hyperfocus3d on ‘originals’. But rarely are the artists the ones getting rich off their art, especially now. I hate capitalism as much as anyone but if your motivation for making art is money you’re in the wrong business and your art probably isn’t that good anyway.