• ChicoSuave@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    4
    ·
    11 months ago

    Does this AI use the same process for piecing together things as LLMs do for art and writing? Is this a drug we have known about but not yet applied as an antibiotic or a whole new compound?

    • krellor@kbin.social
      link
      fedilink
      arrow-up
      12
      ·
      11 months ago

      It doesn’t sound like it but they don’t have enough detail in the article to say.

      It sounds likey they are using a classification model that takes a vectorized text representation of molecules and classifies or scores them by their expected properties/reactivity. They took 39,000 molecules with known reactivity to MRSA to train the model, I assume to classify the structures. Once trained they can feed in arbitrary molecules into the trained model and see which ones are predicted to have antibiotic properties, which they can verify with bench work.

      They likely fed in molecules from classes of likely candidate structures, and the model helped focus and direct the wet work.

      I’m not up on the latest, but years ago I helped a similar project using FPGAs running statistical models to direct lab work.

      • Jerkface@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        11 months ago

        I’d be interested to know why FPGAs were selected for this application. I’m not especially familiar with their use cases.

        • krellor@kbin.social
          link
          fedilink
          arrow-up
          6
          ·
          11 months ago

          This was years ago before GPU processing really took off, and we wanted the performance, but also, wanted to see if we could develop an affordable discrete lab device that could be placed in labs to aid in computationally directed bench work. So effectively, testing the efficacy of the models and designing ASICs to perform lab tests.

        • 1rre
          link
          fedilink
          English
          arrow-up
          3
          ·
          edit-2
          11 months ago

          With a CPU or even a GPU, there is a bunch of inefficiencies for every task as they’re designed to be able to do pretty much anything - your H265 media decoder isn’t going to be doing much when you’re keeping a running sum of the number of a certain type of bond in a list of chemicals

          With ASICs and a lesser extent FPGAs, you can make it so every single transistor is being used at every moment which makes them wildly efficient for doing a single repetitive task, such as running statistical analysis on a huge dataset. This is because rather than being limited by the multiprocessing ability of the CPU or GPU, you can design the “program” to run with as much multiprocessing ability as is possible based on the program, meaning if you stream one input per clock cycle, after a delay you will get one input per clock cycle out, including your update function so long as it’s simple enough (eg moving average, running sum or even just writing to memory)

          This is one specific application of FPGAs (static streaming) but it’s the one that’s relevant here

          • Jerkface@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            11 months ago

            So it sounds like we’re designing the instruction pipeline for maximum parallelism for our task. I was surprised to learn that the first commercial FPGAs were available as early as the '80s. I can see how this would have been an extremely effective option before CUDA became available.

    • Willy@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      3
      ·
      11 months ago

      llms have progressed beyond cut and paste way more than a year ago. they have shown understanding of what items are and how they behave and interact. I know it’s popular here to call it a parrot or whatever but most people don’t have access to the high level stuff and most seem afraid/snobby/parroting things themselves.

    • agissilver@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      11 months ago

      Virtual screening libraries are usually some form of expanded chemical space meaning they contain real and previously unknown compounds. The article says the 12 million compounds screened virtually were commercially available, but I couldn’t see enough of the nature paper to verify. It could be that the virtual screening set was acquired from a private company, but that doesn’t necessarily mean all the compounds are known.