• sugar_in_your_tea@sh.itjust.works
      link
      fedilink
      arrow-up
      31
      ·
      9 months ago

      I can see a few reasons:

      • automated tests on single frames
      • batch renders on a server (e.g. for stills or cutscenes)
      • comparisons across GPU archs - it could essentially be the “standard” for how a scene should be rendered

      And of course, maybe some CPU manufacturer will build in an accelerator so lower end GPUs (say, APUs) could have reasonable raytracing in otherwise GPU limited games (i don’t know enough about modern game pipelines to know if that’s a possibility).

      Or the final reason, which may be the most important of all: why not?

      • RvTV95XBeo@sh.itjust.works
        link
        fedilink
        arrow-up
        19
        ·
        9 months ago

        I’ll add one to this - optimization. A lot of clever optimization techniques tend to come out of projects like this - necessity is the mother of invention.

    • Baŝto
      link
      fedilink
      English
      arrow-up
      10
      ·
      edit-2
      9 months ago

      If the CPUs get strong enough, they could run old raytracing games at some point … especially on hardware platforms that don’t have ray tracing GPUs available for them.