• sugar_in_your_tea@sh.itjust.works
    link
    fedilink
    arrow-up
    0
    ·
    9 months ago

    I can see a few reasons:

    • automated tests on single frames
    • batch renders on a server (e.g. for stills or cutscenes)
    • comparisons across GPU archs - it could essentially be the “standard” for how a scene should be rendered

    And of course, maybe some CPU manufacturer will build in an accelerator so lower end GPUs (say, APUs) could have reasonable raytracing in otherwise GPU limited games (i don’t know enough about modern game pipelines to know if that’s a possibility).

    Or the final reason, which may be the most important of all: why not?

    • RvTV95XBeo@sh.itjust.works
      link
      fedilink
      arrow-up
      0
      ·
      9 months ago

      I’ll add one to this - optimization. A lot of clever optimization techniques tend to come out of projects like this - necessity is the mother of invention.