batch renders on a server (e.g. for stills or cutscenes)
comparisons across GPU archs - it could essentially be the “standard” for how a scene should be rendered
And of course, maybe some CPU manufacturer will build in an accelerator so lower end GPUs (say, APUs) could have reasonable raytracing in otherwise GPU limited games (i don’t know enough about modern game pipelines to know if that’s a possibility).
Or the final reason, which may be the most important of all: why not?
I’ll add one to this - optimization. A lot of clever optimization techniques tend to come out of projects like this - necessity is the mother of invention.
I can see a few reasons:
And of course, maybe some CPU manufacturer will build in an accelerator so lower end GPUs (say, APUs) could have reasonable raytracing in otherwise GPU limited games (i don’t know enough about modern game pipelines to know if that’s a possibility).
Or the final reason, which may be the most important of all: why not?
I’ll add one to this - optimization. A lot of clever optimization techniques tend to come out of projects like this - necessity is the mother of invention.