Practical temporal and stereoscopic filtering
for real-time ray tracing

Henrik Philippi1,2, Jeppe Revall Frisvad2, and Henrik Wann Jensen1

1Luxion, Denmark
2Technical University of Denmark

Abstract
We present a practical method for temporal and stereoscopic filtering that generates stereo-consistent rendering. Existing methods for stereoscopic rendering often reuse samples from one eye for the other or do averaging between the two eyes. These approaches fail in the presence of ray tracing effects such as specular reflections and refractions. We derive a new blending strategy that leverages variance to compute per pixel blending weights for both temporal and stereoscopic rendering. In the temporal domain, our method works well in a low noise context and is robust in the presence of inconsistent motion vectors, where existing methods such as temporal anti-aliasing (TAA) and deep learning super sampling (DLSS) produce artifacts. In the stereoscopic domain, our method provides a new way to ensure consistency between the left and right eyes. The stereoscopic version of our method can be used with our new temporal method or with existing methods such as DLSS and TAA. In all combinations, it reduces the error and significantly increases the consistency between the eyes making it practical for real-time settings such as virtual reality (VR).

CCS Concepts
· Computing methodologies→Rendering; Ray tracing; Antialiasing; Virtual reality

Videos


Reference

Philippi, H., Frisvad, J. R., and Jensen, H. W. Practical temporal and stereoscopic filtering for real-time ray tracing. In Eurographics Symposium on Rendering (EGSR), pp. 111-118. The Eurographics Association, June 2023. [abstract] [lowres pdf]