Skip to content

Latest commit

 

History

History
107 lines (59 loc) · 10 KB

README.md

File metadata and controls

107 lines (59 loc) · 10 KB

CUDA Denoiser For CUDA Path Tracer

Before denoising (2 samples) After denoising

University of Pennsylvania, CIS 565: GPU Programming and Architecture, Project 4

  • Tested on: Windows 10, AMD Ryzen 5 3600, Geforce RTX 2060 Super (personal)

Introduction

One of the main challenges to ray tracing is getting the full picture without the noise. When using naïve implementations, this usually takes many iterations of the algorithm, which quickly becomes an issue for real-time applications. As using the GPU has already allowed us to get great results in near real-time, see the previous project, our goal now is limit the amount of iterations needed to get an image without the noise.

The way to do this is with the use of denoisers. Denoisers help produce a smoother appearance in a pathtraced image with fewer samples-per-pixel/iterations. This is done by blurring pixels. However, a simple blur is usually never what we want. For instance, take a look at this simple cornell box below and see the effect of applying a gaussian blur.

Cornell box (10 samples) Cornell box (5000 samples) Cornell box (simple blur)

Obviously it doesn't come quite close to the cornell box with 5000 samples, but we do see an improvement: less grainy-ness. However, this comes at a great cost as we lose the sharp edges as we blur. To solve this, we need a new filter that preserves the edges. Hence, we implement a technique mentioned in the paper "Edge-Avoiding A-Trous Wavelet Transform for fast Global Illumination Filtering" that does just that.

A-Trous Wavelet Transform Filtering

To understand how this algorithm works, we must understand how any filter, like the gaussian, works. For every pixel in our image, we look at the neighboring pixels, and sum the color value of each one, multiplied by a factor at every value, so that we end up we a normalized value. These factors can be stored in a matrix, commonly known as a kernel, allowing us to take the convolution and get the blurred image. A gaussian blur is simply the the kernel with values based on the gaussian distribution. The wider this kernel is, the more of a blurred effect we get.

filter size = 20x20 filter size = 40x40 filter size = 80x80

note: in the paper and in this implementation, our kernel size is fixed to 5x5, however, we apply an offset such that each pixel is 2^i pixels apart, where i is the level, starting from 0. So, 20x20 translates to 2 iterations, 40x40 to 3 iterations, 80x80 to 4, and so forth. This allows us to use the same values to get the effect of a larger kernel. This helps us with performance (see the following section).

At the core of this algorithm, the denoiser uses what's called a geometry buffer to guide its smoothing filter. This buffer stores information about the scene, namely its color, position in space, and surface normal (See visualization below). As opposed to simply multiplying each pixel by the factor in the kernel, we also multiply by a weight. This weight is the derived by the multiplication of three other weights, one for the difference in the raytraced colors between the neighbor pixel and the center pixel, one for the difference in normal, and one for the difference in position.

per-pixel normals per-pixel positions (scaled down)

To get a better sense of what these weights do to our scene, below is a visualization of the effect of weighting for each individual weight.

Color-based blur Normal-based blur Position-based blur

With all three components combined, we get an image that looks like this:

Cornell box (10 samples) Cornell box (5000 samples) Cornell box (A-trous filter)

Although there is definitely room for improvement, specifically on the parameter fine-tuning, we can already see the improvements that denoising gives us. We can even get great results with as low as 2 samples per path.

Before denoising (2 samples) After denoising

All of this, however, doesn't come for free. There is, of course, a performance and memory hit when using this algorithm. We have to now store an entire G-Buffer for the scene, and we have add this denoising step into our render loop. In most cases, however, by reducing the number of iterations needed to get an image that converges, ray tracing be greatly improved for real-time applications.

Performance Analysis

To understand our performance hit of the denoiser, let's take a look at all the extra steps that occur between the denoised and standard path tracer. When denoising, we have to create a G-buffer, and for this implementation, it's just the first depth where we store the normal and position. This mainly only adds up the memory we use. It only needs to be done in one iteration, as we aren't anti-aliasing. Then, at the end of the last iteration, we pass the final color to the gbuffer for processing.

During the processing stage, i.e., blurring, that's where the majority of the time is taken. To understand how long blurring takes, I used CUDA events to track the time it takes to run the denoiser with varying kernel sizes:

Using a filter of size 80, which translates to 4 iterations of the algorithm, adds about 16 milliseconds to the end of our render. To maintain an fps of 30-60 frames per second, this is a considerable amount of time.

However, as mentioned previously, the number of iteration needed to produce a smooth image lowers substantially. To get a reasonably smooth image required 5000 samples, and that takes roughly 18 seconds on my machine, compared to 0.86 seconds needed to get comparably smooth image with 100 samples, and 0.26 seconds with 10 samples. There is no question that any real-time application will want to incorporate denoising into their rendering engine.

As noted earlier, denoising does not change at different iterations since all the denoising happens at the end.

See below for the visual results for difference in kernel size:

filter size = 20x20 filter size = 40x40 filter size = 80x80

This effectiveness of this algorithm and/or this implementation differs based on material. For specular objects, the details aren't as fine-grained. See below

reference denoised

If we were to increase the size of the light, more rays will reach the light, requiring less samples. This helps with our denoiser:

smaller light (10 iterations, denoised) larger light (10 iterations, denoised)