Self-Supervised Post-Correction for Monte Carlo Denoising

Jonghee Back 1 , Binh-Son Hua 2 , Toshiya Hachisuka 3 , Bochang Moon 1

GIST 1 , VinAI Research 2 , University of Waterloo 3

In Proceedings of ACM SIGGRAPH 2022

Self-Supervised Post-Correction for Monte Carlo Denoising
Our post-correction result for a state-of-the-art image denoiser (AFGSA [Yu et al. 2021]). Our self-supervised optimization uses only test images, and it visually and numerically improves the existing learning-based technique by restoring the fine details. We used the relative L2 (relL2) [Rousselle et al. 2011] as an error metric.

Abstract

Using a network trained by a large dataset is becoming popular for denoising Monte Carlo rendering. Such a denoising approach based on supervised learning is currently considered the best approach in terms of quality. Nevertheless, this approach may fail when the image to be rendered (i.e., the test data) has very different characteristics than the images included in the training dataset. A pre-trained network may not properly denoise such an image since it is unseen data from a supervised learning perspective. To address this fundamental issue, we introduce a post-processing network that improves the performance of supervised learning denoisers. The key idea behind our approach is to train this post-processing network with self-supervised learning. In contrast to supervised learning, our self-supervised model does not need a reference image in its training process. We can thus use a noisy test image and self-correct the model on the fly to improve denoising performance. Our main contribution is a self-supervised loss that can guide the post-correction network to optimize its parameters without relying on the reference. Our work is the first to apply this self-supervised learning concept in denoising Monte Carlo rendered estimates. We demonstrate that our post-correction framework can boost supervised denoising via our self-supervised optimization.

Contents

  • Main Report - Preprint (pdf, 9.6 MB) (TBA)
  • Main Report - Publisher (TBA)
  • Supplementary Report - Preprint (pdf, 8.2 MB) (TBA)
  • Code (GitHub)(TBA)
  • Interactive Viewer (TBA)
  • Presentation Slides (pdf, 4.4 MB) (TBA)
  • BibTex (TBA)