Tizian Zeltner

Radiance Surfaces: Optimizing Surface Representations with a 5D Radiance Field Loss

To appear at SIGGRAPH 2025 (Conference Track)

Our method reconstructs surfaces with the speed and robustness of NeRF-style methods. Left: In contrast to volume-based methods that minimize 2D image losses, as shown in (a), we adopt a spatio-directional radiance field loss formulation, as shown in (b). At each step, our method considers a distribution of optically independent surfaces, increasing the confidence of candidates that agree with the reference imagery. Right: A meaningful surface can be extracted at any iteration during optimization.

Abstract

We present a fast and simple technique to convert images into a radiance surface-based scene representation. Building on existing radiance volume reconstruction algorithms, we introduce a subtle yet impactful modification of the loss function requiring changes to only a few lines of code: instead of integrating the radiance field along rays and supervising the resulting images, we project the training images into the scene to directly supervise the spatio-directional radiance field. The primary outcome of this change is the complete removal of alpha blending and ray marching from the image formation model, instead moving these steps into the loss computation. In addition to promoting convergence to surfaces, this formulation assigns explicit semantic meaning to 2D subsets of the radiance field, turning them into well-defined radiance surfaces. We finally extract a level set from this representation, which results in a high-quality radiance surface model. Our method retains much of the speed and quality of the baseline algorithm. For instance, a suitably modified variant of Instant NGP maintains comparable computational efficiency, while achieving an average PSNR that is only 0.1 dB lower. Most importantly, our method generates explicit surfaces in place of an exponential volume, doing so with a level of simplicity not seen in prior work.

Downloads

Publication