“Compositional Neural Scene Representations for Shading Inference” by Granskog, Rousselle, Papas and Novák

  • ©Jonathan Granskog, Fabrice Rousselle, Marios Papas, and Jan Novák




    Compositional Neural Scene Representations for Shading Inference

Session/Category Title: Appearance Acquisition and Inverse Rendering



    We present a technique for adaptively partitioning neural scene representations. Our method disentangles lighting, material, and geometric information yielding a scene representation that preserves the orthogonality of these components, improves interpretability of the model, and allows compositing new scenes by mixing components of existing ones. The proposed adaptive partitioning respects the uneven entropy of individual components and permits compressing the scene representation to lower its memory footprint and potentially reduce the evaluation cost of the model. Furthermore, the partitioned representation enables an in-depth analysis of existing image generators. We compare the flow of information through individual partitions, and by contrasting it to the impact of additional inputs (G-buffer), we are able to identify the roots of undesired visual artifacts, and propose one possible solution to remedy the poor performance. We also demonstrate the benefits of complementing traditional forward renderers by neural representations and synthesis, e.g. to infer expensive shading effects, and show how these could improve production rendering in the future if developed further.

ACM Digital Library Publication:

Overview Page: