Authors

Stavros Diolatzis, Jan Novák, Fabrice Rouselle, Jonathan Granskog, Miika Aittala, Ravi Ramamoorthi, George Drettakis

Paper Project Website

Teaser

Abstract

We introduce MesoGAN, a model for generative 3D neural textures. This new graphics primitive represents mesoscale appearance by combining the strengths of generative adversarial networks (StyleGAN) and volumetric neural field rendering. The primitive can be applied to surfaces as a neural reflectance shell; a thin volumetric layer above the surface with appearance parameters defined by a neural network. To construct the neural shell, we first generate a 2D feature texture using StyleGAN with carefully randomized Fourier features to support arbitrarily sized textures without repeating artifacts. We augment the 2D feature texture with a learned height feature, which aids the neural field renderer in producing volumetric parameters from the 2D texture. To facilitate filtering, and to enable end-to-end training within memory constraints of current hardware, we utilize a hierarchical texturing approach and train our model on multi-scale synthetic datasets of 3D mesoscale structures. We propose one possible approach for conditioning MesoGAN on artistic parameters (e.g., fiber length, density of strands, lighting direction) and demonstrate and discuss integration into physically based renderers.


Citation

@Article{DNRGARD23,
  author       = "Diolatzis, Stavros and Novak, Jan and Rousselle, Fabrice and Granskog, Jonathan and Aittala, Miika and Ramamoorthi, Ravi and Drettakis, George",
  title        = "MesoGAN: Generative Neural Reflectance Shells",
  journal      = "Computer Graphics Forum",
  year         = "2023",
  url          = "http://www-sop.inria.fr/reves/Basilic/2023/DNRGARD23"
}