Skip to content
Massachusetts Institute of Technology
  • on: June 17, 2020
  • in: NeurIPS
  • ✨Oral

Implicit Neural Representations with Periodic Activation Functions

  • Vincent Sitzmann *
  • Julien N. P. Martel *
  • Alexander W. Bergman
  • David B. Lindell
  • Gordon Wetzstein
*
shared first author
@inproceedings{sitzmann2019siren,
  author = {Sitzmann, Vincent
            and Martel, Julien N.P.
            and Bergman, Alexander W.
            and Lindell, David B.
            and Wetzstein, Gordon},
  title = {Implicit Neural Representations
  with Periodic Activation Functions},
  booktitle = {Proc. NeurIPS},
  year={2020}
}
  • Copy to Clipboard

Implicitly defined, continuous, differentiable signal representations parameterized by neural networks have emerged as a powerful paradigm, offering many possible benefits over conventional representations. However, current network architectures for such implicit neural representations are incapable of modeling signals with fine detail, and fail to represent a signal’s spatial and temporal derivatives, despite the fact that these are essential to many physical signals defined implicitly as the solution to partial differential equations. We propose to leverage periodic activation functions for implicit neural representations and demonstrate that these networks, dubbed sinusoidal representation networks or SIREN, are ideally suited for representing complex natural signals and their derivatives. We analyze SIREN activation statistics to propose a principled initialization scheme and demonstrate the representation of images, wavefields, video, sound, and their derivatives. Further, we show how SIREN s can be leveraged to solve challenging boundary value problems, such as particular Eikonal equations (yielding signed distance functions), the Poisson equation, and the Helmholtz and wave equations. Lastly, we combine SIREN with hypernetworks to learn priors over the space of SIREN functions.