Learning Spatially Collaged Fourier Bases for Implicit Neural Representation

Authors

  • Jason Chun Lok Li The University of Hong Kong
  • Chang Liu The University of Hong Kong
  • Binxiao Huang The University of Hong Kong
  • Ngai Wong The University of Hong Kong

DOI:

https://doi.org/10.1609/aaai.v38i12.29252

Keywords:

ML: Representation Learning, CV: Representation Learning for Vision

Abstract

Existing approaches to Implicit Neural Representation (INR) can be interpreted as a global scene representation via a linear combination of Fourier bases of different frequencies. However, such universal basis functions can limit the representation capability in local regions where a specific component is unnecessary, resulting in unpleasant artifacts. To this end, we introduce a learnable spatial mask that effectively dispatches distinct Fourier bases into respective regions. This translates into collaging Fourier patches, thus enabling an accurate representation of complex signals. Comprehensive experiments demonstrate the superior reconstruction quality of the proposed approach over existing baselines across various INR tasks, including image fitting, video representation, and 3D shape representation. Our method outperforms all other baselines, improving the image fitting PSNR by over 3dB and 3D reconstruction to 98.81 IoU and 0.0011 Chamfer Distance.

Published

2024-03-24

How to Cite

Li, J. C. L., Liu, C., Huang, B., & Wong, N. (2024). Learning Spatially Collaged Fourier Bases for Implicit Neural Representation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(12), 13492-13499. https://doi.org/10.1609/aaai.v38i12.29252

Issue

Section

AAAI Technical Track on Machine Learning III