Seeing Dark Videos via Self-Learned Bottleneck Neural Representation

Authors

  • Haofeng Huang Peking University
  • Wenhan Yang Peng Cheng Laboratory
  • Lingyu Duan Peking University
  • Jiaying Liu Peking University

DOI:

https://doi.org/10.1609/aaai.v38i3.28006

Keywords:

CV: Low Level & Physics-based Vision, CV: Computational Photography, Image & Video Synthesis

Abstract

Enhancing low-light videos in a supervised style presents a set of challenges, including limited data diversity, misalignment, and the domain gap introduced through the dataset construction pipeline. Our paper tackles these challenges by constructing a self-learned enhancement approach that gets rid of the reliance on any external training data. The challenge of self-supervised learning lies in fitting high-quality signal representations solely from input signals. Our work designs a bottleneck neural representation mechanism that extracts those signals. More in detail, we encode the frame-wise representation with a compact deep embedding and utilize a neural network to parameterize the video-level manifold consistently. Then, an entropy constraint is applied to the enhanced results based on the adjacent spatial-temporal context to filter out the degraded visual signals, e.g. noise and frame inconsistency. Last, a novel Chromatic Retinex decomposition is proposed to effectively align the reflectance distribution temporally. It benefits the entropy control on different components of each frame and facilitates noise-to-noise training, successfully suppressing the temporal flicker. Extensive experiments demonstrate the robustness and superior effectiveness of our proposed method. Our project is publicly available at: https://huangerbai.github.io/SLBNR/.

Published

2024-03-24

How to Cite

Huang, H., Yang, W., Duan, L., & Liu, J. (2024). Seeing Dark Videos via Self-Learned Bottleneck Neural Representation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(3), 2321-2329. https://doi.org/10.1609/aaai.v38i3.28006

Issue

Section

AAAI Technical Track on Computer Vision II