Snapshot video through dynamic scattering medium based on deep learning

Guzmán, Felipe; Vera, Esteban; Horisaki, Ryoichi

Abstract

We present an end-to-end deep learning model designed to reconstruct up to eight frames from a single snapshot of a dynamic object passing through an unknown, time-varying scattering medium. Our approach integrates a coded aperture compressive temporal imaging system with a specially designed transformer-based convolutional neural network (CNN), optimized for effective demultiplexing and reconstruction. Both simulation and experimental results demonstrate a successful compression ratio of up to 8X, while maintaining high reconstruction quality. Furthermore, ablation studies reveal that our dual-input CNN model, which utilizes both speckle patterns and their autocor relations, significantly improves reconstruction accuracy. (c) 2025 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

Más información

Título según WOS: Snapshot video through dynamic scattering medium based on deep learning
Volumen: 33
Número: 7
Fecha de publicación: 2025
Página de inicio: 15991
Página final: 16002
Idioma: English
URL: https://doi.org/10.1364/OE.545510
DOI:

10.1364/OE.545510

Notas: ISI