Snapshot video through dynamic scattering medium based on deep learning
Abstract
We present an end-to-end deep learning model designed to reconstruct up to eight frames from a single snapshot of a dynamic object passing through an unknown, time-varying scattering medium. Our approach integrates a coded aperture compressive temporal imaging system with a specially designed transformer-based convolutional neural network (CNN), optimized for effective demultiplexing and reconstruction. Both simulation and experimental results demonstrate a successful compression ratio of up to 8X, while maintaining high reconstruction quality. Furthermore, ablation studies reveal that our dual-input CNN model, which utilizes both speckle patterns and their autocor relations, significantly improves reconstruction accuracy. (c) 2025 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement
Más información
Título según WOS: | Snapshot video through dynamic scattering medium based on deep learning |
Volumen: | 33 |
Número: | 7 |
Fecha de publicación: | 2025 |
Página de inicio: | 15991 |
Página final: | 16002 |
Idioma: | English |
URL: | https://doi.org/10.1364/OE.545510 |
DOI: |
10.1364/OE.545510 |
Notas: | ISI |