TY - JOUR
T1 - A new approach for training a physics-based dehazing network using synthetic images
AU - Del Gallego, Neil Patrick
AU - Ilao, Joel
AU - Cordel, Macario
AU - Ruiz, Conrado
N1 - Funding Information:
We would like to acknowledge De La Salle University (DLSU), Department of Science and Technology (DOST), and the Google Cloud Research program, for funding this research.
Publisher Copyright:
© 2022 Elsevier B.V.
PY - 2022/10
Y1 - 2022/10
N2 - In this study, we propose a new approach for training a physics-based dehazing network, using RGB images and depth maps gathered from a 3D urban virtual environment, with simulated global illumination and physically-based shaded materials. Since 3D scenes are rendered with depth buffers, full image depth can be extracted based on this information, using a custom shader, unlike the extraction of real-world depth maps, which tend to be sparse. Our proposed physics-based dehazing network uses generated transmission and atmospheric maps from RGB images and depth maps from the virtual environment. To make our network compatible with real-world images, we incorporate a novel strategy of using unlit image priors during training, which can also be extracted from the virtual environment. We formulate the training as a supervised image-to-image translation task, using our own DLSU-SYNSIDE (SYNthetic Single Image Dehazing Dataset), which consists of clear images, unlit image priors, transmission, and atmospheric maps. Our approach makes training stable and easier as compared to unsupervised approaches. Experimental results demonstrate the competitiveness of our approach against state-of-the-art dehazing works, using known benchmarking datasets such as I-Haze, O-Haze, and RESIDE, without our network seeing any real-world images during training. The DLSU-SYNSIDE dataset and source code can be accessed through this link: https://neildg.github.io/SynthDehazing/.
AB - In this study, we propose a new approach for training a physics-based dehazing network, using RGB images and depth maps gathered from a 3D urban virtual environment, with simulated global illumination and physically-based shaded materials. Since 3D scenes are rendered with depth buffers, full image depth can be extracted based on this information, using a custom shader, unlike the extraction of real-world depth maps, which tend to be sparse. Our proposed physics-based dehazing network uses generated transmission and atmospheric maps from RGB images and depth maps from the virtual environment. To make our network compatible with real-world images, we incorporate a novel strategy of using unlit image priors during training, which can also be extracted from the virtual environment. We formulate the training as a supervised image-to-image translation task, using our own DLSU-SYNSIDE (SYNthetic Single Image Dehazing Dataset), which consists of clear images, unlit image priors, transmission, and atmospheric maps. Our approach makes training stable and easier as compared to unsupervised approaches. Experimental results demonstrate the competitiveness of our approach against state-of-the-art dehazing works, using known benchmarking datasets such as I-Haze, O-Haze, and RESIDE, without our network seeing any real-world images during training. The DLSU-SYNSIDE dataset and source code can be accessed through this link: https://neildg.github.io/SynthDehazing/.
KW - Deep neural network
KW - Image dehazing
KW - Physics-based dehazing
KW - Unlit image priors
UR - http://www.scopus.com/inward/record.url?scp=85131091211&partnerID=8YFLogxK
U2 - 10.1016/j.sigpro.2022.108631
DO - 10.1016/j.sigpro.2022.108631
M3 - Article
AN - SCOPUS:85131091211
SN - 0165-1684
VL - 199
JO - Signal Processing
JF - Signal Processing
M1 - 108631
ER -