Image fusion based on generative adversarial network consistent with perception
Fu, Yu and Wu, Xiao-Jun and Durrani, Tariq (2021) Image fusion based on generative adversarial network consistent with perception. Information Fusion, 72. pp. 110-125. ISSN 1566-2535 (https://doi.org/10.1016/j.inffus.2021.02.019)
Preview |
Text.
Filename: Fu_etal_IF_2021_Image_fusion_based_on_generative_adversarial_network_consistent_with_perception.pdf
Accepted Author Manuscript License: Download (10MB)| Preview |
Abstract
Deep learning is a rapidly developing approach in the field of infrared and visible image fusion. In this context, the use of dense blocks in deep networks significantly improves the utilization of shallow information, and the combination of the Generative Adversarial Network (GAN) also improves the fusion performance of two source images. We propose a new method based on dense blocks and GANs, and we directly insert the input image-visible light image in each layer of the entire network. We use structural similarity and gradient loss functions that are more consistent with perception instead of mean square error loss. After the adversarial training between the generator and the discriminator, we show that a trained end-to-end fusion network – the generator network – is finally obtained. Our experiments show that the fused images obtained by our approach achieve good score based on multiple evaluation indicators. Further, our fused images have better visual effects in multiple sets of contrasts, which are more satisfying to human visual perception.
-
-
Item type: Article ID code: 75880 Dates: DateEvent31 August 2021Published27 February 2021Published Online22 February 2021AcceptedSubjects: Technology > Electrical engineering. Electronics Nuclear engineering Department: Faculty of Engineering > Electronic and Electrical Engineering Depositing user: Pure Administrator Date deposited: 22 Mar 2021 10:06 Last modified: 21 Dec 2024 01:23 Related URLs: URI: https://strathprints.strath.ac.uk/id/eprint/75880