Abstract
In the field of optics, infrared and visible images are often required for use in different situations. In particular, in the field of optical sensors, infrared and visible sensors are mainly used to obtain different band images, which are applied to improve the comprehensive information content of images and to improve the quality and availability of images. In existing infrared and visible image fusion methods, the focus is often placed on retaining the background information from visible images and the salient targets from infrared images. To address this issue, we introduce a multi-scale dilated attention module into the encoder-decoder structure of the generator. By applying dilated convolution and selfattention mechanisms, this module improves the perceptual capability of the model, thereby improving performance without increasing network complexity. This design emphasizes gradient information and detailed features in visible images. Experimental results on the public TNO dataset demonstrate that our method achieves superior visual quality and preserves the most abundant image information. Moreover, experiments on spacecraft images validate the robustness and applicability of our approach. Simultaneously, our method also provides significant technical support for the optical field.
| Original language | English |
|---|---|
| Pages (from-to) | 120-129 |
| Number of pages | 10 |
| Journal | Current Optics and Photonics |
| Volume | 9 |
| Issue number | 2 |
| DOIs | |
| State | Published - 1 Apr 2025 |
| Externally published | Yes |
Keywords
- Generative adversarial network
- Image fusion
- Infrared and visible image
- Multi-scale dilated attention
- Optics
Fingerprint
Dive into the research topics of 'Multi-scale Dilated Attention-based Generative Adversarial Network for Infrared and Visible Image Fusion'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver