89
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Transformer-based dual path cross fusion for pansharpening remote sensing images

, ORCID Icon, ORCID Icon &
Pages 1170-1200 | Received 16 Sep 2023, Accepted 05 Jan 2024, Published online: 02 Feb 2024
 

ABSTRACT

The purpose of pan-sharpening is to generate high-resolution multispectral (HRMS) images by combining multispectral images (MS) and panchromatic images (PAN), which have become an important part of remote sensing image processing. Therefore, how to better extract complete feature information from MS images and PAN images has become the focus of our attention. In this paper, we propose a new pansharpening method, called Transformer-based Dual-path cross fusion network (TDF) for Pan-sharpening remote sensing images, which aims to extract the spatial details of PAN images while maintaining the spectral fidelity of MS images. The whole network structure can be divided into two parts: in the encoder part, we adopt the Swin-Transformer module for the downsampling operation, which expands the sensory field of the network to the feature map, and then extracts the global information. However, since the Swin-Transformer module is not good at extracting pixel-level details, we introduce the Base Feature Extraction (BFE) and the Invertible Neural Network Block (INNB) modules for the interaction between the local and the global feature information. we also introduce the Edge-Enhancement Block (EEB) to further enhance the feature extraction at multi-scales during the image fusion process. In the decoder section, we once again employ the Swin Transformer module for downsampling. After the convolution operation and activation function, we utilize the Sub-Pixel Convolutional Neural Network for upsampling to generate the ultimate high-resolution multispectral images. Simulation experiments and real experiments are conducted on QuickBird (QB) and WorldView2 (WV2) datasets, which demonstrated our method are superior to the current methods.

Acknowledgements

The authors acknowledge the National Natural Science Foundation of China (Grant nos. 61772319, 62002200, 62202268 and 61972235), the Shandong Natural Science Foundation of China (Grant no. ZR2021MF107, ZR2022MA076) Youth Innovation Technology Project of Higher School in Shandong Province under (Grant No. 2021KJ069, 2019KJN042) and Yantai science and technology innovation development plan(2022JCYJ031).

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data Availability statement

Data available on request from the authors.

Additional information

Funding

The work was supported by the National Natural Science Foundation of China [61772319, 62002200, 62202268 and 61972235].

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.