Next Article in Journal
Effects of Meteorology Changes on Inter-Annual Variations of Aerosol Optical Depth and Surface PM2.5 in China—Implications for PM2.5 Remote Sensing
Next Article in Special Issue
Remote Sensing Image-Change Detection with Pre-Generation of Depthwise-Separable Change-Salient Maps
Previous Article in Journal
Shallow Crustal Structure of S-Wave Velocities in the Coastal Area of South China Constrained by Receiver Function Amplitudes
Previous Article in Special Issue
A Network Combining a Transformer and a Convolutional Neural Network for Remote Sensing Image Change Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MSAC-Net: 3D Multi-Scale Attention Convolutional Network for Multi-Spectral Imagery Pansharpening

1
School of Information Engineering, Northwest A&F University, Xi’an 712100, China
2
School of Information Science and Technology, Northwest University, Xi’an 710127, China
3
Shaanxi Province Silk Road Digital Protection and Inheritance of Cultural Heritage Collaborative Innovation Center, Xi’an 710127, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(12), 2761; https://doi.org/10.3390/rs14122761
Submission received: 22 April 2022 / Revised: 24 May 2022 / Accepted: 6 June 2022 / Published: 8 June 2022
(This article belongs to the Special Issue Deep Reinforcement Learning in Remote Sensing Image Processing)

Abstract

:
Pansharpening fuses spectral information from the multi-spectral image and spatial information from the panchromatic image, generating super-resolution multi-spectral images with high spatial resolution. In this paper, we proposed a novel 3D multi-scale attention convolutional network (MSAC-Net) based on the typical U-Net framework for multi-spectral imagery pansharpening. MSAC-Net is designed via 3D convolution, and the attention mechanism replaces the skip connection between the contraction and expansion pathways. Multiple pansharpening layers at the expansion pathway are designed to calculate the reconstruction results for preserving multi-scale spatial information. The MSAC-Net performance is verified on the IKONOS and QuickBird satellites’ datasets, proving that MSAC-Net achieves comparable or superior performance to the state-of-the-art methods. Additionally, 2D and 3D convolution are compared, and the influences of the number of convolutions in the convolution block, the weight of multi-scale information, and the network’s depth on the network performance are analyzed.

Graphical Abstract

1. Introduction

Multi-spectral (MS) and panchromatic (PAN) images are two remote sensing image types acquired by optical satellites. While they often represent similar scenes, their spectral and spatial resolutions differ. PAN images have high spatial resolution (HR) but low spectral resolution, whereas MS images have high spectral resolution and low spatial resolution (LR). Pansharpening fuses LR-MS and HR-PAN, generating super-resolution MS images with high spatial resolution (HR-MS). This can provide higher quality remote sensing images for such as target detection [1,2,3], distribution estimation [4] and change detection [5,6,7].
The existing pansharpening techniques can be divided into four categories: component substitution [8,9], multi-resolution analysis [10,11], model-based optimization [12,13] and deep learning [14]. Component substitution-based methods achieve pansharpening by replacing MS images’ partial spectral components with spatial information from PAN images. However, this technique causes spectral distortion [15]. Multi-resolution analysis methods [16,17] decompose the source and then synthesize the HR-MS images through fusion and inverse transform. Although this technique can maintain good spectral characteristics, it causes spatial distortion due to decomposition. Model-based optimization methods are limited by their dependence on appropriate prior knowledge and hyper-parameters.
In recent years, deep-learning-based research has achieved great success in image processing [18,19,20,21,22,23,24]. Masi et al. [18] were the first to propose a CNN-based pansharpening method (PNN), whose structure is based on the super-resolution CNN [25]. Yuan et al. [19] designed a multi-scale and multi-depth CNN (MSDCNN), which introduces multi-scale information through two branches and a different number of learning blocks for deep feature learning. Yang et al. [26] proposed a residual network for pansharpening. However, most of the existing approaches focus on spatial feature extraction while paying less attention to the spectral information and spatial scale information of fusion images. Consequently, the fusion process is often characterized by spectral information loss or spatial feature redundancy.
The 3D convolution shows promise in volume data analysis [27,28,29]. For example, Mei et al. [30] used 3D CNN to extract spectral features of remote sensing super-resolution images. Compared to traditional 2D CNN methods, 3D CNN methods emphasize extracting spectral features while preserving spatial features [28]. Therefore, 3D convolution characteristics promote generating images with high spatial and spectral resolution.
Inspired by the human visual system [31], the attention mechanism was proved to have a positive effect on image understanding [32] and has been widely applied in image processing due to its focus on local information [33]. For example, Wang et al. [34] presented a model with several attention blocks and combined these with the residual structure to enable focusing specific areas while reducing the number of calculations. Mei et al. [32] utilized attention mechanisms to study spatial and spectral correlations between adjacent pixels. U-Net (Figure 1) [35], named by its U-shaped structure, has a high capability to extract and represent features and combine low- and high-scale semantic information for semantic segmentation of input images. Guo et al. [36] replaced the basic convolution block of the decoder in the U-Net architecture with the residual channel attention block [37], consequently improving the model capability. Oktay et al. [38] proposed the attention gate (AG) module based on the U-Net architecture to enable automated focus on target structures of different shapes and sizes while suppressing irrelevant areas.
At present, most MS pansharpening methods focus on the injection of spatial details, but ignore the multi-scale features and spectral features of multi-spectral images. This work proposes a novel 3D multi-scale attention deep convolutional network (MSAC-Net) for MS imagery pansharpening. Following the U-Net framework, MSAC-Net consists of a contraction path that extracts high-scale features from LR-MS and HR-PAN (Figure 2, left) and an expansion path that fuses spatial and spectral information (Figure 2, right).The attention mechanism is introduced in the contraction and expansion paths instead of skip connections, supporting focusing spatial details of the feature maps. Furthermore, the deep supervision mechanism enables MSAC-Net to utilize multi-scale spatial information by adding a pansharpening layer at each scale of the expansion path. The results demonstrate that MSAC-Net achieves high performance in both spatial and spectral dimensions. In summary, this work:
  • Designs a 3D CNN to probe the spectral correlation of adjacent band images, thus reducing the spectral distortion in MS pansharpening;
  • Uses a deep supervision mechanism that utilizes multi-scale spatial information to solve the spatial detail missing problem;
  • Applies the AG mechanism instead of the skip connection in U-Net structure and presents experiments demonstrating its advantages in MS pansharpening.
The rest of the manuscript is organized as follows. Section 2 presents the related work, while Section 3 introduces the proposed MSAC-Net. Section 4 describes and analyzes the experimental results. Finally, Section 5 concludes the paper with a short overview of the contributions.

2. Related Work

Ronneberger et al. [35] first proposed the U-Net network for medical image segmentation. Because its structure is to combine low-level and high-level semantic information for semantic segmentation of the input image, U-Net has high feature representation and extraction ability. Therefore, in this section, the various components of U-Net and improvements based on U-Net are introduced.
According to the difference of feature information extracted by U-Net, its composition can be divided into three parts.
1. The contraction pathway (i.e., the encoder). It is mainly used for shallow feature extraction and coding. For the improvement of this component, Milletari et al. [39] referred to residual learning for the design of each convolution block on the basis of U-Net, which makes V-Net learn image features more fully. Wang et al. [40] used ResNet-50 pre-trained on the ImageNet dataset as its encoder to accelerate the convergence and achieve better performance via transfer learning.
2. The expansion pathway (i.e., the decoder). The purpose of this design is to restore or enlarge the size of feature map. Therefore, the design directly affects the final result of feature recovery. Guo et al. [36] learned an additional dual regression mapping after expansion pathway to estimate the down-sampling kernel and reconstruct LR images, which forms a closed-loop to provide additional supervision. Banerjee et al. [41] designed decoder-side dense skip pathways, making the features under the final scale contain the feature results of all scales. In addition, Khalel et al. [42] realized multi-task learning with the design of dual encoder. Ni et al. [43] extracted and fused the pre-output features and deep features of the network again, thus realizing the multi-task learning of the network.
3. The skip connection. The skip connections transfer the shallow feature maps of the network to the deep layer, which is helpful for the training of the deep network. Zhou et al. [44] proposed U-Net++, in which each scale feature is designed to carry out feature transfer to a larger scale, thus forming a dense connection structure at skip connections. Ni et al. [43] also added sparable gate unit at the skip connection to improve the accuracy of image spatial feature extraction. Rundo et al. [45] introduced squeeze-and-excitation blocks instead of the skip connection to expect an increased representational power from modeling the channel-wise dependencies of convolutional features [46]. Wang et al. [47] designed a module to generate attentional feature maps by paying attention to the H, W and C of feature maps to replace the skip connection.
In the U-Net, feature maps are learned by “compression-expansion”. In addition, there are studies targeting U-Net as a whole for improvement. Yang et al. [48] transformed U-Net as a spatial attention module and inserted it into the network as a branch to extract spatial features of images. Wei et al. [49] replaced each convolution block in the U-Net structure with a small U-Net structure to achieve multi-scale feature extraction. Xiao et al. [50] adopted a dual U-Net structure to inject features extracted from the external U-Net into the internal U-Net, achieving the effect of multi-stage detail injection.

3. Method

This section will introduce the overall design of MSAC-Net, including its structure, AG module and in-depth monitoring mechanism.
First, we will introduce the pansharpening model based on CNN. Let the LR-MS image with the size h × w × c be denoted M L R R h × w × c , Similarly, denote the HR-PAN with H × W size as P H R R H × W and HR-MS as M H R R H × W × c . Taking LR-MS and HR-PAN as inputs, the pansharpening task of generating the HR-MS can be expressed as:
( θ ) = M ( M L R , P H R ; θ ) M H R
where M ( · ) represents the mapping from the CNN’s input to the out, θ denotes parameters to be optimized and · is a loss function. The CNN can learn the involved knowledge from input data, offering the possibility for MS pansharpening. Table 1 shows the design of all modules in MSAC-Net.The following subsections introduce the details on the representation of these elements within MSAC-Net.

3.1. The MSAC-Net’s Structure

As shown in Figure 2, MSAC-Net consists of three parts: data preprocessing, the contraction pathway (left sub-network) and the expansion pathway (right sub-network). The contraction pathway and the expansion pathway have the same effect as U-Net. Placed between the two paths, the AG mechanism replaces the skip connection, thus improving the local detail feature representation ability. Moreover, MSAC-Net uses multi-scale information pansharpening layers to resolve the MS pansharpening problems.
The traditional 2D CNN approaches commonly cascade P H R with the band dimension of M L R to obtain the input data X R H × W × ( c + 1 ) . In contrast, the proposed MSAC-Net uses 3D cube data as input. As shown in Figure 2, data preprocessing converts the PAN image P H R to P H R R H × W × c , where P H R i = P H R , P H R i { i | i c , R H × W × c } . Similarly, the interpolation algorithm converts M L R to M H R R H × W × c . Finally, the input image X R 2 × H × W × c is obtained by R ( P H R M H R ) , where ⊕ is the cascade operation, R ( · ) represents the resizing operation and the number “2” stems from the two features, MS and PAN.
In the contraction pathway, the i-th low-scale feature F L i is obtained as:
F L i = B l o c k ( M a x ( F L i 1 ) )
where B l o c k ( · ) is is composed of two groups of kernels with a rectified linear unit (ReLU) as an activation function, (i.e., 3 × 3 × 3 kernel + ReLU), where M a x ( · ) denotes the max-pooling layer with a 1 × 2 × 2 kernel. The 1 × 2 × 2 kernel is chosen to down-sampled H and W with a factor of two while keeping the number of bands ( c ) unchanged.
In the expansion pathway, the i-th high-scale feature F H i is obtained as follows:
F H i = T r a n ( B l o c k ( F H i + 1 A G ( F L i + 1 , F H i + 1 ) ) )
where T r a n ( · ) represents the transposed convolution with a factor of two, and A G ( · ) represents stands for the AG module. The pansharpening layer is set after each scale of the convolution block. The next sections introduce further details on the AG module and the pansharpening layer will be introduced in a later section.

3.2. The Attention Gate (AG) Module

Figure 3 depicts the details of the AG module’s structure. The inputs on the i-th AG module are F H i R F i × c × H i × W i and F L i R F i × c × H i × W i . First, the F H i and F L i change the numbers of channels from F i to F i 2 using a 3 × 3 × 3 kernel and are then cascade. Next, the gate feature ( q a t t i ) is obtained via the ReLU function and 1 × 1 × 1 kernel. Finally, q a t t i is activated using a sigmoid function and multiplied by F L i to obtain the feature x ^ u p i . In contrast to [37], this work uses gate features instead of global pooling because the extracted gate features are more consistent with the PAN features. The AG module’s operations can be formalized as:
q a t t i = C o n v ( σ 1 ( F L i F H i ) ; θ )
x ^ u p i = σ 2 ( q a t t i ) F L i
where σ 1 is the ReLU activate function, σ 2 denotes the sigmoid activation, θ is a parameter and ⊗ stands for pixel-by-pixel multiplication on each feature map. Finally, F H i 1 is obtained by the convolution block using the transposed convolution.
F H i 1 = T r a n ( B l o c k ( x ^ u p i ) )
Utilizing the AG module instead of the skip connection enables MSAC-Net to pay more attention to each scale’s local spatial details, consequently improving the spatial performance of pansharpening.

3.3. Pansharpening Layer and Multi-Scale Cost Function

MSAC-Net has S scales in the expansion pathway, i.e., S scale spaces. The pansharpening layer at each scale reconstructs the high-scale feature using a 1 × 1 × 1 convolution kernel, obtaining Y ^ s i . The formula is:
Y ^ s i = R ( B l o c k ( x ^ u p i ) )
where R ( · ) is the i-th scale pansharpening layer. Accordingly, Y s i is generated using the bicubic interpolation, given as:
Y s i = D ( Y s i 1 )
where Y s 1 represents the ground truth and D ( · ) is the bicubic interpolation with a factor of 2.
The 1 -norm loss is used to constrain Y s i and Y ^ s i at the ith scale. The formula is expressed as:
s i = Y ^ s i Y s i 1
where s i denotes the loss at the i-th scale.
Finally, the proposed MSAC-Net’s multi-scale cost function is calculated as (see Figure 2):
L = 1 + λ i = 2 S s i = Y ^ s 1 Y s 1 1 + λ i = 2 S Y ^ s i Y s i 1
where λ is the weight of multi-scale information.

4. Results

4.1. Experimental Setup

4.1.1. Datasets & Parameter Settings

To test the effectiveness of MSAC-Net, we used datasets collected by IKONOS (http://carterraonline.spaceimaging.com/cgi-bin/Carterra/phtml/login.phtml, accessed on 20 December 2018) and QuickBird (http://www.digitalglobe.com/product-samples, accessed on 20 December 2018) satellites (Figure 4). Both datasets contain four standard colors (R: red, G: green, B:blue, N:near infrared). In order to ensure the availability of the ground truth, Wald’s protocol [51] was used to obtain baseline images in training and simulation tests.
The steps of obtaining simulation data are as follows:
(1)
The original HR-PAN and LR-MS images were down-sampled with a factor of 4;
(2)
The down-sampled HR-PAN was used as the input PAN, and the down-sampled LR-MS was used as the input LR-MS;
(3)
The original LR-MS was used as ground truth in the simulation experiment.
In the real datasets, we only normalized the original image. Therefore, the fusion image obtained from real datasets does not have ground truth.
The dataset information is shown in Table 2. The selected dataset contains rich texture information, such as rivers, roads, mountains, etc. We cropped images randomly from each dataset to 3000 PAN/MS data pairs for a total of 6000 data pairs. The PAN size of 256 × 256 × 1 and MS size of 64 × 64 × 4 . Empirically, all data pairs were divided into a training set, validation set and test set in a ratio of 6:3:1. In the experiment of the real dataset, 50 data pairs with PAN image size of 1024 × 1024 × 1 and MS image size of 256 × 256 × 4 were randomly acquired for each dataset.
The proposed method was implemented in Python 3.6 using the Pytorch-1.7 framework and trained and tested on an NVIDIA 1080 GPU. The stochastic gradient descent algorithm was used to converge during training; its parameter settings are shown in Table 3. Among them, the learning rate decay was halved every 2000 iterations. During the training stage, we saved the best model, which achieved the best performance on the validation dataset, and used it for the test.

4.1.2. Compared Methods

The proposed method is validated through a comparison with several pansharpening methods, including GS [52], Indusion [10], SR [13], PNN [18], PanNet [26], MSDCNN [19], MIPSM [20] and GTP-PNet [24]. GS is based on component substitution. Indusion is a multi-resolution analysis method. Furthermore, SR is based on sparse representation learning technology, whereas PNN is based on a three-layered CNN. PanNet is a residual network based on high-pass filtering. MSDCNN introduces residual learning and constructs a multi-scale and multi-depth feature extraction based on CNN. MIPSM is a CNN fusion model that uses dual branches to extract features. Lastly, GTP-PNet is a residual learning network based on gradient transformation prior. PNN, PanNet, MSDCNN, MIPSM and GTP-PNet are the most advanced deep learning methods presented in recent literature.

4.1.3. Performance Metrics

The performance of the proposed pansharpening method is analyzed through quantitative and visual assessments. We selected six reference evaluation indicators and one non-reference evaluation indicator. All symbols in the reference evaluation indicators are explained in Table 4, and the reference evaluation indicators are as follows:
(1)
The correlation coefficient (CC) [53]: CC reflects the similarity of spectral features between the fused image and the ground truth. CC [ 0 , 1 ] , with 1 being the best attainable value. CC can be expressed as follows:
C C ( X ^ , X ) = 1 M m = 1 M n = 1 N ( X ^ n m X ^ ¯ ) ( X n m X ¯ ) n = 1 N ( X ^ n m X ^ ¯ ) 2 ( X n m X ¯ ) 2
(2)
Peak signal-to-noise ratio (PSNR) [54]: PSNR is an objective measure of the information contained in an image. A larger value demonstrates that there is less distortion between the two images. PSNR can be expressed as follows:
P S N R ( X ^ , X ) = 10 lg L 2 M S E ( X ^ , X )
(3)
Spectral angle mapper (SAM) [55]: SAM calculates the overall spectral distortion between the fused image and the ground truth. S A M [ 0 , 1 ] , with 0 being the best attainable value, is defined as follows:
S A M ( X ^ , X ) = 1 N n = 1 N arccos x ^ n , x n x ^ n x n
(4)
Root mean square error (RMSE) [56]: RMSE measures the deviation between the fused image and the ground truth. R M S E [ 0 , 1 ] , with 0 being the best attainable value, is defined as follows:
R M S E ( X ^ , X ) = M S E ( X ^ , X )
(5)
Erreur relative globale adimensionnelle de synthèse (ERGAS) [57]: ERGAS represents the difference between the fused image and the ground truth. E R G A S [ 0 , 1 ] , with 0 being the best attainable value, can be expressed as follows:
E R G A S ( X ^ , X ) = 100 d 1 M m = 1 M R M S E ( X ^ m , X m ) X ¯ m 2
(6)
Structural similarity index measurement (SSIM) [54]: SSIM measures the similarity between the fusion image and the ground truth image. S S I M [ 0 , 1 ] , with 1 being the best attainable value, is defined as follows:
S S I M ( X ^ , X ) = 2 X ^ ¯ X ¯ + C 1 X ^ ¯ 2 + X ¯ 2 + C 1 α 2 σ X ^ σ X + C 2 σ X ^ 2 + σ X 2 + C 2 β σ X ^ X + C 3 σ X ^ σ X + C 3 γ
(7)
Quality without reference (QNR) [58]: As a non-reference evaluation indicator, QNR compares the brightness, contrast and local correlation of the fused image with the original image. Q N R [ 0 , 1 ] , with 1 being the best attainable value, is defined as follows:
Q N R = ( 1 D λ ) a ( 1 D S ) b
where usually a = b = 1 and the spatial distortion index D s and the spectral distortion index D λ are based on universal image quality index (Q) [59]. Furthermore, D s , D λ [ 0 , 1 ] , with 0 being the best attainable value. Q is defined as:
Q ( X ^ , X ) = σ X ^ X σ X ^ σ X × 2 σ X ^ σ X σ X ^ 2 + σ X 2 × 2 X ^ ¯ X ¯ X ^ ¯ 2 + X ¯ 2
thus, D s and D λ are defined as:
D s = 1 M m = 1 M Q ( X ^ m , P A N ) Q ( X m , P A N L ) p p
D λ = 1 M ( M 1 ) m = 1 M r = 1 r m M Q ( X ^ m , X ^ r ) Q ( M S L m , M S L r ) q q

4.2. The Influences of Multi-Scale Fusion and Attention Gate Mechanism

This section discusses MSAC-Net’s and other methods’ results and analyzes the benefits of each MSAC-Net module. Considering that the image has the characteristics of scale invariance and U-Net will scale the image at multiple scales in the learning process, MSAC-Net designed a deep supervision mechanism to restrain the multi-scale fusion process. The design of the deep supervision mechanism is to ensure that in the fusion process of each scale, its features are constrained to ensure that the main features of the image are learned by the network. The AG mechanism can gradually suppress the feature regions of unrelated regions, thus highlighting the feature regions [38]. For image fusion, because MS images contain a lot of noise [60], MSAC-Net suppressed the noise area through the AG mechanism, to highlight the feature area. In addition, MSAC-Net uses 3D convolution to preserve the spectral information of MS data and injects the PAN images’ spatial information into MS images. These settings allow extracting better performance indicators while retaining the spectral correlation between bands.

4.2.1. The Influence of the Pansharpening Layer

Figure 5 visualizes the reconstructed images Figure 5f–j and reference images Figure 5a–e at each scale. It can be observed that as the scale increases, the details of the reconstructed images Figure 5f–j become more similar to the intermediate reference images Figure 5a–e, indicating that the details of Y ^ s i match closer the details of Y s i at each scale. These findings demonstrate that adding the multi-scale cost function enables MSAC-Net to learn the features of each scale easily.

4.2.2. The Influence of the AG Module

In Figure 6, the feature maps of the first scale with and without the AG module are visualized. More precisely, Figure 6b,e are represent the feature maps without the AG module, and Figure 6c,f show the feature maps with the AG module. One can notice that the details in Figure 6c are significantly enhanced compared to those in Figure 6b. The detailed features regarding “rivers” and “land” in Figure 6c are extracted. Figure 6e has lower contrast and blurry edges, whereas Figure 6f has sharper contrast and sharper edges. Compared with F L i , F H i contains more local spatial information, and the texture information is enhanced, yielding more accurate results. Therefore, these results demonstrates the AG module’s utility in advancing the MSAC-Net’s learning of detailed spatial features.

4.2.3. The Influence of Different Structures

Figure 7 compares the pansharpening results of four methods with different structures.
From the perspective of image visual perception, Figure 7b–e do not differ significantly. As can be observed from Figure 6 and Table 5, the spatial details can be extracted more fully by adding the AG mechanism, albeit with a slight loss of the spectral information. In Table 5, SAM of the “U-Net + scale” method is significantly reduced, and SSIM and PSNR are slightly improved when multi-scale information is introduced. Compared with Figure 7f, the spatial detail error of Figure 7h is reduced. These results show that the reconstruction of multi-scale information can make use of the U-Net’s hierarchical relationships to improve the network’s expression ability. As compared with Figure 7f–h, Figure 7i shows the smallest difference between the reference image and the method’s output. Table 5 demonstrates the MSAC-Net’s superiority stemming from its use of the AG module and multi-scale information.

4.3. Comparison of 2D and 3D Convolutional Networks

Within this subsection, 2D and 3D convolution are used in MSAC-Net, and their indicators are compared to assess the advantages of employing 3D convolution in MS imagery pansharpening. The experimental results are shown in Figure 8 and Table 6.
The visual assessment indicates that Figure 8c has significant spatial distortion, whereas Figure 8d is more similar to the ground truth. For example, the road in Figure 8d is clearer than in Figure 8c. The reason may lay in the fact that 3D convolution considers the spectral characteristics of the MS image, thus enabling the feature maps to obtain more spectral information during the feature extraction process.
In Table 6, the MSAC-Net using 3D convolution shows a 2∼7% improvement regarding most indicators.

4.4. Comparison with the State-of-the-Art Methods

This section compares the proposed method to the eight state-of-the-art methods on the simulated and real dataset. Due to the lack of LR- and HR-MS image pairs, the spatial resolution of both PAN and MS images was reduced four times for training and testing, in accordance with Wald’s protocol [51]. During the verification phase, the original MS image and PAN image were used as input and were compared with the whole image and the partial image visually.

4.4.1. Experiments on the Simulated Dataset

The IKONOS and QuickBird simulation data and results are presented in Figure 9 and Figure 10. In Figure 9d–l and Figure 10d–l, the fusion results obtained with GS, Indusion, SR, PNN, PanNet, MSDCNN, MIPSM GTP-PNet and the proposed MSAC-Net method are presented, respectively. Figure 9a–c and Figure 10a–c are LR-MS, PAN and ground truth, respectively. As seen in Figure 9d–h suffer from significant spectral distortions. In contrast, the proposed method yields a result quite similar to the reference image in terms of visual perception, especially regarding the spatial details. In Figure 10, the resultant images of (d), (e), (j), and (k) are too sharp compared with the reference image, while (h) has a color deviation and (i) is blurred.
In Table 7 and Table 8, the methods’ results on IKONOS and the QuickBird dataset are listed. The proposed method achieves competitive results on both datasets. With respect to the best values obtained by other methods (as reported in Table 7), although MSAC-Net does not perform well in IKONOS, it is superior only in ERGAS and SSIM. However, other indicators of MSAC-Net in Table 7 are still competitive. With the exception of SAM, Table 8 shows that the proposed method’s indicators are superior to those of the current state-of-the-art methods. The results establish the proposed method as satisfactory.

4.4.2. Experiments on the Real Dataset

The results obtained using the real datasets IKONOS and QuickBird are presented in Figure 11 and Figure 12. In Figure 11a–i and Figure 12a–i, the fusion results obtained with GS, Indusion, SR, PNN, PanNet, MSDCNN, MIPSM, GTP-PNet and the proposed MSAC-Net method are presented, respectively. Figure 11j and Figure 12j are the reference PAN images. As seen in Figure 11, the images generated using (a), (b), (e), and (g) are too sharp at the “river” edge. Thus, Figure 11h has an error in the “river” edge. Moreover, the synthetic chroma of Figure 11d deviates, and (f) is blurred. On the other hand, MSAC-Net enhances the spectral information and refines the spatial details.
In Table 9 and Table 10, the QNR, D s and D λ of the eight compared methods are listed for the real IKONOS and QuickBird dataset. The proposed method outperforms other methods regarding all indicators. The indicators in both datasets shown demonstrate that the proposed method can enhances detail extraction and reduces spectral distortion.
Overall, the experiments show the proposed method’s capability to retain spectral and spatial information from the original images in the real dataset.

5. Discussion

5.1. The Effect of Convolution Times

The number of convolution in the convolution block will be discussed here. According to experience, the higher the number of convolution times, the richer the feature expression of the network should be. However, as shown in Figure 13a, when the number of convolution starts from 2 and keeps increasing, PSNR gradually decreases. This may be due to the excessive times of convolution, which makes the network pay too much attention to the form of the samples, ignoring the characteristics of the data, resulting in over-fitting of the network. Therefore, two 3 × 3 × 3 kernels are used for in training.

5.2. The Effect of Multi-Scale Information Weight λ

This subsection discusses the effect of weight λ in Equation (10). Figure 13b shows that the increase in weight λ from 0.01 to 1 gradually strengthens the constraints on scale information in MSAC-Net and the resultant feature maps comprise sufficient details at each scale to enhance the results. When λ increases to 10, the network pays more attention to the scale information than the fusion result, affecting the MSAC-Net’s final result. Therefore, to balance the scale information and fusion result, λ is set as to 1.

5.3. The Effect of Network Depth

A final parameter effect that needs to be investigated is the influence of the MSAC-Net’s depth.
As is known to all, increasing the network depth can improve the effect of feature extraction, but a too deep network will lead to degradation. Therefore, based on experience, we discuss the depth of MSAC-Net between three and five layers. Figure 13c shows that as the network depth increases, the average PSNR of the MSAC-Net on the test set gradually improves. Therefore, according to the results in Figure 13c, the depth of MSAC-Net is set to five layers.
Following these experiments, the weight λ was set to 1, the convolution times was 2 and the depth was 5 in MASC-Net.

6. Conclusions

This work introduces a novel 3D multi-scale attention deep convolutional network (MSAC-Net) method for MS imagery pansharpening. The proposed MSAC-Net method utilizes a 3D deep convolutional network appropriate for the MS images’ characteristics. Moreover, the method integrates the attention mechanism and deep supervision mechanism for spectral and spatial information preservation and extraction. The conducted experiments show that MSAC-Net using 3D convolution achieves better quantitative and visual pansharpening performance than the network with 2D convolution. Exhaustive experiments investigated and analyzed the effects of the designed attention module, multi-scale cost function and three critical MSAC-Net factors. The experimental results demonstrate that every designed module positively affects spatial and spectral information extraction, enabling MSAC-Net to achieve the best performance in the appropriate parameters’ range. Compared to the state-of-the-art pansharpening methods, the proposed MSAC-Net achieved comparable or even superior performance on the real IKONOS and QuickBird satellites’ datasets. Building on the results reported in this study, one can conclude that MSAC-Net is a promising multi-spectral imagery pansharpening method. In future work, we will explore how to make more efficient use of the attention mechanism to obtain smaller CNN while ensuring the quality of the fusion image.

Author Contributions

E.Z.: Conceptualization, Writing—review and editing, Supervision. Y.F.: Methodology, Software, Formal analysis, Writing—original draft. J.W.: Conceptualization, Writing—review and editing. L.L.: Supervision, Investigation, Data curation. K.Y.: Resources. J.P.: Visualization, Project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. 62006188, 62101446), the Xi’an Key Laboratory of Intelligent Perception and Cultural Inheritance (No. 201-9219614SYS011CG033), Key Research and Development Program of Shaanxi (No. 2021ZDLSF06-05), the International Science and Technology Cooperation Research Plan of Shaanxi (No. 2022KW-08), the Program for Chang-jiang Scholars and Innovative Research Team in University (No. IRT_17R87) and the QinChuangyuan high-level innovation and entrepreneurship talent program of Shaanxi (2021QCYRC4-50).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

MSMulti-spectral
PANPanchromatic
HRHigh resolution
LRLow resolution
CNNConvolutional neural network
PNNPansharpening neural network
MSDCNNMulti-scale and multi-depth CNN
AGAttention gate
GSGram–Schmidt
CCCorrelation coefficient
PSNRPeak signal-to-noise ratio
SAMSpectral angle mapper
RMSERoot mean square error
ERGASErreur relative globale adimensionnelle de synthèse
SSIMStructural similarity index measurement
QNRQuality without reference
QImage quality index

References

  1. Liu, Y.; Li, Q.; Yuan, Y.; Du, Q.; Wang, Q. ABNet: Adaptive Balanced Network for Multi-scale Object Detection in Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5614914. [Google Scholar] [CrossRef]
  2. Zhang, W.; Liljedahl, A.K.; Kanevskiy, M.; Epstein, H.E.; Jones, B.M.; Jorgenson, M.T.; Kent, K. Transferability of the deep learning mask R-CNN model for automated mapping of ice-wedge polygons in high-resolution satellite and UAV images. Remote Sens. 2020, 12, 1085. [Google Scholar] [CrossRef] [Green Version]
  3. Witharana, C.; Bhuiyan, M.A.E.; Liljedahl, A.K.; Kanevskiy, M.; Epstein, H.E.; Jones, B.M.; Daanen, R.; Griffin, C.G.; Kent, K.; Jones, M.K.W. Understanding the synergies of deep learning and data fusion of multi-spectral and panchromatic high resolution commercial satellite imagery for automated ice-wedge polygon detection. ISPRS J. Photogramm. Remote Sens. 2020, 170, 174–191. [Google Scholar] [CrossRef]
  4. Tan, K.; Ma, W.; Chen, L.; Wang, H.; Du, Q.; Du, P.; Yan, B.; Liu, R.; Li, H. Estimating the distribution trend of soil heavy metals in mining area from HyMap airborne hyperspectral imagery based on ensemble learning. J. Hazard. Mater. 2021, 401, 123288. [Google Scholar] [CrossRef]
  5. Tan, K.; Jin, X.; Plaza, A.; Wang, X.; Xiao, L.; Du, P. Automatic change detection in high-resolution remote sensing images by using a multiple classifier system and spectral–spatial features. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 3439–3451. [Google Scholar] [CrossRef]
  6. Tan, K.; Zhang, Y.; Du, Q.; Du, P.; Jin, X.; Li, J. Change Detection based on Stacked Generalization System with Segmentation Constraint. Photogramm. Eng. Remote Sens. 2018, 84, 733–741. [Google Scholar] [CrossRef]
  7. Lei, J.; Gu, Y.; Xie, W.; Li, Y.; Du, Q. Boundary Extraction Constrained Siamese Network for Remote Sensing Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5621613. [Google Scholar] [CrossRef]
  8. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS + Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  9. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE pan sharpening of very high resolution multispectral images. IEEE Trans. Geosci. Remote Sens. 2007, 46, 228–236. [Google Scholar] [CrossRef]
  10. Khan, M.M.; Chanussot, J.; Condat, L.; Montanvert, A. Indusion: Fusion of multispectral and panchromatic images using the induction scaling technique. IEEE Geosci. Remote Sens. Lett. 2008, 5, 98–102. [Google Scholar] [CrossRef] [Green Version]
  11. Ranchin, T.; Aiazzi, B.; Alparone, L.; Baronti, S.; Wald, L. Image fusion—The ARSIS concept and some successful implementation schemes. ISPRS J. Photogramm. Remote Sens. 2003, 58, 4–18. [Google Scholar] [CrossRef] [Green Version]
  12. Palsson, F.; Ulfarsson, M.O.; Sveinsson, J.R. Model-based reduced-rank pansharpening. IEEE Geosci. Remote Sens. Lett. 2019, 17, 656–660. [Google Scholar] [CrossRef]
  13. Wang, J.; Liu, L.; Ai, N.; Peng, J.; Li, X. Pansharpening based on details injection model and online sparse dictionary learning. In Proceedings of the 2018 13th IEEE Conference on Industrial Electronics and Applications (ICIEA), Wuhan, China, 31 May–2 June 2018; pp. 1939–1944. [Google Scholar]
  14. Peng, J.; Liu, L.; Wang, J.; Zhang, E.; Zhu, X.; Zhang, Y.; Feng, J.; Jiao, L. PSMD-Net: A Novel Pan-Sharpening Method Based on a Multiscale Dense Network. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4957–4971. [Google Scholar] [CrossRef]
  15. Thomas, C.; Ranchin, T.; Wald, L.; Chanussot, J. Synthesis of multispectral images to high spatial resolution: A critical review of fusion methods based on remote sensing physics. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1301–1312. [Google Scholar] [CrossRef] [Green Version]
  16. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. In Fundamental Papers in Wavelet Theory; Princeton University Press: Princeton, NJ, USA, 2009; pp. 494–513. [Google Scholar]
  17. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A critical comparison among pansharpening algorithms. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2565–2586. [Google Scholar] [CrossRef]
  18. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by convolutional neural networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef] [Green Version]
  19. Yuan, Q.; Wei, Y.; Meng, X.; Shen, H.; Zhang, L. A multiscale and multidepth convolutional neural network for remote sensing imagery pan-sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 978–989. [Google Scholar] [CrossRef] [Green Version]
  20. Liu, L.; Wang, J.; Zhang, E.; Li, B.; Zhu, X.; Zhang, Y.; Peng, J. Shallow–deep convolutional network and spectral-discrimination-based detail injection for multispectral imagery pan-sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1772–1783. [Google Scholar] [CrossRef]
  21. Jiang, H.; Peng, M.; Zhong, Y.; Xie, H.; Hao, Z.; Lin, J.; Ma, X.; Hu, X. A Survey on Deep Learning-Based Change Detection from High-Resolution Remote Sensing Images. Remote Sens. 2022, 14, 1552. [Google Scholar] [CrossRef]
  22. Jin, Z.R.; Zhuo, Y.W.; Zhang, T.J.; Jin, X.X.; Jing, S.; Deng, L.J. Remote Sensing Pansharpening by Full-Depth Feature Fusion. Remote Sens. 2022, 14, 466. [Google Scholar] [CrossRef]
  23. Zhou, M.; Fu, X.; Huang, J.; Zhao, F.; Liu, A.; Wang, R. Effective Pan-Sharpening with Transformer and Invertible Neural Network. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5406815. [Google Scholar] [CrossRef]
  24. Zhang, H.; Ma, J. GTP-PNet: A residual learning network based on gradient transformation prior for pansharpening. ISPRS J. Photogramm. Remote Sens. 2021, 172, 223–239. [Google Scholar] [CrossRef]
  25. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [Green Version]
  26. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A deep network architecture for pan-sharpening. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5449–5457. [Google Scholar]
  27. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised spatial–spectral feature learning by 3D convolutional autoencoder for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  28. Wang, L.; Bi, T.; Shi, Y. A Frequency-Separated 3D-CNN for Hyperspectral Image Super-Resolution. IEEE Access 2020, 8, 86367–86379. [Google Scholar] [CrossRef]
  29. Shi, C.; Liao, D.; Zhang, T.; Wang, L. Hyperspectral Image Classification Based on 3D Coordination Attention Mechanism Network. Remote Sens. 2022, 14, 608. [Google Scholar] [CrossRef]
  30. Mei, S.; Yuan, X.; Ji, J.; Zhang, Y.; Wan, S.; Du, Q. Hyperspectral image spatial super-resolution via 3D full convolutional neural network. Remote Sens. 2017, 9, 1139. [Google Scholar] [CrossRef] [Green Version]
  31. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent models of visual attention. arXiv 2014, arXiv:1406.6247. [Google Scholar]
  32. Mei, X.; Pan, E.; Ma, Y.; Dai, X.; Huang, J.; Fan, F.; Du, Q.; Zheng, H.; Ma, J. Spectral-spatial attention networks for hyperspectral image classification. Remote Sens. 2019, 11, 963. [Google Scholar] [CrossRef] [Green Version]
  33. Fu, J.; Zheng, H.; Mei, T. Look closer to see better: Recurrent attention convolutional neural network for fine-grained image recognition. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4438–4446. [Google Scholar]
  34. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  35. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin, Germany, 2015; pp. 234–241. [Google Scholar]
  36. Guo, Y.; Chen, J.; Wang, J.; Chen, Q.; Cao, J.; Deng, Z.; Xu, Y.; Tan, M. Closed-loop matters: Dual regression networks for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5407–5416. [Google Scholar]
  37. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  38. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  39. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  40. Wang, Y.; Peng, Y.; Liu, X.; Li, W.; Alexandropoulos, G.C.; Yu, J.; Ge, D.; Xiang, W. DDU-Net: Dual-Decoder-U-Net for Road Extraction Using High-Resolution Remote Sensing Images. arXiv 2022, arXiv:2201.06750. [Google Scholar]
  41. Banerjee, S.; Lyu, J.; Huang, Z.; Leung, F.H.; Lee, T.; Yang, D.; Su, S.; Zheng, Y.; Ling, S.H. Ultrasound spine image segmentation using multi-scale feature fusion skip-inception U-Net (SIU-Net). Biocybern. Biomed. Eng. 2022, 42, 341–361. [Google Scholar] [CrossRef]
  42. Khalel, A.; Tasar, O.; Charpiat, G.; Tarabalka, Y. Multi-task deep learning for satellite image pansharpening and segmentation. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 4869–4872. [Google Scholar]
  43. Ni, Y.; Xie, Z.; Zheng, D.; Yang, Y.; Wang, W. Two-stage multitask U-Net construction for pulmonary nodule segmentation and malignancy risk prediction. Quant. Imaging Med. Surg. 2022, 12, 292. [Google Scholar] [CrossRef] [PubMed]
  44. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin, Germany, 2018; pp. 3–11. [Google Scholar]
  45. Rundo, L.; Han, C.; Nagano, Y.; Zhang, J.; Hataya, R.; Militello, C.; Tangherloni, A.; Nobile, M.S.; Ferretti, C.; Besozzi, D.; et al. USE-Net: Incorporating Squeeze-and-Excitation blocks into U-Net for prostate zonal segmentation of multi-institutional MRI datasets. Neurocomputing 2019, 365, 31–43. [Google Scholar] [CrossRef] [Green Version]
  46. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  47. Wang, Y.; Kong, J.; Zhang, H. U-Net: A Smart Application with Multidimensional Attention Network for Remote Sensing Images. Sci. Program. 2022, 2022, 1603273. [Google Scholar] [CrossRef]
  48. Yang, Q.; Xu, Y.; Wu, Z.; Wei, Z. Hyperspectral and multispectral image fusion based on deep attention network. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar]
  49. Wei zhou, W.; Guo wu, Y.; Hao, W. A multi-focus image fusion method based on nested U-Net. In Proceedings of the 2021 the 5th International Conference on Video and Image Processing, Hayward, CA, USA, 22–25 December 2021; pp. 69–75. [Google Scholar]
  50. Xiao, J.; Li, J.; Yuan, Q.; Zhang, L. A Dual-UNet with Multistage Details Injection for Hyperspectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5515313. [Google Scholar] [CrossRef]
  51. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  52. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6,011,875, 4 January 2000. [Google Scholar]
  53. Alparone, L.; Wald, L.; Chanussot, J.; Thomas, C.; Gamba, P.; Bruce, L.M. Comparison of pansharpening algorithms: Outcome of the 2006 GRS-S data-fusion contest. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3012–3021. [Google Scholar] [CrossRef] [Green Version]
  54. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  55. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Proceedings of the Summaries 3rd Annual JPL Airborne Geoscience Workshop, Pasadena, CA, USA, 1–5 June 1992; Volume 1, pp. 147–149. [Google Scholar]
  56. Yang, Y.; Wan, W.; Huang, S.; Lin, P.; Que, Y. A novel pan-sharpening framework based on matting model and multiscale transform. Remote Sens. 2017, 9, 391. [Google Scholar] [CrossRef] [Green Version]
  57. Wald, L. Data Fusion: Definitions and Architectures: Fusion of Images of Different Spatial Resolutions; Presses des MINES: Paris, France, 2002. [Google Scholar]
  58. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef] [Green Version]
  59. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  60. Sun, W.; Ren, K.; Meng, X.; Yang, G.; Xiao, C.; Peng, J.; Huang, J. MLR-DBPFN: A Multi-scale Low Rank Deep Back Projection Fusion Network for Anti-noise Hyperspectral and Multispectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522914. [Google Scholar] [CrossRef]
Figure 1. Classic U-Net design.
Figure 1. Classic U-Net design.
Remotesensing 14 02761 g001
Figure 2. The proposed MSAC-Net architecture.
Figure 2. The proposed MSAC-Net architecture.
Remotesensing 14 02761 g002
Figure 3. The attention gate (AG) module.
Figure 3. The attention gate (AG) module.
Remotesensing 14 02761 g003
Figure 4. Display images in datasets. (ad) are the IKONOS dataset’s images. (eh) are the QuickBird dataset’s images.
Figure 4. Display images in datasets. (ad) are the IKONOS dataset’s images. (eh) are the QuickBird dataset’s images.
Remotesensing 14 02761 g004
Figure 5. The reference image and pansharpening image are compared at different scales. The first line is the reference image for each scale, and the second line is pansharpening image for each scale. (ae) are the reference images of the first to fifth layers respectively, and corresponding, (fj) are the reconstructed images of the first to fifth layers respectively.
Figure 5. The reference image and pansharpening image are compared at different scales. The first line is the reference image for each scale, and the second line is pansharpening image for each scale. (ae) are the reference images of the first to fifth layers respectively, and corresponding, (fj) are the reconstructed images of the first to fifth layers respectively.
Remotesensing 14 02761 g005
Figure 6. The comparison of feature maps with and without the AG module. (a) The ground truth. (b) The 53rd feature map in F L 1 . (c) The 53rd feature map in F H 1 . (d) the ground truth. (e) The 23rd feature map in F L 1 . (f) The 23rd feature map in F H 1 .
Figure 6. The comparison of feature maps with and without the AG module. (a) The ground truth. (b) The 53rd feature map in F L 1 . (c) The 53rd feature map in F H 1 . (d) the ground truth. (e) The 23rd feature map in F L 1 . (f) The 23rd feature map in F H 1 .
Remotesensing 14 02761 g006
Figure 7. The comparison of the results for different structures. (a) The ground truth. (b) U-net. (c) U-net + AG: U-net with AG on the skip connections. (d) U-net + scale: U-Net with multi-scale cost function. (e) MSAC-Net. (fi) are the spectral distortion maps corresponding to (be).
Figure 7. The comparison of the results for different structures. (a) The ground truth. (b) U-net. (c) U-net + AG: U-net with AG on the skip connections. (d) U-net + scale: U-Net with multi-scale cost function. (e) MSAC-Net. (fi) are the spectral distortion maps corresponding to (be).
Remotesensing 14 02761 g007
Figure 8. The comparison of 2D and 3D convolution results on the QuickBird dataset. (a) PAN. (b) The ground truth. (c) The 2D convolutional method. (d) The 3D convolutional method.
Figure 8. The comparison of 2D and 3D convolution results on the QuickBird dataset. (a) PAN. (b) The ground truth. (c) The 2D convolutional method. (d) The 3D convolutional method.
Remotesensing 14 02761 g008
Figure 9. The fused results on the simulated IKONOS dataset. (a) LR-MS. (b) PAN. (c) The ground truth. (d) GS. (e) Indusion. (f) SR. (g) PNN. (h) PanNet. (i) MSDCNN. (j) MIPSM. (k) GTP-PNet. (l) MSAC-Net.
Figure 9. The fused results on the simulated IKONOS dataset. (a) LR-MS. (b) PAN. (c) The ground truth. (d) GS. (e) Indusion. (f) SR. (g) PNN. (h) PanNet. (i) MSDCNN. (j) MIPSM. (k) GTP-PNet. (l) MSAC-Net.
Remotesensing 14 02761 g009
Figure 10. The fused results on the simulated QuickBird dataset. (a) LR-MS. (b) PAN. (c) The ground truth. (d) GS. (e) Indusion. (f) SR. (g) PNN. (h) PanNet. (i) MSDCNN. (j) MIPSM. (k) GTP-PNet. (l) MSAC-Net.
Figure 10. The fused results on the simulated QuickBird dataset. (a) LR-MS. (b) PAN. (c) The ground truth. (d) GS. (e) Indusion. (f) SR. (g) PNN. (h) PanNet. (i) MSDCNN. (j) MIPSM. (k) GTP-PNet. (l) MSAC-Net.
Remotesensing 14 02761 g010
Figure 11. The real results on the IKONOS dataset. (a) GS. (b) Indusion. (c) SR. (d) PNN. (e) PanNet. (f) MSDCNN. (g) MIPSM. (h) GTP-PNet. (i) MSAC-Net. (j) PAN.
Figure 11. The real results on the IKONOS dataset. (a) GS. (b) Indusion. (c) SR. (d) PNN. (e) PanNet. (f) MSDCNN. (g) MIPSM. (h) GTP-PNet. (i) MSAC-Net. (j) PAN.
Remotesensing 14 02761 g011
Figure 12. The real results on the QuickBird dataset. (a) GS. (b) Indusion. (c) SR. (d) PNN. (e) PanNet. (f) MSDCNN. (g) MIPSM. (h) GTP-PNet. (i) MSAC-Net. (j) PAN.
Figure 12. The real results on the QuickBird dataset. (a) GS. (b) Indusion. (c) SR. (d) PNN. (e) PanNet. (f) MSDCNN. (g) MIPSM. (h) GTP-PNet. (i) MSAC-Net. (j) PAN.
Remotesensing 14 02761 g012
Figure 13. Discuss the effects of parameters (a) The effect of convolution times. (b) The effect of multi-scale information weight λ . (c) The effect of network depth.
Figure 13. Discuss the effects of parameters (a) The effect of convolution times. (b) The effect of multi-scale information weight λ . (c) The effect of network depth.
Remotesensing 14 02761 g013
Table 1. The design of the MSAC-Net.
Table 1. The design of the MSAC-Net.
ModuleInput SizeLayer (Filter Size) or MethodFiltersOutput Size
Down 1 2 × c × H 1 × W 1 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 64 64 × c × H 1 × W 1
64 × c × H 1 × W 1 M a x p o o l ( 1 , 2 , 2 ) 64 × c × H 2 × W 2
Down 2 64 × c × H 2 × W 2 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 128 128 × c × H 2 × W 2
128 × c × H 2 × W 2 M a x p o o l ( 1 , 2 , 2 ) 128 × c × H 3 × W 3
Down 3 128 × c × H 3 × W 3 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 256 256 × c × H 3 × W 3
256 × c × H 3 × W 3 M a x p o o l ( 1 , 2 , 2 ) 256 × c × H 4 × W 4
Down 4 256 × c × H 4 × W 4 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 512 512 × c × H 4 × W 4
512 × c × H 4 × W 4 M a x p o o l ( 1 , 2 , 2 ) 512 × c × H 5 × W 5
Down 5 512 × c × H 5 × W 5 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 1024 1024 × c × H 5 × W 5
1024 × c × H 5 × W 5 T r a n ( [ 1 , 4 , 4 ] , s t r i d e = [ 1 , 2 , 2 ] ) 512 512 × c × H 4 × W 4
Up 5 512 × c × H 4 × W 4 A G 5 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 512 512 × c × H 4 × W 4
512 × c × H 4 × W 4 T r a n ( [ 1 , 4 , 4 ] , s t r i d e = [ 1 , 2 , 2 ] ) 256 256 × c × H 3 × W 3
Up 4 256 × c × H 3 × W 3 A G 4 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 256 256 × c × H 3 × W 3
256 × c × H 3 × W 3 T r a n ( [ 1 , 4 , 4 ] , s t r i d e = [ 1 , 2 , 2 ] ) 128 128 × c × H 2 × W 2
Up 3 128 × c × H 2 × W 2 A G 3 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 128 128 × c × H 2 × W 2
128 × c × H 2 × W 2 T r a n ( [ 1 , 4 , 4 ] , s t r i d e = [ 1 , 2 , 2 ] ) 64 64 × c × H 1 × W 1
Up 2 64 × c × H 1 × W 1 A G 2 ( C o n v 3 D ( 3 , 3 , 3 ) , R e L U ) × 2 64 64 × c × H 1 × W 1
Output 64 × c × H 4 × W 4 C o n v 3 D ( 1 , 1 , 1 ) 1 1 × c × H 1 × W 1
Reconstruction F i × c × H i × W i C o n v 3 D ( 1 , 1 , 1 ) 1 1 × c × H i × W i
AG i + 1 ( D o w n i ) F i × c × H i × W i C o n v 3 D ( 1 , 1 , 1 ) F i / 2 ( x ) F i / 2 × c × H i × W i
( U p i + 1 ) F i × c × H i × W i ( g ) F i / 2 × c × H i × W i
( x ) F i / 2 × c × H i × W i R e L U ( x + g ) ) F i × c × H i × W i
( g ) F i / 2 × c × H i × W i
F i × c × H i × W i C o n v 3 D ( 1 , 1 , 1 ) + s i g m o i d 1 1 × c × H i × W i
1 × c × H i × W i F i × c × H i × W i
( D o w n i ) F i × c × H i × W i
Table 2. Dataset information.
Table 2. Dataset information.
SatellitePanBlueGreenRedNIRResolution (PAN/MS)
IKONOS450–900450–530520–610640–720760–8601 m/4 m
QuickBird450–900450–520520–600630–690760–9000.7 m/2.8 m
Table 3. Parameter list.
Table 3. Parameter list.
Learning RateWeight DecayMomentumBatch SizeIteration
0.0110 3 0.916 2 × 10 4
Table 4. Formula symbol table.
Table 4. Formula symbol table.
SymbolMeansSymbolMeans
X ^ Input imageXGround Truth
X ^ ¯ Sample means of X ^ X ¯ Sample means of X
x ^ A pixel in X ^ xA pixel in X
NNumber of pixelsMNumber of band
M S E ( ) Mean square errorLGray level
α Brightness parameter β Contrast parameter
γ Structural parameter C i Infinitely small constant
σ X ^ Variances of X ^ σ X Variances of X
σ X ^ X Covariance of X ^ and X p , q Positive integer
dPAN image spatial resolution/MS image spatial resolution
Table 5. The comparison of the results for different structures on IKONOS.
Table 5. The comparison of the results for different structures on IKONOS.
StructuresCCPSNRSAMERGASRMSESSIM
U-Net0.895921.52005.16707.82630.08390.6525
U-Net + AG0.894921.52475.27217.84070.08390.6543
U-Net + scale0.895521.57544.81877.64030.08370.6624
Our0.896821.61844.42817.43030.08300.6755
Bold means the best results.
Table 6. The comparison of 2D and 3D convolution results on QuickBird.
Table 6. The comparison of 2D and 3D convolution results on QuickBird.
KernelCCPSNRSAMERGASRMSESSIM
2D0.866019.50255.44456.16380.10590.6628
3D0.887020.21685.12695.74460.09750.7107
Bold means the best results.
Table 7. Quality metrics of the different methods on the simulated IKONOS image.
Table 7. Quality metrics of the different methods on the simulated IKONOS image.
MethodsCCPSNRSAMERGASRMSESSIM
GS0.532014.39405.752020.55270.19480.2964
Indusion0.833019.92356.275011.36090.10040.4136
SR0.820618.836312.214715.65790.13980.4513
PNN0.895217.748524.147712.25160.09650.5553
PanNet0.851418.74746.829312.66930.11280.3931
MSDCNN0.906321.05955.60259.80530.07580.5728
MIPSM0.907021.07464.42219.78400.08960.5855
GTP-PNet0.909623.44676.07608.18880.07630.6020
MSAC-Net0.896821.61844.42817.43030.08300.6755
Bold means the best results
Table 8. Quality metrics of the different methods on the simulated QuickBird image.
Table 8. Quality metrics of the different methods on the simulated QuickBird image.
MothedsCCPSNRSAMERGASRMSESSIM
GS0.886619.63922.91789.59280.10310.6973
Indusion0.917821.61062.69697.56480.08220.7116
SR0.911518.39913.559810.70520.11790.5897
PNN0.929721.57445.77137.24830.07780.7398
PanNet0.914121.07933.81428.08590.08830.7010
MSDCNN0.926921.93663.80537.25140.07940.7330
MIPSM0.940322.77244.33596.64320.07240.7300
GTP-PNet0.931121.91807.90217.86610.09620.7109
MSAC-Net0.951523.08005.31206.22970.07010.7966
Bold means the best results.
Table 9. Quality metrics of the different methods on the real IKONOS dataset.
Table 9. Quality metrics of the different methods on the real IKONOS dataset.
MethodsQNR D s D λ
GS0.24790.73660.0588
Indusion0.67850.31500.0096
SR0.59230.39340.0236
PNN0.74430.15790.1162
PanNet0.92510.06790.0075
MSDCNN0.80160.14930.0577
MIPSM0.95020.02670.0237
GTP-PNet0.94940.02030.0309
MSAC-Net0.95290.01940.0282
Bold means the best results.
Table 10. Quality metrics of the different methods on the real QuickBird dataset.
Table 10. Quality metrics of the different methods on the real QuickBird dataset.
MethodsQNR D s D λ
GS0.36890.44460.3358
Indusion0.66220.18700.1854
SR0.74680.22500.0364
PNN0.71390.13290.1767
PanNet0.79250.04690.1685
MSDCNN0.71840.20980.0908
MIPSM0.88160.04200.0797
GTP-PNet0.80650.01410.1820
MSAC-Net0.95460.02670.0192
Bold means the best results.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, E.; Fu, Y.; Wang, J.; Liu, L.; Yu, K.; Peng, J. MSAC-Net: 3D Multi-Scale Attention Convolutional Network for Multi-Spectral Imagery Pansharpening. Remote Sens. 2022, 14, 2761. https://doi.org/10.3390/rs14122761

AMA Style

Zhang E, Fu Y, Wang J, Liu L, Yu K, Peng J. MSAC-Net: 3D Multi-Scale Attention Convolutional Network for Multi-Spectral Imagery Pansharpening. Remote Sensing. 2022; 14(12):2761. https://doi.org/10.3390/rs14122761

Chicago/Turabian Style

Zhang, Erlei, Yihao Fu, Jun Wang, Lu Liu, Kai Yu, and Jinye Peng. 2022. "MSAC-Net: 3D Multi-Scale Attention Convolutional Network for Multi-Spectral Imagery Pansharpening" Remote Sensing 14, no. 12: 2761. https://doi.org/10.3390/rs14122761

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop