Next Article in Journal
Urban Building Type Mapping Using Geospatial Data: A Case Study of Beijing, China
Next Article in Special Issue
Modality-Free Feature Detector and Descriptor for Multimodal Remote Sensing Image Registration
Previous Article in Journal
Landsat-8 and Sentinel-2 Canopy Water Content Estimation in Croplands through Radiative Transfer Model Inversion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PWNet: An Adaptive Weight Network for the Fusion of Panchromatic and Multispectral Images

School of Mathematics and Statistics, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(17), 2804; https://doi.org/10.3390/rs12172804
Submission received: 31 July 2020 / Revised: 22 August 2020 / Accepted: 24 August 2020 / Published: 29 August 2020
(This article belongs to the Special Issue Multi-Sensor Systems and Data Fusion in Remote Sensing)

Abstract

:
Pansharpening is a typical image fusion problem, which aims to produce a high resolution multispectral (HRMS) image by integrating a high spatial resolution panchromatic (PAN) image with a low spatial resolution multispectral (MS) image. Prior arts have used either component substitution (CS)-based methods or multiresolution analysis (MRA)-based methods for this propose. Although they are simple and easy to implement, they usually suffer from spatial or spectral distortions and could not fully exploit the spatial and/or spectral information existed in PAN and MS images. By considering their complementary performances and with the goal of combining their advantages, we propose a pansharpening weight network (PWNet) to adaptively average the fusion results obtained by different methods. The proposed PWNet works by learning adaptive weight maps for different CS-based and MRA-based methods through an end-to-end trainable neural network (NN). As a result, the proposed PWN inherits the data adaptability or flexibility of NN, while maintaining the advantages of traditional methods. Extensive experiments on data sets acquired by three different kinds of satellites demonstrate the superiority of the proposed PWNet and its competitiveness with the state-of-the-art methods.

Graphical Abstract

1. Introduction

Due to technical limitations [1], current satellites, such as QuickBird, IKONOS, WorldView-2, GeoEye-1, can not obtain the high spatial resolution multispectral (MS) images, but only acquire an image pair with complementary features, i.e., a high spatial resolution panchromatic (PAN) image and a low spatial resolution MS image with rich spectral information. To get high-quality products, pansharpening is proposed with the goal of fusing MS and PAN images to generate high resolution multispectral (HRMS) image with the same spatial resolution of the PAN image and the spectral resolution of the MS image [2,3]. It can be cast as a typical kind of image fusion [4] or super-resolution [5] problems and has a wide range of real-world applications, such as enhancing the visual interpretation, monitoring the land cover change [6], object recognition [7], and so on.
Over decades of studies, a large number of pansharpening methods have been proposed in the literature of remote sensing [3]. Most of them can be categorized into the following two main classes [2,3,8]: (1) component substitution (CS)-based methods and (2) multiresolution analysis (MRA)-based methods. The CS class first transforms the original MS image into a new space and then substitutes one component of the transformed MS image by the histogram matched PAN image. The representative methods of the CS class are Intensity-Hue-Saturation (IHS) [9], generalized IHS (GIHS) [10], principal component analysis (PCA) [11], Brovey [12], among many others [13,14,15,16,17]. The MRA-based class is also known as the class of spatial methods, which extracts the high spatial frequencies of the high resolution PAN image through multiresolution analysis tools (e.g., wavelets or Laplacian pyramids) to enhance the spatial information of MS image. The representative methods belonging to the MRA-based class are high-pass filtering (HPF) [18], smoothing filter-based intensity modulation (SFIM) [19], the generalized Laplacian pyramid (GLP) [20,21], among many others [22,23]. The two class methods are fast and easy to implement. However, for the CS-based methods, local dissimilarities between PAN and MS images can not be eliminated, resulting in spectral distortion, and for the MRA-based methods, they have a relatively less spectral distortion but with limited spatial enhancement. From the above, the CS-based and the MRA-based methods usually have complementary performances in improving the spatial quality of MS images while maintaining the corresponding spectral information.
To balance the trade-off performances of the CS-based and MRA-based methods, the hybrid methods by combining both of these two classes have been proposed in recent years. For example, the additive wavelet luminance proportional (AWLP) [24] method is proposed by Otazu et al. via implementing the “à trous” wavelet transform in the IHS space. Shah et al. [25] proposed a method by combining an adaptive PCA method with the discrete contourlet transform. Liao et al. [26] proposed an framework, called guided filter PCA (GFPCA), which performs a guided filter in the PCA domain. Although the hybrid methods have an enhanced performance to the CS-based or MRA-based methods, these improvements are limited due to their hand-crafted design.
Recently, significant progress on improving the spatial and spectral qualities of the fused images for the classical methods has been achieved by variational optimization (VO)-based methods [27,28,29,30,31] and learning-based methods, among which convolution neural network (CNN)-based methods are the most popular, due to their powerful capability and the end-to-end learning strategy. For instance, Masi et al. introduced a CNN architecture with three layers in [32] for the pansharpening problem. Another novel CNN-based model, which is focused on preserving spatial and spectral information, is designed by Yang et al. in [33]. Inspired by these work, Liu et al. [34] proposed a two-stream CNN architecture with 1 -norm loss function to further improve the spatial quality. Zheng et al. [35] proposed a CNN-based method by using deep hyperspectral prior and dual-attention residual network to deal with the problem of that the discriminative ability of CNNs is sometimes hindered. Though having great ability of automatically extracting features and the state-of-the-art performances, CNN-based methods usually require intensive computational resources [36]. In addition, unlike the CS-based and MRA-based methods, CNN-based methods are lack of interpretability and are more like a black-box game. A detailed summary and relevant works for the VO-based methods can be found in [2]. We do not discuss the VO class for more since this paper focuses on a combination of the other three classes.
In this paper, we propose a pansharpening weight network (PWNet) to bridge the classical methods (i.e., CS-based and MRA-based methods) and the learning-based methods (typically the CNN-based methods). On one hand, similar to the hybrid methods, PWNet can combine the merits of the CS-based and the MRA-based methods. On the other hand, similar to learning-based methods, PWNet is data-driven and is very effective and efficient. To achieve this, PWNet uses the CS-based and MRA-based methods as inference modules and utilizes CNN to learn adaptive weight maps for weighting the results of the classical methods. Unlike the above hybrid methods with hand-crafted design, the PWNet can be seen as an automatic and data-driven hybrid method for pansharpening. In addition, the structure of PWNet is very simple to ease training and save computational time.
The main contributions of this work are as follows:
  • A model average network, called pansharpening weight network (PWNet), is proposed. The PWNet can be trained and is the first attempt to combining the classical methods via an end-to-end trainable network.
  • PWNet integrates the complementary characteristics of the CS-based and MRA-based methods and the flexibility of the learning-based (typically the CNN-based) methods, providing an avenue to bridge the gap between them.
  • PWNet is data-driven, and can automatically weight the contributions of different CS-based and MRA-based methods on different data sets. By visualizing the weight maps, we prove that the PWNet is adaptive and robust to different data sets.
  • Extensive experiments on three kinds of data sets have been conducted and shown that the fusion results obtained by PWNet achieve state-of-the-art performance compared with the CS-based, MRA-based methods and other CNN-based methods.
The paper is organized as follows. In Section 2, we briefly introduce the background of the CS-based, MRA-based and learning-based methods. Section 3 introduces the motivation, network architecture, and other details of PWNet. In Section 4, we conduct the experiments, analyze the parameter setting and time complexity and present the comparisons with the-state-of-art methods at the reduced and full scales. Finally, we draw the conclusion in Section 5.

2. Related Work

Notations. We denote the low resolution multispectral (LRMS) image by M S R H × W × N , where H , W , and N are the width, the height, and the number of spectral bands of the LRMS image, respectively. We denote the high resolution PAN image by P R r H × r W , where r is the spatial resolution ratio between MS and PAN, denote by M S ^ R r H × r W × N the reconstructed HRMS image. We let M S k to represent the kth band of the LRMS image, where k = 1 , , N , and let M S k ˜ R r H × r W to represent the upsampled version of M S k by ratio r. For notational simplicity, we also denote by P the histogram matched PAN image. Based on these symbols, we next briefly introduce the main idea of the CS-based, MRA-based and learning-based methods.

2.1. The CS-Based Methods

The CS-based methods are based on the assumption that the spatial and spectral information of LRMS image can be separated by a projection or transformation of the original LRMS image [3,37]. The CS class usually has four steps: (1) upsample the LRMS image to the size of the PAN image; (2) use a linear transformation to project the upsampled LRMS image into another space; (3) replace the component containing the spatial information with the PAN image; (4) perform an inverse transformation to bring the transformed MS data back to their original space and then get the pansharpened MS image (i.e., the estimated HRMS). Due to the changes in low spatial frequencies of the MS image, the substitution procedure usually suffers from spectral distortion. Thus, spectral matching procedure (i.e., histogram matching) is often applied before the substitution.
Mathematically, above fusion process can be simplified without the calculation of the forward and backward transformation as shown in Figure 1, which leads the CS class to have the following equivalent form as
M S k ^ = M S k ˜ + g k ( P I L ) , k = 1 , , N ,
where g 1 , , g N are the injection gains, and I L is a linear combination of the upsampled LRMS image bands and often called intensity component, defined as
I L = k = 1 N w k M S k ˜ ,
where w 1 , , w N usually correspond to the first row of the forward transformation matrix, which is used to measure the degrees of spectral overlap between the MS and PAN channels.
Numerous CS-based methods have been proposed to sharpen the LRMS images according to Equation (1) and flowchart in Figure 1. The CS class includes IHS [9] which exploits the transformation into the IHS color space and its generalized version GIHS [10], PCA [11] based on the statistical irrelevance of each principal component, Brovey [12] based on a multiplicative injection scheme, Gram-Schmidt (GS) [13] which conducts the Gram-Schmidt orthogonalization procedure and by a weighted average of the MS bands minimizing the mean square error (MSE) with respect to a low-pass filtered version PAN image in the adaptive GS (GSA) [15], band-dependent spatial detail (BDSD) [14] and its enhanced version (i.e., BDSD with physical constraints: BDSD-PC) [16], partial replacement adaptive component substitute (PRACS) [17] based on the concept of partial replacement of the intensity component and so on. Each method differs from the others by the different projections of the MS images used in the process and by the different designs of injection gains. Although they show extreme performances in improving the spatial qualities of LRMS images, they usually suffer from heavily spectral distortions in some scenarios due to local dissimilarity or the not well-separated spatial structure with the spectral information. Refer to [3] for more detailed discussions about this.

2.2. The MRA-Based Methods

Unlike the CS-based methods, the MRA class is based on the operator of multi-scale decomposition or low-pass filter (equal to a single scale of decomposition) over the PAN image [3,37]. They first extract the spatial details over a wide range of scales from the high resolution PAN image or from the difference between the PAN image and its low-pass filtered version P L , and then inject the extracted spatial details into each band of upsampled LRMS image. Figure 2 shows the general flowchart of the MRA-based methods.
Generally, for each band k = 1 , 2 , , N , the MRA-based methods can be formulated as
M S k ^ = M S k ˜ + g k ( P P L ) .
As we can see from above Equation (3), different MRA-based methods can be distinguished by the way of obtaining P L and by the design of injection gains g 1 , g 2 , , g N . Several methods belonging to this class have been proposed, such as HPF [18] using the box mask and additive injection, the SFIM [19], decimated Wavelet transform using additive injection model (Indusion) [23], the AWLP [24], GLP with modulation transfer function (MTF)-matched filter (denoted by MTF-GLP) [21], its HPM injection version (MTF-GLP-HPM) [22] and context-based decision version (MTF-GLP-CBD) [38], a trous wavelet transform using the model 3 (ATWT-TM3) [39], and so on.
The MRA-based methods highlight the extraction of multi-scale and local details from the PAN image, well in reducing the spectral distortion but compromising the spatial enhancement. To make up this problem, many approches have been proposed by the utilization of different decomposition schemes (e.g., morphological filters [40]) and the optimization of the injection gains.

2.3. The Learning-Based Method

Apart from the traditional CS-based and MRA-based methods, the learning-based methods have been proposed or applied to the pansharpening, among which the CNN-based methods are the most popular [41]. The CNN-based methods are very flexible, and one can design a CNN with different architectures. Due to the end-to-end and data-driven properties, they achieve the state-of-the-art performances in some studies [32,33,34,35]. After a network architecture design, training image pairs with low resolution MS (LRMS) images as network input and high resolution MS (HRMS) images as network output, are needed to learn the network parameters θ . The learning procedure is based on the choice of loss function and optimization method, and the effect of learning is different from each other according to these choices of loss function and optimization strategy. However, these ideal image pairs are unavailable, and usually simulated based on scale invariant assumption by properly downsampling both PAN and the original MS images to a reduced resolution. Then, the resolution reduced MS images and the original MS can be used as an input-output pair.
Given the input-output MS pairs of M S ˜ i with low resolution and Y i with high resolution, and aided by the auxiliary PAN image P i , i = 1 , 2 , , n , the CNN-based methods optimize the parameter by minimizing the following cost function
L ( θ ) = i = 1 n | | f ( P i , M S ˜ i ; θ ) Y i | | F 2
where f ( P i , M S ˜ i ; θ ) denotes a neural network which takes θ as parameters, and | | · | | F is the Frobenius norm, which is defined as the square root of the sum of the absolute squares of the elements.
To further improve the performances of CNN-based methods, recent work mainly resorts to the deep residual architecture [42] or to increase the depth of the model to extract multi-level abstract features [43]. However, these will require large number of network parameters and burden computation [36]. Unlike the above CNN-based methods that aim at generating the HRMS images or their residual images, we here to reduce the number of parameters and reduce the requirements on the computation capacity of the computer by learning weight maps for the CS-based and the MRA-based methods. Refer to the following section for more detailed discussions about this.

3. The Proposed PWNet Method

3.1. Motivation and Main Idea

According to the above analysis, the CS-based and MRA-based methods are simple and usually have complementary performances, i.e., the CS-based methods are good at spatial rendering but sometimes suffer from severe spectral distortions, while the MRA-based methods performance well in keeping the spectral information of the MS images but may have limited spatial enhancements. And the performances of the CS-based and MRA-based methods show data uncertainty, i.e., they have different fusion performances on different scenarios. The learning-based methods, especially the CNN-based methods, perform well in reducing spatial and spectral distortions due to their powerful feature extraction capabilities and data-driven training scheme. However, they usually need an extremely large data set to train the model parameters and are difficult to be interpretable.
Is there a way to make full use of the complementary performances of the CS-based and MRA-based methods at the same time reducing their data uncertainty? A straightforward idea is to firstly generate multiple fusion results by multiple methods (i.e., the CS-based and MRA-based methods), and then automatically combine them with weights based on performances within different scenarios to boost the fusion result. This may be realized by using a trainable CNN since it is data-driven and has strong abilities in the field of image processing.
Motivated by the above, we propose a novel model average method, referred to as pansharpening weight netowrk (PWNet), for the pansharpening. Specifically, rather than generating only one estimated HRMS image at a time, we use multiple inference modules to generate distinct estimated HRMS images at a time. Each inference module produces a distinct estimate of HRMS with bias, and multiple estimates have the positive and negative deviations. And then the biases can be complemented by averaging the multiple results, thus leading to the distortions of average are smaller than that of a single estimate. In order to make use of the simplicity and complementary characteristics of the CS-based and MRA-based methods, we choose them as inference modules, i.e., use each CS-based or MRA-based method as an inference module, and then design an end-to-end trainable network with the original MS and PAN images as input to simultaneously obtain weight maps for all fusion results obtained by the CS and MRA inference modules. Based on the powerful capability and data-driven training scheme of nerual network, the output weight maps are context and method dependent. Finally, we get an estimated HRMS image through adaptively averaging all the fusion results obtained by the CS-based and MRA-based methods. Figure 3 depicts the main procedures of the proposed PWNet for pansharpening.

3.2. Network Architecture

Pansharpening Weight Network (PWNet): The key of model average is to assign proper weights to different models depending on their performances. Unlike traditional model average methods that are based on hand-crafted design to assign weights, we resort the neural network to adaptively generate weights. The proposed PWNet is composed of two subnet: the CS weight network and the MRA weight network. For each subnet, the original LRMS and/or PAN images are took as input. Similar to [33], high-pass filter operation is first implemented in order to preserve edges and details. For the CS weight subnet, the high-pass filtered MS image is upsampled though a transpose convolution and concatenated with the high-pass filtered PAN image, and then fed into four residual blocks [44]. Different from this, the MRA weight subnet only passes the high-pass filtered PAN image into the subsequent residual blocks. Each of the residual blocks consists of three convolutional layers with each having a learnable filter of size 3 × 3 and a rectified-linear unite (ReLU) activation function [45]. To generate weight maps, the activation function of the output layer is set to be a softmax function. By considering computation efficiency and also for keeping the proportions between each pair of the MS bands unchanged, the PWNet only outputs one weigh map for each CS-based or MRA-based method to achieve a pixel-wise aggregation, which will be discussed below. The detailed architecture of PWNet for generating the weight maps is shown in Figure 4.
The CS-based Result Average: Suppose we have n C S kinds of CS-based methods for our PWNet, and then, for the ith CS-based method, we have its estimated HRMS image as
[ M S ^ C S i 1 , M S ^ C S i 2 , , M S ^ C S i N ] = C S i ( P , M S ˜ )
i = 1 , 2 , , n C S
where M S ^ C S i k , k = 1 , 2 , , N , is the kth MS band and C S i ( · ) denotes the ith CS-based method. According to the CS weight network, we get adaptive weight maps for the used CS-based methods as:
W C S 1 , W C S 2 , , W C S n C S = f ( G ( P ) , G ( M S ) ; θ C S )
where W C S i is the ith output weight map generated by the CS weight network, f ( G ( P ) , G ( M S ) ; θ C S ) denotes the CS weight network, θ C S is the parameter set, the function G ( · ) is the high-pass filter. At last, we conduct pixel-wise multiplication for each estimated HRMS image and its corresponding weight map W C S i and sum the multiplication results to get the CS-based result average, i.e., for each band, we have the averaged result as
M S ^ C S k = i = 1 n C S M S ^ C S i k W C S i ,
k = 1 , . . . , N ,
where M S ^ C S k denotes the result of the kth band of CS-based method module and ⊙ denotes the point-wise multiplication.
The MRA-based Result Average: Consistent with the procedures of the CS-based method module, the averaged results for the n M R A MRA-based methods can be given as
M S ^ M R A k = i = 1 n M R A M S ^ M R A i k W M R A i
k = 1 , . . . , N ,
where M S ^ M R A i k and W M R A i are the kth HRMS band obtained by the ith MRA-based method and the ith output weight map generated by the MRA weight network, which respectively are given as
W M R A 1 , W M R A 2 , , W M R A n M R A = g ( G ( P ) ; θ M R A )
and
M S ^ M R A i 1 , M S ^ M R A i 2 , , M S ^ M R A i N = M R A i ( P , M S ˜ ) , i = 1 , 2 , , n M R A
where g ( G ( P ) ; θ M R A ) denotes the MRA weight network, θ M R A is the parameter set, and M R A i ( · ) is the ith MRA-based method. Note that, different from the CS-based weight network, we only take the PAN image as the input of the MRA-based weight network since the MRA-base methods extract the spatial details only rely on the PAN image.
The Final Result Aggregation: After we have obtained the averaged results of the CS-based method module, M S ^ C S = [ M S ^ C S 1 , , M S ^ C S N ] , and the averaged results of the MRA-based method module, M S ^ M R A = [ M S ^ M R A 1 , , M S ^ M R A N ] , then we can aggregate them for compensating their spatial and/or spectral distortions. The final estimated HRMS image M S ^ can be given as
M S ^ = M S ^ C S + α M S ^ M R A
where α is a factor for balancing the contributions between the CS-based methods and the MRA-based methods.
The Loss function: To learn the model parameters θ = { θ C S , θ M R A } , we would like to minimize the reconstruction errors between the estimated HRMS image, M S ^ , and its corresponding ideal one, Y, i.e.,
min θ 1 n i = 1 n | | M S ^ i Y i | | F 2
where n is the number of training samples.

4. Experiments

4.1. Data Sets and Implementation Details

We conduct several experiments using three data sets respectively collected by the GeoEye-1, WorldView-2, and QuickBird satellites. Each data set is split into two nonoverlapping subset: a training data set and a testing data set. Each sample in the data set consists of an MS and PAN image pairs with the PAN image of size 64 × 64 and the MS image of size 16 × 16 . In order to verify the generalization of the proposed method, we also perform pansharpening on a scene taken on other days for QuickBird satellite at the full resolution experiment. More detailed information of the three data sets is reported in Table 1.
As for the training of our proposed PWNet, due to the unavailable of the ideal HRMS images, similar to the other CNN-based pansharpening methods [32,33,34], we first follow the Wald’s protocol [46] to generate the training input MS and PAN pairs by downsampling both the original PAN and MS images with scale factor r = 4 (i.e., the resolution of the MS and PAN images is reduced by applying the MTF-matched low-pass filters [21]), and then the the original MS images are treated as target outputs.
The PWNet is implemented in Tensorflow and trained on a Intel(R) Core(TM) i5-4210U CPU. We use the Adam algorithm [47] with an initial learning rate of 0.001 to optimize the network parameters. And we set the maximum number of epoch to 1000 and mini-batch sample size to 32. It takes about 1 h to train our network.
We first evaluate the methods at a reduced resolution, in addition to visual analysis on the experimental results, the proposed PWNet and other compared methods are also evaluated by five widely used quantitative metrics, namely, universal image quality index [48] averaged over the bands (Q_avg) and its four band extension, Q4 [3], spectral angle mapper (SAM) [49], Erreur Relative Globale Adimensionnelle de Synth e ´ se (ERAGS) [50] and the spatial correlation coefficient (SCC) [46]. The closer to one the Q_avg, Q4, and SCC, the better the quality of fused results, while the lower the SAM and ERGAS, the better the fusion quality.
We also evaluate the methods at a full resolution. In this case, the quality with no-reference index (QNR) [51] and its spatial index ( D S ) and spectral index ( D λ ) are employed for the quantitative assessment. It should be pointed out that the quantitative assessment at full resolution is challenging since these indexes (i.e., QNR, D S and D λ ) are not computed with unattainable ground truth, but rely heavily on the original MS and PAN images [43]. This tends to quantify the similarity of certain components in the fused images to the low-resolution observations, which will lead biases in these indexes estimation. Due to this reason, some methods can generate images with high QNR values but poor image qualities [52].
In the following, we have carried out five sets of experiments to perform comprehensive analysis on the proposed PWNet, typically the effect of the hyperparameter α , the number of CS-based and MRA-based methods, the weight maps channels, and the quantitative, visual and running time comparisons with the CS-based, MRA-based and learning-based methods at reduced resolution and full resolution.

4.2. Analysis to the Hyper-Parameters α

There is a hyper-parameter α in our proposed PWNet, which is to balance the contributions of the CS-based and MRA-based methods. In this experiment, we will analyze this parameter to optimize the performance of PWNet. We fix the number of CS-based and MRA-based methods to six and change α from 0.1 to 1 with interval 0.1. The results obtained are shown in Table 2. As we can see from it, the PWNet attains constantly good performances when α varies from 0.7 to 0.9. Specially, the best results can be obtained for α = 0.7 . It is worth noting that when α goes to 1, the quantitative indexes seem to become worse. Thus, α = 0.7 can be a relatively good choice in the following experiments.

4.3. Impact of the Number of the CS-Based and MRA-Based Methods

This experiment shows what influences would be produced by the different number of the CS-based and MRA-based methods under the condition of n C S = n M R A for simplicity and also for keeping the balance between the CS class and the MRA class methods. The number of methods to be averaged is very important to balance the spectral and spatial information from the LRMS and high resolution PAN images. Too few methods might extract features incompletely and result in a poor performance, while too many would suffer from computational burden during testing. We would like to find a trade-off value according to the performance and running time. Therefore, we limit the range of n C S (or n M R A ) to between 5 and 8.
Table 3 gives the quantitative results and the running time (in second) when the number of averaged methods varies from 5 to 8. Note that, n M R A and n C S equal 5, which means that the selected CS-based are PCA [11], GIHS [10], Brovey [12], GS [13], GSA [15] and the selected MRA-based methods are HPF [18], SFIM [19], AWLP [24], MTF-GLP-HPM [22], MTF-GLP-CBD [38], respectively. When both n C S and n M R A are equal to 6, the PRACS [17] and the ATWT-M3 [39] methods are added into the CS-based and MRA-based methods, respectively. When n M R A and n C S are equal to 7, we add the BDSD [14] into the CS weight network and the Indusion [23] into the MRA weight network, and the BDSD-PC [16] and the MTF-GLP [21] are added into the CS weight network and the MRA weight network as method modules when n M R A and n C S are equal to 8. As reported in Table 3, the performance of the proposed PWNet is improved when the number of averaged methods increases from 5 to 7, while it decays when n C S and n M R A are equal to 8. This reveals that the use of less number of averaged methods will reduce the performances of the PWNet and increasing the number of averaged methods will not continuously bring improvements but will suffer from more computation. In principle, our proposed PWNet is data-driven and thus can automatically weight different kinds of CS-based and/or MRA-based methods. In practice, we suggest two criteria for the selection of CS-based and MRA-based methods for the proposed PWNet. First, the number of CS-based methods and the number of MRA-based methods should be equal in order to keep the balance contribution of the CS class and MRA class. And then, to further improve the performances and robustness of the PWNet, we suggest selecting the CS-based and MRA-based methods according to their performances reported in [3].

4.4. Impact on the Number of Weight Map Channels

In order to reduce the number of parameters and the computational cost, we set the weight map for each CS-based or MRA-based method to be one channel. That is, each MS band of a HRMS image obtained by a CS-based or MRA-based method share the same weight map. In general, the model capacity will be increased with the number of model parameters. Thus, we conduct the experiments based on different output weight map channels to verify whether the capacity of our model has suffered from the reduction of channels.
The results of PWNet with one shared weight map channel and with four different weight map channels for each CS-based or MRA-based method are reported in Table 4. As we can see from it, the PWNet with one shared weight map channel attains constantly good performances in terms of the five commonly used metrics on three different kinds of satellite data sets, while the PWNet with four different weight map channels for each method has a relatively poor performances with the same training conditions. This may due to that an under-fitting phenomenon caused by excessive parameters has happened in the PWNet with four weight map channels. It further verifies the advantages of the PWNet with one shared weight map channel, which can lower the training difficulty.

4.5. Comparison with the CS-Based and MRA-Based Methods

One key issue of the proposed PWNet method is whether its fusion result is better than that of each participating method. Only when the answer to this question is yes, we can claim that the proposed PWNet can produce appropriate weight maps for each CS-based or MRA-based method, so that the methods involved in model average can complement each other and the result can be improved. Here we set n M R A and n C S to 7 as we have proved in above that this setting has a good balance for performances and running times. The compared pansharpening methods include seven methods belonging to the CS class, namely, PCA [11], GIHS [10], Brovey [12], BDSD [14], GS [13], GSA [15], PRACS [17], and seven methods belonging to the MRA class such as HPF [18], SFIM [19], Indusion [23], AWLP [24], ATWT-M3 [39], MTF-GLP-HPM [22], MTF-GLP-CBD [38]. All methods follow the experimental settings recommended by the authors.
We first inspect the visual quality of the pansharpening results. Figure 5, Figure 6 and Figure 7 present the pansharpened images on all the three data sets, obtained by our proposed PWNet and the other fourteen methods. As we can see from theses figures, the CS-based methods produce a relatively sharper spatial features in Figure 5a–g, but they suffer from spectral distortions, as highlighted in the small window, where the trees around buildings in Figure 5 and the bare soil in Figure 6 are a little darker than that of ground truth. In contrast, less spectral distortions are appeared in the results of the MRA-based methods, however, they show poor spatial rendering as they present a little blurring in Figure 6h–n, especially for the results of AWLP and ATWT-M3. Compared with the CS-based and MRA-based methods, the proposed PWNet can achieve more similar results to the ground truth. From the enlarged area in the upper left corner of the Figure 7o, we can see that PWNet has the best performance in both improving the spatial details and keeping spectral fidelity of the roads and trees. In summary of the visual analysis, the proposed PWNet method can debias the spectral and spatial distortions in the CS-based and the MRA-based methods, and can effectively combine the advantages of these two types of methods, thus shows better visual performances.
Besides visual inspection, we apply numeric metrics to assess the quality of pansharpened images. Table 5, Table 6 and Table 7 report the comparison results of the CS-based methods, MRA-based methods, and the proposed PWNet method on the three data sets. As we can find from these tables, for the WordView-2 and GeoEye-1 data sets, the BDSD method shows the best performances among the fourteen traditional methods, the AWLP achieves the best performance among the CS-based and MRA-based methods for the QuickBird data set. None of the CS-based and MRA-based methods systematically obtain the best performances for all the three data sets. The proposed PWNet yields results with the best spatial and spectral accuracy over the CS-based and MRA-based methods on all the three data sets. This proves once again that the proposed method can combine the advantages of the two types of methods to produce an optimal result.
To be interpretable, we also visualize the some weight maps of the selected traditional methods used in the PWNet, as shown in Figure 8, Figure 9 and Figure 10. It can be seen from these figures that, the edges of the road and the buildings are extracted by the weight maps of both the CS-based and MRA-based methods. Typically, we can see that, for the WorldView-2 and GeoEye-1 data sets, the BDSD method plays an important role as its weight map is clearer than any others, as showin in Figure 8b and Figure 9b, while the AWLP and MTF-GLP-CBD methods show a little greater contribution to the averaged results of the PWNet for the QuickBird data set, as can be see from Figure 10d,e. As for the PCA method, the weight maps are all black, which means that PCA method almost makes no contribution to the final result on all the three tested data sets. This conclusion is consistent with the previous visual inspection in Figure 5, Figure 6 and Figure 7 and quantitative results reported in Table 5, Table 6 and Table 7. This proves the adaptive characteristic of our PWNet as it considers different performance of the selected CS-based and MRA-based methods. From these experimental results, we can conclude that the proposed PWNet are adaptive and robust to different data sets.

4.6. Comparison with the CNN-Based Methods

Currently, the proposed PWNet has shown its priority over the selected traditional CS-based and MRA-based methods. In this subsection, we are going to compare it with the CNN-based methods to verify its effectiveness. The other four stat-of-the-art (SOTA) methods including pansharpening by convolutional neural networks (PNN) [32], deep residual pan-sharpening neural network (DRPNN) [42], multiscale and multidepth convolutional neural network (MSDCNN) [43], are used as alternative methods for comparison. All the compared methods follow the experimental setting of their original papers. Note that the source codes of PNN is provided by the original authors and the codes of DRPNN, MSDCNN are available at https://github.com/Decri.
Figure 11, Figure 12 and Figure 13 show some example regions selected from the pansharpened images on the three test data sets. In Figure 11, by magnifying the selected area in the image three times, it can be obviously seen that the other CNN-based methods have a little blurring to the ground truth, while the edges produced by our proposed PWNet method are more clear and natural as shown in zoomed areas. Although MSDCNN, DRPNN and PNN can produce better results with less spatial distortions, they sometimes suffer from a little spectral distortions, as shown in Figure 12b–d, where the bare soil is a little darker than the reference. This can also be seen in Figure 13b where the buildings in this scene is dark yellow while they are white in the ground truth as shown in Figure 13f. Compared with other CNN-based methods, the proposed PWNet shows a good balance between the injected spatial details and the maintain of original spectral information, this is clearly visible on the vegetable areas and textures (e.g., edges of the roof and road), as shown in Figure 11e, Figure 12e and Figure 13e.
In addition, Table 8 shows the quantitative results for the three tested data sets obtained by the compared CNN-based methods and our proposed PWNet. It should be pointed out that, for each test experiment, we would choose one test sample randomly from the test data set rather than a cherry-picked sample, thus the results listed in Table 8 and Table 5, Table 6 and Table 7 are based on different PAN and MS image pairs and have different quantitative results. For better comparison, the best results among the four methods are highlighted in boldface. According to this table, one can see that performances of the proposed PWNet is better than the other three CNN-based methods in terms of the five indexes.

4.7. Comparison at Full Resolution

The comparison results on three tested images at full resolution are shown in Figure 14, Figure 15 and Figure 16 and Table 9. As we can see from the table that, for the WordView-2 and GeoEye-1 data sets, the DRPNN and PNN method respectively show the best performances, while the proposed PWNet holds the second best position for all the three data sets. On a whole, the CNN-based methods perform better than the traditional methods (i.e., the CS-based and MRA-based methods). By a visual inspection, the PNN, DRPNN, and MSDCNN methods tend to produce blurring results while the proposed PWNet is able to enhancing the spatial quality and shows clearly sharper fusion results, as shown in Figure 14, Figure 15 and Figure 16r. As a summary, compared to the other methods at the full resolution, the proposed PWNet could consistently reconstruct sharper HRMS image with less spectral and spatial distortion.

4.8. Running Time Analysis

In this subsection, we compare the running time of the proposed method with the others on a 64 × 64 LRMS and 256 × 256 PAN image pair. The experiments are performed by MATLAB R2016b on the same platform with Core i5-4210U/1.7 GHz/4G. The running times of different methods are listed in Table 10, in which the time is measured in second. From this table, it can be found that DRPNN is the most time-consuming method, because the number of hidden layers within DRPNN is more than the other CNN-based methods. In addition, the MSDCNN needs a little more time to obtain the fusion result than that of the proposed PWNet. In a word, the proposed PWNet method is more efficient than the CNN-based methods due to less hidden layers and that it only outputs weight maps rather than directly producing an estimated HRMS image.

5. Conclusions

In this paper, we presented a novel model average network for pansharpening, and is referred to as PWNet. The proposed PWNet attempts to integrate the complementary characteristics of the CS-based and MRA-based methods through an end-to-end trainable neural networks, and thus it is data-driven and able to adaptively weight the results of the classical methods depending on their performances. Experiments on several data sets collected by three kinds of satellites demonstrate that the pansharpened HRMS images by the proposed PWNet can not only enhance the spatial qualities but also can keep the spectral information of the original MS images. In addition, the proposed PWNet has some distribution structures. Thus, we will extend the proposed model to a distribution version by using the techniques of distributed processing [53] to further reduce the running time while maintaining the quality of the results.

Author Contributions

J.L. and C.Z. (Chunxia Zhang) structured and drafted the manuscript with assistance from C.Z. (Changsheng Zhou); Y.F. conceived the experiments and generated the graphics with assistance from C.Z. (Changsheng Zhou). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China under Grant 2018AAA0102201, and by the National Natural Science Foundation of China grant number 61877049.

Acknowledgments

The authors would like to thank Gemine Vivone for sharing the pansharpening Matlab toolbox [3], to Giuseppe Scarpa for providing their codes of PNN in [32], to Wei and Yuan in [42,43] for sharing their CNN-based codes for pansharpening.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shaw, G.; Burke, H.K. Spectral imaging for remote sensing. Linc. Lab. J. 2003, 14, 3–28. [Google Scholar]
  2. Meng, X.; Shen, H.; Li, H.; Zhang, L.; Fu, R. Review of the pansharpening methods for remote sensing images based on the idea of meta-analysis: Practical discussion and challenges. Inf. Fusion 2019, 46, 102–113. [Google Scholar] [CrossRef]
  3. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A critical comparison among pansharpenig algorithms. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2565–2586. [Google Scholar] [CrossRef]
  4. Ehlers, M.; Klonus, S.; Astrand, P.; Rosso, P. Multi-sensor image fusion for pansharpening in remote sensing. Int. J. Image Data Fusion 2010, 1, 25–45. [Google Scholar] [CrossRef]
  5. Yue, L.; Shen, H.; Li, J.; Yuan, Q.; Zhang, H.; Zhang, L. Image super-resolution: The techniques, applications, and future. Signal Process. 2016, 128, 389–408. [Google Scholar] [CrossRef]
  6. Souza, C.; Firestone, L.; Silva, M.; Roberts, D. Mapping forest degradation in the Eastern Amazon from SPOT 4 through spectral mixture models. Remote Sens. Environ. 2003, 87, 494–506. [Google Scholar] [CrossRef]
  7. Mohammadzadeh, A.; Tavakoli, A.; Valadan Zoej, M.J. Synthesis of multispectral images to high spatial resolution: Road extraction based on fuzzy logic and mathematical morphology from pansharpened IKONOS images. Photogramm. Rec. 2006, 21, 44–60. [Google Scholar] [CrossRef]
  8. Loncan, L.; De Almeida, L.B.; Bioucas-Dias, J.M.; Briottet, X.; Chanussot, J.; Dobigeon, N.; Fabre, S.; Liao, W.; Licciardi, G.A.; Simoes, M.; et al. Hyperspectral pansharpening: A review. IEEE Geosci. Remote Sens. Mag. 2015, 3, 27–46. [Google Scholar] [CrossRef] [Green Version]
  9. Carper, W.; Lillesand, T.; Kiefer, R. Synthesis of multispectral images to high spatial resolution: The use of intensity-huesaturation transformations for merging SPOT panchromatic and multispectral image data. Photogramm. Eng. Remote Sens. 1990, 56, 459–467. [Google Scholar]
  10. Tu, T.; Su, S.C.; Shyu, H.C.; Huang, P.S. A new look at IHS-like image fusion methods. Inf. Fusion 2001, 2, 177–186. [Google Scholar] [CrossRef]
  11. Chavez, P.S.; Kwarteng, A.W. Extracting spectral contrast in Landsat thematic mapper image data using selective principal component analysis. Photogramm. Eng. Remote Sens. 1989, 55, 339–348. [Google Scholar]
  12. Gillespie, A.; Kahle, A.B.; Walker, R.E. Color enhancement of highly correlated images-II. Channel ration and “Chromaticity” Transform techniques. Remote Sens. Environ. 1987, 22, 343–365. [Google Scholar] [CrossRef]
  13. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6011875, 4 January 2000. [Google Scholar]
  14. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE pan sharpening of very high resolution multispectral images. IEEE Trans. Geosci. Remote Sens. 2008, 46, 228–236. [Google Scholar] [CrossRef]
  15. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS+Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  16. Vivone, G. Robust band-dependent spatial-detail approaches for panchromatic sharpening. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6421–6433. [Google Scholar] [CrossRef]
  17. Choi, J.; Yu, K.; Kim, Y. Synthesis of multispectral images to high spatial resolution: A new adaptive component-substitution based satellite image fusion by using partial replacement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 295–309. [Google Scholar] [CrossRef]
  18. Chavez, P.S., Jr.; Sides, S.C.; Anderson, A. Comparison of three different methods to merge multiresolution and multispectral data: Landsat TM and SPOT panchromatic. Photogramm. Eng. Remote Sens. 1991, 57, 295–303. [Google Scholar]
  19. Liu, J. Smoothing filter-based intensity modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  20. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. An MTF based spectral distortion minimizing model for pan-sharpening of very high resolution multispectral images of urban areas. In Proceedings of the 2nd GRSS/ ISPRS Joint Workshop Remote Sensing and Data Fusion URBAN Areas, Berlin, Germany, 22–23 May 2003; pp. 90–94. [Google Scholar]
  21. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. MTF tailored multiscale fusion of high-resolution MS and Pan imagery. Photogramm. Eng. Remote Sens. 2006, 72, 591–596. [Google Scholar] [CrossRef]
  22. Vivone, G.; Restaino, R.; Dalla Mura, M.; Licciardi, G.; Chanussot, J. Contrast and error-based fusion schemes for multispectral image pansharpening. IIEEE Trans. Geosci. Remote Sens. 2014, 11, 930–934. [Google Scholar] [CrossRef] [Green Version]
  23. Khan, M.M.; Chanussot, J.; Condat, L.; Montavert, A. Indusion: Fusion of multispectral and panchromatic images using the induction scaling technique. IEEE Trans. Geosci. Remote Sens. 2008, 5, 98–102. [Google Scholar] [CrossRef] [Green Version]
  24. Otazu, X.; González-Audìcana, M.; Fors, O.; Núñez, J. Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef] [Green Version]
  25. Shah, V.P.; Younan, N.H.; King, R.L. An efficient pan-sharpening method via a combined adaptive PCA approach and contourlets. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1323–1335. [Google Scholar] [CrossRef]
  26. Liao, W.; Huang, X.; Van Coillie, F.; Gautama, S.; Pižurica, A.; Philips, W.; Liu, H.; Zhu, T.; Shimoni, M.; Moser, G.; et al. Processing of Multiresolution Thermal Hyperspectral and Digital Color Data: Outcome of the 2014 IEEE GRSS Data Fusion Contest. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 2984–2996. [Google Scholar] [CrossRef]
  27. Ballester, C.; Vselles, V.; Igual, L.; Verdera, J.; Rouge, B. A variational model for P+XS image fusion. Int. J. Comput. Vis. 2006, 69, 43–58. [Google Scholar]
  28. Moller, M.; Wittman, T.; Bertozzi, A.; Berger, M. A variational approach for sharpening high dimensional images. SIAM J. Imaging Sci. 2012, 5, 150–178. [Google Scholar] [CrossRef] [Green Version]
  29. Zhu, X.; Bamler, R. A sparse image fusion algorithm with application to pansharpening. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2827–2836. [Google Scholar] [CrossRef]
  30. Deng, L.; Vivone, G.; Guo, W.; Mura, M.; Chanussot, J. A variational pansharpening approach based on reproducible kernel hilbert space and heaviside function. IEEE Trans. Image Process. 2018, 27, 4330–4344. [Google Scholar] [CrossRef]
  31. Palsson, F.; Ulfarsson, M.; Sveinsson, J. Model-based reduced-rank pansharpening. IEEE Geosci. Remote Sens. Lett. 2020, 17, 656–660. [Google Scholar] [CrossRef]
  32. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by convolutional neural networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef] [Green Version]
  33. Yang, J.; Fu, X.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A deep network architecture for pan-sharpening. In Proceedings of the International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5449–5457. [Google Scholar]
  34. Liu, X.; Wang, Y.; Liu, Q. Synthesis of multispectral images to high spatial resolution: Remote sensing image fusion based on two-stream fusion network. IEEE Trans. Geosci. Remote Sens. 2018, 46, 428–439. [Google Scholar]
  35. Zheng, Y.; Li, J.; Li, Y.; Guo, J.; Wu, X.; Chanussot, J. Hyperspectral Pansharpening Using Deep Prior and Dual Attention Residual Network. IEEE Trans. Geosci. Remote Sens. 2020, 10, 1–18. [Google Scholar] [CrossRef]
  36. Han, S.; Pool, J.; Tran, J.; Dally, W. Learning both weights and connections for efficient neural network. In Proceedings of the Advances in Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; pp. 1135–1143. [Google Scholar]
  37. Ghamisi, P.; Rasti, B.; Yokoya, N.; Wang, Q.; Hofle, B.; Bruzzone, L.; Bovolo, F.; Chi, M.; Anders, K.; Gloaguen, R.; et al. Multisource and multitemporal data fusion in remote sensing: A comprehensive review of the state of the art. IEEE Geosci. Remote Sens. Mag. 2019, 7, 6–39. [Google Scholar] [CrossRef] [Green Version]
  38. Alparone, L.; Wald, L.; Chanussot, J.; Thomas, C.; Gamba, P.; Bruce, L. Comparison of pansharpening algorithms: Outcome of the 2006 GRS-S DataFusion Contest. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3012–3021. [Google Scholar] [CrossRef] [Green Version]
  39. Ranchin, T.; Wald, L. Fusion of high spatial and spectral resolution images: The ARSIS concept and its implementation. Photogramm. Eng. Remote Sens. 2000, 66, 49–61. [Google Scholar]
  40. Restaino, R.; Vivone, G.; Dalla Mura, M.; Chanussot, J. Fusion of multispectral and panchromatic images based on morphological operators. IEEE Trans. Image Process. 2016, 25, 2882–2895. [Google Scholar] [CrossRef] [Green Version]
  41. Zhong, J.; Yang, B.; Huang, G. Remote sensing image fusion with convolutional neural network. Sens. Imaging 2016, 17, 10. [Google Scholar] [CrossRef]
  42. Wei, Y.; Yuan, Q.; Shen, H.; Zhang, L. Boosting the Accuracy of Multispectral Image Pansharpening by Learning a Deep Residual Network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1795–1799. [Google Scholar] [CrossRef] [Green Version]
  43. Yuan, Q.; Wei, Y.; Meng, X.; Shen, H.; Zhang, L. A Multiscale and Multidepth Convolutional Neural Network for Remote Sensing Imagery Pan-Sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 3, 978–989. [Google Scholar] [CrossRef] [Green Version]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  45. Nair, N.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  46. Wald, L.; Ranchin, T.; Mangolini, M. Data fusion: Definitions and architectures: Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  47. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  48. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  49. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the Spectral Angle Mapper (sam) algorithm. In Proceedings of the Summaries of the Third Annual JPL Airborne Geoscience Workshop, Pasadena, CA, USA, 1–5 June 1992; pp. 147–149. [Google Scholar]
  50. Zhou, J.; Civco, D.; Silander, J. Chanussot, A wavelet transform method to merge landsat tm and spot panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  51. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef] [Green Version]
  52. Qu, Y.; Baghbaderani, R.K.; Qi, H.; Kwan, C. Unsupervised Pansharpening Based on Self-Attention Mechanism. arXiv 2020, arXiv:2006.09303v1. [Google Scholar] [CrossRef]
  53. Silva, G.; Medeiros, R.; Jaimes, B.; Takahashi, C.C.; Vieira, D.; Braga, A. CUDA-Based Parallelization of Power Iteration Clustering for Large Datasets. IEEE Access 2017, 5, 27263–27271. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the CS-based methods for pansharpening.
Figure 1. Flowchart of the CS-based methods for pansharpening.
Remotesensing 12 02804 g001
Figure 2. Flowchart of the MRA-based methods for pansharpening.
Figure 2. Flowchart of the MRA-based methods for pansharpening.
Remotesensing 12 02804 g002
Figure 3. Blockdiagram of the proposed PWNet for pansharpening. The proposed method can be divided into three independent part, i.e., the weight maps network, the CS inference modules and the MRA inference modules. The weight maps network takes the reduced MS and PAN image as inputs, and outputs weight maps for each of the CS-based and MRA-based methods. At the same time, the CS inference modules and the MRA inference modules generate HRMS images according to specific pansharpening methods. Finally, the estimated HRMS image is obtained by averaging all HRMS images estimated by the selected CS-based and MRA-based methods with the weight maps generated by the weight maps network.
Figure 3. Blockdiagram of the proposed PWNet for pansharpening. The proposed method can be divided into three independent part, i.e., the weight maps network, the CS inference modules and the MRA inference modules. The weight maps network takes the reduced MS and PAN image as inputs, and outputs weight maps for each of the CS-based and MRA-based methods. At the same time, the CS inference modules and the MRA inference modules generate HRMS images according to specific pansharpening methods. Finally, the estimated HRMS image is obtained by averaging all HRMS images estimated by the selected CS-based and MRA-based methods with the weight maps generated by the weight maps network.
Remotesensing 12 02804 g003
Figure 4. The architecture of proposed PWNet for generating the weight maps.
Figure 4. The architecture of proposed PWNet for generating the weight maps.
Remotesensing 12 02804 g004
Figure 5. Visual comparison of the CS-based and MRA-based methods and the proposed PWNet method on the WorldView-2 images, (a) PAN; (b) LRMS; (c) PCA; (d) GIHS; (e) Brovey; (f) BDSD; (g) GS; (h) GSA; (i) PRACS; (j) HPF; (k) SFIM; (l) Indusion; (m) AWLP; (n) ATWT-M3; (o) MTF-GLP-HPM; (p) MTF-GLP-CBD; (q) PWNet (ours); (r) Ground Truth.
Figure 5. Visual comparison of the CS-based and MRA-based methods and the proposed PWNet method on the WorldView-2 images, (a) PAN; (b) LRMS; (c) PCA; (d) GIHS; (e) Brovey; (f) BDSD; (g) GS; (h) GSA; (i) PRACS; (j) HPF; (k) SFIM; (l) Indusion; (m) AWLP; (n) ATWT-M3; (o) MTF-GLP-HPM; (p) MTF-GLP-CBD; (q) PWNet (ours); (r) Ground Truth.
Remotesensing 12 02804 g005
Figure 6. Visual comparison of the CS-based and MRA-based methods and the proposed PWNet method on the GeoEye-1 images, (a) PAN; (b) LRMS; (c) PCA; (d) GIHS; (e) Brovey; (f) BDSD; (g) GS; (h) GSA; (i) PRACS; (j) HPF; (k) SFIM; (l) Indusion; (m) AWLP; (n) ATWT-M3; (o) MTF-GLP-HPM; (p) MTF-GLP-CBD; (q) PWNet (ours); (r) Ground Truth.
Figure 6. Visual comparison of the CS-based and MRA-based methods and the proposed PWNet method on the GeoEye-1 images, (a) PAN; (b) LRMS; (c) PCA; (d) GIHS; (e) Brovey; (f) BDSD; (g) GS; (h) GSA; (i) PRACS; (j) HPF; (k) SFIM; (l) Indusion; (m) AWLP; (n) ATWT-M3; (o) MTF-GLP-HPM; (p) MTF-GLP-CBD; (q) PWNet (ours); (r) Ground Truth.
Remotesensing 12 02804 g006
Figure 7. Visual comparison of the CS-based and MRA-based methods and the proposed PWNet method on the QuickBird images, (a) PAN; (b) LRMS; (c) PCA; (d) GIHS; (e) Brovey; (f) BDSD; (g) GS; (h) GSA; (i) PRACS; (j) HPF; (k) SFIM; (l) Indusion; (m) AWLP; (n) ATWT-M3; (o) MTF-GLP-HPM; (p) MTF-GLP-CBD; (q) PWNet (ours); (r) Ground Truth.
Figure 7. Visual comparison of the CS-based and MRA-based methods and the proposed PWNet method on the QuickBird images, (a) PAN; (b) LRMS; (c) PCA; (d) GIHS; (e) Brovey; (f) BDSD; (g) GS; (h) GSA; (i) PRACS; (j) HPF; (k) SFIM; (l) Indusion; (m) AWLP; (n) ATWT-M3; (o) MTF-GLP-HPM; (p) MTF-GLP-CBD; (q) PWNet (ours); (r) Ground Truth.
Remotesensing 12 02804 g007
Figure 8. Visualization of weight maps on the WorldView-2 images, (a) PCA; (b) BDSD; (c) PRACS; (d) AWLP; (e) MTF-GLP-HPM; (f) MTF-GLP-CBD.
Figure 8. Visualization of weight maps on the WorldView-2 images, (a) PCA; (b) BDSD; (c) PRACS; (d) AWLP; (e) MTF-GLP-HPM; (f) MTF-GLP-CBD.
Remotesensing 12 02804 g008
Figure 9. Visualization of weight maps on the GeoEye-1 images, (a) PCA; (b) BDSD; (c) PRACS; (d) AWLP; (e) MTF-GLP-HPM; (f) MTF-GLP-CBD.
Figure 9. Visualization of weight maps on the GeoEye-1 images, (a) PCA; (b) BDSD; (c) PRACS; (d) AWLP; (e) MTF-GLP-HPM; (f) MTF-GLP-CBD.
Remotesensing 12 02804 g009
Figure 10. Visualization of weight maps on the QuickBird images, (a) PCA; (b) BDSD; (c) PRACS; (d) AWLP; (e) MTF-GLP-HPM; (f) MTF-GLP-CBD.
Figure 10. Visualization of weight maps on the QuickBird images, (a) PCA; (b) BDSD; (c) PRACS; (d) AWLP; (e) MTF-GLP-HPM; (f) MTF-GLP-CBD.
Remotesensing 12 02804 g010
Figure 11. Visual comparison of the CNN-based methods on the WorldVie-2 data set, (a) LRMS; (b) PNN; (c) DRPNN; (d) MSDCNN; (e) PWNet (ours); (f) Ground Truth.
Figure 11. Visual comparison of the CNN-based methods on the WorldVie-2 data set, (a) LRMS; (b) PNN; (c) DRPNN; (d) MSDCNN; (e) PWNet (ours); (f) Ground Truth.
Remotesensing 12 02804 g011
Figure 12. Visual comparison of the CNN-based methods on the GeoEye-1 data set, (a) LRMS; (b) PNN; (c) DRPNN; (d) MSDCNN; (e) PWNet (ours); (f) Ground Truth.
Figure 12. Visual comparison of the CNN-based methods on the GeoEye-1 data set, (a) LRMS; (b) PNN; (c) DRPNN; (d) MSDCNN; (e) PWNet (ours); (f) Ground Truth.
Remotesensing 12 02804 g012
Figure 13. Visual comparison of the CNN-based methods on the QuickBird data set, (a) LRMS; (b) PNN; (c) DRPNN; (d) MSDCNN; (e) PWNet(ours); (f) Ground Truth.
Figure 13. Visual comparison of the CNN-based methods on the QuickBird data set, (a) LRMS; (b) PNN; (c) DRPNN; (d) MSDCNN; (e) PWNet(ours); (f) Ground Truth.
Remotesensing 12 02804 g013
Figure 14. Visual comparison of different methods on the WorldVie-2 data set at full resolution, (a) PCA; (b) GIHS; (c) Brovey; (d) BDSD; (e) GS; (f) GSA; (g) PRACS; (h) HPF; (i) SFIM; (j) Indusion; (k) AWLP; (l) ATWT-M3; (m) MTF-GLP-HPM; (n) MTF-GLP-CBD; (o) PNN; (p) DRPNN; (q) MSDCNN; (r) PWNet (ours).
Figure 14. Visual comparison of different methods on the WorldVie-2 data set at full resolution, (a) PCA; (b) GIHS; (c) Brovey; (d) BDSD; (e) GS; (f) GSA; (g) PRACS; (h) HPF; (i) SFIM; (j) Indusion; (k) AWLP; (l) ATWT-M3; (m) MTF-GLP-HPM; (n) MTF-GLP-CBD; (o) PNN; (p) DRPNN; (q) MSDCNN; (r) PWNet (ours).
Remotesensing 12 02804 g014
Figure 15. Visual comparison of different methods on the GeoEye-1 data set at full resolution, (a) PCA; (b) GIHS; (c) Brovey; (d) BDSD; (e) GS; (f) GSA; (g) PRACS; (h) HPF; (i) SFIM; (j) Indusion; (k) AWLP; (l) ATWT-M3; (m) MTF-GLP-HPM; (n) MTF-GLP-CBD; (o) PNN; (p) DRPNN; (q) MSDCNN; (r) PWNet (ours).
Figure 15. Visual comparison of different methods on the GeoEye-1 data set at full resolution, (a) PCA; (b) GIHS; (c) Brovey; (d) BDSD; (e) GS; (f) GSA; (g) PRACS; (h) HPF; (i) SFIM; (j) Indusion; (k) AWLP; (l) ATWT-M3; (m) MTF-GLP-HPM; (n) MTF-GLP-CBD; (o) PNN; (p) DRPNN; (q) MSDCNN; (r) PWNet (ours).
Remotesensing 12 02804 g015
Figure 16. Visual comparison of different methods on the QuickBird data set at full resolution, (a) PCA; (b) GIHS; (c) Brovey; (d) BDSD; (e) GS; (f) GSA; (g) PRACS; (h) HPF; (i) SFIM; (j) Indusion; (k) AWLP; (l) ATWT-M3; (m) MTF-GLP-HPM; (n) MTF-GLP-CBD; (o) PNN; (p) DRPNN; (q) MSDCNN; (r) PWNet (ours).
Figure 16. Visual comparison of different methods on the QuickBird data set at full resolution, (a) PCA; (b) GIHS; (c) Brovey; (d) BDSD; (e) GS; (f) GSA; (g) PRACS; (h) HPF; (i) SFIM; (j) Indusion; (k) AWLP; (l) ATWT-M3; (m) MTF-GLP-HPM; (n) MTF-GLP-CBD; (o) PNN; (p) DRPNN; (q) MSDCNN; (r) PWNet (ours).
Remotesensing 12 02804 g016
Table 1. Information of the three data set.
Table 1. Information of the three data set.
SatelliteResolution of PANResolution of MSNumber of Training Data SetNumber of Test Data Set
GeoEye-10.5 m2.0 m18221
WordView-20.5 m2.0 m22432
QuickBird0.6 m2.4 m52960
Table 2. Quantiative results obtained by PWNet with different α .
Table 2. Quantiative results obtained by PWNet with different α .
α Q_avgSAMERGASSCCQ4
0.10.98073.02502.50260.98310.9836
0.20.98252.99462.31400.98550.9854
0.30.98243.02162.29940.98580.9851
0.40.98312.99012.29470.98610.9856
0.50.98113.21682.30720.98540.9852
0.60.98073.35812.39350.98430.9854
0.70.98343.02762.21910.98660.9863
0.80.98313.04292.25040.98670.9859
0.90.98323.05432.25360.98620.9863
10.94933.27633.96630.95820.9546
Table 3. Quantitative results and running time with different number of the averaged methods.
Table 3. Quantitative results and running time with different number of the averaged methods.
n CS and n MRA Q_avgSAMERGASSCCQ4Time
50.87815.92143.51790.86090.87610.3800
60.90795.71743.06160.90730.90620.5900
70.92095.16542.89860.91310.91650.9200
80.87815.45493.69420.86330.86151.6200
Table 4. Results of the proposed PWNet with one shared weight map channel or four different weight map channels for each CS-based or MRA-based method. The best results are highlighted in bold.
Table 4. Results of the proposed PWNet with one shared weight map channel or four different weight map channels for each CS-based or MRA-based method. The best results are highlighted in bold.
SatelliteNumber of ChannelsIndex
Q_avgSAMERGASSCCQ4
WordView-21 channel0.98742.75822.38300.98130.9805
4 channels0.339014.737914.27550.83780.8007
GeoEye-11 channel0.92074.21202.68680.93840.9117
4 channels0.358519.945114.53000.81410.5585
QuickBird1 channel0.89412.82521.81020.94800.8227
4 channels0.344219.610714.67450.88670.1462
Table 5. Quantitative comparison of the CS-based and MRA-based methods and the proposed PWNet method on the WorldView-2 images. The best and second best results are highlighted in bold.
Table 5. Quantitative comparison of the CS-based and MRA-based methods and the proposed PWNet method on the WorldView-2 images. The best and second best results are highlighted in bold.
MethodQ_avgSAMERGASSCCQ4
CSPCA0.88855.39745.86500.94190.8705
GIHS0.88585.68205.99370.92890.8777
Brovey0.88525.30375.90380.93570.8748
BDSD0.97094.42743.27010.96410.9628
GS0.89245.21955.75360.94280.8817
GSA0.96754.03823.66930.95420.9581
PRACS0.92244.52985.24800.92480.9143
MRAHPF0.94544.15244.41250.95210.9372
SFIM0.94984.20464.20150.95780.9426
Indusion0.86515.31626.46960.90250.8502
AWLP0.96173.72993.65080.95930.9529
ATWT-M30.87236.18986.55920.90670.8696
MTF-GLP-HPM0.96763.97313.52270.96070.9589
MTF-GLP-CBD0.96204.03973.44180.95750.9620
PWNet0.98363.54892.45840.98010.9723
Table 6. Quantitative comparison of the CS-based and MRA-based methods and the proposed PWNet method on the GeoEye-1 images. The best and second best results are highlighted in bold.
Table 6. Quantitative comparison of the CS-based and MRA-based methods and the proposed PWNet method on the GeoEye-1 images. The best and second best results are highlighted in bold.
MethodQ_avgSAMERGASSCCQ4
CSPCA0.81106.34144.87200.85440.8175
GIHS0.81256.31894.85340.85390.8197
Brovey0.80636.51034.95930.84960.8120
BDSD0.92115.35283.81330.87210.9309
GS0.81196.31994.86500.85440.8188
GSA0.89596.31574.04250.85980.9079
PRACS0.82746.14864.72540.84830.8342
MRAHPF0.85626.30384.46500.85960.8638
SFIM0.85966.36884.41520.85990.8655
Indusion0.75187.12865.81350.78490.7579
AWLP0.87236.60964.39180.85480.8835
ATWT-M30.71937.96945.79940.81180.7218
MTF-GLP-HPM0.90166.33403.99310.86390.9108
MTF-GLP-CBD0.91986.32073.91220.86380.9198
PWNet0.95435.32452.84050.93800.9593
Table 7. Quantitative comparison of the the CS-based and MRA-based methods and the proposed PWNet method on the QuickBird images. The best and second best results are highlighted in bold.
Table 7. Quantitative comparison of the the CS-based and MRA-based methods and the proposed PWNet method on the QuickBird images. The best and second best results are highlighted in bold.
MethodQ_avgSAMERGASSCCQ4
CSPCA0.57547.46974.59310.79110.7165
GIHS0.79654.21202.94880.90670.7906
Brovey0.81693.90592.79370.91870.8044
BDSD0.89774.29542.69790.93980.8906
GS0.78564.73823.17000.88740.7838
GSA0.88604.02252.53590.92520.8814
PRACS0.85103.89352.62580.90760.8319
MRAHPF0.88563.88582.41560.92780.8778
SFIM0.89023.80652.36170.93160.8818
Indusion0.71314.39433.56500.86640.7004
AWLP0.90683.57752.21820.94250.8982
ATWT-M30.80594.50732.90600.89820.7912
MTF-GLP-HPM0.90563.75432.33660.93780.8985
MTF-GLP-CBD0.88774.03652.53600.92610.8877
PWNet0.91963.52752.16350.94310.9109
Table 8. Quantitative comparison of the CNN-based methods on three test data sets. The best results are highlighted in bold.
Table 8. Quantitative comparison of the CNN-based methods on three test data sets. The best results are highlighted in bold.
MethodQ_avgSAMERGASSCCQ4
WorldView-2PNN0.94795.21805.10400.89750.9256
DRPNN0.92924.47655.30720.88870.9030
MSDCNN0.94434.41356.84780.94070.8512
PWNet0.98742.75822.38300.98130.9805
GeoEye-1PNN0.86394.81903.52240.88880.8435
DRPNN0.79344.86853.77460.89310.7733
MSDCNN0.84624.97073.31150.90930.8518
PWNet0.92074.21202.68680.93840.9117
QuickBirdPNN0.70473.31132.69800.92580.5950
DRPNN0.76482.42891.75440.94730.7750
MSDCNN0.77272.70341.84290.94500.7845
PWNet0.82582.21631.55250.96440.8234
Table 9. Performance comparison on three test data sets at full resolution. The best and second best results are highlighted in bold and underlined, respectively.
Table 9. Performance comparison on three test data sets at full resolution. The best and second best results are highlighted in bold and underlined, respectively.
GeoEye-1WorldView-2QuickBird
Methed D λ D S QNR D λ D S QNR D λ D S QNR
PCA0.10200.15420.75950.02650.21430.76490.05270.10360.8491
GIHS0.03470.18120.79030.07670.08450.84530.08190.08150.8432
Brovey0.03370.17690.79540.03770.10080.86530.05450.05770.8910
BDSD0.21030.38870.48270.10160.25690.66760.27480.32500.4895
GS0.03790.18160.78730.01670.10490.88020.04430.06130.8971
GSA0.29250.12260.62070.10790.14440.76330.09400.11260.8040
PRACS0.05870.15330.79700.02310.09080.88820.04920.08450.8704
HPF0.15000.20230.67800.05330.10280.84940.06030.05150.8913
SFIM0.14690.19710.68490.05320.09740.85460.05870.05170.8926
Indusion0.08670.11260.81050.03120.05440.91610.06910.04440.8895
AWLP0.16210.19140.67760.04370.09080.86950.05750.05830.8875
ATWT-M30.11980.17820.72340.06170.08180.86150.05880.08060.8653
MTF-GLP-HPM0.18100.19860.65630.06040.10960.83660.08270.06750.8554
MTF-GLP-CBD0.21990.20090.62340.07520.12820.80620.08000.06730.8580
PNN0.10420.04040.85960.06360.03220.90620.03060.04900.9218
DRPNN0.02680.07640.89880.03590.06360.90270.01990.09540.8866
MSDCNN0.06990.07950.85620.02820.07340.90040.09460.10680.8087
PWNet0.02890.11270.86160.01720.07000.91390.04280.05590.9037
Table 10. Running time comparison of different methods (in second).
Table 10. Running time comparison of different methods (in second).
MethodTimeMethedTime
PCA0.0900Indusion0.0600
GIHS0.0100AWLP0.0400
Brovey0.0010ATWTM30.1100
BDSD0.2700MTF_GLP_HPM0.0400
GS0.0300MTF_GLP_CBD0.0500
GSA0.1000PNN0.5900
PRACS0.1000DRPNN6.5100
HPF0.0100MSDCNN1.9900
SFIM0.0100PWNet1.0200

Share and Cite

MDPI and ACS Style

Liu, J.; Feng, Y.; Zhou, C.; Zhang, C. PWNet: An Adaptive Weight Network for the Fusion of Panchromatic and Multispectral Images. Remote Sens. 2020, 12, 2804. https://doi.org/10.3390/rs12172804

AMA Style

Liu J, Feng Y, Zhou C, Zhang C. PWNet: An Adaptive Weight Network for the Fusion of Panchromatic and Multispectral Images. Remote Sensing. 2020; 12(17):2804. https://doi.org/10.3390/rs12172804

Chicago/Turabian Style

Liu, Junmin, Yunqiao Feng, Changsheng Zhou, and Chunxia Zhang. 2020. "PWNet: An Adaptive Weight Network for the Fusion of Panchromatic and Multispectral Images" Remote Sensing 12, no. 17: 2804. https://doi.org/10.3390/rs12172804

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop