Next Article in Journal
The Frequency Selective Effect of Radar Backscattering from Multiscale Sea Surface
Previous Article in Journal
A Multisensor Approach to Satellite Monitoring of Trends in Lake Area, Water Level, and Volume
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Images Classification Based on Dense Convolutional Networks with Spectral-Wise Attention Mechanism

1
School of Computer Science, Northwestern Polytechnical University, Xi’an 710129, China
2
Department of Electronics and Informatics, Vrije Universiteit Brussel, Brussel 1050, Belgium
3
National Key Laboratory of Science and Technology on Space Microwave, Xi’an 710000, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(2), 159; https://doi.org/10.3390/rs11020159
Submission received: 29 November 2018 / Revised: 4 January 2019 / Accepted: 12 January 2019 / Published: 16 January 2019
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Hyperspectral images (HSIs) data that is typically presented in 3-D format offers an opportunity for 3-D networks to extract spectral and spatial features simultaneously. In this paper, we propose a novel end-to-end 3-D dense convolutional network with spectral-wise attention mechanism (MSDN-SA) for HSI classification. The proposed MSDN-SA exploits 3-D dilated convolutions to simultaneously capture the spectral and spatial features at different scales, and densely connects all 3-D feature maps with each other. In addition, a spectral-wise attention mechanism is introduced to enhance the distinguishability of spectral features, which improves the classification performance of the trained models. Experimental results on three HSI datasets demonstrate that our MSDN-SA achieves competitive performance for HSI classification.

Graphical Abstract

1. Introduction

Hyperspectral images (HSIs) have hundreds of continuous observation bands throughout the electromagnetic spectrum with high spectral resolution [1]. Based on such abundant spectral bands, HSIs have been widely used in various applications, including agriculture development [2], mineral resource exploitation [3], and environmental earth sciences [4]. Supervised land cover classification is one of the most significant topics in hyperspectral remote sensing. However, the redundancy of spectral band information combined with limited training samples [5,6] often poses a challenge to HSI classification.
Conventional HSI supervised classification models are often based on spectral information. Typical classifiers include those based on distance measure [7], k-nearest-neighbors [8], maximum likelihood criterion [9], and logistic regression [10]. To improve classification performance, Random Forests (RF) [11], and AdaBoost [12], which are ensemble learning or multiple classifier methods, have been found to be effective for HSI classification.
However, classification algorithms based on spectral information exploiting only the spectral information fail to capture the important spatial variability perceived for high-resolution data. Furthermore, as HSIs are typically presented in the format of 3-D cubes, it is reasonable to combine the abundant spectral features and spatial features in complementary form to improve the performance of HSI classification. For example, spectral-spatial combined features can be extracted from the HSI, at different frequencies and scales, by a series of 3-D discrete wavelet filters [13,14], 3-D Gabor filters [15,16], or 3-D scattering wavelet filers [17]. In this way, a large number of feature cubes can be created, which contain important information about local signal changes in spectrum, space, and joint spectral-spatial correlations. This information is essential for tackling challenging classification tasks.
In recent years, deep convolutional neural networks with hierarchical feature learning capability have become the mainstream machine learning methods in the field of computer vision, and have achieved gratifying results in different tasks [18,19]. Compared with traditional manually engineered features, deep learning technology automatically learns hierarchical features from raw input data [20,21]. The great success of convolutional neural networks (CNNs) and their extensions have motivated remarkable efforts in spectral-spatial HSI classification [22]. For instance, Yang et al. [23] proposed a two-channel CNN framework, which extracts jointly the spectral features and spatial features from HSI. Chen et al. [24] and Li et al. [25] have utilized a 3-D convolutional kernel to learn the discriminative spectral-spatial features and classifications are performed in an end-to-end structure. Note that the method in [25] used relatively smaller spatial sized cubes as input and has fewer parameters to tune. Gao et al. [26] proposed a CNN based architecture, which benefits from the multiple inputs corresponding to various image features, and exploited the both spectral and spatial contextual information concurrently for HSI classification. Yang et al. [27] advocated a recurrent 3-D CNN (R-3-D-CNN) model, which can often outperform other models and converge faster because of its 3-D convolutional operators and the recurrent network structure. However, it is reported that R-3-D-CNN requires more training samples than the traditional machine learning methods [27].
Although deep learning models have finally shown promising performance in HSI classification, they have an insatiable hunger for larger and larger data sets, while the available labeled samples are rather limited in the HSIs. Thus, the problem of small training samples restricts the deep learning based HSI classification approaches to obtain better performance. To deal with this problem, Ma et al. [28] proposed a spatially updated deep auto-encoder for spectral-spatial feature extraction, by adding a sample similarity regularization mechanism and combining it with the collaborative representation-based classification to deal with the problem of small training sets. In [23], when training samples are limited, transfer learning is used, where low-level and mid-level features are transferred from other scenes. Pan et al. [29] proposed an ensemble deep learning based method, multi-grained network (MugNet) for limited training samples. To take full advantage of abundant unlabeled samples, they adopted a semi-supervised manner in the process of generating convolution kernels. Again, to deal with the problem of limited training samples, the concept of generative adversarial networks (GAN) has been extended to be a conditional model with semi-supervised classification methods by He et al. [30]. The authors trained a generator and a discriminator on spatial-spectral features obtained from HSI. Then, they added a softmax layer to the discriminator network at the end and fine-tuning the network to perform classification. In addition, Zhong et al. [31] proposed a supervised spectral-spatial residual network (SSRN) with consecutive spectral and spatial residual blocks to extract spectral and spatial features from HSI. It is reported that SSRN is more effective in the case of small training samples, which is due to the fact that SSRN contains residual connections [19] between each of the other convolutional layers so that residual blocks are constructed. The success of residual connections has demonstrated that combining the features of the lower layers can capture finer features.
In this paper, we propose a method to distill the dense connectivity [32] of the network and construct a novel learning architecture with dense connectivity for automated classification from 3-D HSI. In addition, considering the redundancy of the spectral bands in HSI, a new spectral-wise attention mechanism is added to the proposed network. As a result, this paper contributes in four major respects:
  • We introduce a network architecture specifically designed for the 3-D patches of HSI. The network uses dilated convolutions to capture features at various patch scales, thereby obtaining multiple scales within a single layer. Dense connectivity connects 3-D feature maps learned from different layers, increasing the diversity of inputs in subsequent layers.
  • A new spectral-wise attention mechanism is aiming to selectively emphasize informative spectral features and suppress less useful spectral features. The spectral-wise attention mechanism that applies soft weights on features is well suited and more efficient for the following HSI classification tasks. To the best of our knowledge, this is the first time an attention mechanism has been introduced for HSI classification.
  • Experimental results on three HSI datasets demonstrate that our novel end-to-end 3-D dense convolutional network with spectral-wise attention mechanism (MSDN-SA) method outperforms the state-of-the-art methods.
The remainder of this paper is organized as follows. In Section 2, we briefly introduce the different connections and the spectral-attention mechanism in CNNs. In Section 3, we describe our MSDN-SA method in detail. In Section 4 and Section 5, we present and discuss the experimental results. Finally, in Section 6, the paper is summarized with an outline suggestion for future work.

2. Related Work

2.1. Residual Connections and Dense Connectivity in CNNs

Previous research [33,34,35] has shown that utilizing multi-level features in CNNs through skip-connections is effective for various vision tasks. Residual connections and dense connectivity act as two different connectivity patterns: both have been successfully used in challenging natural image processing tasks. Here, we briefly introduce these two connection concepts.
In principle, a residual connection (Figure 1a) adds a skip connection that bypasses the nonlinear transformations with an identity mapping. Reference layer inputs explicitly represent the layers as learning residual functions [19]. A residual connection can be formally expressed as:
x l = h ( x l 1 ) + F ( x l 1 ) ,
where x l 1 and x l refer to the input and output of the l-th layer, respectively. The h ( x l 1 ) = x l 1 is an identity mapping function and the function F ( ) represents a non-linear transformation which can be a composite function of operations such as Convolution (Conv), Batch Normalization (BN) [36], Rectified Linear Units (ReLU) [37], or Pooling [38]. By using residual connections, the gradient can flow directly from later to earlier layers through the identity function [19].
In order to maximize the flow of information between network layers, Gao et al. [32] proposed a series of dense connectivity from any layer to all subsequent layers (Figure 1b). Differing from the residual connections, which combine features through summation, dense connectivity combines features by concatenating them. Specifically, all previous feature maps of layers x 0 , , x l 1 , can be used to compute the output of the l-th layer:
x l = F ( { x 0 , , x l 1 } ) ,
where { x 0 , , x l 1 } is the concatenation of all previous feature maps.
Each layer has direct access to the gradients from the loss function and the original input signal, leading to an implicit deep supervision. The work of [32] also allows features to be reused, while adding only a small set of feature maps to the network [32]. In addition, dense connectivity has a regularization parameter that reduces overfitting on tasks with small training data. Therefore, dense connectivity can be beneficial to perform HSI classification, especially with small training data.

2.2. Attention Mechanism

Evidence from human perception process [39] demonstrates the importance of an attention mechanism, which usually uses top-level information to guide a bottom-up feedforward process. An attention mechanism can be viewed as a tool to bias the allocation of available processing resources towards the most informative components of an input signal.
Recently, researchers have applied attention mechanism in deep neural networks. Generally, there are two categories of such work: spatial-based attention and channel-wise attention. Spatial-based attention mostly focuses on specific location scenes. Such mechanisms were the focus of early systems for image categorization [40], and were later shown to yield significant improvements for Visual Question Answering (VQA) and captioning [41,42,43]. Channel-wise attention is a complementary form of attention, which involves learning a task-specific modulation that is selectively applied to individual feature maps across the entire scene. Among them, Squeeze-and-Excitation (SE) network (recognized as ImageNet Large Scale Visual Recognition Competition (ILSVRC) winner), produces significant performance improvements for state-of-the-art deep architectures at slightly greater computational cost. SE block [44] is a novel architectural unit designed to improve the representational capacity of a network by enabling it to perform dynamic channel-wise feature recalibration.
On the basis of attention mechanism, many deep neural network structures have been proposed and widely used in various applications [43,44,45,46,47]. For example, Fu et al. [45] introduced a recurrent attention CNN, which can locate the discriminative region recurrently for fine-grained image recognition performance. Li et al. [46] introduced a global attention upsample module to guide the integration of low- and high-level features in semantic segmentation. Wang et al. [47] proposed a residual attention network, which is built by trunk-and-mask [48] attention mechanism to generate attention-aware features for Image Classification. It is worth mentioning that SE network is a lightweight gating mechanism [44], specialized to model channel-wise relationships in a computationally efficient manner and designed to enhance the representational power of modules throughout the network.
In the case of HSI, hundreds of spectral bands are directly used as input data for convolution, which inevitably carries some noise bands. Therefore, we are more concerned with the correlation of the spectral-wise features from HSI which are based on the 3-D feature maps. Inspired by the SE network, we propose a spectral-wise attention mechanism which will be discussed in detail in the next section.

3. Proposed Methods

In this section, we describe a novel 3-D network for HSI classification. There are two key components in our proposed method: dense convolutional network with dilated convolution and spectral-wise attention mechanism. The first part is dilated convolution [49] based on 3-D patches of an HSI using dense connectivity to simultaneously extract spectral-spatial features. The dense connectivity is used to derive multi-level features in networks. In the second part, inspired by the successful application of attention mechanism in deep neural networks, a spectral-wise attention mechanism is added to the proposed network.
The proposed framework is illustrated schematically in Figure 2. Firstly, we extract the S × S × D neighborhoods of the center pixel within each spectral band together with its corresponding category label as samples, where S × S denotes the neighborhood space size, and D is the spectral depth. Once 3-D samples are extracted from an HSI, they are fed into the MSDN-SA model to obtain the classification results. There are seven convolutional layers in the MSDN-SA. In Figure 2, the colored lines in the convolutional layers represent 3 × 3 × 7 dilated convolutions, with each color representing a different channel through layers of different dilation. Note that a “channel” in this paper refers to a filter such that the total number of the “channels” stands for the dimensionality of the output space, i.e., the number of output filters in the convolution. Features are refined at each layer by a spectral-wise attention mechanism. Then, an average pooling layer and a fully connected (FC) layer follow. Finally, a soft-max activation function is used in the final output layer.

3.1. Dense Convolutional Network with Dilated Convolution

To take advantage of the capability of 3-D spatial filtering, we propose a novel learning architecture with dense connectivity for automated classification from 3-D HSI. In the inspiring work of [32] the dense connectivity was used within layers at a single scale, with transition layers to acquire information at different scales. Differing from the dense convolutional network (DenseNet) as per [32], we combine dense connectivity between the multi-scale feature maps, enabling dense connectivity between the feature maps of the entire network. This facilitates more efficient use of all feature maps and an even larger reduction of the number of required parameters. The presented module uses 3-D dilated convolutions to systematically aggregate multi-scale contextual information without losing spatial resolution.
Instead of using transitional layers to capture features at different scales, the proposed MSDN-SA uses dilated convolutions. A dilated convolution [49] W h , d with dilation d + uses a dilated kernel h that is nonzero only at distances that are a multiple of d pixels from the center. In the multi-scale approach, each individual channel of a feature map within a single layer operates at a different scale. Specifically, we associate the convolution operations for each channel of the output image of a certain layer with a different dilation. The setting of dilations is shown in Section 4.2. Formally, the output of x l is a dilated convolution of the j -th feature cube with the k -th kernel of the l -th layer, given by:
g l j ( x l ) = k = 0 c l 1 W h l j k , d l j x l k ,
where g ( ) is a dilated convolution operation, x l k denote as the feature maps of the l -th layer are convolved with the k -th kernel and c l is the number of kernels in the l -th layer.
When using the dilated convolution, the multi-scale approach has an additional advantage compared with traditional scaling. All feature maps have the same number of rows and columns as the input and output image, for all layers, and hence, when computing a feature map for a specific layer, it is not restricted to use only the output of the previous layer. Instead, we use all previously computed feature maps by densely connecting a network, as described by Equation (2). Thus, we change the dilated convolutional with dense connectivity operation Equation (3) to:
g l j ( { x 0 , , x l 1 } ) = i = 0 l 1 k = 0 c i 1 W h l j k i , d l j x i k ,
where i = 0 l 1 index the previous layers.

3.2. Spectral-Wise Attention Mechanism

For an HSI classification based on 3-D convolution based network, hundreds of spectral bands are directly used as input data for convolution, which inevitably carries some noise bands. To mitigate this problem, we use the SE block [44] to recalibrate spectral-wise feature responses by modelling interdependencies between spectral features. We model the interdependencies between spectral-wise features based on all the bands of each 3-D feature map, which we call spectral-wise attention mechanism. This mechanism aims to selectively emphasize informative spectral features and suppress less useful spectral features. The basic structure of the spectral-wise attention mechanism is illustrated in Figure 3. Here, the neighborhood space size and spectral depth are denoted with S × S and D , respectively, such that the input layer d + . Our starting point is the spectral-wise attention, which is denoted as F s p e c t r a l in our model and yields the spectral feature attention vector v 1 × 1 × D .
It is worth mentioning that the attention mechanism is independently applied to eight channels which are the outputs of the convolutional layers in our algorithm. Following the definition of the channel-wise attention model, the spectral-wise attention model on each 3-D feature map can be informally described as follows. “Summary statistics” are calculated per-band, and then transformations are applied to first shrink and then expand the dimensionality of these statistics.
Formally, summary statistics are computed with a global average pooling applied to individual spectral-wise feature channels U = [ u k ] k = 1 D , yielding the vector P = ( p k ) k = 1 D . This is followed by first applying a shrinking operation to the vector P with the operator S shrink D r × D compressing it into a lower dimensional space, and then followed by an expansion operation S expand D × D r mapping it back to the original, higher dimensional space:
v = σ ( S expand ( δ ( S shrink ( P ) ) ) ) ,
where δ refers to the ReLU function and σ refers to a sigmoid activation, with “reduction ratio” r of the shrinking operation empirically set to be 4. The final output of the spectral-wise attention U is obtained by rescaling the transformation output U with the activations:
U = F scale ( u k , v k ) = v k u k .
To illustrate the application of the spectral-wise attention mechanism to our dense network, Figure 4 depicts the schema of the proposed approach. The spectral-wise attention is the weight added after the 3-D dilated convolution operation, but before the connection operation. For each feature map, a global average pooling layer transforms the S × S × D -sized feature map to a 1 × 1 × D -sized feature vector, which corresponds to P = ( p k ) k = 1 D in Equation (5). Next, a fully connected layer generates an output vector 1 × 1 × D r . This process is a shrinking operation, which corresponds to S shrink in Equation (5). After applying ReLU function which corresponds to δ in Equation (5), an expansion operation is performed. The second fully connected layer generates an output vector 1 × 1 × D , this operation corresponds to S expand in Equation (5). Lastly, a sigmoid activation is employed, which corresponds to σ in Equation (5).

3.3. Network Implementation Details

By combining dense convolutional network and spectral-wise attention mechanism, a new network is formed. Details of the layers of the proposed MSDN-SA are described in Table 1. The implementation of MSDN-SA is given as follows.
Taking the Indian Pines dataset as an example, the 3-D samples with size 13 × 13 × 200 are used as the input data. The MSDN-SA has seven layers. Each feature map is the result of applying the dense connection operations given by Equation (4) to all previous feature maps: 3-D dilated convolutions (3-D-DConv) with 3 × 3 × 7 pixel filters and a channel-specific dilation followed with batch normalization. We represent this step operation as DConv-BN. Following each dense connection layer, the spectral-wise attention mechanism is applied to each 3-D feature map and added in accordance to Figure 4. It is worth mentioning that “1 ×1 × 200, 8” denotes as eight attention weights obtained by eight independent channels. Finally, an average pooling layer and a fully connected (FC) layer transforms a 5 × 5 × 48 spectral-spatial feature into a 1 × 1 × L output feature vector, L represents the number of neurons. In the Indian Pines dataset, we select L = 360. Note that all layers in our MSDN-SA, including convolutional and average pooling, are implemented in a 3-D manner. Therefore, when extracting features and making predictions, the MSDN-SA can completely retain and utilize the 3-D spectral-spatial information.
Network implementation details for other datasets are carried out in a similar manner and hence, are omitted.

4. Experiments Results

In order to evaluate the effectiveness of the proposed method, we tested it on three hyperspectral datasets. Class accuracy, overall accuracy (OA), average accuracy (AA), and kappa coefficient ( κ ) were adopted to assess the classification results. We implemented 10 trials of hold-out cross validation for each dataset: the mean values and standard deviations are reported for each dataset. For each trial, a limited number of training samples were randomly selected from each class, and the remaining samples were used as a blind test. The training sample sizes are set to a minimal level to make the classification task more challenging than otherwise [50].

4.1. Datasets

To evaluate the performance of the proposed method for HSI classification, we use the following three datasets:
(1)
Indian Pines Dataset
The Indian Pines image was gathered by the AVIRIS sensor during a flight over the Indian Pines site in Northwestern Indiana, including 16 vegetation classes. It contains 145 × 145 pixels and 220 spectral bands in the range of 0.4–2.5 μm. Due to water absorption, 20 spectral bands were removed, and the remaining 200 spectral bands were used for classification.
(2)
University of Pavia Dataset
The University of Pavia image was recorded by the ROSIS sensor over Pavia, northern Italy, including 16 urban land-cover classes and having 610 × 340 pixels. It contains 115 spectral reflectance bands, at the wavelength range 0.43–0.86 μm. Twelve spectral bands were removed due to noise, and the remaining 103 spectral bands were used for the experiments.
(3)
University of Houston Dataset
The University of Houston dataset was acquired by the NSF-funded National Center for Airborne Laser Mapping (NCALM) over the University of Houston campus and the neighboring urban areas using the ITRES-CASI (Compact Airborne Spectrographic Imager) 1500 hyperspectral imager in 2012. It contains 15 land cover classes with 349 × 1905 pixels and 144 bands are used for assessment with wavelength ranging from 0.36 to 1.05 μm.

4.2. Experimental Setting

We followed a previous study [51] and adopted the same weight initialization method. In all experiments, dilations were evenly distributed s i j [ 1 , 10 ] by setting the dilation of channel j of layer i equal to s i j = ( ( i w + j ) mod   10 ) + 1 , where w is the number of kernels in the convolutional layers. The soft-max activation function was used in the final output layer and the Nesterov Stochastic Gradient Descent (SGD) optimization method was employed during training to minimize the cross-entropy between labels of samples and network outputs. We set the batch size to 16, with the network trained over 100 epochs on three HSI datasets (60 epochs with learning rate 0.01 and 40 epochs with learning rate 0.001). Then, we analyzed two factors that control the training and classification performance of MSDN-SA: (1) number of kernels in the convolutional layers, and (2) size of input spatial cubes.
First, we experimentally verified the number of kernels in the convolutional layers. We tested different kernel numbers from 4 to 20 with fixed intervals of four in each convolutional layer. Classifications were performed on three datasets with only 20 training samples per class using a different number of kernels. The results are shown in Figure 5a. The network with eight kernels in each convolutional layer obtained the best performance in the Indian Pines dataset and University of Pavia dataset, and the network with 20 kernels achieved the highest classification accuracy in the University of Houston dataset, though only marginally higher than the results with eight kernels. For the sake of consistency, we used eight kernels for all datasets. Note that this verifies that dense connectivity allows features to be reused, and a small number of kernels are sufficient.
Second, to obtain an optimal size of the spatial neighborhood in the MSDN-SA, we assessed 5 × 5, 9 × 9, 13 × 13, 17 × 17, and 21 × 21 neighborhoods. Figure 5b shows the classification performance of three HSI datasets using different spatial neighborhood sizes with only 20 samples per class as training samples. We can see from Figure 5b that initially, as the spatial size increases, the accuracy increases rapidly, however when the spatial size reaches 13 × 13, the accuracy stabilizes. Therefore, to balance between the accuracy of classification and the amount of data involved in computation, we chose empirically 13 × 13 as the spatial neighborhood size.
In our experiments, the performance of MSDN and MSDN-SA are compared with four recently proposed supervised HSI classification methods. The algorithms compared in this paper are summarized as follows:
(1)
CCF [52]: Canonical Correlation Forests based on spectral feature with 100 trees.
(2)
SVM-3DG [14]: An SVM-based classification method by applying the 3-D discrete wavelet transform and Markov random field (MRF).
(3)
CNN-transfer [23]: A CNN with two-branch architecture based on spectral-spatial feature, where a transfer learning strategy is used. Specifically, the source datasets of Indian Pines for pretraining are Salinas Valley, which were collected by the same sensor AVIRIS, and the source datasets of Pavia University for pretraining are Pavia Center which were collected by the same sensor ROSIS.
(4)
3D-CNN [25]: The 3D-CNN network framework has two 3-D convolutional layers and a fully connected layer. The network structure is set as given in [25].
(5)
SSRN [31]: The architecture of the SSRN is set out in [31]. The spectral feature learning part includes two convolutional layers and two spectral residual blocks, the spatial feature learning part comprises of one 3-D convolutional layer and two spatial residual blocks. Finally, there is an average pooling layer and a fully connected layer to output the results.
Below we compare the above algorithms with our proposed method. For the three HSI datasets, for a fair comparison, the network structures were set to the same width and depth. Additionally, we set the same input volume size of 13 × 13 × D for our proposed method on all datasets.

4.3. Results of Indian Pines Dataset

The 10-time average classification accuracies and the corresponding standard deviations of the Indian Pines dataset are reported in Table 2 and the classification maps of different methods are shown in Figure 6c–g. For this dataset, SSRN and MSDN-SA outperform other methods, with MSDN-SA achieving an advantage of approximately 1% to 12%. Note that the numbers of class samples in this dataset are quite unbalanced. In particular, those of the classes Alfalfa, Grass-pasture-mowed and Oats are very few. Except for the SSRN, all compared methods perform well in these classes. Although in terms of classification accuracy, SSRN is the best competitor and its performance is close to the proposed method, it does not do well with classes with a small number of samples. On the contrary, the stability and performance of our algorithm are obvious.

4.4. Results of University of Pavia Dataset

The results of University of Pavia dataset are reported in Table 3 and the classification maps of different methods are shown in Figure 7. With nine classes and 20 samples per class in this dataset, a total of 180 pixels were used for training. This number is smaller than that used for the other two datasets (320 in Indian Pines dataset and 300 in University of Houston dataset). Among all five methods only SSRN and MSDN-SA achieve more than 90% OA, AA and κ . With limited training samples, only well-suited deep features are able to exploit the spectral space of 115 dimensionality. From this table we can see that MSDN-SA reports at least 87% accuracy for all classes, with the AA significantly higher than that achievable by the compared methods. Note that SSRN also performs well in this dataset with more adequate training samples available for each class. Compared with SSRN, the proposed MSDN-SA still performs marginally better in both OA and κ .

4.5. Results of University of Houston Dataset

The results of the experiments of University of Houston dataset are listed in Table 4. As this data set is too large, given the space limitations, we show only the two algorithms with the best classification results and the classification maps are shown in Figure 8c,d. MSDN-SA works well again with few training samples. The advantage of MSDN-SA for this dataset is more significant as compared to the other two algorithms, especially regarding OA and κ . In terms of class accuracy, MSDN-SA performs well in all classes and it achieves the highest accuracy in nine classes.
Overall, for three HSI datasets, the proposed MSDN-SA has achieved better performance than those methods compared. We observe that the skip-connections networks used in both SSRN and MSDN-SA show good results, indicating that this connection mechanism strategy has a positive effect on feature propagation while training with a very small number of samples.

5. Analysis and Discussion

5.1. Effect of Training Samples

The above experimental results have shown that the proposed MSDN-SA method performs well in HSI classifications, especially in the case of having smaller training samples. In this part, we would like to further investigate the scenarios of extremely scarce training samples. The curves of AA with respect to a different number of training samples are shown in Figure 9.
As expected, as the number of training samples increases, the accuracy increases. We can see from Figure 9 that MSDN-SA outperforms other methods in most cases. Regarding Indian Pines and University of Pavia datasets, using only five training samples per class, MSDN-SA has achieved an average accuracy of more than 80% and 83% respectively. Although classification of University of Houston dataset is more challenging, on 10–50 training samples per class MSDN-SA scores significantly higher than other compared methods.

5.2. Effect of Spectral-Wise Attention Mechanism

To validate the effectiveness of the spectral-wise attention mechanism, we tested and compared the proposed network with and without the spectral-wise attention mechanism. The effectiveness of spectral-wise attention can be demonstrated in Figure 10. It shows the OA of three datasets with 20 training samples per class. It is obvious that spectral-wise attention improves the classification results for all three datasets, with performance boosted by a larger margin on Indian Pines dataset and University of Houston dataset, than on the University of Pavia dataset. The effect of spectral-wise attention is thought to be related to the redundancy of the input bands. Then, we also investigated the weights generated by the spectral-wise attention mechanism in different layers. Take Indian Pines as an example: the average weights generated by the spectral-wise attention mechanism in the first layer and the penultimate layer are shown in Figure 11, respectively. From Figure 11, we can see that the attention mechanism has less influence on the shallow features, and the weights are concentrated near 0.5. As the number of layers increases, the weights generated in the attention mechanism have more guidance on the deep features, thereby emphasizing informative spectral features and suppressing less useful spectral features, as illustrated in Figure 11.

5.3. Effect of Dilated Convolution

In this part, we will validate the effectiveness of the dilated convolution. First, we replace each dilated convolution of the proposed method to a traditional three-dimensional convolution, and we represent this model as DN-SA. Then, we compare DN-SA with SVM-3DG, SSRN, and MSDA-SA, which are the top three performing methods in Section 4. The effectiveness of dilated convolution can be demonstrated in Figure 12. It shows the OA of three datasets with 20 training samples per class. It is obvious that dilated convolutions improve the classification results for all three datasets, with performance boosted by a larger margin on University of Pavia and the University of Houston datasets, than on the Indian Pines dataset. Our work shows that the dilated convolution operator is particularly suited to dense prediction due to its ability to expand the receptive field without losing resolution.

6. Conclusions

In this paper, we have proposed a network architecture specifically designed for 3-D patches of hyperspectral datasets. Specifically, we have proposed a novel dense convolutional network that uses dilated convolutions instead of traditional scaling operations to learn features at different scales. It uses multiple scales in each layer, and computes the feature map of each layer using all the feature maps of earlier layers, resulting in a densely connected network. Furthermore, a spectral-wise attention mechanism, adding soft weights on features, was proposed to enhance the distinguishability of spectral features. By combing the dense convolutional network with dilated convolution and spectral-wise attention, the resulting MSDN-SA network architecture enables accurate training with relatively small training sets. Experimental results on three popular HSI benchmark datasets demonstrate that MSDN-SA performs consistently, offering the highest classification accuracy.
In terms of future research, we plan to research how to select effective samples as the training samples, which are potentially more effective for training the network, which remains active research.

Author Contributions

All the authors made significant contributions to this work. B.F. and Y.L. devised the approach and analyzed the data; J.C.-W.C. helped design the experiments and provided advice for the preparation and revision of the work; Bei Fang performed the experiments; and H.Z. helped with the experiments.

Funding

This work was supported in part by the National Natural Science Foundation of China (61871460, 61876152), Foundation Project for Advanced Research Field (614023804016HK03002), Shaanxi International Scientific and Technological Cooperation Project (2017KW-006), the National Key Laboratory of Science and Technology on Space Microwave (6142411040404), and the Innovation Foundation for Doctor Dissertation of Northwestern Polytechnical University (CX201816).

Acknowledgments

The authors would like to thank P. Gamba from the University of Pavia, Pavia, Italy, for providing the reflective optics system imaging spectrometer data and corresponding reference information. The authors would also like to thank the National Center for Airborne Laser Mapping for providing the Houston dataset.

Conflicts of Interest

The authors declare no competing financial interests. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

References

  1. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y.F. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  2. Liang, L.; Di, L.; Zhang, L.; Deng, M.; Qin, Z.; Zhao, S.; Lin, H. Estimation of crop LAI using hyperspectral vegetation indices and a hybrid inversion method. Remote Sens. Environ. 2015, 165, 123–134. [Google Scholar] [CrossRef]
  3. Yokoya, N.; Chan, J.; Segl, K. Potential of resolution-enhanced hyperspectral data for mineral mapping using simulated enmap and sentinel-2 images. Remote Sens. 2016, 8, 172. [Google Scholar] [CrossRef]
  4. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef]
  5. Landgrebe, D.A. Signal Theory Methods in Multispectral Remote Sensing; Wiley: Hoboken, NJ, USA, 2003; Chapter 3. [Google Scholar]
  6. He, L.; Li, J.; Liu, C.; Li, S. Recent advances on spectral-spatial hyperspectral image classification: An overview and new guidelines. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1579–1597. [Google Scholar] [CrossRef]
  7. Du, Q.; Chang, C. A linear constrained distance-based discriminant analysis for hyperspectral image classification. Pattern Recognit. 2001, 34, 361–373. [Google Scholar] [CrossRef]
  8. Samaniego, L.; Bardossy, A.; Schulz, K. Supervised classification of remotely sensed imagery using a modified k-NN technique. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2112–2125. [Google Scholar] [CrossRef]
  9. Ediriwickrema, J.; Khorram, S. Hierarchical maximum-likelihood classification for improved accuracies. IEEE Trans. Geosci. Remote Sens. 1997, 35, 810–816. [Google Scholar] [CrossRef]
  10. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised hyperspectral image segmentation using multinomial logistic regression with active learning. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4085–4098. [Google Scholar] [CrossRef]
  11. Xia, J.; Bombrun, L.; Berthoumieu, Y.; Germain, C.; Du, P. Spectral–spatial rotation forest for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 10, 4605–4613. [Google Scholar] [CrossRef]
  12. Chan, C.W.; Paelinckx, D. Evaluation of random forest and adaboost tree-based ensemble classification and spectral band selection for ecotope mapping using airborne hyperspectral imagery. Remote Sens. Environ. 2008, 112, 2999–3011. [Google Scholar] [CrossRef]
  13. Qian, Y.; Ye, M.; Zhou, J. Hyperspectral image classification based on structured sparse logistic regression and three-dimensional wavelet texture features. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2276–2291. [Google Scholar] [CrossRef]
  14. Cao, X.; Xu, L.; Meng, D.; Zhao, Q.; Xu, Z. Integration of 3-dimensional discrete wavelet transform and Markov random field for hyperspectral image classification. Neurocomputing 2017, 226, 90–100. [Google Scholar] [CrossRef]
  15. Jia, S.; Shen, L.; Li, Q. Gabor feature-based collaborative representation for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1118–1129. [Google Scholar]
  16. Shen, L.; Jia, S. Three-dimensional gabor wavelets for pixel-based hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2011, 49, 5039–5046. [Google Scholar] [CrossRef]
  17. Tang, Y.Y.; Lu, Y.; Yuan, H. Hyperspectral image classification based on three-dimensional scattering wavelet transform. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2467–2480. [Google Scholar] [CrossRef]
  18. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  20. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  21. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  22. Petersson, H.; Gustafsson, D.; Bergstrom, D. Hyperspectral image analysis using deep learning—A review. In Proceedings of the 2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA), Oulu, Finland, 12–15 December 2016; pp. 1–6. [Google Scholar]
  23. Yang, J.; Zhao, Y.Q.; Chan, C.W. Learning and transferring deep joint spectral–spatial features for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4729–4742. [Google Scholar] [CrossRef]
  24. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  25. Li, Y.; Zhang, H.; Shen, Q. Spectral-spatial classification of hyperspectral imagery with 3d convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  26. Gao, Q.; Lim, S.; Jia, X. Hyperspectral image classification using convolutional neural networks and multiple feature learning. Remote Sens. 2018, 10, 299. [Google Scholar] [CrossRef]
  27. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.K.; Zhang, X. Hyperspectral image classification with deep learning models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  28. Ma, X.; Wang, H.; Geng, J. Spectral-spatial classification of hyperspectral image based on deep auto-encoder. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4073–4085. [Google Scholar] [CrossRef]
  29. Pan, B.; Shi, Z.; Xu, X. MugNet: Deep learning for hyperspectral image classification using limited samples. ISPRS J. Photogramm. Remote Sens. 2018, 145, 108–119. [Google Scholar] [CrossRef]
  30. He, Z.; Liu, H.; Wang, Y.; Hu, J. Generative adversarial networks based semi-supervised learning for hyperspectral image classification. Remote Sens. 2017, 9, 1042. [Google Scholar] [CrossRef]
  31. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-d deep learning framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  32. Huang, G.; Liu, Z.; Laurens, V.D.M.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  33. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 39, 640–651. [Google Scholar]
  34. Sermanet, P.; Kavukcuoglu, K.; Chintala, S.; Lecun, Y. Pedestrian Detection with Unsupervised Multi-stage Feature Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  35. Hariharan, B.; Arbelaez, P.; Girshick, R.; Malik, J. Hypercolumns for object segmentation and fine-grained localization. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  36. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning (ICML 2015), Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  37. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the AISTATS, Ft. Lauderdale, FL, USA, 11–13 April 2011; p. 3. [Google Scholar]
  38. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  39. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent models of visual attention. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, CA, USA, 8–13 December 2014. [Google Scholar]
  40. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhudinov, R.; Zemel, R.; Bengio, Y. Show, attend and tell: Neural image caption generation with visual attention. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 2048–2057. [Google Scholar]
  41. Zhu, Y.; Groth, O.; Bernstein, M.; Fei-Fei, L. Visual7w: Grounded question answering in images. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 4995–5004. [Google Scholar]
  42. Yang, Z.; He, X.; Gao, J.; Deng, L.; Smola, A. Stacked attention networks for image question answering. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 21–29. [Google Scholar]
  43. Nam, H.; Ha, J.W.; Kim, J. Dual attention networks for multimodal reasoning and matching. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2156–2164. [Google Scholar]
  44. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation net-works. arXiv 2017, arXiv:1709.01507. [Google Scholar]
  45. Fu, J.; Zheng, H.; Mei, T. Look closer to see better: Recurrent attention convolutional neural network for fine-grained image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4476–4484. [Google Scholar]
  46. Li, H.; Xiong, P.; An, J.; Wang, L. Pyramid attention network for semantic segmentation. arXiv 2018, arXiv:1805.10180v2. [Google Scholar]
  47. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  48. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 8–16 October 2016. [Google Scholar]
  49. Yu, F.; Koltun, V. Multi-Scale Context Aggregation by Dilated Convolutions. In Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  50. Li, F.; Xu, L.; Siva, P.; Wong, A.; Clausi, D.A. Hyperspectral image classification with limited labeled training samples using enhanced ensemble learning and conditional random fields. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 1–12. [Google Scholar] [CrossRef]
  51. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  52. Xia, J.; Yokoya, N.; Iwasaki, A. Hyperspectral image classification with canonical correlation forests. IEEE Trans. Geosci. Remote Sens. 2017, 55, 1–11. [Google Scholar] [CrossRef]
Figure 1. Schema of (a) residual connections module and (b) dense connectivity module.
Figure 1. Schema of (a) residual connections module and (b) dense connectivity module.
Remotesensing 11 00159 g001
Figure 2. Proposed framework for HSI classification. The colored lines in the convolutional layers represent 3 × 3 × 7 dilated convolutions, with each color representing a different channel through layers of different dilation.
Figure 2. Proposed framework for HSI classification. The colored lines in the convolutional layers represent 3 × 3 × 7 dilated convolutions, with each color representing a different channel through layers of different dilation.
Remotesensing 11 00159 g002
Figure 3. Basic structure of spectral-wise attention mechanism.
Figure 3. Basic structure of spectral-wise attention mechanism.
Remotesensing 11 00159 g003
Figure 4. Dense network with spectral-wise attention mechanism.
Figure 4. Dense network with spectral-wise attention mechanism.
Remotesensing 11 00159 g004
Figure 5. Effect of network hyper-parameters. (a) overall accuracy (OA) of different kernel numbers. (b) OA of different spatial size.
Figure 5. Effect of network hyper-parameters. (a) overall accuracy (OA) of different kernel numbers. (b) OA of different spatial size.
Remotesensing 11 00159 g005
Figure 6. Classification maps of Indian Pines dataset. (a) False color image. (b) Reference image. (c) CCF. (d) SVM-3DG. (e) 2D-CNN-transfer. (f) 3D-CNN. (g) SSRN. (h) MSDN. (i) MSDN-SA.
Figure 6. Classification maps of Indian Pines dataset. (a) False color image. (b) Reference image. (c) CCF. (d) SVM-3DG. (e) 2D-CNN-transfer. (f) 3D-CNN. (g) SSRN. (h) MSDN. (i) MSDN-SA.
Remotesensing 11 00159 g006
Figure 7. Classification maps of University of Pavia dataset. (a) True color image. (b) Reference image. (c) CCF. (d) SVM-3DG. (e) 2D-CNN-transfer. (f) 3D-CNN. (g) SSRN. (h) MSDN. (i) MSDN-SA.
Figure 7. Classification maps of University of Pavia dataset. (a) True color image. (b) Reference image. (c) CCF. (d) SVM-3DG. (e) 2D-CNN-transfer. (f) 3D-CNN. (g) SSRN. (h) MSDN. (i) MSDN-SA.
Remotesensing 11 00159 g007
Figure 8. Classification maps of University of Houston dataset. (a) True color image. (b) Reference image. (c) SSRN. (d) MSDN-SA.
Figure 8. Classification maps of University of Houston dataset. (a) True color image. (b) Reference image. (c) SSRN. (d) MSDN-SA.
Remotesensing 11 00159 g008
Figure 9. Effect of number of training samples on: (a) Indian Pines dataset. (b) University of Pavia dataset. (c) University of Houston dataset.
Figure 9. Effect of number of training samples on: (a) Indian Pines dataset. (b) University of Pavia dataset. (c) University of Houston dataset.
Remotesensing 11 00159 g009
Figure 10. Effect of spectral-wise attention on: Indian Pines, University of Pavia and University of Houston datasets.
Figure 10. Effect of spectral-wise attention on: Indian Pines, University of Pavia and University of Houston datasets.
Remotesensing 11 00159 g010
Figure 11. Average weights generated by spectral-wise attention mechanism on first layer and penultimate layer on Indian Pines dataset.
Figure 11. Average weights generated by spectral-wise attention mechanism on first layer and penultimate layer on Indian Pines dataset.
Remotesensing 11 00159 g011
Figure 12. Effect of dilated convolution on: Indian Pines, University of Pavia and University of Houston datasets.
Figure 12. Effect of dilated convolution on: Indian Pines, University of Pavia and University of Houston datasets.
Remotesensing 11 00159 g012
Table 1. Network architecture details of proposed novel end-to-end 3-D dense convolutional network with spectral-wise attention mechanism (MSDN-SA) for Indian Pines Dataset.
Table 1. Network architecture details of proposed novel end-to-end 3-D dense convolutional network with spectral-wise attention mechanism (MSDN-SA) for Indian Pines Dataset.
LayerKernel sizeNetworkOutput Size
Inputs--13 × 13 × 200
3-D-DConv13 × 3 × 7DConv-BN13 × 13 × 200, 8
Attention mechanism--1 × 1 × 200, 8
3-D-DConv23 × 3 × 7DConv-BN13 × 13 × 200, 8
Attention mechanism--1 × 1 × 200, 8
Dense concatenate--13 × 13 × 200, 16
3-D-DConv33 × 3 × 7DConv-BN13 × 13 × 200, 8
Attention mechanism--1 × 1 × 200, 8
Dense concatenate--13 × 13 × 200, 24
3-D-DConv43 × 3 × 7DConv-BN13 × 13 × 200, 8
Attention mechanism--1 × 1 × 200, 8
Dense concatenate--13 × 13 × 200, 32
3-D-DConv53 × 3 × 7DConv-BN13 × 13 × 200, 8
Attention mechanism--1 × 1 × 200, 8
Dense concatenate--13 × 13 × 200, 40
3-D-DConv63 × 3 × 7DConv-BN13 × 13 × 200, 8
Attention mechanism--1 × 1 × 200, 8
Dense concatenate--13 × 13 × 200, 48
3-D-Average Pooling3 × 3 × 8stride 25 × 5 × 48, 8
FC, Soft-max 360
Table 2. Network architecture details of proposed MSDN-SA for Indian Pines Dataset.
Table 2. Network architecture details of proposed MSDN-SA for Indian Pines Dataset.
ClassSamplesMethods
Train/TestCCFSVM-3DGCNN-Transfer3D-CNNSSRNMSDNMSDN-SA
120/2695.77 ± 2.8497.44 ± 2.2297.95 ± 2.4498.08 ± 2.7286.59 ± 7.3195.65 ± 2.9495.62 ± 1.60
220/140867.12 ± 6.6770.12 ± 7.5465.21 ± 6.5664.42 ± 6.4393.80 ± 3.8371.77 ± 1.9678.31 ± 2.83
320/81067.14 ± 5.2971.73 ± 17.8867.10 ± 4.1765.72 ± 2.1593.75 ± 4.4878.46 ± 5.1889.20 ± 3.02
420/21789.54 ± 4.0591.71 ± 5.7788.51 ± 3.1988.02 ± 1.9580.82 ± 6.2394.65 ± 2.0192.68 ± 4.75
520/46387.58 ± 3.4188.26 ± 8.0488.06 ± 2.6188.39 ± 0.9598.64 ± 1.7593.76 ± 3.2787.79 ± 0.58
620/71093.21 ± 2.5697.46 ± 1.9494.28 ± 1.3994.65 ± 0.5999.49 ± 0.5297.57 ± 1.6994.27 ± 0.38
714/1495.00 ± 6.78100 ± 0.0095.54 ± 5.8096.43 ± 5.0566.46 ± 10.5892.86 ± 5.8396.47 ± 4.74
820/45898.19 ± 0.5099.41 ± 0.8391.13 ± 0.8286.79 ± 3.2499.96 ± 0.1097.43 ± 1.1098.18 ± 0.09
910/10100 ± 0.00100 ± 0.00100 ± 0.00100 ± 0.0061.67 ± 7.39100 ± 0.00100 ± 0.00
1020/95281.16 ± 6.4174.37 ± 6.1578.69 ± 4.1476.24 ± 2.6968.86 ± 17.7080.48 ± 4.6381.28 ± 2.00
1120/243558.73 ± 4.4174.74 ± 4.6059.55 ± 2.1660.14 ± 1.8983.28 ± 1.8274.43 ± 3.8579.99 ± 1.09
1220/57380.38 ± 4.8991.04 ± 6.9278.59 ± 3.6461.43 ± 14.0783.58 ± 7.0189.69 ± 1.7984.82 ± 3.20
1320/18599.03 ± 0.3499.10 ± 0.3198.27 ± 0.6098.92 ± 0.2196.90 ± 2.1898.21 ± 0.4197.23 ± 0.61
1420/124590.25 ± 4.8386.43 ± 7.8790.24 ± 0.5391.41 ± 0.1499.98 ± 0.0488.43 ± 2.4995.80 ± 0.21
1520/36662.90 ± 4.4094.72 ± 9.1577.93 ± 3.8385.22 ± 10.2260.16 ± 1.9379.46 ± 4.9564.97 ± 1.62
1620/7395.75 ± 2.7796.35 ± 2.8595.97 ± 1.75100 ± 0.0079.96 ± 2.9995.34 ± 1.2496.71 ± 1.97
OA(%)75.60 ± 1.0481.43 ± 1.0575.18 ± 1.0274.51 ± 1.1084.35 ± 4.1983.62 ± 3.9586.62 ± 2.36 1
AA(%)85.11 ± 0.5889.56 ± 0.6085.59 ± 1.0184.74 ± 1.7683.24 ± 3.2689.26 ± 2.6689.58 ± 1.82
κ × 10072.50 ± 1.1478.93 ± 1.1072.80 ± 1.1071.21 ± 1.3082.20 ± 4.6880.96 ± 2.0785.16 ± 2.02
1 Results that surpass all competing methods are bold.
Table 3. Network architecture details of proposed MSDN-SA for University of Pavia Dataset.
Table 3. Network architecture details of proposed MSDN-SA for University of Pavia Dataset.
ClassSamplesMethods
Train/TestCCFSVM-3DGCNN-Transfer3D-CNNSSRNMSDNMSDN-SA
120/661173.21 ± 6.4791.99 ± 4.8770.62 ± 3.5968.05 ± 3.9898.95 ± 0.6296.76 ± 2.7793.31 ± 2.02
220/1862979.88 ± 6.7990.74 ± 5.4875.41 ± 5.5066.58 ± 4.8099.85 ± 0.0991.81 ± 2.0898.88 ± 1.36
320/207980.15 ± 6.2581.84 ± 9.8478.08 ± 4.1275.47 ± 3.9487.58 ± 2.7485.33 ± 3.2187.93 ± 1.97
420/304492.81 ± 5.1589.99 ± 3.8291.44 ± 1.2692.62 ± 0.9482.48 ± 7.5692.74 ± 1.7391.33 ± 3.32
520/132599.53 ± 0.4396.54 ± 1.8498.87 ± 1.8798.15 ± 2.6299.98 ± 0.0598.26 ± 0.2299.97 ± 0.11
620/500982.65 ± 2.8684.12 ± 10.6378.19 ± 2.8871.49 ± 5.1671.46 ± 3.2381.81 ± 2.1387.29 ± 1.78
720/131093.96 ± 2.3490.06 ± 3.8691.30 ± 2.3088.09 ± 2.1689.74 ± 4.4690.46 ± 2.4791.68 ± 2.41
820/366277.96 ± 7.2390.83 ± 5.7581.65 ± 5.0987.56 ± 2.5384.18 ± 5.6988.35 ± 3.1789.14 ± 3.32
920/92799.81 ± 0.1099.98 ± 0.0599.10 ± 0.8799.63 ± 0.0898.48 ± 0.4299.01 ± 0.8999.08 ± 0.39
OA(%)81.42 ± 2.8890.04 ± 1.3675.48 ± 1.5473.97 ± 0.3592.30 ± 1.9791.01 ± 2.5392.99 ± 2.02
AA(%)86.66 ± 1.0690.68 ± 1.6584.99 ± 1.3083.07 ± 1.2591.88 ± 1.9091.39 ± 1.5692.98 ± 1.04
κ × 10076.22 ± 3.3686.95 ± 1.6173.14 ± 3.1767.52 ± 0.0990.01 ± 2.4988.29 ± 2.2190.98 ± 2.94
Table 4. Network architecture details of proposed MSDN-SA for University of Houston Dataset.
Table 4. Network architecture details of proposed MSDN-SA for University of Houston Dataset.
ClassSamplesMethods
Train/TestCCFSVM-3DG3D-CNNSSRNMSDNMSDN-SA
120/123174.76 ± 4.5877.01 ± 6.6889.65 ± 4.9871.04 ± 4.1183.50 ± 3.7985.04 ± 3.72
220/123469.30 ± 5.0778.53 ± 1.7665.24 ± 0.7987.09 ± 2.0586.76 ± 4.7588.11 ± 0.80
320/67779.07 ± 6.3187.99 ± 12.2689.96 ± 4.5998.81 ± 0.9593.69 ± 4.2995.80 ± 4.17
420/122462.01 ± 3.3569.91 ± 4.2962.14 ± 2.2678.88 ± 5.7487.61 ± 3.1988.50 ± 0.03
520/122290.39 ± 1.8793.64 ± 2.7292.76 ± 2.4994.36 ± 1.6790.65 ± 0.9692.93 ± 1.72
620/30566.02 ± 5.9678.69 ± 6.6359.51 ± 8.5887.52 ± 9.0976.85 ± 4.2969.87 ± 1.59
720/124838.69 ± 5.9684.43 ± 3.0948.48 ± 7.1479.42 ± 3.9780.41 ± 4.7389.30 ± 4.52
820/122456.95 ± 5.7861.41 ± 4.8951.96 ± 5.9696.35 ± 4.7090.08 ± 5.5794.16 ± 3.10
920/123247.95 ± 5.1368.78 ± 3.7974.51 ± 1.0372.26 ± 3.3570.73 ± 4.2881.96 ± 4.52
1020/120771.48 ± 6.9972.74 ± 7.2150.04 ± 6.2183.70 ± 4.3380.31 ± 5.3388.10 ± 7.87
1120/121554.93 ± 6.6965.27 ± 6.3339.88 ± 6.5892.94 ± 5.2282.22 ± 5.4389.54 ± 3.45
1220/121371.27 ± 7.6977.96 ± 6.6867.07 ± 6.0077.40 ± 3.5380.77 ± 6.1488.43 ± 0.44
1320/44960.51 ± 4.4587.97 ± 6.3945.77 ± 8.6679.84 ± 12.4978.43 ± 3.9088.69 ± 3.53
1420/40884.39 ± 2.3594.28 ± 6.0384.19 ± 0.1790.23 ± 1.8090.92 ± 3.0891.92 ± 0.95
1520/64067.34 ± 7.5988.33 ± 4.4284.93 ± 2.7689.47 ± 4.8392.39 ± 4.9992.83 ± 3.45
OA(%)65.09 ± 1.6077.17 ± 0.7666.17 ± 0.9983.21 ± 0.9884.69 ± 0.8088.32 ± 0.34
AA(%)66.34 ± 1.4279.13 ± 1.0967.07 ± 1.6285.29 ± 1.3384.35 ± 0.7988.34 ± 0.28
κ × 10062.31 ± 1.7175.32 ± 0.8463.44 ± 1.0581.85 ± 1.0583.13 ± 0.8687.37 ± 0.37

Share and Cite

MDPI and ACS Style

Fang, B.; Li, Y.; Zhang, H.; Chan, J.C.-W. Hyperspectral Images Classification Based on Dense Convolutional Networks with Spectral-Wise Attention Mechanism. Remote Sens. 2019, 11, 159. https://doi.org/10.3390/rs11020159

AMA Style

Fang B, Li Y, Zhang H, Chan JC-W. Hyperspectral Images Classification Based on Dense Convolutional Networks with Spectral-Wise Attention Mechanism. Remote Sensing. 2019; 11(2):159. https://doi.org/10.3390/rs11020159

Chicago/Turabian Style

Fang, Bei, Ying Li, Haokui Zhang, and Jonathan Cheung-Wai Chan. 2019. "Hyperspectral Images Classification Based on Dense Convolutional Networks with Spectral-Wise Attention Mechanism" Remote Sensing 11, no. 2: 159. https://doi.org/10.3390/rs11020159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop