Next Article in Journal
Human Activity Recognition in a Free-Living Environment Using an Ear-Worn Motion Sensor
Previous Article in Journal
Reengineering Indoor Air Quality Monitoring Systems to Improve End-User Experience
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Transfer Learning-Based Hyperspectral Image Classification Using Residual Dense Connection Networks

1
School of Information Science and Engineering, Yunnan University, Kunming 650504, China
2
Yunnan Power Grid Co., Ltd., Kunming 650011, China
*
Author to whom correspondence should be addressed.
Submission received: 28 February 2024 / Revised: 14 April 2024 / Accepted: 16 April 2024 / Published: 23 April 2024
(This article belongs to the Topic Hyperspectral Imaging and Signal Processing)

Abstract

:
The extraction of effective classification features from high-dimensional hyperspectral images, impeded by the scarcity of labeled samples and uneven sample distribution, represents a formidable challenge within hyperspectral image classification. Traditional few-shot learning methods confront the dual dilemma of limited annotated samples and the necessity for deeper, more effective features from complex hyperspectral data, often resulting in suboptimal outcomes. The prohibitive cost of sample annotation further exacerbates the challenge, making it difficult to rely on a scant number of annotated samples for effective feature extraction. Prevailing high-accuracy algorithms require abundant annotated samples and falter in deriving deep, discriminative features from limited data, compromising classification performance for complex substances. This paper advocates for an integration of advanced spectral–spatial feature extraction with meta-transfer learning to address the classification of hyperspectral signals amidst insufficient labeled samples. Initially trained on a source domain dataset with ample labels, the model undergoes transference to a target domain with minimal samples, utilizing dense connection blocks and tree-dimensional convolutional residual connections to enhance feature extraction and maximize spatial and spectral information retrieval. This approach, validated on three diverse hyperspectral datasets—IP, UP, and Salinas—significantly surpasses existing classification algorithms and small-sample techniques in accuracy, demonstrating its applicability to high-dimensional signal classification under label constraints.

1. Introduction

Hyperspectral imaging (HSI) systems amass extensive spatial and spectral data across a broad array of spectral bands, presenting a rich tapestry of information [1,2]. This bounty has catalyzed advancements across varied domains, such as precision agriculture [3], environmental surveillance [4,5], and disaster mitigation [6,7], signifying its interdisciplinary impact. The realm of hyperspectral image classification, a pivotal segment of hyperspectral analysis, has elicited considerable scholarly interest [8,9]. Yet, the classification endeavors for hyperspectral remote-sensing imagery confront persistent obstacles. A critical imperative lies in the more profound exploration of the intrinsic deep features within hyperspectral images. Addressing the paucity of training samples and enhancing classification efficacy in high-dimensional contexts with limited data remain pressing challenges. These hurdles underscore the substantial prospects for continued research and advancements in the field.
In traditional classification methods, the classification of hyperspectral images has focused on manual feature extraction [9,10,11,12] and the use of traditional shallow classifiers, including K-Nearest Neighbor (KNN) [13], Support Vector Machine (SVM) [14], logistic regression [15], the manifold learning method [16], among others. These conventional methods can only extract shallow feature information and neglect deep feature information. Classification performance relies significantly on prior knowledge, manual parameter adjustments, and feature selection. However, this approach lacks the adaptability required to address classification tasks in complex scenarios.
Deep learning methods possess the ability to acquire discriminative features from extensive annotated data and apply these features to classification tasks. As a result, deep learning methods have emerged as a promising approach to hyperspectral image (HSI) classification, offering substantial advantages over traditional methods. Chen et al. [17] utilized deep stacked autoencoders to extract spatial and spectral features from hyperspectral images. This approach effectively captured contextual spatial information and spectral information from HSIs, leading to a successful classification and good performance. To address the distinct characteristics of hyperspectral image data cubes, Li et al. [18] employed a 3D convolutional neural network (3D-CNN) for hyperspectral image classification. Thompson et al. used deep belief networks to extract features at a deep level for hyperspectral image classification [19]. Zhong et al. [20] introduced a supervised spectral–spatial residual network (SSRN) to iteratively acquire discriminative features from the abundant spectral characteristics and spatial contexts of hyperspectral images (HSI). The goal was to extract integrated spatial–spectral information and identify significant spectral–spatial features for classification purposes.
The performance of conventional supervised deep learning methods is based on a significant number of labeled samples for model training. Nevertheless, the exorbitant cost of annotation leads to a severely restricted number of labeled samples for hyperspectral images as a whole. Therefore, using traditional deep learning models for hyperspectral image classification with insufficient training samples can easily lead to overfitting and suboptimal classification performance. To overcome this challenge, researchers have proposed various approaches to tackle the issue of hyperspectral image classification in scenarios with limited sample sizes. Some approaches [21,22] employ data augmentation to generate additional training samples for deep learning models such as CNNs, thus expanding data size and improving the model’s generalizability. Several semi-supervised approaches [23,24] involve the combination of a limited number of labeled samples with unlabeled samples during training. These methods leverage the information from unlabeled samples to obtain feature representations that are more robust and highly generalized. Transfer learning-based approaches [25,26] employ a model that has been pre-trained on a large-scale dataset. The weights of the pre-trained model serve as initialization parameters, which are subsequently fine-tuned on a small sample dataset. By harnessing the feature extraction capabilities of the pre-trained model, this approach effectively enhances the classification performance using small-sample datasets.
Taking into account the challenges in hyperspectral image classification, the limited availability of labeled training samples in hyperspectral images poses a significant constraint on the learning and feature extraction capacity of deep neural network models. Furthermore, the high-dimensional characteristics of hyperspectral images present a challenge to models trained on a small number of annotated samples regarding the extraction of an adequate set of features. As a consequence, the extraction of intrinsic deep-level features from hyperspectral images becomes arduous, leading to a diminished classification accuracy in hyperspectral image classification tasks. Therefore, the construction of deep neural network models for hyperspectral image classification in scenarios with limited training samples poses a significant research challenge. We took into account that ResNet, through residual blocks, enables inter-layer connections that reinforce feature reuse and alleviate the vanishing gradient problem, and in DenseNet structures, each layer is directly connected to all subsequent layers, allowing the extraction of deeper features, further mitigating the vanishing gradient problem and effectively extracting deep features. Therefore, to more effectively extract deep features from the spectral and spatial dimensions of hyperspectral images under conditions of limited samples and to enhance the performance of hyperspectral classification, this paper presents a meta-transfer framework for few-shot hyperspectral image classification based on a three-dimensional Residual Dense Connection Network (ResDenseNet). The primary contributions of this paper are summarized as follows.
(1) The proposition of a meta-transfer few-shot learning classification (MFSC) method aimed at surmounting the hurdle of scarce annotated samples: The method employs a meta-learning training strategy to harmonize data from disparate class samples within a unified feature space, facilitating the prediction of categories for unlabeled samples through similarity between the support set and query set within this feature domain.
(2) The introduction of a novel hyperspectral image classification network, dubbed ResDenseNet, designed to address the underutilization of spectral and spatial information within hyperspectral images: This architecture synergizes the DenseNet (Densely Connected Convolutional Networks) [27,28] and ResNet (Residual Network) [29] frameworks. An enhanced spectral dense block is deployed for the assimilation of spatial–spectral features, complemented by a three-dimensional residual block for the further extraction of spatial and spectral attributes. Classification is achieved through a multilayer perceptron (MLP). The ResDenseNet architecture comprehensively mines deep features within the proximal space of samples, extracting more discriminative attributes to bolster the classification acumen of hyperspectral images.
The remainder of this study is structured as follows: Section 2 provides an overview of the existing cross-domain few-shot hyperspectral classification algorithm for transfer learning. In Section 3, we present the framework of our proposed MFSC approach, which aims to tackle the issue of limited labeled samples in hyperspectral images. Section 4 presents the experimental results of our methods, along with our analysis. Finally, Section 5 concludes our work.

2. Related Work

In the context of transfer learning [30,31,32], a model is initially trained on a source dataset, which comprises abundant annotated data from multiple classes known as source classes. Subsequently, the model parameters and features are then adapted to the target dataset with a limited number of labeled samples, where the classes are non-overlapping. This process allows the model to be transferred and adjusted to handle the target dataset, which contains only a small number of labeled samples. Koch et al. [32] proposed an early technique known as Deep Convolutional Siamese Networks. This method performs feature extraction on a pair of samples using the same network and employs the Euclidean distance to measure similarity for classification. However, despite its simplicity and intuitiveness, this approach often fails to achieve satisfactory results in complex scenarios. Based on this, Vinyals et al. [33] introduced Matching Networks, which integrate bidirectional LSTM networks with feature metric learning. By calculating the cosine distance between output features, it captures the similarity between support set and query set images, thereby achieving the classification objective. Nevertheless, this approach encounters difficulties when dealing with intricate and irregular spatial structures. Although this method performs well when the distribution of the source domain data is close to that of the target domain data, existing transfer learning methods struggle to effectively generalize the model from the source domain to the target domain when there is a significant difference in data distributions. Therefore, research is conducted on cross-domain small-sample classification techniques for situations where the source and target domain data distributions differ, aiming to bolster the transfer learning model’s capacity for generalization.
To address the challenges posed by cross-task learning, researchers have proposed a range of meta-learning techniques [34,35,36,37,38,39], which can be classified into two main categories: metric-based and optimization-based approaches. Metric-based methods focus on acquiring a robust feature space by employing the Euclidean distance to gauge the likeness between unlabeled samples and labeled samples of each class. Conversely, optimization-based meta-learning strategies aim to train a universal model capable of swiftly converging to an effective solution for new tasks through a limited number of gradient descent iterations. Nevertheless, when dealing with scant training samples, these methods are susceptible to overfitting, and their weight-update process tends to be relatively sluggish. Consequently, there is a pressing need to enhance and refine these meta-learning techniques to ensure their practicality and efficacy within the realm of few-shot learning.
On the other hand, given the high-dimensional characteristics of hyperspectral images, combining more efficient hyperspectral feature extraction methods with small-sample learning techniques has become a pivotal approach to tackle the challenge of limited annotated samples in hyperspectral data. Liu et al. [39] introduced a Deep Few-shot Learning (DFSL) method that explores the impact of various feature extraction methods on the metric space for classification outcomes. However, this approach still faces limitations when dealing with similarity issues within the metric space. Reference [40] proposes a novel and compact framework based on the Transformer, called the Spectral Coordinate Transformer (SCFormer), which employs two mask patterns (Random Mask and Sequential Mask) in SCFormer-R and SCFormer-S, respectively, aiming to generate more distinguishable spectral features using the existing spectral priors.
To tackle the challenges posed by the characteristics of high-dimensional data features in hyperspectral images and the limited number of labeled training samples, which make it difficult to thoroughly explore the deep-level features of hyperspectral images and subsequently result in suboptimal classification accuracy, this paper proposes a novel approach: the meta-transfer few-shot classification method. Furthermore, to enhance the classification of hyperspectral images, a residual dense connection network is introduced. On the one hand, this method facilitates the transfer of the transferable knowledge acquired from a source domain dataset to the target domain with a limited number of samples. This addresses the issue of restricted training samples that hinder the accuracy of classification in deep learning models. On the other hand, by taking advantage of the capabilities of the residual dense connection network, features are used more effectively, and the exchange of features between convolutional layers is intensified, ultimately contributing to an overall improvement in classification accuracy.

3. Proposed Meta-Transfer Hyperspectral Image Few-Shot Classification

3.1. Proposed MFSC Framework

The entire process flow diagram is shown in Figure 1. It comprises two main components: the cross-domain few-shot learning strategy and the residual dense connection feature extraction and classification network. Arrows indicate the flow of feature vectors in the algorithm, with red arrows representing feature vectors originating from the target domain, while black arrows represent feature vectors coming from the source domain.
The few-shot learning strategy, based on metric learning-based meta-transfer, leverages the transferrable feature knowledge trained from the source domain dataset and transfers this knowledge to the target domain with a small number of labeled samples. These two types of small-sample learning are conducted simultaneously. Model weights trained on the source domain dataset are used to initialize the weights of the feature extraction network. This is performed to enhance the hyperspectral image (HSI) classification accuracy, addressing the issue of limited training samples that constrain the classification accuracy in deep learning models.
By utilizing the mapping layer and the residual dense connection network, features from the source domain and the target domain are mapped to a feature space. This ensures that samples from the same class have a similar distribution in the feature space, while samples from different classes are distributed as far apart as possible in the feature space. The residual dense connection network allows for the more comprehensive extraction of spatial–spectral features and enhances direct feature transfer between convolutional layers, thus improving classification accuracy.

3.2. Cross-Domain Few-Shot Learning and Training Strategy

The entire process flowchart for the few-shot learning is shown in Figure 2. Training of the few-shot learning model consists of two stages. First, a set of data called source class data are used to train the model, with this class having an abundant number of samples. Then, training and testing are carried out on the target class data, where the classes do not overlap and only a small number of labeled samples are available. These two stages alternate until the model converges.
From the original HSI datasets of the source and target classes, C classes are randomly selected from each, with each class containing K labeled samples to create the source domain support set S s = { ( x s i , y s i ) } i = 1 C × K and the target domain support set S t = { ( x t i , y t i ) } i = 1 C × K . Then, N unlabeled samples are randomly selected from the remaining data in both the source and target domains to create the source domain query set, Q s = ( x s j , y s j ) j = 1 C × N , and the target domain query set, Q t = ( x t j , y t j ) j = 1 C × N . This entire selection process is referred to as a C-way K-shot task. Each time the support and query sets are selected for model training, it constitutes an episode.
In each training episode, during the training cycle, the model is first trained on the source domain dataset. The source domain support set, S s = { ( x s i , y s i ) } i = 1 C × K , is fed into the network to extract features, and the feature vectors, c s k , for the k-th class in the support set in the feature space are computed. The source domain query set samples, x s j , are then passed through the feature network to extract embedded features, f φ ( x s j ) . The Euclidean distance, d f φ x s j , c s k , between the feature vectors of the query set samples, x s j , and the feature vector, c s k , of the class to which the support set samples belong in the feature space is calculated [41]. Subsequently, the probability that a query set sample, x s j , belongs to class k in the support set is computed using the following the SoftMax function:
P y j = k x s j Q S = exp d f φ x s j , c s k k = 1 C exp d f φ x s j , c s k
In each episode, during the training process, f φ represents a mapping layer and a spatial–spectral feature extraction network with learnable parameters denoted as φ , y j represents the true class labels of the samples x s j , and C is the number of classes in each episode. The training loss in each episode is calculated as the sum of the negative log probabilities between all query set samples and their corresponding true class labels:
L s = x s j , y s j Q s log p φ y s j = k | x s j
Then, the model continues training using the target domain data. The support set data, S t = { ( x t i , y t i ) } i = 1 C × K , from the target domain dataset are fed into the model trained on the source domain data. This calculates the feature vector, ctk, for the k-th class in the feature space. Similarly, the samples xtj from the target domain query set, Q t = ( x t j , y t j ) j = 1 C × N , are input into the feature extraction network, extracting embedded features, f φ ( x t j ) , for the query set samples. The Euclidean distance, d f φ x t j , c t k , between the query sample xtj and the feature vectors of the samples belonging to class k in the feature space is computed. The probability that the query sample xtj belongs to class k is calculated through the SoftMax function. On this basis, the loss value for the query sample is also computed.
L t = x t j , y t j Q t log p φ y t j = k | x t j
The data from the source domain and the target domain are randomly selected to form a training dataset that includes support and query sets. The model is trained by minimizing the loss function and optimizing the parameters of the model. This ensures that the features f φ ( x s j ) and f φ ( x t i ) of the query samples from the source domain and target domain, respectively, are as close as possible to the corresponding support set features, c s k and c t k , for that sample. The minimization of the loss function, J ( φ ) , is calculated using Equation (4).
J ( φ ) = l o g P ( y j = k | x j Q s ) = d ( f φ ( x j ) , c k ) + l o g k = 1 C e ( d ( f φ ( x j ) , c k ) )
After multiple rounds of training with multiple episodes and models, when the loss function in the target domain meets the termination condition, the training is concluded.

3.3. Spatial–Spectral Feature Extraction Module Based on ResDenseNet Network

The proposed algorithm workflow is illustrated in Figure 1, which shows the MFSC framework. It mainly consists of three parts: the mapping layer module, the ResDenseNet feature extractor, and the multilayer perceptron module.

3.3.1. Mapping Layer Module

In the mapping module, first 9 × 9 × S c data cubes, D S , are selected from the source dataset as the network’s input, where 9 × 9 represents the spatial dimensions, and S c represents the number of spectral bands. For the target domain dataset, 9 × 9 × T C data cubes, D T , are selected as input for network testing, where T C represents the number of spectral bands. Mapping layers are used to reduce the dimensionality of the input samples, ensuring that the input dimensions are the same. Due to the large number of spectral bands in HSI and strong correlations between adjacent bands, mapping layers use a 1 × 1 × 100 convolutional kernel to reduce the number of spectral bands in both the source and target domains, reducing the data to 100 dimensions for convenience in subsequent convolution calculations. The final output of the mapping layer is a support feature vector or a query set feature vector with a size of 9 × 9 × 100 .

3.3.2. ResDenseNet Feature Extractor

The ResDenseNet feature extractor is used as the spatial–spectral feature extraction network; it consists mainly of a DenseNet module and ResidualNet module. In order to address the loss of feature information due to gradient vanishing, amplify feature propagation, and extract feature vectors more effectively, the algorithm initially employs DenseNet module for model training.
The spectral dense block consists of four sets of convolutional kernels, with each set containing 8 filters of size 3 × 3 × 3 . These are combined with Mish activation functions and batch normalization (BN) to perform non-linear transformations on the feature maps. In DenseNet, each layer is concatenated with all preceding layers along the channel dimension, combining feature maps from all previous layers as input for the next layer to achieve feature reuse and enhance efficiency:
D X l = D H D X 0 , D X 1 , D X l 1
where D H is a non-linear transformation function, which uses the structure of Convolution 3 × 3 × 3 (Conv), batch normalization (BN), Mish, and concatenation operations. The subscript l denotes the layer number. The ReLU function causes some neurons to have an output of 0, resulting in network sparsity, and the Mish [42] function, f x = x t a n h l n 1 + e x , unlike the ReLU function, has a softer zero boundary and smoother characteristics, allowing for a better flow of information into deep neural networks and better preservation of information, thus producing enhanced accuracy and generalization. The output of the function is not affected by saturation, and positive values can reach arbitrarily high values, avoiding saturation due to a cap. Therefore, Mish is used as the activation function in this paper. The output feature map from the last layer of dense connection block undergoes average pooling, yielding a vector, DenseFV, of dimensions 8 × 7 × 7 × 100 . Subsequently, this vector is fed into the three-dimensional ResidualNet module.
In the ResidualNet module, there are four sets of non-linear transformation functions. Each set of non-linear transformation functions includes 16 filters of size 3 × 3 × 3, batch normalization (BN), and Mish activation. It employs a shortcut connection structure, creating a skip connection between the input of the first layer and the output of the last layer. This design allows the network to concentrate on learning the disparity between input and output, streamlining the learning objectives and challenges. The output feature map of the residual block is of size 16 × 7 × 7 × 100. After undergoing average pooling, max pooling, and a set of 32 filters of size 3 × 3 × 3 , the feature map is flattened to a 1 × 1 × 160 vector (ResidualFV). This vector is then processed through a fully connected layer and a SoftMax activation function. Additionally, it undergoes a multilinear mapping as input to the MLP. The number of nodes in the fully connected layer corresponds to the number of classes in the dataset.

3.3.3. Multilayer Perceptron Module

The ultimately extracted feature vector from the multilinear mapping is fed into the MLP for classification. This MLP consists of five fully connected layers, with the first four layers each containing 1024 nodes. The final fully connected layer has only one node. ReLU activation functions and dropout are incorporated between adjacent fully connected layers. The ultimate output of the multilayer perceptron is employed to compute the loss value following Formula (4), after which the classification process is executed.
Through training, the loss function of the spatial–spectral feature extraction network model is minimized. This optimization of parameters in the residual dense connection module allows it to extract features from the input sample data, mapping them into feature space. In this space, the feature vectors of samples with the same class are closer to each other, resulting in smaller interclass distances, while the feature vectors of samples from different classes are farther apart, leading to larger interclass distances.

4. Experiments

4.1. Experimental Dataset

To validate the effectiveness of our approach, we utilized the hyperspectral Chikusei dataset as the source domain dataset, and the Indian Pines, Pavia University, and Salinas datasets [43,44] as the target domain datasets. The pseudo-color images and real land cover maps of this experimental dataset are shown in Figure 3 and Figure 4.
The Chikusei dataset has a spectral wavelength range of 343–1080 nm, a spatial resolution of approximately 2.5 m, and a data size of 2571 × 2335 pixels. It consists of 128 spectral bands and includes 77,592 ground pixels, categorized into 19 distinct land cover classes.
The Indian Pines dataset covers a spectral wavelength range of 400–2500 nm, with a spatial resolution of about 20 m. The image data size is 145 × 145 pixels and comprises 200 spectral bands. It encompasses a total of 16 land cover classes. The Salinas dataset has a spectral wavelength range of 400–2500 nm and a spatial resolution of approximately 3.7 m. The image size for this dataset is 512 × 217 pixels and includes 224 spectral bands. However, due to the impact of water vapor absorption on certain bands, only 204 bands are retained. This dataset covers 16 different categories of agricultural land cover, including, but not limited to, corn, wheat, soybeans, grasslands, and vineyards. The Pavia University dataset’s spectral wavelength range is 430–860 nm, with a spatial resolution of approximately 1.3 m. After preprocessing, the dataset has a total of 115 spectral bands, with 13 noisy bands removed. Land cover types in this region consist of nine classes, including asphalt roads, meadows, gravel, trees, metals, bare land, asphalt roofs, bricks, and shadows.

4.2. Experimental Settings

To evaluate the effectiveness of the MFSC method, 9 × 9 × C data cubes were selected as input for the network from the Chikusei source domain dataset, where 9 × 9 represents the spatial dimensions, and C is the number of spectral bands. For the target domain datasets, namely Indian Pines, Pavia University, and Salinas, 9 × 9 × L cubes were chosen as the input for testing, where L is the number of spectral bands. The model was trained for 10,000 episodes, and for each episode iteration, following the few-shot training method, 1 labeled sample and 19 unlabeled samples from each class were randomly selected to form the source dataset for model training. The Adam optimizer was used, and to balance convergence speed and accuracy, the model learning rate was set to 0.001. Furthermore, to account for the impact of random sample selection on model training, all experimental results were averaged over 10 trials. The hardware environment used for this experiment is a laptop equipped with an Intel Core i7-4810MQ 8-core 2.80 GHz processor, 16 GB of memory, and an NVIDIA GeForce RTX 2060 graphics card with 6 GB RAM, while the software environment utilized Python 3.8 and PyTorch 1.7.1 running on Windows 10.

4.3. Experimental Results and Analysis

To validate the effectiveness of the proposed method in the paper, it was compared with non-few-shot learning methods and few-shot learning methods. In experiments comparing the proposed method with non-few-shot learning methods, the proposed method was compared with SVM, 3D-CNN [45], and SSRN [46]. In experiments comparing the proposed method with other few-shot learning methods, the proposed method was compared with the DFSL + NN [37], DFSL + SVM [47,48], RN-FSL [49], Gai-CFSL [50], DPGN [51], DCFSL [52], SCFormer-R, and SCFormer-S [41] methods. In each comparison experiment, the same training approach as the few-shot methods was employed. Five labeled samples from each class in the target domain dataset were randomly selected for transferring the model trained in the source domain to the target domain, with the remaining target domain samples used as test data. For the small-sample learning methods in comparison, we randomly selected 200 labeled source domain samples from each class to learn transferable knowledge, following the same setup for comparison. To verify the effectiveness of the Mish function and batch normalization (BN) added to the model in the paper, a comparative performance analysis was performed using the DCFSL method. In this comparison, the Mish + BN part was removed, while keeping the rest of the network structure consistent, serving as a set of ablation experiments. The results of the ablation experiments are presented in the “MFSC” row of the tables, where the activation function used is the Softmax activation function, consistent with the DCFSL method. In contrast, the experimental data in the “Ours” row were obtained under the MFSC algorithm framework, incorporating Mish + BN and replacing the original Softmax activation function. For the IP, UP, and Salinas datasets, the study compared the classification performance of different methods. The evaluation was carried out using three metrics: overall accuracy (OA), average accuracy (AA), and Kappa coefficient. Specific comparative results are shown in Table 1, Table 2 and Table 3.
Table 1, Table 2 and Table 3 present the results of comparative experiments on the target datasets, IP, UP, and Salinas, with each class having five labeled samples. From the tables, it can be observed that the methods based on few-shot learning achieve higher overall accuracy compared to non-few-shot methods. This indicates that the episodic training strategy is better suited for classification tasks with limited labeled samples. In the IP dataset, the proposed few-shot learning method shows significant improvements over the traditional SVM classification method, with an increase of 25.64% in OA, 21.95% in AA, and a 28.13% increase in Kappa. In the IP, UP, and Salinas datasets, when compared to deep learning-based methods like 3D-CNN and SSRN, the proposed method achieves significant increases in OA when the number of labeled samples is five, with improvements of 16.73%, 19.35% and 6.34% in IP; and 10.13%, 8.83%, and 4.15% in UP and Salinas, respectively. This indicates that the meta-learning training strategy allows the model to learn transferable knowledge and features from the source-class data, thus aiding in predicting the target-class data. The relatively low performance of the non-few-shot learning methods shown in Table 1, Table 2 and Table 3 illustrates that non-small-sample learning methods extract shallow features with weaker discriminative capabilities for different target categories. The limited labeled samples are insufficient for non-small-sample learning methods to effectively train a classification model. However, meta-learning training strategies enable the model to learn transferable knowledge and features from the source-class data, aiding in predicting target-class data.
In the few-shot classification methods, the method proposed in this paper also demonstrates significant improvements in detection accuracy compared to other methods. On the IP, UP, and Salinas datasets, when compared to the DFSL + NN, DFSL + SVM, RN-FSL, Gai-CFSL, DCFSL, SCFormer-R, and SCFormer-S methods, the proposed method achieves improvements in OA of 12.95%, 10.91%, 14.43%, 8.83%, 5.79%, 7.59%, and 7.65% on IP; 8.27%, 6.39%, 5.84%, 2.9%, 2.37%, 3.71%, and 2.19% on UP; and 3.92%, 4.02%, 6.86%, 3.14%, 1.63% 1.67%, and 2.15% on Salinas, respectively, when there are few labeled samples in the target domain. With the presence of a small number of labeled samples in the target domain, the method proposed in this article utilizes the ResDenseNet network to reduce data distribution differences and learn more discriminative feature spaces. Compared to other methods, this approach can obtain a better feature space, which can improve the classification performance of the target domain samples. The classification results on the IP, UP, and Salinas datasets show that the proposed method achieves average accuracy (OA) of 72.60%, 86.02%, and 90.97%, respectively. This strongly confirms the effectiveness and robustness of the ResDenseNet model in the few-shot high-dimensional spectral data classification task. Additionally, the incorporation of the Mish function and batch normalization (BN) not only effectively mitigates the vanishing gradient problem but also enhances the model’s generalization capabilities. Furthermore, compared to the ReLU function, the Mish function is smoother, leading to an improvement in training stability and average accuracy.
Table 4, Table 5 and Table 6 report the detailed classification results of different classification algorithms on the UP, IP, and Salinas datasets, respectively. The last columns of the tables present the classification accuracy and standard deviation for each class in the dataset based on multiple experiments. It can be observed from Table 4 that, compared to other algorithms, the proposed method achieved the highest recognition rates in three of nine categories. It also performed well in accurately classifying the “Bricks”, “Bitumen”, “Metal sheets”, and “Trees” categories, which were challenging for other methods. The proposed method shows a certain gap from the optimal results among the three categories, including “Gravel”, “Meadows”, and “Asphalt” in the UP dataset, when compared to the methods of contrast. The UP dataset has the highest spatial resolution among the three datasets, but it has the lowest spectral resolution. The data for the three categories are the most prone to generating spectrally similar but different substances. The data in Table 5 and Table 6 illustrate that, compared to other algorithms, the method proposed in the paper achieved the highest recognition rates in 11 out of 16 categories and 10 out of 16 categories, respectively. It significantly improved the classification accuracy for categories like “Grapes_untrained”, “Vinyard_untrained”, and “Soil_vinyard_develop” in the Salines dataset, where other methods had relatively lower accuracy. Furthermore, compared to other methods, the proposed method also substantially increased the classification accuracy of categories like “Grass-pasture”, “Corn”, “Corn-mintill”, “Corn-notill”, and "Woods” in the IP dataset.
Figure 5, Figure 6 and Figure 7 display the classification results of the proposed method and comparative methods using the IP, UP, and Salinas datasets. It can be seen from the figures that the method proposed in this paper exhibits fewer misclassifications. On the contrary, the SVM-based method shows more misclassified objects. Compared to the SVM-based method, the 3D-CNN and SSRN methods have fewer misclassifications, mainly due to the stronger representation learning capabilities of deep learning methods. However, deep learning methods require a large number of training samples, and when the number of training samples is reduced, these methods experience a significant decrease in classification accuracy. This indicates that, when labeled samples are limited, the extracted features are not effective enough, leading to lower accuracy when classifying objects with similar spectral characteristics. In the case of few-shot data, using a few-shot learning approach to construct ResDenseNet significantly improves the classification accuracy compared to the SVM method and deep learning methods like 3D-CNN and SSRN.
In complex scenes, objects within a specific area are rarely composed of just one type of material. Typically, there are varying amounts of other material categories present, leading to spectral noise from other categories within the spectral characteristics of the primary material. Additionally, at the boundaries between two different land cover types, there inevitably exists interference from neighboring land cover categories’ spectral feature vectors. This makes it difficult to accurately extract both the spatial and spectral information of land cover, resulting in subtle differences between different types of land cover. In addition, it can lead to significant distinctions between the same types of land cover, causing the misclassification of certain land cover areas at the boundaries. In the case of few-shot data, while methods like DFSL + NN, DFSL + SVM, and RN-FSC consider the scarcity of labeled samples in hyperspectral imagery, their performance in accurately classifying challenging classes still lags behind the method proposed in this paper.
From the experimental results shown in the figures, it can be observed that when land cover features are relatively easy to distinguish and the feature vectors are distinct, the classification method employed in this paper, as well as other few-shot learning methods, can achieve good classification results. For example, in Figure 5, for the IP dataset, classes like “Oats” and “Grass-Trees”; in Figure 6, for the UP dataset, classes like “Asphalt” and “Shadow”; and in Figure 7, for the Salinas dataset, classes like “Celery”, “Stubble”, “Fallow_smooth”, “Lettuce_romaince_5wk”, and “Brocoil_green_weeds_1” have feature vectors in the feature space that are relatively easy to differentiate. In situations with only a small number of labeled samples, traditional machine learning methods, such as SVM, and general few-shot learning methods can also achieve good classification results. On the contrary, deep learning methods that require a large number of training samples are prone to overfitting, leading to a lower classification accuracy.
For land cover categories with similar features and small feature vector distances that tend to produce errors in classification, such as “Meadows” and “Alfalfa” in the UP dataset; “Vinyard_untrained”, “Vinyard_vertical_trellis”, and “Corn_senesced_green_weeds” in the Salinas dataset; and “Stone-Steel-Tower”, “Hay-windrowed”, “Woods”, and “Soybean-mintill” in the IP dataset, the classification results rely more on the effective extraction of land cover features. From the classification results, it can be seen that the method proposed in this paper achieves a relatively good classification accuracy for such categories. MFSC follows, and DCFSL has fewer misclassifications compared to SVM, 3D-CNN, and SSRN. This indicates, on the one hand, that meta-learning training strategies are advantageous for enhanced knowledge transfer and improved classification performance. On the other hand, it also demonstrates that the residual dense connection network designed in this paper can reduce data distribution differences, leading to a better feature space with higher interclass discriminability. Under small-sample training conditions, the training data’s effectiveness and robustness are superior to those of other methods. Furthermore, the method proposed in this paper has fewer misclassification points than DCFSL, indicating that this network model has good generalizability, can extract deeper and more discriminative features, and can achieve better classification results for classes that are difficult to accurately classify.

5. Conclusions

To address the contradiction between the limited number of training samples in HSI (hyperspectral imaging) and the need for a large number of annotated samples for effective deep learning, as well as the trade-off between a small number of labeled samples and the extraction of more effective feature vectors, this paper proposes a hyperspectral image classification method based on the residual dense connection network in the metric learning framework. The main contributions are as follows:
Improved ResDenseNet Network: In comparison to traditional residual networks, this paper introduces a dense connection structure in the three-dimensional convolutional block of the improved ResDenseNet network. This structure can fully explore deep features in the spatial neighborhood of samples, effectively extract spatial and spectral features, and complement the original spectral features. It can obtain more representative features, contributing to hyperspectral images classification.
Activation function and batch normalization: Building on the original network, the ReLU activation function is replaced with the Mish function, and batch normalization (BN) is introduced. This not only effectively alleviates the problem of gradient vanishing; it also enhances the model’s generalization ability. Additionally, compared to the ReLU function, the Mish function is smoother, leading to improvement in training stability and average accuracy.
The experimental results demonstrate that the proposed method, when compared to classical hyperspectral image classification methods and other classic few-shot learning methods, exhibits strong generalization capabilities in deep network models on three datasets: IP, UP, and Salinas. When only a limited number of labeled samples are available, the proposed method achieves a higher recognition accuracy than the algorithms used in the control experiments. Our future work will focus on accurately identifying ground objects in the presence of mixed substances, investigating Transformer learning strategies that can more effectively mine the spatial–spectral features of hyperspectral images, thereby enhancing the classification accuracy of complex ground objects.

Author Contributions

The scheme of this research was performed according to task-sharing among co-authors. H.Z. conceptualized the paper. X.W. wrote the original draft and designed the study. K.X. performed the classification. Y.M. performed the statistical analysis. G.Y. conducted the literature survey. All authors managed the analyses and discussions of the data and read, corrected, and approved the final manuscript before submission. G.Y. served as the corresponding author. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Yunnan Province Special Fund for Key Programs in Science and Technology, China, grant number 202202AD080004; and the Natural Science Foundation of China (Grant Nos. 62061049, 12263008).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

Author Yu Ma was employed by Yunnan Power Grid Co., Ltd. The remaining authors delcare that the research was conducted in the absence of any commercial or financial relationships that could be contrued as a potential confilict of interest.

References

  1. Zhang, G.; Cao, W.; Wei, Y. Spatial perception correntropy matrix for hyperspectral image classification. Appl. Sci. 2022, 12, 6797. [Google Scholar] [CrossRef]
  2. Wan, S.; Gong, C.; Zhong, P.; Du, B.; Zhang, L.; Yang, J. Multiscale dynamic graph convolutional network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3162–3177. [Google Scholar] [CrossRef]
  3. Jia, S.; Jiang, S.; Lin, Z.; Li, N.; Xu, M.; Yu, S. A survey: Deep learning for hyperspectral image classification with few labeled samples. Neurocomputing 2021, 448, 179–204. [Google Scholar] [CrossRef]
  4. Liang, L.; Di, L.; Zhang, L.; Deng, M.; Qin, Z.; Zhao, S.; Lin, H. Estimation of crop LAI using hyperspectral vegetation indices and a hybrid inversion method. Remote Sens. Environ. 2015, 165, 123–134. [Google Scholar] [CrossRef]
  5. Yang, X.; Yu, Y. Estimating soil salinity under various moisture conditions: An experimental study. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2525–2533. [Google Scholar] [CrossRef]
  6. Li, S.; Dian, R.; Fang, L.; Bioucas-Dias, J.M. Fusing hyperspectral and multispectral images via coupled sparse tensor factorization. IEEE Trans. Image Process. 2018, 27, 4118–4130. [Google Scholar] [CrossRef] [PubMed]
  7. Zhang, S.; Li, J.; Wu, Z.; Plaza, A. Spatial discontinuity-weighted sparse unmixing of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5767–5779. [Google Scholar] [CrossRef]
  8. Tong, Q.X.; Zhang, B.; Zhang, L.F. Current progress of hyperspectral remote sensing in China. J. Remote Sens. 2016, 20, 689–707. [Google Scholar]
  9. Jia, S.; Hu, J.; Zhu, J.; Jia, X.; Li, Q. Three-dimensional local binary patterns for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2399–2413. [Google Scholar] [CrossRef]
  10. Li, Y.; Li, Q.; Liu, Y.; Xie, W. A spatial-spectral SIFT for hyperspectral image matching and classification. Pattern Recognit. Lett. 2019, 127, 18–26. [Google Scholar] [CrossRef]
  11. Wang, Q.; Zhang, F.; Li, X. Optimal clustering framework for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5910–5922. [Google Scholar] [CrossRef]
  12. Wang, Q.; He, X.; Li, X. Locality and structure regularized low rank representation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 911–923. [Google Scholar] [CrossRef]
  13. Ma, L.; Crawford, M.M.; Tian, J. Local manifold learning-based k-nearest-neighbor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4099–4109. [Google Scholar] [CrossRef]
  14. Kuo, B.C.; Huang, C.S.; Hung, C.C.; Liu, Y.L.; Chen, I.L. Spatial information based support vector machine for hyperspectral image classification. In Proceedings of the 2010 IEEE International Geoscience and Remote Sensing Symposium, IEEE, Honolulu, HI, USA, 25–30 July 2010; pp. 832–835. [Google Scholar]
  15. Ren, Y.; Zhang, Y.; Li, L. A spectral-spatial hyperspectral data classification approach using random forest with label constraints. In Proceedings of the 2014 IEEE Workshop on Electronics, Computer and Applications, Ottawa, ON, Canada, 8–9 May 2014; pp. 344–347. [Google Scholar]
  16. Wang, J.X.; Chen, S.B.; Ding, C.H.; Tang, J.; Luo, B. RanPaste: Paste consistency and pseudo label for semi-supervised remote sensing image semantic segmentation. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  17. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  18. Li, Y.; Zhang, H.; Shen, Q. Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  19. Thompson, S.; Teixeira-Dias, F.; Paulino, M.; Hamilton, A. Ballistic response of armour plates using generative adversarial networks. Def. Technol. 2022, 18, 1513–1522. [Google Scholar] [CrossRef]
  20. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  21. Huang, L.; Chen, Y. Dual-path siamese CNN for hyperspectral image classification with limited training samples. IEEE Geosci. Remote Sens. Lett. 2020, 18, 518–522. [Google Scholar] [CrossRef]
  22. Shendryk, Y.; Rist, Y.; Ticehurst, C.; Thorburn, P. Deep learning for multi-modal classification of cloud, shadow and land cover scenes in PlanetScope and Sentinel-2 imagery. ISPRS J. Photogramm. Remote Sens. 2019, 157, 124–136. [Google Scholar] [CrossRef]
  23. Zheng, X.; Jia, J.; Chen, J.; Guo, S.; Sun, L.; Zhou, C.; Wang, Y. Hyperspectral image classification with imbalanced data based on semi-supervised learning. Appl. Sci. 2022, 12, 3943. [Google Scholar] [CrossRef]
  24. Yang, Y.; Tang, X.; Zhang, X.; Ma, J.; Liu, F.; Jia, X.; Jiao, L. Semi-supervised multiscale dynamic graph convolution network for hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–15. [Google Scholar] [CrossRef] [PubMed]
  25. Datta, D.; Mallick, P.K.; Bhoi, A.K.; Ijaz, M.F.; Shafi, J.; Choi, J. Hyperspectral image classification: Potentials, challenges, and future directions. Comput. Intell. Neurosci. 2022, 2022, 3854635. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, X.; Tan, K.; Du, P.; Pan, C.; Ding, J. A unified multiscale learning framework for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4508319. [Google Scholar] [CrossRef]
  27. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  28. Wang, X.; Fan, Y. Multiscale densely connected attention network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 1617–1628. [Google Scholar] [CrossRef]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. Dhillon, G.S.; Chaudhari, P.; Ravichandran, A.; Soatto, S. A baseline for few-shot image classification. arXiv 2019, arXiv:1909.02729. [Google Scholar]
  31. Mathivanan, P.; Maran, P. Color image encryption based on novel kolam scrambling and modified 2D logistic cascade map (2D LCM). J. Supercomput. 2024, 80, 2164–2195. [Google Scholar] [CrossRef]
  32. Devabathini, N.J.; Mathivanan, P. Sign Language Recognition Through Video Frame Feature Extraction using Transfer Learning and Neural Networks. In Proceedings of the 2023 International Conference on Next Generation Electronics (NEleX), Vellore, Tamil Nadu, India, 14–16 December 2023; pp. 1–6. [Google Scholar]
  33. Koch, G.; Zemel, R.; Salakhutdinov, R. Siamese neural networks for one-shot image recognition. In Proceedings of the ICML Deep Learning Workshop, Lille, France, 6–11 July 2015; Volume 2. [Google Scholar]
  34. Vinyals, O.; Blundell, C.; Lillicrap, T.; Wierstra, D. Matching networks for one shot learning. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
  35. Ren, M.; Triantafillou, E.; Ravi, S.; Snell, J.; Swersky, K.; Tenenbaum, J.B.; Larochelle, H.; Zemel, R.S. Meta-learning for semi-supervised few-shot classification. arXiv 2018, arXiv:1803.00676. [Google Scholar]
  36. Munkhdalai, T.; Yu, H. Meta networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 2554–2563. [Google Scholar]
  37. Sun, Q.; Liu, Y.; Chua, T.S.; Schiele, B. Meta-transfer learning for few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 403–412. [Google Scholar]
  38. Yu, M.; Guo, X.; Yi, J.; Chang, S.; Potdar, S.; Cheng, Y.; Tesauro, G.; Wang, H.; Zhou, B. Diverse few-shot text classification with multiple metrics. arXiv 2018, arXiv:1805.07513. [Google Scholar]
  39. Liu, Y.; Sun, Q.; Liu, A.A.; Su, Y.; Schiele, B.; Chua, T.S. LCC: Learning to customize and combine neural networks for few-shot learning. arXiv 2019, arXiv:1904.08479. [Google Scholar]
  40. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep few-shot learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 2290–2304. [Google Scholar] [CrossRef]
  41. Li, J.; Zhang, Z.; Song, R.; Li, Y.; Du, Q. SCFormer: Spectral Coordinate Transformer for Cross-Domain Few-Shot Hyperspectral Image Classification. IEEE Trans. Image Process. 2024, 33, 840–855. [Google Scholar] [CrossRef] [PubMed]
  42. Zhang, S.; Chen, Z.; Wang, D.; Wang, Z.J. Cross-Domain Few-Shot Contrastive Learning for Hyperspectral Images Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5514505. [Google Scholar] [CrossRef]
  43. Misra, D. Mish: A self regularized non-monotonic activation function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
  44. Indian Pines Dataset. Available online: https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 17 November 2023).
  45. Yokoya, N.; Iwasaki, A. Airborne Hyperspectral Data over Chikusei; SAL-2016-05-27; Technical Report; Space Application Laboratory, The University of Tokyo: Tokyo, Japan, 2016. [Google Scholar]
  46. Lee, H.; Kwon, H. Going deeper with contextual CNN for hyperspectral image classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef] [PubMed]
  47. Liu, B.; Yu, A.; Gao, K.; Wang, Y.; Yu, X.; Zhang, P. Multiscale nested U-Net for small sample classification of hyperspectral images. J. Appl. Remote Sens. 2022, 16, 016506. [Google Scholar] [CrossRef]
  48. Peng, Y.; Liu, Y.; Tu, B.; Zhang, Y. Convolutional Transformer-Based Few-Shot Learning for Cross-Domain Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 1335–1349. [Google Scholar] [CrossRef]
  49. Gao, K.; Liu, B.; Yu, X.; Qin, J.; Zhang, P.; Tan, X. Deep relation network for hyperspectral image few-shot classification. Remote Sens. 2020, 12, 923. [Google Scholar] [CrossRef]
  50. Zhang, Y.; Li, W.; Zhang, M.; Wang, S.; Tao, R.; Du, Q. Graph information aggregation cross-domain few-shot learning for hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 1912–1925. [Google Scholar] [CrossRef]
  51. Yang, L.; Li, L.; Zhang, Z.; Zhou, X.; Zhou, E.; Liu, Y. Dpgn: Distribution propagation graph network for few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13390–13399. [Google Scholar]
  52. Li, Z.; Liu, M.; Chen, Y.; Xu, Y.; Li, W.; Du, Q. Deep Cross-Domain Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5501618. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed MFSC.
Figure 1. Framework of the proposed MFSC.
Sensors 24 02664 g001
Figure 2. Flowchart of the cross-domain few-shot learning algorithm.
Figure 2. Flowchart of the cross-domain few-shot learning algorithm.
Sensors 24 02664 g002
Figure 3. Chikusei and Indian Pines dataset. (a) False color image of the Chikusei dataset. (b) Ground-truth map of the Chikusei dataset. (c) False color image of Indian Pines dataset. (d) Ground-truth map of Indian Pines dataset.
Figure 3. Chikusei and Indian Pines dataset. (a) False color image of the Chikusei dataset. (b) Ground-truth map of the Chikusei dataset. (c) False color image of Indian Pines dataset. (d) Ground-truth map of Indian Pines dataset.
Sensors 24 02664 g003
Figure 4. Pavia University and Salinas dataset. (a) False color image of the Pavia University dataset. (b) Ground-truth map of the Pavia University dataset. (c) False color image of the Salinas dataset. (d) Ground truth map of the Salinas dataset.
Figure 4. Pavia University and Salinas dataset. (a) False color image of the Pavia University dataset. (b) Ground-truth map of the Pavia University dataset. (c) False color image of the Salinas dataset. (d) Ground truth map of the Salinas dataset.
Sensors 24 02664 g004
Figure 5. Classification results of different methods in Indian Pines dataset: (a) Gai-CFSL; (b) SVM; (c) 3D-CNN; (d) SSRN; (e) DFSL + NN; (f) DFSL + SVM; (g) RN-FSC; (h) DCFSL; (i) MFSC; and (j) MFSC, (Mish + BN) ours.
Figure 5. Classification results of different methods in Indian Pines dataset: (a) Gai-CFSL; (b) SVM; (c) 3D-CNN; (d) SSRN; (e) DFSL + NN; (f) DFSL + SVM; (g) RN-FSC; (h) DCFSL; (i) MFSC; and (j) MFSC, (Mish + BN) ours.
Sensors 24 02664 g005
Figure 6. Classification results of different methods in the Pavia University dataset: (a) Gai-CFSL; (b) SVM; (c) 3D-CNN; (d) SSRN; (e) DFSL + NN; (f) DFSL + SVM; (g) RN-FSC; (h) DCFSL; (i) MFSC; and (j) MFSC (Mish + BN), ours.
Figure 6. Classification results of different methods in the Pavia University dataset: (a) Gai-CFSL; (b) SVM; (c) 3D-CNN; (d) SSRN; (e) DFSL + NN; (f) DFSL + SVM; (g) RN-FSC; (h) DCFSL; (i) MFSC; and (j) MFSC (Mish + BN), ours.
Sensors 24 02664 g006
Figure 7. Classification results of different methods in the Salinas dataset: (a) Gai-CFSL; (b) SVM; (c) 3D-CNN; (d) SSRN; (e) DFSL + NN; (f) DFSL + SVM; (g) RN-FSC; (h) DCFSL; (i) MFSC; and (j) MFSC (Mish + BN), ours.
Figure 7. Classification results of different methods in the Salinas dataset: (a) Gai-CFSL; (b) SVM; (c) 3D-CNN; (d) SSRN; (e) DFSL + NN; (f) DFSL + SVM; (g) RN-FSC; (h) DCFSL; (i) MFSC; and (j) MFSC (Mish + BN), ours.
Sensors 24 02664 g007
Table 1. Comparison of the classification performance of different methods in Indian Pines datasets at number of labeled samples K = 5.
Table 1. Comparison of the classification performance of different methods in Indian Pines datasets at number of labeled samples K = 5.
MethodsOA (%)AA (%)Kappa × 100
Non-few-shot learningSVM45.8559.2439.68
3D-CNN54.7663.9348.72
SSRN61.3659.7556.91
Few-shot learningDFSL + NN59.6572.2454.55
DFSL + SVM61.6973.0556.78
RN-FSC58.1769.9052.52
Gai-CFSL63.7774.9859.20
DCFSL66.8177.8962.64
SCFormer-R65.0174.6560.20
SCFormer-S64.9575.5960.31
MFSC71.4981.1967.81
MFSC (Mish + BN) Ours72.6081.6269.16
Table 2. Comparison of the classification performance of different methods in Pavia University datasets at number of labeled samples K = 5.
Table 2. Comparison of the classification performance of different methods in Pavia University datasets at number of labeled samples K = 5.
MethodsOA (%)AA (%)Kappa × 100
Non-few-shot learningSVM64.1268.1855.59
3D-CNN65.7473.7257.37
SSRN76.2679.5170.56
Few-shot learningDFSL + NN77.7572.2454.55
DFSL + SVM79.6376.4173.05
RN-FSC80.1877.1273.73
Gai-CFSL83.1282.3577.96
DCFSL83.6583.7778.70
SCFormer-R82.3182.2576.55
SCFormer-S83.8382.4778.47
MFSC85.0986.6980.78
MFSC (Mish + BN) Ours86.0288.2181.93
Table 3. Comparison of the classification performance of different methods in Salinas datasets at number of labeled samples K = 5.
Table 3. Comparison of the classification performance of different methods in Salinas datasets at number of labeled samples K = 5.
MethodsOA (%)AA (%)Kappa × 100
Non-few-shot learningSVM80.7187.5878.61
3D-CNN84.2089.5682.46
SSRN86.3993.2484.95
Few-shot learningDFSL + NN87.0591.0185.63
DFSL + SVM86.9590.0885.51
RN-FSC84.1188.8382.38
Gai-CFSL87.8392.4186.48
DCFSL89.3494.0488.17
SCFormer-R89.3093.8988.10
SCFormer-S88.8294.1387.57
MFSC90.5494.4989.47
MFSC (Mish + BN) Ours90.9795.3689.98
Table 4. Class-specific classification accuracy (%) of different methods for the target-scene UP datasets (five labeled samples from TD).
Table 4. Class-specific classification accuracy (%) of different methods for the target-scene UP datasets (five labeled samples from TD).
ClassSVM3DCNNSSRNDFSL + NNGai-CFSLRN-FSCDCFSLSCFormer-RSCFormer-SOurs
Shadow99.1335.5798.0896.9279.6299.1998.6699.0895.4698.52 ± 0.32
Bricks68.1757.2785.3458.1388.5963.4866.7374.2178.1589.22 ± 0.64
Bitumen40.6287.6460.0770.6262.0670.0481.1886.9781.5587.28 ± 1.47
Bare soil37.1263.4053.5671.2390.6657.9977.3257.5062.9687.14 ± 2.45
Metal Sheet95.4490.7798.3410098.5499.4399.4998.9099.01100 ± 0.0
Trees60.2277.3178.0289.9974.6592.1593.4586.5480.9294.53 ± 1.59
Gravel39.9868.9155.2357.4777.7449.8167.4669.2671.3266.19 ± 0.24
Meadows83.9163.0595.1384.6371.4993.4487.7490.9292.0484.54 ± 1.13
Asphalt88.9859.8291.8469.1997.7768.5582.2076.9280.7889.42 ± 2.78
Table 5. Class-specific classification accuracy (%) of different methods for the Salinas target scene datasets (five labeled samples from TD).
Table 5. Class-specific classification accuracy (%) of different methods for the Salinas target scene datasets (five labeled samples from TD).
ClassSVMRes-3D-CNNSS-CNNGai-CFSLDPGNRN-FSCDCFSLSCFormer-RSCFormer-SOurs
Brocoli_green_weeds_185.6039.4793.0299.5987.7296.4599.5598.9698.2299.91 ± 0.20
Brocoli_green_weeds_298.5474.0293.5198.8199.4999.1599.7199.8799.9599.92 ± 0.14
Fallow65.3849.3384.3190.1879.7685.8593.6893.5498.6298.55 ± 1.23
Fallow_rough_plow95.8288.7186.4398.2398.3498.4999.4598.6996.2699.89 ± 0.06
Fallow_smooth95.8377.5090.9186.7580.1382.6790.3992.8696.6793.48 ± 1.39
Stubble99.9297.5299.5599.2199.9297.2999.2799.9199.9599.47 ± 0.60
Celery95.2961.5397.5498.5899.8699.3999.0498.6499.1199.94 ± 0.04
Grapes_untrained57.0068.9373.5274.2350.8471.5972.6176.6372.0078.77 ± 2.28
Soil_vinyard_develop90.6492.8393.8197.7489.0388.1699.7499.5899.5699.99 ± 0.01
Corn_sensced_green_weeds85.8769.3377.2180.5481.2469.7284.5181.5286.7284.10 ± 2.31
Lettuce_romaince_4wk38.3259.0742.3796.4389.4689.2998.1797.8496.5999.15 ± 0.71
Lettuce_romaince_5wk87.5670.5995.8599.1399.1794.0399.0499.5699.6599.97 ± 0.04
Lettuce_romaince_6wk88.6675.3899.2398.6199.5699.4598.9799.6499.4299.12 ± 0.69
Lettuce_romaince_7wk87.8789.1292.9897.9598.8796.5897.7798.4098.1999.04 ± 0.73
Vinyard_untrained33.1847.6250.3773.8559.7569.3074.1273.9072.6580.40 ± 4.08
Vinyard_vertical_trellis81.6488.9080.5488.7577.6981.8690.6291.3492.5287.26 ± 4.4
Table 6. Class-specific classification accuracy (%) of different methods for the Indian Pines datasets from the target scene (five TD labeled samples).
Table 6. Class-specific classification accuracy (%) of different methods for the Indian Pines datasets from the target scene (five TD labeled samples).
ClassSVM3D-CNNDFSLGai-CFSLDPGNRN-FSCDCFSLSCFormer-RSCFormer-SOurs
Alfalfa41.3092.6897.4491.6095.1296.6595.6187.8093.41100.00 ± 0.0
Corn-notill42.8638.4438.3450.4647.1545.9550.4446.1350.2555.02 ± 5.06
Corn-mintill39.0444.8543.3544.8827.0341.2548.4245.6851.8466.91 ± 4.99
Corn59.4936.6468.4581.6156.4759.0679.5764.0161.7797.24 ± 2.16
Grass-pasture59.6371.1370.2170.7639.7565.9073.8973.8777.8584.77 ± 2.57
Grass-Tree84.7972.6976.3884.2561.3869.5188.2687.5490.4184.50 ± 8.58
Grass-pasture-moved92.8610099.7797.1010099.6599.5796.9699.1399.13 ± 1.74
Hay-windrowed90.7983.9375.6791.1292.3976.9188.4486.159078.73 ± 3.46
Oats90.0033.3399.0099.2610010010098.6798100.00 ± 0.0
Soybean-notill34.5764.8447.9062.6857.9126.0561.7158.4256.0672.37 ± 1.47
Soybean-mintill0.0058.0457.8066.5441.1865.3657.8264.4857.8766.47 ± 4.85
Soybean-clean15.0123.6438.1342.0647.9626.3140.3434.4634.9745.99 ± 6.59
Wheat89.7691.0098.0497.1189.0099.2899.2598.2095.60100.00 ± 0.0
Woods90.9153..9783.0887.1078.6575.6687.2685.3586.9097.17 ± 0.56
Building_Grass-Trees_drives17.3656.9662.8668.7446.7269.9068.7166.8565.6267.19 ± 7.36
Stone-Stell_Towers86.0210099.9497.4798.8699.8898.5299.8999.77100.00 ± 0.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, H.; Wang, X.; Xia, K.; Ma, Y.; Yuan, G. Transfer Learning-Based Hyperspectral Image Classification Using Residual Dense Connection Networks. Sensors 2024, 24, 2664. https://doi.org/10.3390/s24092664

AMA Style

Zhou H, Wang X, Xia K, Ma Y, Yuan G. Transfer Learning-Based Hyperspectral Image Classification Using Residual Dense Connection Networks. Sensors. 2024; 24(9):2664. https://doi.org/10.3390/s24092664

Chicago/Turabian Style

Zhou, Hao, Xianwang Wang, Kunming Xia, Yi Ma, and Guowu Yuan. 2024. "Transfer Learning-Based Hyperspectral Image Classification Using Residual Dense Connection Networks" Sensors 24, no. 9: 2664. https://doi.org/10.3390/s24092664

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop