Next Article in Journal
Entity Linking Model Based on Cascading Attention and Dynamic Graph
Previous Article in Journal
Multi-Objective Combinatorial Optimization Algorithm Based on Asynchronous Advantage Actor–Critic and Graph Transformer Networks
Previous Article in Special Issue
Attention-Guided and Topology-Enhanced Shift Graph Convolutional Network for Skeleton-Based Action Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pixel-Level Decision Fusion for Land Cover Classification Using PolSAR Data and Local Pattern Differences

by
Spiros Papadopoulos
,
Vassilis Anastassopoulos
* and
Georgia Koukiou
Electronics Laboratory, Physics Department, University of Patras, 26504 Patras, Greece
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(19), 3846; https://doi.org/10.3390/electronics13193846 (registering DOI)
Submission received: 27 August 2024 / Revised: 24 September 2024 / Accepted: 26 September 2024 / Published: 28 September 2024
(This article belongs to the Special Issue Artificial Intelligence in Image Processing and Computer Vision)

Abstract

:
Combining various viewpoints to produce coherent and cohesive results requires decision fusion. These methodologies are essential for synthesizing data from multiple sensors in remote sensing classification in order to make conclusive decisions. Using fully polarimetric Synthetic Aperture Radar (PolSAR) imagery, our study combines the benefits of both approaches for detection by extracting Pauli’s and Krogager’s decomposition components. The Local Pattern Differences (LPD) method was employed on every decomposition component for pixel-level texture feature extraction. These extracted features were utilized to train three independent classifiers. Ultimately, these findings were handled as independent decisions for each land cover type and were fused together using a decision fusion rule to produce complete and enhanced classification results. As part of our approach, after a thorough examination, the most appropriate classifiers and decision rules were exploited, as well as the mathematical foundations required for effective decision fusion. Incorporating qualitative and quantitative information into the decision fusion process ensures robust and reliable classification results. The innovation of our approach lies in the dual use of decomposition methods and the application of a simple but effective decision fusion strategy.

1. Introduction

Remote sensing technologies have profoundly changed our ability to collect information about the Earth’s surface, allowing for the monitoring and classification of land cover and land use in a variety of scenarios. The combination of diverse datasets, including PolSAR, optical, etc. has considerably increased the possibility for urban land cover classification, ecological land mapping, and overall sea and land monitoring. Using remote sensing data for land cover classification is critical for addressing a variety of environmental and urban planning concerns. This introduction provides a detailed summary of recent research initiatives that investigate the use of many data sources and decision-level approaches to improve the precision and resilience of land cover classification.
Urban areas are dynamic environments, and monitoring land cover changes in these regions is essential for urban planning and development. An approach by González-Santiago et al. [1] explored deep self-supervised hyperspectral-LiDAR fusion, leveraging self-supervised learning to enhance land cover classification. Hyperspectral image classification also benefits from innovative fusion techniques. Tu et al. [2] proposed a superpixel-pixel-subpixel multilevel network, addressing the challenges of mixed spectral features and noise in hyperspectral images, resulting in superior classification performance. The use of machine learning algorithms in remote sensing was explored by Arpitha et al. [3], who employed various classifiers within Google Earth Engine for comprehensive land use and land cover mapping. Moreover, the fusion of optical and SAR data showed great potential in improving classification outcomes. Liu et al. [4] developed a dual-input model utilizing image-level fusion for SAR-optical cross-modal feature learning, significantly enhancing classification accuracy. Recent research has explored various advanced methods for improving PolSAR image classification. For instance, Hua et al. [5] proposed a Multi-Modal Contrastive Fully Convolutional Network (MCFCN) that integrates multi-modal features and contrastive learning, which effectively addresses the challenges of speckle noise and enhances classification accuracy with limited labeled data. Lv et al. [6] proposed a nonparametric sample augmentation approach for hyperspectral image classification, improving classification performance through iterative sample augmentation. Quan et al. [7] presented a multimodal fusion strategy that integrates SAR and optical data at the feature level, significantly improving land cover classification results. In another study, Chen et al. [8] investigated the complementary strengths of fully polarimetric SAR and optical imaging. Their method utilized polarimetric decomposition techniques and object-based decision tree classification, resulting in enhanced accuracy by combining data from these two sources.
A study conducted by Bui and Mucsi [9] compared two fusion methods, layer-stacking and Dempster–Shafer (D-S) theory-based approaches, using Sentinel-1 and Sentinel-2 data. They found that decision-level fusion with the D-S theory provided the best mapping accuracy for urban land cover mapping. Another study by Jin et al. [10] introduced a Bayesian decision-level fusion approach for multi-sensor data, significantly improving the classification accuracy by considering detailed spectral and phenomenology information. This probabilistic framework allowed for substantial improvements in classification accuracy, especially when combining multi-sensor data with distinct characteristics such as spectral and temporal features. However, a potential drawback of this method is its computational complexity and the need for extensive prior information. Bayesian inference can be demanding in terms of computational resources, especially when large datasets or numerous variables are involved. The integration of SAR and optical data has been a focal point in several studies. Maggiolo et al. [11] proposed a decision fusion technique combining optical and SAR data through Markov Random Fields (MRFs). The strength of this method lies in its ability to account for spatial dependencies between neighboring pixels, which is particularly advantageous in large-scale applications. By optimizing classification through spatial correlation, this approach helps mitigate noise and improve the consistency of the classification output over wide areas. However, the approach may be sensitive to noise in the spatial relationships, meaning that in areas with highly variable pixel values or complex textures, the model might not perform as well, leading to suboptimal classifications. Zhu et al. [12] developed a more advanced decision-level fusion method that integrates multi-band SAR images with the Dempster–Shafer (D-S) evidence theory and convolutional neural networks (CNNs). This hybrid approach merges the feature extraction capabilities of CNNs with the uncertainty-handling strength of the D-S theory, resulting in a highly robust classification system. The use of CNNs automates feature extraction, while the D-S theory ensures more reliable decision-making when dealing with conflicting or uncertain data from multiple sensors. For instance, the training of CNNs requires a large amount of labeled data, and the process can be time-consuming and computationally intensive. Papadopoulos et al. [13] introduced a correlated decision fusion method that integrates fully polarimetric SAR data and thermal infrared images. By focusing on the transmission of quality bits, this approach improves the classification accuracy by leveraging data quality. However, one potential disadvantage is that fully polarimetric SAR data can be challenging to acquire, as it often requires specialized equipment and expertise. Furthermore, the correlated decision fusion method might face limitations when the data sources (SAR and thermal) are not perfectly aligned or when the correlation between these modalities is weaker than expected, which could hinder the method’s effectiveness in certain environments. Chen et al. [14] presented a decision-level fusion method by integrating Landsat 8 and Sentinel-1 data through decision-level fusion (DLF). Their research highlighted that DLF significantly improved crop classification accuracy, illustrating the effectiveness of data fusion in agricultural applications.
Browsing through the bibliography, Local Binary Patterns (LBP) were broadly used not only to extract features for land cover classification but also for the identification of drunk people by extracting patterns on their forehead vessels using thermal infrared imagery, as discussed in papers [15,16]. Additionally, the specific feature extraction methodology was also used for the analysis of Ground Penetrating Radar (GPR) [17] data to highlight the hyperbolic peaks that represent buried objects or more generally various subsurface structures.
In this study, there are two objectives. First, we aim to learn how the implementation of a Local Pattern Differences (LPD) descriptor to the scattering components of Pauli and Krogager on PolSAR data will affect the accuracy of the three classifiers, and second, we aim to see how much we can improve the overall accuracy combining these three classifiers using the Bayesian decision fusion. To achieve this objective, the first step is to preprocess the acquired data. The specific procedure is composed of calibration, polarimetric decomposition, terrain correction, and finally registration. By accurately aligning the images, we establish a consistent spatial reference for subsequent analysis and classification. After preprocessing, we examine how we can overcome the speckle noise of the datasets by choosing suitable variables such as the optimal thresholds and the quantization window size for a more sufficient feature extraction. Furthermore, our objective is to find patterns between the scatterers that we can use as features and characterize the land cover types.
The novelty of our method lies in the implementation of two decomposition techniques as a base for feature extraction for the LPD descriptor, which has previously only been implemented in raw SAR data. Then, we fed with these features the three classifiers, such as a simple Neural Network (NN), a Decision Tree (DT) and a Random Forest (RF), whose outputs are local decisions for each land cover type. To address the possible vulnerability of each classifier, we fused them using a modified Bayesian decision fusion that uses their accuracies as a weight to highlight the more efficient model.
In the subsequent sections, we delve deeper into our study. Section 2 outlines the study area and materials utilized. Section 3 elaborates on the preprocessing of PolSAR data, and a reference is made to the decomposition methods we use. In Section 4, we analyze the LPD descriptor that we use for feature extraction. In Section 5, we refer to the classifiers we use and the results of each individual classifier, while in Section 6, we quote the mathematics behind our decision fusion method and the overall experimental results in the fusion center. In Section 7, the results of our study, the possible imperfections, the reasons behind them, and the thoughts of how to improve this work are discussed.

2. Study Area and Materials

The broader area of Vancouver was chosen as the study area, which is located in the polygon from the coordinates 123 ° 16 27 W to 122 ° 57 29 W Longitude and 49 ° 21 10 N to 49 ° 08 48 N   L a t i t u d e . The study area consists of five main types of land cover including Urban, Forest, Sea, Crops and Industrial. The location of the study area is depicted in Figure 1. Also, we used the data from satellite ALOS, which has an absolute orbit of 16,982 and an incidence angle (near-far) of 22.73 ° and 24.97 ° degrees, respectively. An ALOS PALSAR P1.1 Single Look Complex (SLC) product was acquired on 2 April 2009, with an L band as the center frequency, a PLR beam mode, and 30 m in spatial resolution. VV, VH, HV and HH polarizations were used in our study. Images of ALOS PALSAR were freely downloaded from the Alaska Satellite facility data search (https://search.asf.alaska.edu, accessed on 31 July 2023).

3. Preprocessing: Fully PolSAR

As it was mentioned previously, polarimetric SAR decomposition techniques were used in this work as a basis for our LPD feature extraction. In particular, it was Pauli’s and Krogager’s decompositions used together that gave us the opportunity for the high accuracy detection of both natural and man-made land cover types. Further along in this section, we will cover the presentation of the techniques we used so that the reader can easily understand the proposed approach as well as a brief presentation of the pre-processing steps.

3.1. Pauli’s Decomposition

The core concept of Pauli’s decomposition is to represent the scattering matrix [S] of a pixel into a sum of elementary scattering matrices, each expressing specific deterministic scattering mechanisms [18,19,20]. In the context of the conventional orthogonal linear ( h , v ) basis, assuming S h v = S v h , the Pauli basis S a ,   S b ,   S c can be described using the following three 2 × 2 matrices:
      S a = 1 2 1 0 0 1
S b = 1 2 1 0 0 1
S c = 1 2 0 1 1 0
Consequently, given a measured scattering matrix [ S ] , it can be represented as follows:
S = S h h S h v S h v S v v = α S a + β S b + γ S c
where: S h h represents the horizontal-to-horizontal scattering polarization, S h v represents a transition or scattering from a horizontal to vertical polarization, which is equal to S v h , and S v v represents the vertical to vertical polarization. Also, scattering coefficients are calculated as:
α = S h h + S v v 2
β = S h h S v v 2
γ = 2 S h v
The matrix S a corresponds to the scattering matrix of a sphere, a plate, or a trihedral reflector. In this context, the intensity of the coefficient α indicates the power scattered by targets characterized by single- or odd-bounce scattering. The second matrix, S b , represents the scattering mechanism of a dihedral oriented at 0 degrees, with β indicating the power scattered by such targets. Lastly, the third matrix, S c , pertains to the scattering mechanism of a diplane oriented at 45 degrees, where the coefficient γ is associated with scatterers capable of returning orthogonal polarization. Volume scattering is a prime example of this type of scattering. This correspondence is illustrated in Table 1.

3.2. Pauli Color-Coded Representation

The polarimetric information of the scattering matrix can be depicted by combining the intensities ( S h h 2 , S v v 2 , 2 S h v 2 ) into a single RGB image. However, a significant limitation is the difficulty in physically interpreting the resulting image in terms of S h h 2 , S v v 2 , 2 S h v 2 . To address this, an RGB image can be created using the intensities α 2 , β 2 , γ 2 , which correspond to distinct physical scattering mechanisms, as outlined in Table 1. The most used coding scheme is:
β 2 r e d         γ 2 g r e e n     α 2 b l u e    

3.3. Krogager’s Decomposition

Interpreting a SAR image, especially a fully polarimetric SAR image, is highly challenging [22]. The goal of polarimetric decomposition is to represent the scattering matrix (coherent decomposition) or, when a second-order description is necessary, the covariance matrix (incoherent decomposition) as a mixture of canonical objects. These objects offer a more straightforward physical interpretation.
Let S ( x , y ) denote a 2-by-2 scattering matrix. A coherent polarimetric decomposition can be expressed as:
S x , y = m = 1 M c m S m x , y
In this context, S m x , y represents the response of the m-th canonical object, c m denotes the weight of S m x , y   in the combination that results in S x , y , M is the number of components, and x and y are the spatial coordinates. The Krogager polarimetric decomposition is characterized using the circular polarization scattering matrix S ( R , L ) ( x ,   y ) , where R signifies the right-handed circular component and L represents the left-handed circular component. In monostatic radar systems, such as SAR, the scattering matrix is symmetric. Consequently, the S R , L x ,   y   components can be described in terms of linear polarization components as follows [23].
S R R x ,   y = S H H x ,   y S V V x ,   y 2 + i S H V x ,   y ,
S L L x ,   y = S H H x ,   y S V V x ,   y 2 i S H V x ,   y ,
S R L x ,   y = S L R x ,   y = S H H x ,   y + S V V x ,   y 2 ,
where i denotes the imaginary unit. According to [24], and omitting the x ,   y dependence for simplicity, the Krogager polarimetric decomposition is defined as follows:
S ( R , L ) = S R R S R L S L R S L L = e i ϕ k s e i ϕ s 0 i i 0 + k d e i 2 η 0 0 e i 2 η + k h e i 2 η 0 0 0
where k s , k d , and k h are real-valued quantities representing the scattering coefficients from a sphere, a diplane, and a helix, respectively. Additionally, ϕ is the absolute phase term that depends on the distance between the target and the sensor, ϕ s represents the displacement of the sphere relative to the diplane and helix components, and   η   denotes their orientation angle. The scattering coefficients k s , k d , and k h can be derived from the circular polarization scattering components [25] as follows:
k s = S R L ,
k d = m i n S R R , S L L ,
k h = a b s ( S R R , S L L )
where the symbol · denotes the modulus of a complex quantity and a b s ( ) represents the absolute value.
Previous studies demonstrated that the Krogager decomposition is particularly effective among various coherent polarimetric decompositions in discriminating man-made targets from natural targets [25,26]. However, it lacks the ability to distinguish between different types of man-made targets. On the other hand, Pauli’s decomposition can distinguish natural targets very well [27]. So, the combination of these two will give us complementary information for our study area, which has plenty of mixed land cover types.
The SLC PolSAR data, depicted in Figure 2a, requires thorough preprocessing to extract valuable information. Utilizing the Sentinel Application Platform (SNAP) version 9.0.8, we follow a systematic procedure that includes radiometric calibration [28], Pauli’s and Krogager’s decomposition, and Geometric Doppler terrain correction [29].
Radiometric calibration is essential for converting raw digital values into physically meaningful units. This process adjusts the SAR image so that pixel values accurately represent the radar backscatter of the reflecting surface, while still preserving geometric distortions, as shown in Figure 2b. Next, the two decomposition methods are applied to transform the complex polarimetric matrices into three distinct components for each method (see one of the components in Figure 2c,d). This transformation provides a visual intuitive representation of polarimetric information, making it easier to interpret scattering mechanisms within the radar data. Finally, the Geometric Doppler terrain correction is employed to address geometric distortions caused by a varying topography. Using a Digital Elevation Model (DEM) [30], this correction adjusts for uneven terrain, ensuring that radar reflections align with accurate geographic coordinates. The result is a georeferenced dataset (Figure 2e,f), which is crucial for precise spatial analysis and scientific interpretation.

4. Feature Extraction: LPD

Building on the theoretical framework discussed earlier, we utilized Pauli’s (Figure 3b) and Krogager’s (Figure 3a) scattering components, extracted from the SNAP software (Version 9.0.8). These components represented as α ,   β ,   γ and k s , k d , k h   correspond to the intensities of the scattering coefficients and were measured in decibels. Since negative decibel values cannot be directly used in color mapping, we applied a normalization process to each component. This involved adjusting their histograms to scale the values between 0 and 255 to see the compatibility with color representation.
In SAR images, as reported in previous research [31], various land cover types exhibit distinct textures due to differences in surface roughness [32]. However, these textures comprise different local structures at certain quantization levels. Our approach focused on the patterns of textures that were created by the combination of scatterers in each land cover type in SAR image. Figure 3 illustrates the products that we were working on, i.e., three bands from each decomposition method depicted as RGB images.
Observing Figure 3, it is evident that the local structures vary between different classes in terms of contrast. To capture these local structures, we quantize each band of the two decomposition methods into five intervals using a contrast technique. Choosing this number of intervals gave us the opportunity to capture local structures as effectively as possible and at the same time maintain computational efficiency and mitigate the impact of speckle noise. This approximation sufficiently captures the essential variations in local structures without introducing excessive complexity or noise sensitivity that might arise from using a higher number of levels.
After quantizing the bands, we constructed the LPD for each pixel by extracting statistical features from the local structures. Finally, pixels were classified using three different classifiers with the proposed LPD.
To implement the image quantization, a widely used contrast technique derived from the recent local binary pattern method [33,34] was applied. First, all pixel intensities within a moving window were quantized into five levels based on the difference between the central pixel and the surrounding pixels within the window. Let g c   represent the intensity of the central pixel. The quantization procedure is then formulated as follows:
s i =       2       1 g i > g c + t + 20 g c + t + 20 > g i > g c + t     0                                                 g i g c t 1 2 g c t > g i > g c t 20 g i < g c t 20           ,   i 1 , , h 2
where   g i   is the intensity of pixel i in the window, and t is a threshold. An example of the quantization process with t = 5 is shown in Figure 4. After quantization, connected components of the same quantization level, which correspond to local patterns, can represent local structures. For instance, pixels with the value “2” and “1” capture sharp edges in urban areas, those with the value “0” describe homogeneous regions, and those with the values “−2”and “−1” can capture dark primitives. During the quantization stage, a soft threshold was employed to mitigate the impact of speckle noise.
To obtain more detailed information and accurately characterize the texture, it is necessary to use multiple thresholds. Utilizing multiple thresholds allows for capturing global information while also resisting speckle noise. Given that speckle is multiplicative, we recommend increasing the thresholds by multiplying them by a constant. For example, we selected thresholds of t = [5, 10, 15] because with low values of t , we can use all the information given from this study, and we also succeeded in avoiding mixed land covered areas with the label of “0” i.e., the homogenous region.
For a single threshold, following the quantization stage, we can identify three types of local structures. To characterize these local structures, we then used local mean and variance:
      P a v e = 1 N 2 j = 1 N 2 g j
P v a r = 1 N 2 j = 1 N 2 g j P a v e 2  
The mean feature extracted from the local structures captures local intensity fluctuations, making noisy pixels less noticeable [35]. Variance is an excellent measure for detecting boundaries and edges [35]. It is important to note that other effective measures can also be used to characterize local structures. Two parameters need consideration: the size of the moving window (h) and the thresholds (t). According to the investigation in [36], the optimal h depends on the image resolution and the classification task. In this work, a window size of h = 5 was selected as a balanced choice. A smaller window could lead to over-localized feature extraction, which might help detect subtle changes in the data but could also introduce noise or lead to overfitting. On the other hand, a larger window might overlook important localized variations, reducing performance. Once the parameters are set, the LPD feature vector was constructed for each pixel, with the following form:
L P D = P a v e , 1 , P v a r , 1 , , P a v e , n , P v a r , n
with n as the number of thresholds that are suitable for each problem.

5. Classification: Experimental Results

As discussed in the previous section, the LPD descriptor was applied across the entire study area to extract the appropriate features for each pixel. After the LPD descriptor, each pixel had a unique identity, which consists of 36 features. For computational efficiency, we selected four 20 × 20 pixel windows to represent the four land cover types in our study. Specifically, we used red for urban areas, blue for the sea, yellow for crop, and green for forests, as illustrated in Figure 5.
For classification, we used 80% of the pixels as a training dataset and 20% as a testing dataset. These datasets were then “fed” to a simple 2-layer NN [37], an RF classifier [38], and a DT classifier [39], as shown in Table 2.
In Table 3, Table 4 and Table 5, the confusion matrices of each classifier are presented to understand how many of the predicted classes of the testing pixels were correctly classified with the true class labels and how many were incorrectly classified as a different class.
The analysis of the confusion matrices reveals that while each classifier has its strengths, the random forest consistently provided the best overall performance as we observe in Table 6, particularly in distinguishing sea and urban classes with the accuracies of 95% and 86.3%, respectively. The decision tree showed weaknesses in separating similar vegetative classes, such as crop and forest, with notable misclassifications due to overlapping spectral features or boundary effects. The neural network improved classification for urban and forest classes but still faced challenges with distinguishing sea from crop. Misclassifications are primarily due to spectral similarities and mixed pixels in boundary areas.

6. Decision Fusion: Experimental Results

The outputs from the feature extraction stage are fed, as we discussed above, into three classifiers: a simple neural network (NN), a random forest (RF), and a decision tree (DT) classifier. Based on Duda’s [40] pattern classification framework, we employed a Bayesian model suitable for our problem.
We have equal prior probabilities for each class, denoted as P C 1 = P C 1 = = P C k , where k ranges from 1 to 4. Each classifier provides k   conditional probabilities (likelihoods) P j ( X = c C k ) with j = 1 ,   2 ,   3 ,   representing the probability of each class C k given the evidence X (predictions from the classifiers).
From [41], it can be observed that the class posterior probability from multiple classifiers is given by
P ( C k X = c ) = P C k   j = 1 3 P j X = c C k × w j k 3 P C k   j = 1 3 P j X = c C k × w j
where P C k is the prior probability of class C k , P j X = c C k is the likelihood of the evidence X (predictions from the classifiers) given class C k for classifier j , and w j is the reliability weight (accuracy) of classifier j .
Equation (21) essentially shows that the class posterior is proportional to the product of the conditional probabilities of class C k across the classifiers, each weighted by the classifier’s reliability and adjusted by the prior probabilities, divided by the evidence.
To avoid instabilities and to create a more efficient model, we proceeded with the log-posterior probabilities. After calculations, the class posterior probability equation became:
P C k X = c = exp log P C k + j log P j X = c C k × w j   k exp log P C k + j log P j X = c C k × w j   ,
As a next step, the log-posterior probabilities must be normalized. To avoid overflow when exponentiating, we computed the maximum log-posterior probability for each class, subtracted this maximum value from each log-posterior probability, and converted the adjusted log-posterior probabilities back to regular probabilities. This ensures numerical stability and accurate calculations.
The final class predictions were determined by selecting the class with the highest posterior probability i.e.,
m a x k = 1 4 { P C k X = c }
Although the initial classification results were promising, with all classifiers achieving over 70% accuracy—specifically 72.9% for the decision tree (DT) classifier, 79.1% for the neural network (NN) classifier, and 85.9% for the random forest (RF) classifier—we were confident that implementing the Bayesian-based decision fusion model would lead to an even better performance. Our expectations were confirmed, as we achieved an impressive overall accuracy of 98.1% at the fusion center. This represents an improvement of almost 12.2% over the “strongest” individual classifier. After decision fusion, the accuracy of each land cover type was configured, as shown in Table 7. In the following, Table 8 is briefly compared to our study with similar research. Recognizing the difficulty of directly comparing two methodologies due to the different approaches used, we selected the research most closely aligned with ours.

7. Discussion: Conclusions

7.1. Discussion

In this article, we proposed a novel land cover classification approach based on features that were extracted from two decomposition methodologies using a Local Pattern Differences descriptor to exploit as much as possible the advantages of both decompositions in target detection. Then, we employed three individual classifiers to categorize four land cover types using these features for training and finally we created a more robust and accurate model by “feeding” the local decisions into a Bayesian decision fusion model.
Our study area is the broader area of Vancouver and consists of four main land cover types: urban, sea, crop, and forests. We employed data from the ALOS satellite. The preprocessing steps involved radiometric calibration, Pauli’s and Krogager’s decomposition, and geometric Doppler terrain correction for PolSAR data. Feature extraction included the revealing of patterns that were hiding into the scattering coefficients.
As we can observe above from the three classifiers (Table 3, Table 4 and Table 5), we had a relatively large percentage of pixel misclassification in pixels that represents crop. More specifically, the DT classifier categorized 18 and 16 crop pixels as sea and forest, respectively. Τhe same phenomenon, but to a lesser extent, is also observed in the other two classifiers. Also, a lower percentage of forest pixels were mistakenly labeled as urban or crop in all classifiers.
Using the Fisher Linear Discriminant [42], we can visualize the feature space, projected onto the subspace generated by the eigenvectors with the largest eigenvalues (Figure 6a,b). Both images in Figure 6a,b depict the cluster separability of four land cover types for training and testing datasets. These errors were expected since we have a huge area without a 100% clear land cover type, as we inspect in Figure 6a,b below. There are urban areas mixed with trees, crops planted with trees, and not low vegetation, or they are over-watered, which was the main reason for the pixel misclassifications. Generally, we observe in Figure 6 that the areas with the greatest overlap are those of the sea, the forest, and crop.
The results demonstrate that random forest outperforms other models, especially in accuracy-challenging classes such as sea and crop, due to its robust handling of complex and noisy data patterns. This knowledge is valuable to researchers and practitioners in urban planning, environmental monitoring, and agricultural management, as it guides the selection of appropriate classifiers for specific land cover types.
We prove that it is possible to address the limitations of individual classifiers and to take advantage of their strengths by employing a Bayesian decision fusion model to combine their local decisions and compute the posterior probabilities for each class. For the fusion to be considered successful, the overall accuracy at the fusion center is needed to surpass that of the best individual classifier. Our experimental results demonstrate this success, with an overall accuracy improvement of 12.1% over the random forest’s 85.9%. Specifically, we achieved 99.7% accuracy for urban pixels, 99.2% for sea, 94.4% for crop, and 98.5% for forest.

7.2. Conclusions

Through this classification and decision fusion process, we not only deepened our understanding but also identified areas for future exploration that could further enhance accuracy. As we believe, the limitations for one study are the cause for further innovations in the scientific world. We try to motivate researchers to possibly work not only with raw data but also with preprocessed datasets that can hide features that could be helpful for classification. Furthermore, identifying and incorporating additional pixel-level features would allow for a clearer distinction of mixed land cover types in datasets with speckle noise or other interferences. Optimization of the classifier’s parameters would be one more important subject for examination, as well as the use of some other classification techniques that would give better separability among the classes. Also, the usage of more land cover types can provide a more generalized approach to the classification.
Looking ahead, the field presents challenges and opportunities. Our goal is addressing the open challenges in this research domain, by using the correlation between the pixels of our study area and refining the integration of quality information for decision fusion. The nature of land cover that frequently changes necessitates the continuous improvement of methodologies.
On the other hand, these challenges also pave the way for innovative breakthroughs. Progress in machine learning techniques, sensor advancements, and increased computational power offers the potential for developing more advanced and precise classification approaches. Tapping into the intersections of new technologies such as remote sensing and artificial intelligence may reveal fresh opportunities to improve land cover analysis.
In conclusion, the future of this research area lies in overcoming obstacles while capitalizing on opportunities for progress. Ongoing investigation, innovation, and the adoption of the latest technologies will be crucial in steering our research forward, ultimately leading to a deeper and more complete understanding of land cover patterns and dynamics.

Author Contributions

Conceptualization, S.P., G.K. and V.A.; methodology, S.P., G.K. and V.A.; resources, S.P. and G.K.; writing—original draft preparation, S.P.; writing—review and editing, G.K. and V.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The PolSAR data was downloaded from ASF Data Search Vertex (https://search.asf.alaska.edu/#/, accessed on 15 November 2022) with the product name ALPSRP169820980-L1.1.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. González-Santiago, J.; Schenkel, F.; Gross, W.; Middelmann, W. Deep Self-Supervised Hyperspectral-Lidar Fusion for Land Cover Classification. In Proceedings of the IGARSS 2023—2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023. [Google Scholar] [CrossRef]
  2. Tu, B.; Ren, Q.; Li, Q.; He, W.; He, W. Hyperspectral Image Classification Using a Superpixel–Pixel–Subpixel Multilevel Network. IEEE Trans. Instrum. Meas. 2023, 72, 5013616. [Google Scholar] [CrossRef]
  3. Arpitha, M.; Ahmed, S.A.; Harishnaika, N. Land Use and Land Cover Classification Using Machine Learning Algorithms in Google Earth Engine. Earth Sci. Inform. 2023, 16, 3057–3073. [Google Scholar] [CrossRef]
  4. Liu, S.; Wang, H.; Hu, Y.; Zhang, M.; Zhu, Y.; Wang, Z.; Li, D.; Yang, M.; Wang, F. Land Use and Land Cover Mapping in China Using Multimodal Fine-Grained Dual Network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4405219. [Google Scholar] [CrossRef]
  5. Hua, W.; Wang, Y.; Yang, S.; Jin, X. PolSAR Image Classification Based on Multi-Modal Contrastive Fully Convolutional Network. Remote Sens. 2024, 16, 296. [Google Scholar] [CrossRef]
  6. Lv, Z.; Zhang, P.; Sun, W.; Benediktsson, J.A.; Lei, T. Novel Land-Cover Classification Approach with Nonparametric Sample Augmentation for Hyperspectral Remote-Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–13. [Google Scholar] [CrossRef]
  7. Quan, Y.; Zhang, R.; Li, J.; Ji, S.; Guo, H.; Yu, A. Learning SAR-Optical Cross Modal Features for Land Cover Classification. Remote Sens. 2024, 16, 431. [Google Scholar] [CrossRef]
  8. Chen, Y.; He, X.; Xu, J.; Guo, L.; Lu, Y.; Zhang, R. Decision Tree-Based Classification in Coastal Area Integrating Polarimetric SAR and Optical Data. Data Technol. Appl. 2021, 56, 342–357. [Google Scholar] [CrossRef]
  9. Bui, D.H.; Mucsi, L. Comparison of Layer-Stacking and Dempster-Shafer Theory-Based Methods Using Sentinel-1 and Sentinel-2 Data Fusion in Urban Land Cover Mapping. Geo-Spat. Inf. Sci. 2022, 25, 1–14. [Google Scholar] [CrossRef]
  10. Jin, Y.; Guan, X.; Ge, Y.; Jia, Y.; Li, W. Improved Spatiotemporal Information Fusion Approach Based on Bayesian Decision Theory for Land Cover Classification. Remote Sens. 2022, 14, 6003. [Google Scholar] [CrossRef]
  11. Maggiolo, L.; Solarna, D.; Moser, G.; Serpico, S.B. Optical-Sar Decision Fusion with Markov Random Fields for High-Resolution Large-Scale Land Cover Mapping. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 5508–5511. [Google Scholar] [CrossRef]
  12. Zhu, J.; Pan, J.; Jiang, W.; Yue, X.; Yin, P. SAR Image Fusion Classification Based on the Decision-Level Combination of Multi-Band Information. Remote Sens. 2022, 14, 2243. [Google Scholar] [CrossRef]
  13. Papadopoulos, S.; Koukiou, G. Vassilis Anastassopoulos Correlated Decision Fusion Accompanied with Quality Information on a Multi-Band Pixel Basis for Land Cover Classification. J. Imaging 2024, 10, 91. [Google Scholar] [CrossRef]
  14. Chen, S.; Useya, J.; Mugiyo, H. Decision-Level Fusion of Sentinel-1 SAR and Landsat 8 OLI Texture Features for Crop Discrimination and Classification: Case of Masvingo, Zimbabwe. Heliyon 2020, 6, e05358. [Google Scholar] [CrossRef] [PubMed]
  15. Koukiou, G.; Anastassopoulos, V. Drunk person identification using local difference patterns. In Proceedings of the 2016 IEEE International Conference on Imaging Systems and Techniques (IST), Chania, Greece, 4–6 October 2016; pp. 401–405. [Google Scholar] [CrossRef]
  16. Koukiou, G.; Anastassopoulos, V. Local difference patterns for drunk person identification. Multimed. Tools Appl. 2018, 77, 9293–9305. [Google Scholar] [CrossRef]
  17. Tassiopoulou, S.; Koukiou, G. Fusing Ground-Penetrating Radar Images for Improving Image Characteristics Fidelity. Appl. Sci. 2024, 14, 6808. [Google Scholar] [CrossRef]
  18. Cloude, S.R.; Pottier, E. A Review of Target Decomposition Theorems in Radar Polarimetry. IEEE Trans. Geosci. Remote Sens. 1996, 34, 498–518. [Google Scholar] [CrossRef]
  19. Chen, S.-W.; Li, Y.; Wang, X.; Xiao, S.; Sato, M. Modeling and Interpretation of Scattering Mechanisms in Polarimetric Synthetic Aperture Radar: Advances and Perspectives. IEEE Signal Process. Mag. 2014, 31, 79–89. [Google Scholar] [CrossRef]
  20. Sun, X.; Song, H.; Wang, R.; Li, N. High-Resolution Polarimetric SAR Image Decomposition of Urban Areas Based on a POA Correction Method. Remote Sens. Lett. 2018, 9, 363–372. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Wu, L.; Geng, W. A New Classifier for Polarimetric SAR Images. Prog. Electromagn. Res. 2009, 94, 83–104. [Google Scholar] [CrossRef]
  22. Gaglione, D.; Clemente, C.; Pallotta, L.; Proudler, I.; De Maio, A.; Soraghan, J.J. Krogager Decomposition and Pseudo-Zernike Moments for Polarimetric Distributed ATR. In Proceedings of the 2014 Sensor Signal Processing for Defence (SSPD), Edinburgh, UK, 8–9 September 2014. [Google Scholar] [CrossRef]
  23. Milan, J.M. Book Review [Review of “Principles of Modern Radar-Basic Principles (Richards, M.A., Eds, et Al; 2010)]. IEEE Aerosp. Electron. Syst. Mag. 2013, 28, 40–42. [Google Scholar] [CrossRef]
  24. Hellmann, M.; Krogager, E. Comparison of Decompositions for Pol-SAR Image Interpretation. Int. Geosci. Remote Sens. Symp. 2002, 3, 1313–1315. [Google Scholar] [CrossRef]
  25. Alberga, V.; Krogager, E.; Chandra, M.; Wanielik, G. Potential of Coherent Decompositions in SAR Polarimetry and Interferometry. In Proceedings of the IGARSS’04. 2004 IEEE International Geoscience and Remote Sensing Symposium, Anchorage, AK, USA, 20–24 September 2004; Volume 3. [Google Scholar] [CrossRef]
  26. Wei, Q.; Chen, J.-j.; Zhao, H.-z.; Feng, Z. Target Decomposition for Fully Polarimetric Wideband Radar System. In Proceedings of the IEEE 10th International Conference on Signal Processing Proceedings (ICSP), Beijing, China, 24–28 October 2010. [Google Scholar] [CrossRef]
  27. Zhang, L.; Zhang, J.; Zou, B.; Zhang, Y. Comparison of Methods for Target Detection and Applications Using Polarimetric SAR Image. Piers Online 2008, 4, 140–145. [Google Scholar]
  28. Kumar, D. Urban Objects Detection from C-Band Synthetic Aperture Radar (SAR) Satellite Images through Simulating Filter Properties. Sci. Rep. 2021, 11, 6241. [Google Scholar] [CrossRef] [PubMed]
  29. Jiang, W.; Yu, A.; Dong, Z.; Wang, Q. Comparison and Analysis of Geometric Correction Models of Spaceborne SAR. Sensors 2016, 16, 973. [Google Scholar] [CrossRef]
  30. Makineci, H.B.; Karabörk, H. Evaluation Digital Elevation Model Generated by Synthetic Aperture Radar Data. ISPRS 2016, XLI-B1, 57–62. [Google Scholar] [CrossRef]
  31. Guan, D.-d.; Tang, T.; Li, Y.; Lu, J. Local Pattern Descriptor for SAR Image Classification. In Proceedings of the IEEE 5th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Singapore, 1–4 September 2015. [Google Scholar] [CrossRef]
  32. Rajesh, K.; Jawahar, C.V.; Sengupta, S.; Sinha, S. Performance Analysis of Textural Features for Characterization and Classification of SAR Images. Int. J. Remote Sens. (Print) 2001, 22, 1555–1569. [Google Scholar] [CrossRef]
  33. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  34. Tan, X.; Triggs, B. Enhanced Local Texture Feature Sets for Face Recognition under Difficult Lighting Conditions. IEEE Trans. Image Process. 2010, 19, 1635–1650. [Google Scholar] [CrossRef]
  35. Lee, J.-S. Digital Image Enhancement and Noise Filtering by Use of Local Statistics. IEEE Trans. Pattern Anal. Mach. Intell. 1980, PAMI-2, 165–168. [Google Scholar] [CrossRef] [PubMed]
  36. Dai, D.; Yang, W.; Sun, H. Multilevel Local Pattern Histogram for SAR Image Classification. IEEE Geosci. Remote Sens. Lett. 2011, 8, 225–229. [Google Scholar] [CrossRef]
  37. Popescu, M.C.; E Balas, V.; Perescu-Popescu, L.; Mastorakis, N. Multilayer Perceptron and Neural Networks. WSEAS Trans. Circuits Syst. 2009, 8, 579–588. [Google Scholar]
  38. Pal, M. Random Forest Classifier for Remote Sensing Classification. Int. J. Remote Sens. 2005, 26, 217–222. [Google Scholar] [CrossRef]
  39. Safavian, S.R.; Landgrebe, D. A Survey of Decision Tree Classifier Methodology. IEEE Trans. Syst. Man. Cybern. 1991, 21, 660–674. [Google Scholar] [CrossRef]
  40. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification; John Wiley & Sons: Hoboken, NJ, USA, 2012; pp. 90–91. ISBN 9781118586006. [Google Scholar]
  41. Kuncheva, L.I. Combining Pattern Classifiers; John Wiley & Sons: Hoboken, NJ, USA, 2004; pp. 126–127. ISBN 9780471660255. [Google Scholar]
  42. Koukiou, G. Short Words for Writer Identification Using Neural Networks. Appl. Sci. 2023, 13, 6841. [Google Scholar] [CrossRef]
Figure 1. Study area: the broader area of Vancouver. Map data ©2024: Google, Landsat/Copernicus.
Figure 1. Study area: the broader area of Vancouver. Map data ©2024: Google, Landsat/Copernicus.
Electronics 13 03846 g001
Figure 2. Correction of geometric distortions in the ALOS ascending image: (a) amplitude of original image, (b) amplitude of calibrated image, (c) Pauli component, (d) Krogager component, (e) georeferenced Pauli component, and (f) georeferenced Krogager components.
Figure 2. Correction of geometric distortions in the ALOS ascending image: (a) amplitude of original image, (b) amplitude of calibrated image, (c) Pauli component, (d) Krogager component, (e) georeferenced Pauli component, and (f) georeferenced Krogager components.
Electronics 13 03846 g002
Figure 3. RGB representation of our study area: (a) Krogager’s scattering components and (b) Pauli’s scattering components.
Figure 3. RGB representation of our study area: (a) Krogager’s scattering components and (b) Pauli’s scattering components.
Electronics 13 03846 g003
Figure 4. Illustration of the quantization process of 5 by 5 pixel window. Each of the neighboring pixel’s ( g i ) intensities compared with the central’s ( g c ) to detect the local patterns. Then, this procedure is repeated for all pixels of our study area.
Figure 4. Illustration of the quantization process of 5 by 5 pixel window. Each of the neighboring pixel’s ( g i ) intensities compared with the central’s ( g c ) to detect the local patterns. Then, this procedure is repeated for all pixels of our study area.
Electronics 13 03846 g004
Figure 5. Windows used for classification in our study area, (a) Krogager and (b) Pauli.
Figure 5. Windows used for classification in our study area, (a) Krogager and (b) Pauli.
Electronics 13 03846 g005
Figure 6. Clusters of datasets: (a) training dataset, (b) testing dataset. Blue spots: sea, red spots: urban, yellow spots: crops, and green spots: forest.
Figure 6. Clusters of datasets: (a) training dataset, (b) testing dataset. Blue spots: sea, red spots: urban, yellow spots: crops, and green spots: forest.
Electronics 13 03846 g006
Table 1. Pauli bases and the corresponding meaning [21].
Table 1. Pauli bases and the corresponding meaning [21].
Pauli BasisMeaning
S a Single- or odd-bounce scattering: this occurs when a radar signal interacts with a target and undergoes a single reflection or bounce before reaching the radar sensor.
S b Double- or even-bounce scattering: This can happen, for instance, when radar waves hit a surface, reflect off, and then reflect again off another surface before returning to the sensor.
S c Volume scattering: This type of scattering is more complex and involves multiple interactions within the target volume, leading to a scattering signal that does not follow a simple direct path (forest canopy).
Table 2. Number of pixels used for each class for training and testing.
Table 2. Number of pixels used for each class for training and testing.
ClassNumber AssignedTraining PixelsTesting Pixels
Urban032080
Sea132080
Crop232080
Forest332080
Table 3. Confusion matrix of decision tree.
Table 3. Confusion matrix of decision tree.
Class0123Total
True Class07121680
126117080
2016461880
3131115580
Total86807579
Predicted Class
Table 4. Confusion matrix of neural network.
Table 4. Confusion matrix of neural network.
Class0123Total
True Class07411480
156211280
2113511580
35186680
Total85777187
Predicted Class
Table 5. Confusion matrix of random forest.
Table 5. Confusion matrix of random forest.
Class0123Total
True Class06911980
10764080
20865780
310056580
Total79857581
Predicted Class
Table 6. Local cover types accuracies of the classifiers for each class and the overall accuracies.
Table 6. Local cover types accuracies of the classifiers for each class and the overall accuracies.
Classifier ClassOverall Accuracy
UrbanSeaCropForest
Decision TreeAccuracy
(%)
88.876.357.568.872.9
Neural Network92.577.563.882.579.1
Random Forest86.39581.281.285.9
Table 7. Posterior probabilities for each class according to our method.
Table 7. Posterior probabilities for each class according to our method.
Class
UrbanSeaCropForestOverall Accuracy
Accuracy (%)99.799.294.998.598.1
Table 8. Comparison of our methodology with the proposed Bayesian fusion methodology of Jin et al.
Table 8. Comparison of our methodology with the proposed Bayesian fusion methodology of Jin et al.
Study/AuthorData Sources/InputsMethodologyClassifiers UsedEvaluation MetricsClassification Results
Jin et al. [10]Multi-source satellite images (high spatial and temporal resolution)
MODIS, LANDSAT
Spatiotemporal information fusion using Bayesian Decision Theory to integrate spatial and temporal data. Preprocessing involved aligning multi-source images spatially and temporally.Support Vector Machine (SVM) (for LANDSAT)
ED-similarity (for MODIS) and PBF
Class-wise and Overall Accuracy (OA)PBF Class-wise Accuracy:
Construction Land: 96%
Crop1: 96%
Crop2: 64%
Gobi: 86%
Grassland: 43%
Slope Field: 57%
Wasteland: 42%
Water: 96%
OA: 75%
Our studyPolSAR data Local Pattern Descriptor (LPD) for local structure analysis of the decomposition components.Quantization of image bands using contrast technique; Local structure capture with LPD; Bayesian Decision Fusion combining Decision Tree (DT), Neural Network (NN), and Random Forest (RF) classifiersDecision Tree, Neural Network, Random Forest (with Bayesian fusion)OA, Class-wise accuracies, Posterior probabilities (after Bayesian fusion)Pre-fusion:
DT: 72.9%
NN: 79.1%
RF: 85.9%
After fusion: OA: 98.1%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Papadopoulos, S.; Anastassopoulos, V.; Koukiou, G. Pixel-Level Decision Fusion for Land Cover Classification Using PolSAR Data and Local Pattern Differences. Electronics 2024, 13, 3846. https://doi.org/10.3390/electronics13193846

AMA Style

Papadopoulos S, Anastassopoulos V, Koukiou G. Pixel-Level Decision Fusion for Land Cover Classification Using PolSAR Data and Local Pattern Differences. Electronics. 2024; 13(19):3846. https://doi.org/10.3390/electronics13193846

Chicago/Turabian Style

Papadopoulos, Spiros, Vassilis Anastassopoulos, and Georgia Koukiou. 2024. "Pixel-Level Decision Fusion for Land Cover Classification Using PolSAR Data and Local Pattern Differences" Electronics 13, no. 19: 3846. https://doi.org/10.3390/electronics13193846

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop