Publication Cover
Canadian Journal of Remote Sensing
Journal canadien de télédétection
Volume 49, 2023 - Issue 1
280
Views
0
CrossRef citations to date
0
Altmetric
Research Article

A Novel Classification Method for PolSAR Image Combining the Deep Learning Model and Adaptive Boosting of Shallow Classifiers

Une nouvelle méthode de classification des images PolSAR combinant le modèle d’apprentissage profond et l’optimisation adaptative des algorithmes traditionnels

ORCID Icon, , &
Article: 2257331 | Received 13 Apr 2023, Accepted 11 Aug 2023, Published online: 15 Sep 2023

Abstract

Polarimetric synthetic aperture radar (PolSAR) images are classified mainly according to the backscattering information of ground objects. For regions with complex backscattering information, misclassification is easy to occur, which leads to challenges in improving the classification accuracy of the PolSAR image. Given this situation, this paper combines the Deep Learning Model and traditional classifiers to classify PolSAR image. First, the Convolution Neural Network (CNN) was used to classify the PolSAR image and according to the category prediction probability of pixels, the key pixels easily misclassified are located. Then, the adaptive boosting (AdaBoost) algorithm combined the three shallow classifiers (the Support Vector Machine (SVM), the Wishart and the Decision Tree classifier) into strong classifiers to reclassify the key pixels. Finally, the labels of key pixels and other pixels are output as the final classification result. Experiments on two PolSAR images show that the proposed method can improve classification performance and obtain better classification results.

Résumé

Les images d’un radar à synthèse d’ouverture polarimétrique (PolSAR) sont classées principalement en fonction des informations de rétrodiffusion des objets au sol. Pour les régions où l’information de rétrodiffusion est complexe, il est facile de produire des erreurs de classification, ce qui pose des défis pour l’amélioration de la précision des classifications des images PolSAR. Dans ce contexte, cet article combine un modèle d’apprentissage profond et des algorithmes traditionnels pour classer l’image PolSAR. Tout d’abord, un réseau neuronal de convolution (CNN) est utilisé pour classer l’image PolSAR et les pixels clés, facilement mal classés, sont localisés selon la probabilité de bonne classification d'une classe donnée. Ensuite, l’algorithme d’optimisation adaptatif (AdaBoost) a combiné les trois algorithmes traditionnels (la machine à vecteurs de support (SVM), Wishart et l’arbre de décision) en des algorithmes puissants pour reclasser les pixels clés. Enfin, les étiquettes des pixels clés et des autres pixels de l’image sont extraites en tant que résultat final de la classification. Des expériences sur deux images PolSAR montrent que la méthode proposée peut améliorer les performances de la classification et obtenir des résultats plus précis.

Introduction

Synthetic aperture radar actively transmits electromagnetic waves and receives echoes to obtain ground information (Lee and Pottier Citation2009). It is not blocked by light and clouds and can work all the time and all the weather (Hong et al. Citation2015). PolSAR has 4 polarization channels. It can receive both horizontal and vertical electromagnetic waves after horizontal transmission and vertical transmission to obtain fully polarized backscattering information of ground objects. It has more advantages in representing the polarization characteristics of ground objects (Wang et al. Citation2022). Therefore, the classification of PolSAR images is one of its important applications.

Traditional PolSAR image classification methods can be divided into unsupervised polarization classification and supervised polarization classification according to whether training samples are needed (Zhang et al. Citation2022). In the absence of prior knowledge, unsupervised classification refers to the method of clustering statistical analysis of images by computers based on cluster theory and establishment of decision rules for classification according to statistical characteristics of characteristic parameters of samples to be classified. For example, Van Zyl (Van Zyl Citation1989) proposed an unsupervised classification method based on the relationship between phase and rotation between the incident wave and the scattered wave. The k-means algorithm (Gadhiya and Roy Citation2020) initializes the clustering center by calculating the maximum likelihood estimate of each cluster. H/α unsupervised classification (Cloude and Pottier Citation1997) classifies targets according to scattering entropy H and mean scattering angle feature parameters. Supervised classification is a process in which training samples with known attribute categories are first used to train classifiers to master the statistical characteristics of each category and then classification recognition is carried out according to classification decision rules. The supervised polarization classification for PolSAR image mainly includes the maximum likelihood (Cheng et al. Citation2013; Shokrollahi and Ebadi Citation2016), the SVM (Aghababaee et al. Citation2013; Zhao et al. Citation2023) and the decision tree (Qi et al. Citation2012; Deng et al. Citation2015). The use of samples makes the precision of the supervised classification method higher than that of the unsupervised classification method to some extent (Deng et al. Citation2015; Santana-Cedres et al. Citation2019).

With the development of machine learning, researchers have found that when the target object has rich meaning, the above classification methods of shallow structures have obvious shortcomings in terms of feature extraction and generalization ability. In recent years, due to the strong learning ability of the deep learning model (Hinton et al. Citation2006), it can directly learn rich features of images, solve complex problems, have good portability and greatly improve the accuracy of image classification, showing good performance in PolSAR image classification (Dong et al. Citation2020; Liu et al. Citation2021). At present, deep learning models such as deep Boltzmann machine, stacked auto-encoder, CNN and capsule networks have been applied in PolSAR image classification. For example, Hua et al. (Hua and Guo Citation2020) proposed a multi-layer Wishart constrained Boltzmann machine model for PolSAR image classification based on the fact that PolSAR image were subject to Wishart distribution to improve classification results. Shang et al. (Shang et al. Citation2019) added the relation between local pixels as the classification feature and proposed a new method of PolSAR classification combining scattering power and the stack Sparse Self-encoder. To greatly reduce the annotation cost and improve the classification performance, Bi et al. (Bi et al. Citation2019) proposed an active deep learning method that combined active learning and CNN to classify minimum supervised PolSAR images. Cheng et al. (Cheng et al. Citation2021) argued that a single neuron in CNN could not represent multiple polarization attributes of land cover, while a capsule network could use a vector instead of a single neuron to represent polarization attributes and proposed a layered capsule network for the classification of PolSAR images. Although the classification accuracy of the deep learning method is high, deep learning requires a large number of calculation, has high requirements on hardware, and the model design is very complex.

From the above analysis, it can be found that any classifier has its advantages and disadvantages. Existing studies have found that the advantages of multiple classifiers can be obtained by integrating multiple classifiers (Jun-Feng and Luo Citation2009; Doğan and Akay Citation2010; Mangai et al. Citation2010). The classification results of multiple classifiers are integrated to obtain better classification results than that of a single classifier (Breiman Citation2001; Polikar Citation2006; Ghimire et al. Citation2012; Maghsoudi et al. Citation2012). For example, Qin et al. (Qin et al. Citation2017) took the layer module Constrained Boltzmann machine of deep belief net as the classifier and built the integrated boosting model with adaptive boosting(AdaBoost), which improved the classification performance of PolSAR image and avoided the requirement of large data volume. He et al. (He et al. Citation2020) proposed a PolSAR image classification method combining nonlinear manifold learning and a full convolutional network. Zhu et al. (Zhu et al. Citation2021) combined deep learning technology with traditional classifiers based on scattering features to propose a new semi-supervised PolSAR image classification method to solve the deficiency of the labeling training data set. Jiao et al. (Jiao and Liu Citation2016) realized the rapid realization of Wishart distance through a special linear transformation, which accelerated the classification speed of the POLSAR image and made the polarized information available for subsequent neural networks. Jamali et al. (Jamali et al. Citation2022) used the Haar wavelet transform to carry out effective feature extraction in deep CNN, to improve the classification accuracy of the PolSAR image.

Some pixels located in regions with complex backscattering information may have features of several land covers at the same time. If these pixels are treated the same as other pixels by the classifier, they are prone to be misclassified. The above classification methods, no matter the traditional shallow classification method, deep learning classification method, or the classification method integrating multiple classifiers, do not focus on the pixels that are easy to be misclassified, affecting the final classification results. To solve this problem, this paper reclassifies pixels that are easy to be misclassified to get the final classification result. In this paper, pixels that are easy to be misclassified are key pixels, and others are general pixels. First, CNN is used to classify the PolSAR image and divides all pixels into key pixels and general pixels according to the category prediction probabilities of each pixel. The labels of general pixels are retained. Then, the AdaBoost algorithm is used to form a strong classifier consisting of the SVM classifier, the Wishart classifier and the decision tree classifier to reclassify key pixels. Finally, the labels of all pixels are combined into the final result.

Materials and methods

Materials

The first PolSAR image was obtained by the NASA/JPL AIRSAR system on August 16, 1989, from the L-band fully polarized 4-view data over Flevoland, Central Netherlands. The image size is 750 pixels ×1024 pixels, the azimuth-oriented resolution is 12.1 m, and the range-oriented resolution is 6.7 m, including 11 land covers, namely bare land, beet, grass, lucerne, pea, potato, rape, soybean, water, wheat and wood.

is the Pauli RGB image of the PolSAR image. JPL Laboratory conducted a detailed survey of this area during the imaging period and obtained the ground truth reference map (Radman et al. Citation2022) shown in , which provides a basis for evaluating classification accuracy.

Figure 1. The first PolSAR image: (a) Pauli RGB image; (b) ground truth reference map.

Figure 1. The first PolSAR image: (a) Pauli RGB image; (b) ground truth reference map.

The second experimental data of this paper is the PolSAR image collected by the EMISAR system of Denmark in Foloum on April 17, 1998. The image has a size of 1100 pixels × 876 pixels, a range resolution of 0.75 m and an azimuth resolution of 1.5 m. The scene is an L-band full PolSAR image containing five lands covers (except the background): Bare land, Broad Leaves Crop, Fine Stem Crop, Forest and Town. shows the Pauli RGB image and the corresponding ground truth reference map (He et al. Citation2020) is shown in .

Figure 2. The second PolSAR image: (a) Pauli RGB image; (b) ground truth reference map.

Figure 2. The second PolSAR image: (a) Pauli RGB image; (b) ground truth reference map.

Methods

The polarimetric scattering matrix S of PolSAR is obtained by measuring the scattering echo in each resolution unit on the ground (1) [S]=[SHHSHVSVHSVV](1)

The scattering matrix S can only describe the so-called coherent or pure scatterers, and cannot describe the so-called distributed scatterers. These scatterers can only be described statistically due to speckle noise. To reduce the influence of speckle noise, only the second-order polarization expression can be used to analyze the distributed scatterers. Coherence matrix T is one of the second-order descriptor factors. (2) [T]=12|SHH+SVV|2(SHH+SVV)(SHHSVV)*2(SHH+SVV)SHV*(SHH+SVV)*(SHHSVV)|SHHSVV|22(SHHSVV)SHV*(SHH+SVV)*SHV(SHHSVV)*SHV|SHV|2(2)

Previous studies have shown that (Duan et al. Citation2019) backscattering information of ground objects is mainly concentrated in the coherence matrix T. Therefore, the 9 elements of coherence matrix T, namely T11, T12_real, T12_imaginary, T13_real, T13_imaginary, T22, T23_real, T23_imaginary and T33, are chosen as the classification features of the method in this paper.

The flowchart of the proposed method is shown in . Firstly, training samples are selected by the random stratified sampling method, and CNN is trained by patches which are centered on the training samples to obtain a classification model. Next, the PolSAR image is clipped into multiple patches through sliding the window. Then the patches are fed into the CNN classification model to divide all pixels into general pixels and key pixels. Subsequently, the AdaBoost algorithm is used to compose a strong classifier based on an SVM classifier, a Wishart classifier and a decision tree classifier to reclassify key pixels. Finally, the labels of key pixels and general pixels are combined into the final classification result.

Figure 3. The flowchart of the proposed method.

Figure 3. The flowchart of the proposed method.

Selecting training samples

The training samples of the proposed method in this paper come from the ground truth reference map corresponding to each PolSAR image, and the number of pixels of each land cover varies greatly. To avoid the adverse impact of the imbalance of the number of training samples of each land cover on the final result, this method adopts the stratified random sampling method (Tessier et al. Citation2023) to select training samples for each land cover. Stratified random sampling is applicable to survey objects with a large number of total units and large internal differences, and has a small sampling error.

The total number of training samples was determined according to the number of land covers. There are 11 land covers in first PolSAR image, so the total number of training samples is 11,000. There are 5 land covers in second PolSAR image, so the total number of training samples is 5000.

Finding key pixels

CNN is a kind of feed forward neural network with deep structure including convolution computation, and is one of the representative algorithms of deep learning. In this paper, CNN is used to locate key pixels. The structure of CNN adopted in the method of this paper is shown in . The probability of dropout is 0.4, the learning rate is 0.001 and the number of iterations is 100.

Figure 4. The structure of CNN used in the proposed method.

Figure 4. The structure of CNN used in the proposed method.

With the training samples selected in Section “Selecting training samples” as the central pixels, the corresponding patches with size of 15 pixels × 15 pixels were constructed to train the CNN, and the classification model was obtained. Pad 7 zeros around the PolSAR image to add 14 pixels to both the rows and columns of the image. By sliding through the window, the PolSAR image was clipped to obtain 15 pixels × 15 pixels patches with each pixel as the center pixel. The patches were fed into the CNN classification model to obtain the output of the last fully connected (FC) layer and the final classification result. FC layers can complete the further fusion of features, so that the features finally seen by the neural network are global. The FC layer outputs a matrix with the size of N × C, where N is the number of pixels in the PolSAR image and C is the number of land covers. The row of this matrix represents the C label prediction probabilities of a pixel. Find out the maximum prediction probability and the second-largest prediction probability of each pixel, and the non-negative difference between the two is the probability difference. If the probability difference of a pixel is less than the given threshold, it is considered as a key pixel; otherwise, it is a general pixel. At the same time, labels of general pixels are reserved. This process is shown in .

Figure 5. The flowchart of locating the key pixels. C is the number of land cover and P1, …, PC are the C prediction probabilities of each pixel.

Figure 5. The flowchart of locating the key pixels. C is the number of land cover and P1, …, PC are the C prediction probabilities of each pixel.

Reclassifying

The SVM classifier can usually obtain higher classification accuracy (Mountrakis et al. Citation2011), the decision tree classifier is good at mining potential relationships between data (Hong et al. Citation2017), and the Wishart classifier is based on polarization coherence matrix for SAR multi-view cases (Lee et al. Citation1994). These three shallow classifiers widely used in the field of PolSAR classification are taken as weak classifiers to generate a strong classifier through the AdaBoost algorithm to reclassify key pixels.

Firstly, the SVM classifier, Wishart classifier and decision tree classifier were trained respectively to obtain three classification models.

SVM is a binary model (Khosravi et al. Citation2021). The basic idea of SVM learning is to solve the separation hyperplane that can correctly partition the training data set and has the largest geometric spacing. The separation hyperplane is wTx+b=0, where x is a value in the hyperspace, w is the hyperplane normal vector, and b is the distance from the hyperplane to the origin. LIBSVM open source library (Chang and Lin Citation2011) were used in the proposed method to realize SVM multi-class classification.

Lee et al. (Lee et al. Citation1994) extended the maximum likelihood (Entezari et al. Citation2012) rule to SAR multi-view cases and developed a supervised classification algorithm of polarimetric coherence matrix based on complex Wishart distribution. The decision rule is pωi, if ωi=ArgmaxL([T]|[Σ̂i]), where ωi is a class; Argmaxf(x) returns the x value corresponding to the maximum value of f(); [Σ̂i] is the maximum likelihood estimation of the coherence matrix. A new decision rule pωi, if ωi=Argmind([T]|[Σ̂i]) can be obtained by taking the minus sign of the above formula and removing the terms unrelated to the research clustering, where d([T][Σ̂i])=ln|[Σ̂i]|+Tr([Σ̂i]1[T]), the Tr is the matrix trace. Pixels are assigned to the class ωi with the smallest distance.

The decision tree classifier (Yin et al. Citation2020) selects the best feature recursively, and divides the training samples according to the feature. The decision tree classifier measures the selection of features by information gain and selects the features with the maximum information gain after splitting. Assumes that the sample set is X and the proportion of the samples with label k is the Pk, information entropy H(X) of X is defined as H(X)=k=1Pklog2Pk. The smaller the value of H(X), the higher the purity of X is. H(X|Y)=k=1P(yk)H(X=xk|Y) is the conditional entropy and represents the information entropy. P(yk) represents the proportion of a feature. The information gain I(X,Y)=H(X)H(X|Y) refers to the degree to which the uncertainty of the whole sample features is reduced after a certain feature is known. Finally, the decision tree classification model is obtained by pruning the generated tree.

AdaBoost (Dou et al. Citation2018) is an iterative algorithm, whose core idea is to use the same training set to train different weak classifiers, and then combine these weak classifiers to form a strong classifier. In this process, the weight of the samples misclassified by the previous weak classifier will increase, while the weight of the samples correctly classified will decrease.

(x1,y1),,(xN,yN) is the training sample set, where yi{1,1},i=1,,N is used to represent the label of the training sample. Initialize the weight distribution D1 of the training sample set. Each training sample is assigned the same weight wi=1/N,i=1,,N, then the initial weight distribution of the training set is: (3) D1=(w1,,wN)=(1/N,,1/N)(3) where wi is the weight of each training sample, and N is the number of training samples.

Then the t-th iteration is implemented, t = 1, 2, 3.

A weak classifier h with the lowest error rate is selected as the Ht basic classifier, and the error of the basic classifier on the distribution Dt is calculated (4) et=P(Ht(xi)yi)=i=1NwtiI(Ht(xi)yi)(4) (5) I={1, Ht(xi)yi0, Ht(xi)=yi(5) where Dt=(w1,,wN) is the weight of the training sample set at the t-th iteration; et is the error rate.

Calculate the weight of the basic classifier in the final strong classifier (6) αt=12ln(1etet)(6) where, αt is the weight of the weak classifier in the t-th iteration.

Update the weight distribution Dt+1 of the training samples: (7) Dt+1=Dt(i)exp(αtyiHt(xi))Zt(7) where Zt=2et(1et) is the normalization constant; when samples are misclassified, yiHt(xi)=1; when samples are correctly classified, yiHt(xi)=1.

Finally, each weak classifier is combined according to its weight: (8) f(x)=t=13αtHt(x)(8)

Through the function sign, a strong classifier is obtained as follows: (9) Hfinal=sign(f(x))=sign(t=13αtHt(x))(9)

The AdaBoost algorithm is aimed at the binary classification problem. Since the PolSAR image used in this paper contains C land covers, the proposed method in this paper realizes classification by iterating the AdaBoost algorithm C-1 times. In each iteration, one land cover was taken as 1 and other land covers were taken as −1, and a strong classifier was constructed to classify key pixels. Key pixels labeled 1 did not participate in the next iteration and their land covers were recorded. After C-1 iterations, new land covers of all key pixels were obtained.

Finally, the land covers of general pixels and key pixels were combined into the classification results of PolSAR.

Results and discussion

The method in this paper was implemented using MATLAB2022b programming, and the computer was configured with 48 GB memory and NVIDIA GeForce RTX 2080TiGPU. To verify the superiority of the proposed method, the following three aspects of comparative experiments were carried out.

Precision comparison before and after reclassification of key pixels

For the first PolSAR image, when the probability difference was 0.7, the overall classification accuracy reached the maximum of 92.10%. For the second PolSAR image, when the probability difference was 0.5, the overall classification accuracy reached the maximum of 91.02%. The key pixels shown in were located. It can be found that most of the key pixels are located at the edge of each land cover, that is, the area with complex backscattering information.

Figure 6. (a) Key pixels of the first PolSAR image when the probability difference is 0.7; (b) key pixels of the second PolSAR image when the probability difference is 0.5.

Figure 6. (a) Key pixels of the first PolSAR image when the probability difference is 0.7; (b) key pixels of the second PolSAR image when the probability difference is 0.5.

The number of key pixels and pixels on the ground truth reference map is shown in and .

Table 1. The number of key pixels and pixels in the ground reference map for the first PolSAR image.

Table 2. The number of key pixels and pixels in the ground reference map for the second PolSAR image.

As shown in and , labels of the key pixels in two PolSAR images before and after reclassification has changed.

Figure 7. The labels of key pixels in the first PolSAR image: (a) before reclassification; (b) after reclassification.

Figure 7. The labels of key pixels in the first PolSAR image: (a) before reclassification; (b) after reclassification.

Figure 8. The labels of key pixels in the second PolSAR image: (a) before reclassification; (b) after reclassification.

Figure 8. The labels of key pixels in the second PolSAR image: (a) before reclassification; (b) after reclassification.

Since the ground truth reference map does not cover the entire PolSAR image, the analysis that follows in this section focuses only on the key pixels in the ground truth reference map, rather than the entire image.

According to the ground truth reference map of the first experimental data shown in and the number of key pixels of each land cover in the ground truth reference map shown in , the classification accuracy of key pixels before and after reclassification is statistically analyzed, and the results are shown in . Among the 11 land covers, the classification accuracy of 10 land covers is improved. The accuracy of wood increases by 0.3%, which is the smallest gain. The accuracy of soybean increases by 38.29%, which is the biggest gain. That the average classification accuracy of key pixels in the first experimental data is improved by 9.44% indicates the method in this paper is effective in improving the classification accuracy of key pixels. However, the classification accuracy of water decreases by 29.63% from 88.89% to 59.26%, and the reasons for this result will be analyzed.

Table 3. Before and after reclassification, the statistics of key pixels in the ground reference map of the first PolSAR image.

As can be seen from , there are only 27 key pixels of water in the ground truth reference map. These pixels are located in the rectangular box of , namely 8 pixels in the upper right corner of the rectangular box and 19 pixels in the lower left corner. As is shown in , before reclassification, eight pixels in the upper right corner are correctly labeled as water; 16 pixels in the lower left corner are correctly labeled as water and three pixels in the lower left corner are incorrectly labeled as Bare Land. As is shown in , after reclassification, six pixels in the upper right corner are correctly labeled as water and two pixels in the upper right are incorrectly labeled as potatoes; 10 pixels in the lower left corner are correctly labeled as water and nine pixels in the lower left corner are incorrectly labeled as Bare Land.

Figure 9. (a) and (c) the labels of key pixels in the ground truth reference map before and after reclassification, and the label of the pixels in the top right black rectangle box is water; (b) and (d) enlarged area inside the black rectangle box in the upper right corner of (a) and (c); (e) the Pauli RGB image of the first PolSAR image; (f) enlarged area inside the red rectangle box in the upper right corner of (e).

Figure 9. (a) and (c) the labels of key pixels in the ground truth reference map before and after reclassification, and the label of the pixels in the top right black rectangle box is water; (b) and (d) enlarged area inside the black rectangle box in the upper right corner of (a) and (c); (e) the Pauli RGB image of the first PolSAR image; (f) enlarged area inside the red rectangle box in the upper right corner of (e).

As can be seen from , the land cover in the rectangle box shown in is water. However, the upper right corner of the rectangle box is an abnormal light color area and the lower left corner is a slightly larger abnormal dark blue area, which is shown in . Before reclassification, the CNN classification method was used, and the input data were patches with a size of 15 pixels ×15 pixels. The feature value of the central pixel in a patch was easily affected by the feature values of the surrounding pixels. Because the upper right corner area was small, all pixels in this area were greatly affected by the feature values of pixels outside the region and easy to be labeled as the land cover of surrounding pixels. While the lower left corner area was large, so some pixels in this area were greatly affected by the feature values of the pixels outside the region. As a result, some pixels were easily labeled as the land cover of the surrounding pixels. When reclassifying, the combination of shallow classifiers was used, and the input data was pixels, which were less affected by the feature values of surrounding pixels. The land cover of pixels depended largely on their features. So, more pixels in the two regions were labeled as water when CNN classification is adopted than when the proposed method was adopted. Although the two areas were obviously different from the surrounding pixels from the visual effect, the ground truth reference map shown in was drawn by hand and the two very small anomalous land covers in the top right rectangular box of had to be labeled water. Besides, there are only 27 key pixels of water, and even a slight change in the number makes the accuracy change before and after the reclassification appears drastic. That is what caused water’s accuracy to drop after reclassification. The above analysis also shows that CNN is inferior to the proposed method in describing the details of land covers.

According to the ground reference map of the second experimental data shown in , the classification accuracy of key pixels before and after reclassification is statistically analyzed. The results are shown in . The classification accuracy of three out of five land covers is improved. The accuracy of Fine Stem Crop increases the least, with 10.96% and that of Forest increases the most, with 30.72%. The average classification accuracy of key pixels in the second experimental data is improved by 9.61%, which also proves the effectiveness of the proposed method in improving the classification accuracy of key pixels. But Broad Leaves Crop’s accuracy dropped slightly from 58.16 to 57.52%, and Town’s dropped sharply from 60.08 to 42.38%. The accuracy of Broad Leaves Crop changes a little before and after reclassification, so next, the reasons for the decline in the accuracy of Town will be analyzed.

Table 4. Before and after reclassification, the statistics of key pixels in the ground reference map of the second PolSAR image.

In the ground reference map, the number of key pixels labeled as Town is 6310 shown in . Before and after reclassification, the labels of these key pixels are recorded, as shown in . After reclassification, the labels of some key pixels are changed from Bare Land and Town to Fine Stem Crop, Broad Leaves Crop and Forest, which can be found in and and . As shown in , Town is fragmented and mixed with Fine Stem Crop, Broad Leaves Crop and Forest. Similar to the case of water in the first PolSAR image, when CNN was used for classification, the fewer land covers were easily ignored, while the proposed method was more likely to find them. However, the hand-painted ground truth reference map shown in could not separate the small land covers from the largest number of land covers and had to ignore these small land covers. This makes the accuracy of Town before reclassification greater than that after classification. The above analysis also shows that the proposed method has more advantages in describing the details of ground objects than CNN ().

Figure 10. (a) Pauli RGB image of the second PolSAR image; (b) the ground truth reference map in the four rectangular boxes; (c) – (f) the four rectangular boxes with 3 times magnification; (g) – (h) the labels of key pixels labeled as Town in the ground truth reference map before and after reclassification.

Figure 10. (a) Pauli RGB image of the second PolSAR image; (b) the ground truth reference map in the four rectangular boxes; (c) – (f) the four rectangular boxes with 3 times magnification; (g) – (h) the labels of key pixels labeled as Town in the ground truth reference map before and after reclassification.

Table 5. The labels of key pixels labeled as town in the ground reference map before and after reclassification.

Comparison of visual effects of various methods

To evaluate the visual effect of the proposed method, five groups of comparison experiments were set up, i.e., the SVM classifier, the Wishart classifier, the decision tree classifier, the CNN and the AdaBoost algorithm (a SVM classifier, a Wishart classifier and a decision tree classifier were taken as weak classifiers).

shows the classification results of the methods described above. With the ground truth map shown in , four elliptical regions were selected for comparison on each classification result. It can be found that, compared with the classification results of the SVM classifier, the Wishart classifier, the decision tree classifier and the AdaBoost algorithm, the classification results of the CNN are purer and less spotted, but the classification results of the CNN are fuzzier and the texture is lost. The texture of the result of the proposed method is clearer than that of the CNN. There are fewer spots in the classification results of the proposed method than in the shallow classifiers and the AdaBoost algorithm.

Figure 11. Classification results of different methods for the first PolSAR image: (a) CNN; (b) SVM; (c) Wishart; (d) decision tree; (e) AdaBoost; (f) the proposed method.

Figure 11. Classification results of different methods for the first PolSAR image: (a) CNN; (b) SVM; (c) Wishart; (d) decision tree; (e) AdaBoost; (f) the proposed method.

As shown in , these are the classification results of the second experimental data respectively using the CNN, the SVM classifier, the Wishart classifier, the decision tree classifier, the AdaBoost algorithm and the proposed method. With the ground truth map shown in , four elliptical regions were selected for comparison on each classification result. It can also be found that, compared with the SVM classifier, the Wishart classifier, the decision tree classifier and the AdaBoost algorithm, the classification results of the CNN are with fewer spots but fuzzier and the texture is lost. The texture of the results of the proposed method is clearer than that of the CNN, while there are much fewer spots in the classification results of the proposed method than those of the shallow classifiers and the AdaBoost algorithm.

Figure 12. Classification results of different methods for the second PolSAR image: (a) CNN; (b) SVM; (c) Wishart; (d) decision tree; (e) AdaBoost; (f) the proposed method.

Figure 12. Classification results of different methods for the second PolSAR image: (a) CNN; (b) SVM; (c) Wishart; (d) decision tree; (e) AdaBoost; (f) the proposed method.

Comparison of classification accuracy of various methods

With the ground reference maps shown in and as references, the classification results of the two experimental data were analyzed from four aspects of overall accuracy (OA), Kappa coefficient (KC), producer accuracy and user accuracy by using a confusion matrix. The classification results of the proposed method are compared with those of the CNN, the SVM classifier, the Wishart classifier, the decision tree classifier and the AdaBoost algorithm.

The OA and KC of the classification results of the first experimental data are shown in . In terms of accuracy, compared with the other 5 classification methods, the OA of the proposed method is improved by 2.22–20.38%, and the KC is increased by 0.03–0.24.

Table 6. The overall accuracy and Kappa coefficient of the first PolSAR image.

The producer accuracy of the classification results of each land cover using different methods for the first experimental data are shown in . The standardized differences (STD) of the producer accuracy of each land cover using these methods are adopted in the last row of the table. They are used for the local and global evaluation respectively. The producer accuracy of each land cover using the proposed method is larger than the minimum and close to the maximum or even greater than the maximum. The STD is used to judge the equilibrium of the producer accuracy of all land covers using each method. The smaller the value, the more balanced the data. As can be seen from the data in , the STD of the proposed method is 0.0569, which is 0.0012 larger than the minimum value. It shows that the producer accuracy of the proposed method is relatively balanced.

Table 7. The producer accuracy of the first PolSAR image.

The user accuracy of each land cover using different methods for the first experimental data is shown in . The STD of the user accuracy of all land cover using each method is listed in the last row of the table. The user accuracy of the proposed method is larger than the minimum value and close to the maximum value. The STD of the proposed method is 0.0482, which is 0.0047 larger than the minimum value and very close to the minimum value.

Table 8. The user accuracy of the first PolSAR image.

Therefore, the proposed method is superior to the other five classification methods in terms of the producer accuracy and the user accuracy, both locally and globally.

The OA and KC of the classification results of the second experimental data are shown in . Compared with the other five classification methods, the OA of the proposed method increases by 1.06–11.53%, and the KC increases by 0.02–0.15.

Table 9. The overall accuracy and Kappa coefficient of the second PolSAR image.

In the second experimental data, the producer accuracy of each land cover using different classification methods is shown in . The last row of the table lists the STD of the user accuracy using each method. It can be found that the producer accuracy of Fine Stem Crop and Forest using the proposed method is the maximum, while that of Bare Land, Broad Leaves Crop and Town using the proposed method is close to the maximum, with the difference values of 0.95, 0.08 and 2.51%, respectively. Moreover, the STD of the proposed method is the smallest, which is 0.0703, indicating that the producer accuracy using the proposed method is the most balanced.

Table 10. The producer accuracy of the second PolSAR image.

The user accuracy of each land cover using different classification methods is shown in for the second experimental data. The last row shows the STD of the user accuracy using each method. It can be found that the user accuracy of Bare land, Fine Stem Crop and Town using the proposed method is the highest, while those of Broad Leaves Crop and Forest are close to the maximum accuracy, with a difference of 0.49 and 0.81%. Moreover, the STD of the proposed method is the smallest, which is 0.0527, indicating that the user accuracy using the proposed method is the most balanced.

Table 11. The user accuracy for the second PolSAR image.

Conclusion

The proposed method in this paper combines the advantages of the deep learning model and shallow classification method to reclassify pixels that are easy to be misclassified and obtain better classification results. In terms of local classification accuracy, overall classification accuracy and visual effect, the proposed classification method performs better than the other methods listed above.

Acknowledgement

The authors would like to thank the anonymous reviewers and the editor for their constructive comments and suggestions.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This work was supported by the Science and Technology Research Project of Hubei Province, Department of Education and Excellent Young and Middle-aged Science and Technology Innovation Team Program for Colleges and Universities in Hubei Province under [Grant Q20203006]; University-level research start-up fund under [Grant ESRC20220046]; and Natural Science Foundation of Hubei Province, [2019CFB827].

References

  • Aghababaee, H., Amini, J., and Tzeng, Y.C. 2013. “Contextual PolSAR image classification using fractal dimension and support vector machines.” European Journal of Remote Sensing, Vol. 46(No. 1): pp. 317–332. doi:10.5721/EuJRS20134618.
  • Bi, H.X., Xu, F., Wei, Z.Q., Xue, Y., and Xu, Z.B. 2019. “An active deep learning approach for minimally supervised PolSAR image classification.” IEEE Transactions on Geoscience and Remote Sensing, Vol. 57(No. 11): pp. 9378–9395. doi:10.1109/TGRS.2019.2926434.
  • Breiman, L. 2001. “Random forests.” Machine Learning, Vol. 45(No. 1): pp. 5–32. doi:10.1023/A:1010933404324.
  • Chang, C.C., and Lin, C.J. 2011. “Libsvm: A library for support vector machines.” ACM Transactions on Intelligent Systems and Technology, Vol. 2(No. 3): pp. 1–27. doi:10.1145/1961189.1961199.
  • Cheng, J.D., Zhang, F., Xiang, D.L., Yin, Q., Zhou, Y.S., and Wang, W. 2021. “PolSAR image land cover classification based on hierarchical capsule network.” Remote Sensing, Vol. 13(No. 16): pp. 3132. doi:10.3390/rs13163132.
  • Cheng, X.G., Huang, W.L., and Gong, J.Y. 2013. “A decomposition-free scattering mechanism classification method for PolSAR images with Neumann’s model.” Remote Sensing Letters, Vol. 4(No. 12): pp. 1176–1184. doi:10.1080/2150704X.2013.858840.
  • Cloude, S.R., and Pottier, E. 1997. “An entropy based classification scheme for land applications of polarimetric SAR.” IEEE Transactions on Geoscience and Remote Sensing, Vol. 35(No. 1): pp. 68–78. doi:10.1109/36.551935.
  • Deng, L., Yan, Y-N., and Sun, C. 2015. “Use of sub-aperture decomposition for supervised PolSAR classification in urban area.” Remote Sensing, Vol. 7(No. 2): pp. 1380–1396. doi:10.3390/rs70201380.
  • Doğan, H., and Akay, O. 2010. “Using AdaBoost classifiers in a hierarchical framework for classifying surface images of marble slabs.” Expert Systems with Applications, Vol. 37(No. 12): pp. 8814–8821. doi:10.1016/j.eswa.2010.06.019.
  • Dong, H.W., Zou, B., Zhang, L.M., and Zhang, S.Y. 2020. “Automatic design of CNNS via differentiable neural architecture search for PolSAR image classification.” IEEE Transactions on Geoscience and Remote Sensing, Vol. 58(No. 9): pp. 6362–6375. doi:10.1109/TGRS.2020.2976694.
  • Dou, P., Chen, Y.B., and Yue, H.Y. 2018. “Remote-sensing imagery classification using multiple classification algorithm-based AdaBoost.” International Journal of Remote Sensing, Vol. 39(No. 3): pp. 619–639. doi:10.1080/01431161.2017.1390276.
  • Duan, Y., Chen, N., and Chen, Y.B. 2019. “A novel PolSAR image classification method based on optimal polarimetric features and contextual information.” Canadian Journal of Remote Sensing, Vol. 45(No. 6): pp. 795–813. doi:10.1080/07038992.2019.1697222.
  • Entezari, I., Motagh, M., and Mansouri, B. 2012. “Comparison of the performance of l-band polarimetric parameters for land cover classification.” Canadian Journal of Remote Sensing, Vol. 38(No. 5): pp. 629–643. doi:10.5589/m12-051.
  • Gadhiya, T., and Roy, A.K. 2020. “Superpixel-driven optimized Wishart network for fast PolSAR image classification using global k-means algorithm.” IEEE Transactions on Geoscience and Remote Sensing, Vol. 58(No. 1): pp. 97–109. doi:10.1109/TGRS.2019.2933483.
  • Ghimire, B., Rogan, J., Galiano, V.R., Panday, P., and Neeti, N. 2012. “An evaluation of bagging, boosting, and random forests for land-cover classification in Cape Cod, Massachusetts, USA.” GIScience & Remote Sensing, Vol. 49(No. 5): pp. 623–643. doi:10.2747/1548-1603.49.5.623.
  • He, C., Tu, M.X., Xiong, D.H., and Liao, M.S. 2020. “Nonlinear manifold learning integrated with fully convolutional networks for PolSAR image classification.” Remote Sensing, Vol. 12(No. 4):pp. 655. pp doi:10.3390/rs12040655.
  • Hinton, G.E., Osindero, S., and Teh, Y.W. 2006. “A fast learning algorithm for deep belief nets.” Neural Computation, Vol. 18(No. 7): pp. 1527–1554. doi:10.1162/NECO.2006.18.7.1527.
  • Hong, G., Wang, S., Li, J., and Huang, J. 2015. “Fully polarimetric synthetic aperture radar (SAR) processing for crop type identification.” Photogrammetric Engineering & Remote Sensing, Vol. 81(No. 2): pp. 109–117. doi:10.14358/PERS.81.2.109.
  • Hong, W., Shao, L., and Yin, Q. 2017. "Decision hierarchical classification by FLD for vegetation application using PolSAR features." IEEE International Geoscience & Remote Sensing Symposium, Fort Worth, TX, July 23–28, 2017.
  • Hua, W.Q., and Guo, Y.H. 2020. “Classification of polarimetric synthetic aperture radar images based on multilayer wishart-restricted Boltzmann machine.” Journal of Applied Remote Sensing, Vol. 14(No. 03): pp. 1–13. doi:10.1117/1.JRS.14.036516.
  • Jamali, A., Mahdianpari, M., Mohammadimanesh, F., Bhattacharya, A., and Homayouni, S. 2022. “PolSAR image classification based on deep convolutional neural networks using wavelet transformation.” IEEE Geoscience and Remote Sensing Letters, Vol. 19(No. No. 2022): pp. 1–5. doi:10.1109/LGRS.2022.3185118.
  • Jiao, L.C., and Liu, F. 2016. “Wishart deep stacking network for fast PolSAR image classification.” IEEE Transactions on Image Processing, Vol. 25(No. 7): pp. 3273–3286. doi:10.1109/TIP.2016.2567069.
  • Jun-Feng, G.E., and., and Luo, Y.-P. 2009. “A comprehensive study for asymmetric AdaBoost and its application in object detection.” Acta Automatica Sinica, Vol. 35(No. 11): pp. 1403–1409. doi:10.1016/S1874-1029(08)60115-9.
  • Khosravi, I., Razoumny, Y., Afkoueieh, J.H., and Alavipanah, S.K. 2021. “Fully polarimetric synthetic aperture radar data classification using probabilistic and non-probabilistic kernel methods.” European Journal of Remote Sensing, Vol. 54(No. 1): pp. 310–317. doi:10.1080/22797254.2021.1924081.
  • Lee, J.S., Grunes, M.R., and Kwok, R. 1994. “Classification of multi-look polarimetric SAR imagery based on complex Wishart distribution.” International Journal of Remote Sensing, Vol. 15(No. 11): pp. 2299–2311. doi:10.1080/01431169408954244.
  • Lee, J. S., and Pottier, E. 2009. Polarimetric Radar Imaging: From Basics to Applications. Boca Raton: CRC press.
  • Liu, S.J., Luo, H.W., and Shi, Q. 2021. “Active ensemble deep learning for polarimetric synthetic aperture radar image classification.” IEEE Geoscience and Remote Sensing Letters, Vol. 18(No. 9): pp. 1580–1584. doi:10.1109/LGRS.2020.3005076.
  • Maghsoudi, Y., Collins, M., and Leckie, D.G. 2012. “Polarimetric classification of boreal forest using nonparametric feature selection and multiple classifiers.” International Journal of Applied Earth Observation and Geoinformation, Vol. 19(No. No. 2012): pp. 139–150. doi:10.1016/j.jag.2012.04.015.
  • Mangai, U.G., Samanta, S., Das, S., and Chowdhury, P.R. 2010. “A survey of decision fusion and feature fusion strategies for pattern classification.” IETE Technical Review, Vol. 27(No. 4): pp. 293–307. doi:10.4103/0256-4602.64604.
  • Mountrakis, G., Im, J., and Ogole, C. 2011. “Support vector machines in remote sensing: A review.” ISPRS Journal of Photogrammetry and Remote Sensing, Vol. 66(No. 3): pp. 247–259. doi:10.1016/j.isprsjprs.2010.11.001.
  • Polikar, R. 2006. “Essemble based systems in decision making.” IEEE Circuits and Systems Magazine, Vol. 6(No. 3): pp. 21–45. doi:10.1109/MCAS.2006.1688199.
  • Qi, Z.X., Yeh, A.G.O., Li, X., and Lin, Z. 2012. “A novel algorithm for land use and land cover classification using radarsat-2 polarimetric SAR data.” Remote Sensing of Environment, Vol. 118(No. 3): pp. 21–39. doi:10.1016/j.rse.2011.11.001.
  • Qin, F., Guo, J., and Sun, W. 2017. “Object-oriented ensemble classification for polarimetric SAR imagery using restricted Boltzmann machines.” Remote Sensing Letters, Vol. 8(No. 3): pp. 204–213. doi:10.1080/2150704X.2016.1258128.
  • Radman, A., Mahdianpari, M., Brisco, B., Salehi, B., and Mohammadimanesh, F. 2022. “Dual-branch fusion of convolutional neural network and graph convolutional network for PolSAR image classification.” Remote Sensing, Vol. 15(No. 1): pp. 75. doi:10.3390/rs15010075.
  • Santana-Cedres, D., Gomez, L., Trujillo, A., Aleman-Flores, M., Deriche, R., and Alvarez, L. 2019. “Supervised classification of fully PolSAR images using active contour models.” IEEE Geoscience and Remote Sensing Letters, Vol. 16(No. 7): pp. 1165–1169. doi:10.1109/LGRS.2019.2892524.
  • Shang, R.H., Liu, Y.K., Wang, J.M., Jiao, L.C., and Stolkin, R. 2019. “Stacked auto-encoder for classification of polarimetric SAR images based on scattering energy.” International Journal of Remote Sensing, Vol. 40(No. 13): pp. 5094–5120. doi:10.1080/01431161.2019.1579378.
  • Shokrollahi, M., and Ebadi, H. 2016. “Improving the accuracy of land cover classification using fusion of polarimetric SAR and hyperspectral images.” Journal of the Indian Society of Remote Sensing, Vol. 44(No. 6): pp. 1017–1024. doi:10.1007/s12524-016-0559-4.
  • Tessier, N., Boissonnot, R., Desvignes, V., Fröchen, M., Merlo, M., Blanchard, O., Chevrier, C., et al. 2023. “Use and storage of pesticides at home in France (the pesti’home survey 2014).” Environmental Research, Vol. 216(No. Pt 2): pp. 114452. doi:10.1016/j.envres.2022.114452.
  • Van Zyl, J.J. 1989. “Unsupervised classification of scattering behavior using radar polarimetry data.” IEEE Transactions on Geoscience and Remote Sensing, Vol. 27(No. 1): pp. 36–45. doi:10.1109/36.20273.
  • Wang, J.L., Hou, B., Ren, B., Zhang, Y.K., Yang, M.J., Wang, S., and Jiao, L.C. 2022. “Parameter selection of Touzi decomposition and a distribution improved autoencoder for PolSAR image classification.” ISPRS Journal of Photogrammetry and Remote Sensing, Vol. 186(No. 2022): pp. 246–266. doi:10.1016/j.isprsjprs.2022.02.003.
  • Yin, Q., Cheng, J.D., Zhang, F., Zhou, Y.S., Shao, L.Y., and Hong, W. 2020. “Interpretable PolSAR image classification based on adaptive-dimension feature space decision tree.” IEEE Access., Vol. 8(No. No. 2020): pp. 173826–173837. doi:10.1109/ACCESS.2020.3023134.
  • Zhang, L., Zhang, S., Zou, B., and Dong, H. 2022. “Unsupervised deep representation learning and few-shot classification of PolSAR images.” IEEE Transactions on Geoscience and Remote Sensing, Vol. 60(No. 2020): pp. 1–16. doi:10.1109/TGRS.2020.3043191.
  • Zhao, M., Cheng, Y., Qin, X., Yu, W., and Wang, P. 2023. “Semi-supervised classification of PolSAR images based on co-training of CNN and SVM with limited labelled samples.” Sensors, Vol. 23(No. 4): pp. 2109. doi:10.3390/s23042109.
  • Zhu, L.K., Ma, X.S., Wu, P.H., and Xu, J.G. 2021. “Multiple classifiers based semi-supervised polarimetric SAR image classification method.” Sensors, Vol. 21(No. 9): pp. 3006. doi:10.3390/s21093006.