A Novel Classification Method for PolSAR Image Combining the Deep Learning Model and Adaptive Boosting of Shallow Classifiers

Abstract Polarimetric synthetic aperture radar (PolSAR) images are classified mainly according to the backscattering information of ground objects. For regions with complex backscattering information, misclassification is easy to occur, which leads to challenges in improving the classification accuracy of the PolSAR image. Given this situation, this paper combines the Deep Learning Model and traditional classifiers to classify PolSAR image. First, the Convolution Neural Network (CNN) was used to classify the PolSAR image and according to the category prediction probability of pixels, the key pixels easily misclassified are located. Then, the adaptive boosting (AdaBoost) algorithm combined the three shallow classifiers (the Support Vector Machine (SVM), the Wishart and the Decision Tree classifier) into strong classifiers to reclassify the key pixels. Finally, the labels of key pixels and other pixels are output as the final classification result. Experiments on two PolSAR images show that the proposed method can improve classification performance and obtain better classification results.


Introduction
Synthetic aperture radar actively transmits electromagnetic waves and receives echoes to obtain ground information (Lee and Pottier 2009).It is not blocked by light and clouds and can work all the time and all the weather (Hong et al. 2015).PolSAR has 4 polarization channels.It can receive both horizontal and vertical electromagnetic waves after horizontal transmission and vertical transmission to obtain fully polarized backscattering information of ground objects.It has more advantages in representing the polarization characteristics of ground objects (Wang et al. 2022).Therefore, the classification of PolSAR images is one of its important applications.
Traditional PolSAR image classification methods can be divided into unsupervised polarization classification and supervised polarization classification according to whether training samples are needed (Zhang et al. 2022).In the absence of prior knowledge, unsupervised classification refers to the method of clustering statistical analysis of images by computers based on cluster theory and establishment of decision rules for classification according to statistical characteristics of characteristic parameters of samples to be classified.For example, Van Zyl (Van Zyl 1989) proposed an unsupervised classification method based on the relationship between phase and rotation between the incident wave and the scattered wave.The k-means algorithm (Gadhiya and Roy 2020) initializes the clustering center by calculating the maximum likelihood estimate of each cluster.H/a unsupervised classification (Cloude and Pottier 1997) classifies targets according to scattering entropy H and mean scattering angle feature parameters.Supervised classification is a process in which training samples with known attribute categories are first used to train classifiers to master the statistical characteristics of each category and then classification recognition is carried out according to classification decision rules.The supervised polarization classification for PolSAR image mainly includes the maximum likelihood (Cheng et al. 2013;Shokrollahi and Ebadi 2016), the SVM (Aghababaee et al. 2013;Zhao et al. 2023) and the decision tree (Qi et al. 2012;Deng et al. 2015).The use of samples makes the precision of the supervised classification method higher than that of the unsupervised classification method to some extent (Deng et al. 2015;Santana-Cedres et al. 2019).
With the development of machine learning, researchers have found that when the target object has rich meaning, the above classification methods of shallow structures have obvious shortcomings in terms of feature extraction and generalization ability.In recent years, due to the strong learning ability of the deep learning model (Hinton et al. 2006), it can directly learn rich features of images, solve complex problems, have good portability and greatly improve the accuracy of image classification, showing good performance in PolSAR image classification (Dong et al. 2020;Liu et al. 2021).At present, deep learning models such as deep Boltzmann machine, stacked auto-encoder, CNN and capsule networks have been applied in PolSAR image classification.For example, Hua et al. (Hua and Guo 2020) proposed a multi-layer Wishart constrained Boltzmann machine model for PolSAR image classification based on the fact that PolSAR image were subject to Wishart distribution to improve classification results.Shang et al. (Shang et al. 2019) added the relation between local pixels as the classification feature and proposed a new method of PolSAR classification combining scattering power and the stack Sparse Self-encoder.To greatly reduce the annotation cost and improve the classification performance, Bi et al. (Bi et al. 2019) proposed an active deep learning method that combined active learning and CNN to classify minimum supervised PolSAR images.Cheng et al. (Cheng et al. 2021) argued that a single neuron in CNN could not represent multiple polarization attributes of land cover, while a capsule network could use a vector instead of a single neuron to represent polarization attributes and proposed a layered capsule network for the classification of PolSAR images.Although the classification accuracy of the deep learning method is high, deep learning requires a large number of calculation, has high requirements on hardware, and the model design is very complex.
From the above analysis, it can be found that any classifier has its advantages and disadvantages.Existing studies have found that the advantages of multiple classifiers can be obtained by integrating multiple classifiers (Jun-Feng and Luo 2009; Do� gan and Akay 2010; Mangai et al. 2010).The classification results of multiple classifiers are integrated to obtain better classification results than that of a single classifier (Breiman 2001;Polikar 2006;Ghimire et al. 2012;Maghsoudi et al. 2012).For example, Qin et al. (Qin et al. 2017)  Some pixels located in regions with complex backscattering information may have features of several land covers at the same time.If these pixels are treated the same as other pixels by the classifier, they are prone to be misclassified.The above classification methods, no matter the traditional shallow classification method, deep learning classification method, or the classification method integrating multiple classifiers, do not focus on the pixels that are easy to be misclassified, affecting the final classification results.To solve this problem, this paper reclassifies pixels that are easy to be misclassified to get the final classification result.In this paper, pixels that are easy to be misclassified are key pixels, and others are general pixels.First, CNN is used to classify the PolSAR image and divides all pixels into key pixels and general pixels according to the category prediction probabilities of each pixel.The labels of general pixels are retained.Then, the AdaBoost algorithm is used to form a strong classifier consisting of the SVM classifier, the Wishart classifier and the decision tree classifier to reclassify key pixels.Finally, the labels of all pixels are combined into the final result.

Materials
The first PolSAR image was obtained by the NASA/JPL AIRSAR system on August 16, 1989, from the L-band fully polarized 4-view data over Flevoland, Central Netherlands.The image size is 750 pixels �1024 pixels, the azimuth-oriented resolution is 12.1 m, and the range-oriented resolution is 6.7 m, including 11 land covers, namely bare land, beet, grass, lucerne, pea, potato, rape, soybean, water, wheat and wood.
Figure 1a is the Pauli RGB image of the PolSAR image.JPL Laboratory conducted a detailed survey of this area during the imaging period and obtained the ground truth reference map (Radman et al. 2022) shown in Figure 1b, which provides a basis for evaluating classification accuracy.
The second experimental data of this paper is the PolSAR image collected by the EMISAR system of Denmark in Foloum on April 17, 1998.The image has a size of 1100 pixels � 876 pixels, a range resolution of 0.75 m and an azimuth resolution of 1.5 m.The scene is an L-band full PolSAR image containing five lands covers (except the background): Bare land, Broad Leaves Crop, Fine Stem Crop, Forest and Town. Figure 2a shows the Pauli RGB image and the corresponding ground truth reference map (He et al. 2020) is shown in Figure 2b.

Methods
The polarimetric scattering matrix S of PolSAR is obtained by measuring the scattering echo in each resolution unit on the ground The scattering matrix S can only describe the socalled coherent or pure scatterers, and cannot describe the so-called distributed scatterers.These scatterers can only be described statistically due to speckle noise.To reduce the influence of speckle noise, only the second-order polarization expression can be used to analyze the distributed scatterers.Coherence matrix T is one of the second-order descriptor factors. (2) Previous studies have shown that (Duan et al. 2019) backscattering information of ground objects is mainly concentrated in the coherence matrix T. Therefore, the 9 elements of coherence matrix T, namely T11, T12_real, T12_imaginary, T13_real, T13_ imaginary, T22, T23_real, T23_imaginary and T33, are chosen as the classification features of the method in this paper.
The flowchart of the proposed method is shown in Figure 3. Firstly, training samples are selected by the random stratified sampling method, and CNN is trained by patches which are centered on the training samples to obtain a classification model.Next, the PolSAR image is clipped into multiple patches through sliding the window.Then the patches are fed into the CNN classification model to divide all pixels into general pixels and key pixels.Subsequently, the AdaBoost algorithm is used to compose a strong classifier based on an SVM classifier, a Wishart classifier and a decision tree classifier to reclassify key pixels.Finally, the labels of key pixels and general pixels are combined into the final classification result.

Selecting training samples
The training samples of the proposed method in this paper come from the ground truth reference map corresponding to each PolSAR image, and the number of pixels of each land cover varies greatly.To avoid the adverse impact of the imbalance of the number of training samples of each land cover on the final result, this method adopts the stratified random sampling method (Tessier et al. 2023) to select training samples for each land cover.Stratified random sampling is applicable to survey objects with a large number of total units and large internal differences, and has a small sampling error.
The total number of training samples was determined according to the number of land covers.There are 11 land covers in first PolSAR image, so the total Finding key pixels CNN is a kind of feed forward neural network with deep structure including convolution computation, and is one of the representative algorithms of deep learning.In this paper, CNN is used to locate key pixels.The structure of CNN adopted in the method of this paper is shown in Figure 4.The probability of dropout is 0.4, the learning rate is 0.001 and the number of iterations is 100.
With the training samples selected in Section "Selecting training samples" as the central pixels, the corresponding patches with size of 15 pixels � 15 pixels were constructed to train the CNN, and the classification model was obtained.Pad 7 zeros around the PolSAR image to add 14 pixels to both the rows and columns of the image.By sliding through the window, the PolSAR image was clipped to obtain 15 pixels � 15 pixels patches with each pixel as the center pixel.The patches were fed into the CNN classification model to obtain the output of the last fully connected (FC) layer and the final classification result.FC layers can complete the further fusion of features, so that the features finally seen by the neural network are global.The FC layer outputs a matrix with the size of N � C, where N is the number of pixels in the PolSAR image and C is the number of land covers.The row of this matrix represents the C label prediction probabilities of a pixel.Find out the maximum prediction probability and the second-largest prediction probability of each pixel, and the non-negative difference between the two is the probability difference.If the probability difference of a pixel is less than the given threshold, it is considered as a key pixel; otherwise, it is a general pixel.At the same time, labels of general pixels are reserved.This process is shown in Figure 5.

Reclassifying
The SVM classifier can usually obtain higher classification accuracy (Mountrakis et al. 2011), the decision tree classifier is good at mining potential relationships between data (Hong et al. 2017), and the Wishart  classifier is based on polarization coherence matrix for SAR multi-view cases (Lee et al. 1994).These three shallow classifiers widely used in the field of PolSAR classification are taken as weak classifiers to generate a strong classifier through the AdaBoost algorithm to reclassify key pixels.
Firstly, the SVM classifier, Wishart classifier and decision tree classifier were trained respectively to obtain three classification models.
SVM is a binary model (Khosravi et al. 2021).The basic idea of SVM learning is to solve the separation hyperplane that can correctly partition the training data set and has the largest geometric spacing.The separation hyperplane is w T x þ b ¼ 0, where x is a value in the hyperspace, w is the hyperplane normal vector, and b is the distance from the hyperplane to the origin.LIBSVM open source library (Chang and Lin 2011) were used in the proposed method to realize SVM multi-class classification.
Lee et al. (Lee et al. 1994) extended the maximum likelihood (Entezari et al. 2012) rule to SAR multiview cases and developed a supervised classification algorithm of polarimetric coherence matrix based on complex Wishart distribution.The decision rule is , where x i is a class; Argmaxf ðxÞ returns the x value corresponding to the maximum value of f ðÞ; Ri � � is the maximum likelihood estimation of the coherence matrix.A new be obtained by taking the minus sign of the above formula and removing the terms unrelated to the research clustering, where , the Tr is the matrix trace.Pixels are assigned to the class x i with the smallest distance.
The decision tree classifier (Yin et al. 2020) selects the best feature recursively, and divides the training samples according to the feature.The decision tree classifier measures the selection of features by information gain and selects the features with the maximum information gain after splitting.Assumes that the sample set is X and the proportion of the samples with label k is the is the conditional entropy and represents the information entropy.P y k ð Þ represents the proportion of a feature.The informa- refers to the degree to which the uncertainty of the whole sample features is reduced after a certain feature is known.Finally, the decision tree classification model is obtained by pruning the generated tree.
AdaBoost (Dou et al. 2018) is an iterative algorithm, whose core idea is to use the same training set to train different weak classifiers, and then combine these weak classifiers to form a strong classifier.In this process, the weight of the samples misclassified by the previous weak classifier will increase, while the weight of the samples correctly classified will decrease.
x 1 , y 1 ð Þ , :::, x N , y N ð Þ is the training sample set, where  distribution D 1 of the training sample set.Each training sample is assigned the same weight w i ¼ 1=N, i ¼ 1, :::, N, then the initial weight distribution of the training set is: where w i is the weight of each training sample, and N is the number of training samples.Then the t-th iteration is implemented, t ¼ 1, 2, 3.
A weak classifier h with the lowest error rate is selected as the H t basic classifier, and the error of the basic classifier on the distribution D t is calculated where D t ¼ w 1 , :::, w N ð Þ is the weight of the training sample set at the t-th iteration; e t is the error rate.
Calculate the weight of the basic classifier in the final strong classifier where, a t is the weight of the weak classifier in the tth iteration.Update the weight distribution D tþ1 of the training samples: where Z t ¼ 2 ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi ffi e t 1 À e t ð Þ p is the normalization constant; when samples are misclassified, y i H t x i ð Þ ¼ À 1; when samples are correctly classified, y i H t x i ð Þ ¼ À 1: Finally, each weak classifier is combined according to its weight: Through the function sign, a strong classifier is obtained as follows: The AdaBoost algorithm is aimed at the binary classification problem.Since the PolSAR image used in this paper contains C land covers, the proposed method in this paper realizes classification by iterating the AdaBoost algorithm C-1 times.In each iteration, one land cover was taken as 1 and other land covers were taken as À 1, and a strong classifier was constructed to classify key pixels.Key pixels labeled 1 did not participate in the next iteration and their land covers were recorded.After C-1 iterations, new land covers of all key pixels were obtained.
Finally, the land covers of general pixels and key pixels were combined into the classification results of PolSAR.

Results and discussion
The method in this paper was implemented using MATLAB2022b programming, and the computer was configured with 48 GB memory and NVIDIA GeForce RTX 2080TiGPU.To verify the superiority of the proposed method, the following three aspects of comparative experiments were carried out.

Precision comparison before and after reclassification of key pixels
For the first PolSAR image, when the probability difference was 0.7, the overall classification accuracy reached the maximum of 92.10%.For the second PolSAR image, when the probability difference was 0.5, the overall classification accuracy reached the maximum of 91.02%.The key pixels shown in Figure 6 were located.It can be found that most of the key pixels are located at the edge of each land cover, that is, the area with complex backscattering information.
The number of key pixels and pixels on the ground truth reference map is shown in Tables 1 and 2.
As shown in Figures 7 and 8, labels of the key pixels in two PolSAR images before and after reclassification has changed.
Since the ground truth reference map does not cover the entire PolSAR image, the analysis that follows in this section focuses only on the key pixels in the ground truth reference map, rather than the entire image.
According to the ground truth reference map of the first experimental data shown in Figure 1b and the number of key pixels of each land cover in the ground truth reference map shown in Table 1, the classification accuracy of key pixels before and after reclassification is statistically analyzed, and the results are shown in Table 3.Among the 11 land covers, the classification accuracy of 10 land covers is improved.The accuracy of wood increases by 0.3%, which is the smallest gain.The accuracy of soybean increases by 38.29%, which is the biggest gain.That the average classification accuracy of key pixels in the first experimental data is improved by 9.44% indicates the method in this paper is effective in improving the classification accuracy of key pixels.However, the classification accuracy of water decreases by 29.63% from 88.89% to 59.26%, and the reasons for this result will be analyzed.
As can be seen from Table 1, there are only 27 key pixels of water in the ground truth reference map.These pixels are located in the rectangular box of Figure 9a, c, namely 8 pixels in the upper right corner of the rectangular box and 19 pixels in the lower left corner.As is shown in Figure 9b, before reclassification, eight pixels in the upper right corner are correctly labeled as water; 16 pixels in the lower left corner are correctly labeled as water and three pixels in the lower left corner are incorrectly labeled as Bare Land.As is shown in Figure 9d, after reclassification, six pixels in the upper right corner are correctly labeled as water and two pixels in the upper right are incorrectly labeled as potatoes; 10 pixels in the lower left corner are correctly labeled as water and nine pixels in the lower left corner are incorrectly labeled as Bare Land.
As can be seen from Figure 1b, the land cover in the rectangle box shown in Figure 9e is water.However, the upper right corner of the rectangle box is an abnormal light color area and the lower left corner is a slightly larger abnormal dark blue area, which is shown in Figure 9f.Before reclassification, the CNN classification method was used, and the input data were patches with a size of 15 pixels �15 pixels.The feature value of the central pixel in a patch was easily affected by the feature values of the surrounding pixels.Because the upper right corner area was small, all pixels in this area were greatly affected by the feature values of pixels outside the region and easy to be labeled as the land cover of surrounding pixels.While the lower left corner area was large, so some pixels in this area were greatly affected by the feature values of the pixels outside the region.As a result, some pixels were easily labeled as the land cover of the surrounding pixels.When reclassifying, the combination of shallow classifiers was used, and the input data was pixels, which were less affected by the feature values of surrounding pixels.The land cover of pixels depended largely on their features.So, more pixels in the two regions were labeled as water when CNN classification is adopted than when the proposed method was adopted.Although the two areas were obviously different from the surrounding pixels from the visual effect, the ground truth reference map shown in Figure 1b was drawn by hand and the two very small anomalous land covers in the top right       CNN is inferior to the proposed method in describing the details of land covers.
According to the ground reference map of the second experimental data shown in Figure 2b, the classification accuracy of key pixels before and after reclassification is statistically analyzed.The results are shown in Table 4.The classification accuracy of three out of five land covers is improved.The accuracy of Fine Stem Crop increases the least, with 10.96% and that of Forest increases the most, with 30.72%.The average classification accuracy of key pixels in the second experimental data is improved by 9.61%, which also proves the effectiveness of the proposed method in improving the classification accuracy of key pixels.But Broad Leaves Crop's accuracy dropped slightly from 58.16 to 57.52%, and Town's dropped sharply from 60.08 to 42.38%.The accuracy of Broad Leaves Crop changes a little before and after reclassification, so next, the reasons for the decline in the accuracy of Town will be analyzed.
In the ground reference map, the number of key pixels labeled as Town is 6310 shown in Table 2. Before and after reclassification, the labels of these key pixels are recorded, as shown in Table 5.After reclassification, the labels of some key pixels are changed from Bare Land and Town to Fine Stem Crop, Broad Leaves Crop and Forest, which can be found in Table 5 and Figure 10g and h.As shown in Figure 10c-f, Town is fragmented and mixed with Fine Stem Crop, Broad Leaves Crop and Forest.Similar to the case of water in the first PolSAR image, when CNN was used for classification, the fewer land covers were easily ignored, while the proposed method was more likely to find them.However, the hand-painted ground truth reference map shown in Figure 2b could not separate the small land covers from the largest number of land covers and had to ignore these small land covers.This makes the accuracy of Town before reclassification greater than that after classification.The above analysis also shows that the proposed method has more advantages in describing the details of ground objects than CNN (Figure 10).

Comparison of visual effects of various methods
To evaluate the visual effect of the proposed method, five groups of comparison experiments were set up, i.e., the SVM classifier, the Wishart classifier, the decision tree classifier, the CNN and the AdaBoost algorithm (a SVM classifier, a Wishart classifier and a decision tree classifier were taken as weak classifiers).
Figure 11 shows the classification results of the methods described above.With the ground truth map shown in Figure 1b, four elliptical regions were selected for comparison on each classification result.It can be found that, compared with the classification results of the SVM classifier, the Wishart classifier, the decision tree classifier and the AdaBoost algorithm, the classification results of the CNN are purer and less spotted, but the classification results of the CNN are fuzzier and the texture is lost.The texture of the result of the proposed method is clearer than that of the CNN.There are fewer spots in the classification results of the proposed method than in the shallow classifiers and the AdaBoost algorithm.
As shown in Figure 12, these are the classification results of the second experimental data respectively using the CNN, the SVM classifier, the Wishart classifier, the decision tree classifier, the AdaBoost     results of the proposed method than those of the shallow classifiers and the AdaBoost algorithm.

Comparison of classification accuracy of various methods
With the ground reference maps shown in Figures 1b  and 2b as references, the classification results of the two experimental data were analyzed from four aspects of overall accuracy (OA), Kappa coefficient (KC), producer accuracy and user accuracy by using a confusion matrix.The classification results of the proposed method are compared with those of the CNN, the SVM classifier, the Wishart classifier, the decision tree classifier and the AdaBoost algorithm.
The OA and KC of the classification results of the first experimental data are shown in Table 6.In terms of accuracy, compared with the other 5 classification methods, the OA of the proposed method is improved by 2.22-20.38%,and the KC is increased by 0.03-0.24.
The producer accuracy of the classification results of each land cover using different methods for the first experimental data are shown in Table 7.The standardized differences (STD) of the producer accuracy of each land cover using these methods are adopted in the last row of the table.They are used for the local and global evaluation respectively.The producer accuracy of each land cover using the proposed method is larger than the minimum and close to the maximum or even greater than the maximum.The STD is used to judge the equilibrium of the producer accuracy of all land covers using each method.The smaller the value, the more balanced the data.As can be seen from the data in Table 7, the STD of the proposed method is 0.0569, which is 0.0012 larger than the minimum value.It shows that the producer accuracy of the proposed method is relatively balanced.
The user accuracy of each land cover using different methods for the first experimental data is shown in Table 8.The STD of the user accuracy of all land cover using each method is listed in the last row of the table.The user accuracy of the proposed method is larger than the minimum value and close to the maximum value.The STD of the proposed method is 0.0482, which is 0.0047 larger than the minimum value and very close to the minimum value.
Therefore, the proposed method is superior to the other five classification methods in terms of the producer accuracy and the user accuracy, both locally and globally.
The OA and KC of the classification results of the second experimental data are shown in Table 9.Compared with the other five classification methods, the OA of the proposed method increases by 1.06-11.53%,and the KC increases by 0.02-0.15.
In the second experimental data, the producer accuracy of each land cover using different classification methods is shown in Table 10.The last row of the table lists the STD of the user accuracy using each method.It can be found that the producer accuracy of Fine Stem Crop and Forest using the proposed method is the maximum, while that of Bare Land, Broad Leaves Crop and Town using the proposed method is close to the maximum, with the difference values of 0.95, 0.08 and 2.51%, respectively.Moreover, the STD of the proposed method is the smallest, which is 0.0703, indicating that the producer accuracy using the proposed method is the most balanced.
The user accuracy of each land cover using different classification methods is shown in Table 11 for the second experimental data.The last row shows the STD of the user accuracy using each method.It can be found that the user accuracy of Bare land, Fine Stem Crop and Town using the proposed method is the highest, while those of Broad Leaves Crop and Forest are close to the maximum accuracy, with a difference of 0.49 and 0.81%.Moreover, the STD of the proposed method is the smallest, which is 0.0527, indicating that the user accuracy using the proposed method is the most balanced.

Conclusion
The proposed method in this paper combines the advantages of the deep learning model and shallow classification method to reclassify pixels that are easy to be misclassified and obtain better classification results.In terms of local classification accuracy, overall classification accuracy and visual effect, the proposed classification method performs better than the other methods listed above.
took the layer module Constrained Boltzmann machine of deep belief net as the classifier and built the integrated boosting model with adaptive boosting(AdaBoost), which improved the classification performance of PolSAR image and avoided the requirement of large data volume.He et al. (He et al. 2020) proposed a PolSAR image classification method combining nonlinear manifold learning and a full convolutional network.Zhu et al. (Zhu et al. 2021) combined deep learning technology with traditional classifiers based on scattering features to propose a new semi-supervised PolSAR image classification method to solve the deficiency of the labeling training data set.Jiao et al. (Jiao and Liu 2016) realized the rapid realization of Wishart distance through a special linear transformation, which accelerated the classification speed of the POLSAR image and made the polarized information available for subsequent neural networks.Jamali et al. (Jamali et al. 2022) used the Haar wavelet transform to carry out effective feature extraction in deep CNN, to improve the classification accuracy of the PolSAR image.

Figure 3 .
Figure 3.The flowchart of the proposed method.

Figure 4 .
Figure 4.The structure of CNN used in the proposed method.
1, :::, N is used to represent the label of the training sample.Initialize the weight

Figure 5 .
Figure 5.The flowchart of locating the key pixels.C is the number of land cover and P1, … , PC are the C prediction probabilities of each pixel.

Figure 6 .
Figure 6.(a) Key pixels of the first PolSAR image when the probability difference is 0.7; (b) key pixels of the second PolSAR image when the probability difference is 0.5.

Figure 7 .
Figure 7.The labels of key pixels in the first PolSAR image: (a) before reclassification; (b) after reclassification.

Figure 8 .
Figure 8.The labels of key pixels in the second PolSAR image: (a) before reclassification; (b) after reclassification.

Figure 9 .
Figure 9. (a) and (c) the labels of key pixels in the ground truth reference map before and after reclassification, and the label of the pixels in the top right black rectangle box is water; (b) and (d) enlarged area inside the black rectangle box in the upper right corner of (a) and (c); (e) the Pauli RGB image of the first PolSAR image; (f) enlarged area inside the red rectangle box in the upper right corner of (e).

Figure 10 .
Figure 10.(a) Pauli RGB image of the second PolSAR image; (b) the ground truth reference map in the four rectangular boxes; (c) -(f) the four rectangular boxes with 3 times magnification; (g) -(h) the labels of key pixels labeled as Town in the ground truth reference map before and after reclassification.
algorithm and the proposed method.With the ground truth map shown in Figure2b, four elliptical regions were selected for comparison on each classification result.It can also be found that, compared with the SVM classifier, the Wishart classifier, the decision tree classifier and the AdaBoost algorithm, the classification results of the CNN are with fewer spots but fuzzier and the texture is lost.The texture of the results of the proposed method is clearer than that of the CNN, while there are much fewer spots in the classification

Table 1 .
The number of key pixels and pixels in the ground reference map for the first PolSAR image.

Table 2 .
The number of key pixels and pixels in the ground reference map for the second PolSAR image.
rectangular box of Figure9ehad to be labeled water.Besides, there are only 27 key pixels of water, and even a slight change in the number makes the accuracy change before and after the reclassification appears drastic.That is what caused water's accuracy to drop after reclassification.The above analysis also shows that

Table 3 .
Before and after reclassification, the statistics of key pixels in the ground reference map of the first PolSAR image.

Table 4 .
Before and after reclassification, the statistics of key pixels in the ground reference map of the second PolSAR image.

Table 5 .
The labels of key pixels labeled as town in the ground reference map before and after reclassification.

Table 6 .
The overall accuracy and Kappa coefficient of the first PolSAR image.

Table 7 .
The producer accuracy of the first PolSAR image.

Table 8 .
The user accuracy of the first PolSAR image.

Table 9 .
The overall accuracy and Kappa coefficient of the second PolSAR image.

Table 10 .
The producer accuracy of the second PolSAR image.

Table 11 .
The user accuracy for the second PolSAR image.