
Hyperspectral images contain abundant spectral and spatial information of the surface of the earth, but there are more difficulties in processing, analyzing, and sample-labeling these hyperspectral images. In this paper, local binary pattern (LBP), sparse representation and mixed logistic regression model are introduced to propose a sample labeling method based on neighborhood information and priority classifier discrimination. A new hyperspectral remote sensing image classification method based on texture features and semi-supervised learning is implemented. The LBP is employed to extract features of spatial texture information from remote sensing images and enrich the feature information of samples. The multivariate logistic regression model is used to select the unlabeled samples with the largest amount of information, and the unlabeled samples with neighborhood information and priority classifier discrimination are selected to obtain the pseudo-labeled samples after learning. By making full use of the advantages of sparse representation and mixed logistic regression model, a new classification method based on semi-supervised learning is proposed to effectively achieve accurate classification of hyperspectral images. The data of Indian Pines, Salinas scene and Pavia University are selected to verify the validity of the proposed method. The experiment results have demonstrated that the proposed classification method is able to gain a higher classification accuracy, a stronger timeliness, and the generalization ability.
Citation: Ansheng Ye, Xiangbing Zhou, Kai Weng, Yu Gong, Fang Miao, Huimin Zhao. Image classification of hyperspectral remote sensing using semi-supervised learning algorithm[J]. Mathematical Biosciences and Engineering, 2023, 20(6): 11502-11527. doi: 10.3934/mbe.2023510
[1] | Yufeng Li, Chengcheng Liu, Weiping Zhao, Yufeng Huang . Multi-spectral remote sensing images feature coverage classification based on improved convolutional neural network. Mathematical Biosciences and Engineering, 2020, 17(5): 4443-4456. doi: 10.3934/mbe.2020245 |
[2] | Haifeng Song, Weiwei Yang, Songsong Dai, Haiyan Yuan . Multi-source remote sensing image classification based on two-channel densely connected convolutional networks. Mathematical Biosciences and Engineering, 2020, 17(6): 7353-7377. doi: 10.3934/mbe.2020376 |
[3] | Hua Wang, Weiwei Li, Wei Huang, Jiqiang Niu, Ke Nie . Research on land use classification of hyperspectral images based on multiscale superpixels. Mathematical Biosciences and Engineering, 2020, 17(5): 5099-5119. doi: 10.3934/mbe.2020275 |
[4] | Eric Ke Wang, Nie Zhe, Yueping Li, Zuodong Liang, Xun Zhang, Juntao Yu, Yunming Ye . A sparse deep learning model for privacy attack on remote sensing images. Mathematical Biosciences and Engineering, 2019, 16(3): 1300-1312. doi: 10.3934/mbe.2019063 |
[5] | Haifeng Song, Weiwei Yang, Songsong Dai, Lei Du, Yongchen Sun . Using dual-channel CNN to classify hyperspectral image based on spatial-spectral information. Mathematical Biosciences and Engineering, 2020, 17(4): 3450-3477. doi: 10.3934/mbe.2020195 |
[6] | Qian Zhang, Haigang Li, Ming Li, Lei Ding . Feature extraction of face image based on LBP and 2-D Gabor wavelet transform. Mathematical Biosciences and Engineering, 2020, 17(2): 1578-1592. doi: 10.3934/mbe.2020082 |
[7] | Eric Ke Wang, Fan Wang, Ruipei Sun, Xi Liu . A new privacy attack network for remote sensing images classification with small training samples. Mathematical Biosciences and Engineering, 2019, 16(5): 4456-4476. doi: 10.3934/mbe.2019222 |
[8] | Qihan Feng, Xinzheng Xu, Zhixiao Wang . Deep learning-based small object detection: A survey. Mathematical Biosciences and Engineering, 2023, 20(4): 6551-6590. doi: 10.3934/mbe.2023282 |
[9] | Qing Yang, Jun Chen, Najla Al-Nabhan . Data representation using robust nonnegative matrix factorization for edge computing. Mathematical Biosciences and Engineering, 2022, 19(2): 2147-2178. doi: 10.3934/mbe.2022100 |
[10] | Lei Yang, Guowu Yuan, Hao Wu, Wenhua Qian . An ultra-lightweight detector with high accuracy and speed for aerial images. Mathematical Biosciences and Engineering, 2023, 20(8): 13947-13973. doi: 10.3934/mbe.2023621 |
Hyperspectral images contain abundant spectral and spatial information of the surface of the earth, but there are more difficulties in processing, analyzing, and sample-labeling these hyperspectral images. In this paper, local binary pattern (LBP), sparse representation and mixed logistic regression model are introduced to propose a sample labeling method based on neighborhood information and priority classifier discrimination. A new hyperspectral remote sensing image classification method based on texture features and semi-supervised learning is implemented. The LBP is employed to extract features of spatial texture information from remote sensing images and enrich the feature information of samples. The multivariate logistic regression model is used to select the unlabeled samples with the largest amount of information, and the unlabeled samples with neighborhood information and priority classifier discrimination are selected to obtain the pseudo-labeled samples after learning. By making full use of the advantages of sparse representation and mixed logistic regression model, a new classification method based on semi-supervised learning is proposed to effectively achieve accurate classification of hyperspectral images. The data of Indian Pines, Salinas scene and Pavia University are selected to verify the validity of the proposed method. The experiment results have demonstrated that the proposed classification method is able to gain a higher classification accuracy, a stronger timeliness, and the generalization ability.
Hyperspectral Image (HSI) is the simultaneous imaging of target areas in dozens to hundreds of continuous spectral bands [1,2,3,4]. It effectively integrates the spatial and spectral information in the imaging scene, with strong target detection ability and better material identification ability [5,6,7,8]. It is widely used in agriculture and forestry, geological exploration, marine exploration, Environmental monitoring and other fields [9,10,11,12]. However, HSI is characterized by high data dimension, large information redundancy and high correlation between bands, which brings great difficulties to its processing and classification [13,14,15,16,17,18,19,20,21,22,23]. Therefore, how to reduce the redundant information of the data, extract the features of the hyperspectral images effectively, and realize the accurate classification of the hyperspectral images are the hot and difficult issues in the current hyperspectral image processing and classification research.
Sample labeling of hyperspectral images often requires expert knowledge and experience, so the cost of sample labeling is high [24,25,26]. When the labeled samples are limited, semi-supervised learning can explore the useful information of the unlabeled samples to participate in the model training and reduce the labeling cost [27,28]. In the field of machine learning, semi-supervised learning acquires knowledge and experience from a small number of labeled samples. The usable information is mined from a large number of unlabeled samples, which can improve the classification accuracy [29,30,31]. Therefore, a large number of scholars have studied the semi-supervised learning for remote sensing images. Camps-Valls et al. [32] proposed a graph-based hyperspectral image classification method, and constructed the graph structure through the graph method. The data context information based on the composite kernel is integrated, and the Nystrom method is used to speed up classification. Yang et al. [33] proposed a semi-supervised band selection technique for hyperspectral image classification. A metric learning method has been used to measure the features of hyperspectral images, and a semi-supervised learning method has been used to select a subset of valid bands from the original bands. Tan et al. [34] proposed a hyperspectral image classification method based on segmentation integration and semi-supervised support vector machine. The spatial information of the tag samples is extracted by using a segmentation algorithm to filter the samples, and classified based on semi-supervised learning. Samiappan et al. [35] combined active learning and co-training to perform semi-supervised classification of hyperspectral images. The initial classification model is trained according to the labeled samples, and the heuristic active learning is performed on the unlabeled samples. Combined with the original data, the labeled samples are divided into views, and the unlabeled samples with high heuristic values are selected to join the training samples. Zhang et al. [36] used a semi-supervised classification method based on hierarchical segmentation and active learning to extract spatial information from hyperspectral images, the training set is updated iteratively by using the information of a large number of unlabeled samples to complete the hyperspectral image classification. In addition, some other methods are also proposed in recent years [37,38,39,40,41,42,43,44,45].
In hyperspectral images, each pixel corresponds to a spectral curve that reflects its inherent physical, chemical and optical properties. The main basis of hyperspectral image classification is to use the feature information of different pixels to label the pixels belonging to different landmarks and obtain the corresponding classification maps [46,47]. Therefore, a large number of scholars have studied hyperspectral image classification. Melgani et al. [48] proposed a hyperspectral image classification method based on support vector machines (SVM). The kernel function is introduced to solve the nonlinear separable problem and avoid the curse of dimensionality. Ratle et al. [49] introduced neural networks into hyperspectral image classification. In the training phase, the loss function is optimized to avoid local optimization. Chen et al. [50] constructed a hyperspectral image classification model based on sparse representation, and compared the classification results of common machine learning methods. In order to improve the shortcomings of sparse representation in dealing with nonlinear problems, Chen et al. [51] proposed a kernel sparse representation technique. In addition, Cui et al. [52] proposed a multiscale sparse representation algorithm for robust hyperspectral image classification. Automatic and adaptive weight allocation schemes based on spectral angle ratio are incorporated into the multi-classifier framework to fuse sparse representation information at all scales. Tang et al. [53] proposed two sparse representation algorithms based on manifolds to solve the instability problem of sparse algorithms. The regularization and local invariance techniques are used, and two manifold-based regularization items are merged into the l1-based objective function. Wang et al. [54] applied the neighborhood-cutting technique to sparse representation, and combined the joint spatial and spectral sparse representation classification method. Wang and Celik [55] improved the classification accuracy of hyperspectral images by combining context information in the sparse coefficient domain. Hu et al. [56] proposed two weighted kernel joint sparse representation methods, which determine the calculation weight by calculating the kernel similarity between adjacent pixels. Xue et al. [57] presented two novel sparse graph regularization methods, SGR and SGR with total variation. Yang et al. [58] studied the effect of the p-norm distance metric on the minimum distance technique, and proposed a supervised-learning p-norm distance metric to optimize the value of p. Zhang et al. [59] proposed a multi-scale dense network for HSI classification that made full use of different scale information in the network structure and combined scale information throughout the network. Liu et al. [60] proposed a class-wise adversarial adaptation method in conjunction with the class-wise probability MMD as the class-wise distribution adaptation network. Wang et al. [61] proposed the graph-based semi-supervised learning with weighted features for HSI classification. In addition, some new other methods are also proposed by a lot of researchers [62,63,64,65,66,67,68,69,70,71].
To sum up, hyperspectral images contain rich spectral and spatial information of earth surface features, which increases the processing and analysis difficulties. In addition, the training samples of actual hyperspectral images are small and there is a sample labeling problem. The local binary pattern, sparse representation and mixed logistic regression model are used in this paper. A new hyperspectral image feature extraction method based on local binary pattern is proposed to obtain texture features of hyperspectral images and enrich hyperspectral image sample information. A sample selection strategy based on active learning is designed to determine the unlabeled samples. Based on these, a new sample labeling method based on neighborhood information and priority classifier discrimination is deeply studied to expand the training samples. And a novel classification method based on texture features and semi-supervised learning is studied to improve the classification accuracy of remote sensing images.
The main contributions of this paper are described as follows.
(ⅰ) A novel a classification method of hyperspectral remote sensing images based on texture features and semi-supervised learning is proposed by using local binary pattern, sparse representation, hybrid logistic regression.
(ⅱ) The local binary pattern is used to effectively extract the features of spatial texture information of remote sensing images and enrich the feature information of samples.
(ⅲ) A multiple logistic regression model is used to optimally select unlabeled samples, which are labeled by using neighbourhood information and priority classifier discrimination to achieve pseudo-labeling of unlabeled samples.
(ⅳ) A novel classification method of hyperspectral remote sensing images based on semi-supervised learning is proposed to effectively achieve accurate classification of hyperspectral images.
In here, the local binary pattern (LBP) and sparse expression are introduced in order to clearly describe the basic theories of these methods.
LBP is a feature extraction method, which can extract spatial texture information of images. Texture, which is widely used in image processing and image analysis, represents the slow change or periodic change of the surface structure of the object [72]. LBP is also widely used in feature extraction of hyperspectral images due to the simple structure and easy calculation. Give the center pixel gc(xc,yc) and the neighborhood pixel gp, which are described as follow.
gp=(xc+Rcos(2ΠpP),yc−Rsin(2ΠpP)) | (1) |
where, gp(p=0,1,...,P−1) represents the coordinate values of P pixels uniformly distributed on the circular domain with gc as the centre and R as the radius. The quantized texture feature form of one region is shown in Figure 1.
LBPgc=2p×∑P−1p=0s(gp−gc) | (2) |
s(x)={1,0,x>0x≤0 | (3) |
Sparse representation means that the signal can be approximately represented by using linear combination of the atoms in the dictionary. Now, X=[X1,X2,...,Xc]∈RD is given as the HSI pixel and D is the number of image bands. In here, Xi=[xi1,xi2,...,xiNc]∈RD, Nc represents the number of samples in the class i.
For samples in the class i, it can be approximated as follow.
y≈xi1α1+xi2α2+...+xiNcαNc = [xi1,xi2,...,xiNc][α1,α2,...,αNc]T=Xiαi | (4) |
where, Xi represents a sparse sub-dictionary of the samples in the class i. αi represents the sparse vector of test samples y, which contains only a few non-zero values.
In order to obtain the sparsest vector αi, the following formula is solved.
˜α=argmin‖αi‖0,s.t.y=Aαi | (5) |
where, ‖ . ‖0 is a l0 norm, which represents the number of non-zero atoms in the vector, also known as sparsity. A is a sparse dictionary. It is a NP-hard problem to solve the Eq (5) directly. Under some conditions, the minimization solving problem (l0) is approximated by the minimization solving problem (l1), which can be relaxed.
˜α=argmin‖αi‖1,s.t.y=Aαi | (6) |
Furthermore, the solution can be converted to the following expression.
˜α=argmin‖αi‖0,s.t.‖Aαi−y‖22<ε | (7) |
where, ε represents the refactoring error. Orthogonal matching pursuit (OMP) algorithm is used to solve the expression (7). After the sparsity coefficient is calculated, the reconstruction residual for each class of the test samples y can be calculated.
ri(y)=‖y−A˜αi‖2 | (8) |
where, there is i∈{1,2,...,C}.
Finally, the reconstruction residuals of all dictionaries are compared. The minimum residual is the classification y.
class(y)=argmin(ri(y)),i=1,2,...C | (9) |
In hyperspectral remote sensing image classification, the attribute characteristics of each ground feature are subordinated to a specific distribution. If the number of training samples does not reach the amount that depicts the distribution of the ground feature, it will affect the subsequent classification results. In practical applications, there are difficulties, such as fewer labeled samples, difficulty in manual labeling, and time-consuming and labor-intensive problems and so on. Therefore, a new sample labeling method based on neighborhood information and priority classifier discrimination is proposed to obtain the learned pseudo labeled samples in order to expand the labeled samples and improve the classification accuracy of model.
Before the samples are labeled, sample selection is required. This is because if all unlabeled samples are labeled directly, all unlabeled samples need to be labeled, which will cost a lot of computation time. Moreover, due to the small number of initial labeled samples and the limited available information, it is difficult to label some samples with a certain accuracy. The mislabeled samples obviously affect the classification accuracy of model. The main objective of the sample selection strategy is to select the unlabeled samples with the largest amount of information. These unlabeled samples can construct a valuable training set after labeling, which can effectively promote the improvement of classification results. Therefore, a sample selection method based on multiple logistic regression is proposed to realize the selection of samples.
The classified probability matrix p(yki|xi) of each sample by using multiple logistic regression model has a large amount of information that can be mined. The multiple logistic regression classifier is modeled by discriminant Bayesian decision model. According to the generalized linear model, it can be obtained as follow.
P(y;δ)=b(y)exp(δTT(y)−a(δ)) | (10) |
The specific expression of multiple logistic regression is described as follow.
p(yi=k∣xi,η)=exp(ηkg(xi))/∑Nk=1exp(ηkg(xi)) | (11) |
where, g(x)=[g1(x),g2(x),…,gf(x)]T is the feature vectors of the input, and η=[ηT1,ηT2,…ηTk] represents the regression parameter vector of the classifier. It is worth noting that the feature vector is often represented by introducing the kernel, which is not only used to improve the indivisibility, but also obtain the better classifier by training samples. Generally, the kernel function is radial basis function (RBF), which is described as follow.
K(xm,xn)=e−||xm−xn||22p2 | (12) |
After the feature vector is determined, the regression parameter η of model is only determined, and the probability matrix p(yki|xi) of each unlabeled sample belonging to each class is determined. The sample amount of information is determined by the Breaking Ties (BT) and the Least Confidence (LC). In this paper, the BT is selected to determine the amount of information.
The BT shows the similarity between the two classifications by comparing the difference between the maximum category probability and the sub-maximum category probability. The difference is smaller, the similarity between the two types of samples is greater. The uncertainty is greater, the amount of information is greater. Si is the similarity between classifications, which is described as follow.
Si=maxp(yki|xi)−secondmax(p(yki|xi)) | (13) |
The Si is finally sorted in ascending order. p(yki|xi) is the obtained probability matrix of each sample by multiple logistic regression classifier. maxp(yki|xi) represents the largest value in the obtained probability matrix, secondmax(p(yki|xi)) represents the second largest value in the obtained probability matrix. k is the number of regression parameters.
The features of hyperspectral remote sensing images have some correlation. The ground objects are closer, the correlation is stronger. In the research of sample labeling, spatial neighborhood information based on training samples is widely used. However, due to the unknown central pixel and the lack of sufficient information, the neighborhood information of unlabeled samples is relatively less in the research of sample labeling. Generally, each pixel label on hyperspectral image must be consistent with one pixel label in its neighborhood. This property can be used to label the unlabeled samples. The label information of training samples around the unlabeled samples can be used to discriminate the unlabeled samples. The labeling discrimination method based on neighborhood information centers on the sample to be labeled. The unlabeled samples are labeled with a block diagram. All the occurrences of sample labels are recorded and denoted as the neighborhood information. The labeled samples are used to train the classifier and classify the unlabeled samples. Determine whether the predicted sample label by the classifier appears in the neighborhood information of the unlabeled samples. If it appears, the predicted label by the classifier is the sample label. Otherwise, the samples are put into the unlabeled samples. One of the most important problems is whether the unlabeled samples which satisfy the neighborhood information can be reliably labeled by the classifier. At present, some studies use multiple classifiers to discriminate together and achieve good classification effect. However, how to determine the determination of labels, when the predicted labels by multiple classifiers are inconsistent, but all appear in the neighborhood information of unlabeled samples.
Therefore, a sample labeling method based on neighborhood information and priority classifier discrimination is proposed in this paper. For unlabeled samples with the neighborhood information, the classifier with the highest priority is used for prediction. If the obtained prediction marker appears in the neighborhood information, its marker is determined. Otherwise, the classifier with the lowest priority is used for prediction. Determine whether the label can be determined until the sample labeling is ended. The sample labeling process based on neighborhood information and priority classifier discrimination is shown in Figure 2.
The sample labeling method is an iterative process. Although it cannot ensure enough training samples around all unlabeled samples at the initial stage of sample labeling, it can ensure that some unlabeled samples are sufficient. The unlabeled samples are labeled and extended to the training set. With each iteration, the training set grows. Those unlabeled samples whose neighborhood training samples are not sufficient may reach the label condition at a certain labeling. The sample labeling method with replacement ensures the sample labeling accuracy to a certain extent, and improves the performance of classifier.
Hyperspectral remote sensing images consist of pairs of continuous spectral bands, which contain rich spectral and spatial information of earth surface features. Some objects that cannot be identified by conventional remote sensing means can be identified in hyperspectral images. However, the abundant data information increases the processing and analysis difficulties, and there are some problems of fewer labeled samples, difficulty in manual labeling and time-consuming and so on. In order to improve the classification accuracy of hyperspectral remote sensing images, a new classification method of hyperspectral remote sensing images based on texture features and semi-supervised learning is proposed in this paper. Firstly, aiming at the problems of high correlation between bands, information redundancy, high dimension and complex processing, LBP is employed to deal with the hyperspectral images. The texture features of hyperspectral images are effectively extracted to enrich the feature information. To solve the problem of limited label samples, a new sample labeling method based on neighborhood information and priority classifier discrimination is proposed in here. Secondly, a sample selection strategy is designed to find some samples from a large number of unlabeled samples. Finally, the sparse representation and mixed logistic regression are applied to achieve a new classification method to effectively achieve accurate classification of hyperspectral images.
The model of hyperspectral image classification method based on texture features and semi-supervised learning is shown in Figure 3.
The detailed implementation steps of hyperspectral image classification method based on texture features and semi-supervised learning are described as follows.
Step 1. Normalize the hyperspectral images. Principal component analysis is used to perform dimensionality reduction.
Step 2. Calculate the texture for each principal component by using LBP. The histogram statistics are performed according to the symmetric rotation invariant equivalence mode. D-dimensional feature data for each layer is used to obtain feature information for each pixel and retain pixel coordinates.
Step 3. The original hyperspectral images are performed by linear discriminant analysis to obtain a projection matrix.
Step 4. The hyperspectral images are mapped into low dimensional images for suitable Euclidean distance analysis by using a projection matrix.
Step 5. Training sample sets and test sample sets are randomly selected from each type of sample according to a certain proportion of low-dimensional images.
Step 6. Calculate the Euclidean distance from all training samples by using Lk=‖yi−Ak‖2 to maintain K training coordinates that are close to the Euclidean distance of each test sample as the sparse dictionary for the test samples.
Step 7. The extracted feature information is loaded into the test samples and corresponding sparse dictionary according to the corresponding coordinates to obtain the test sample set and sparse dictionary.
Step 8. The sparse representation is solved. Orthogonal matching pursuit (OMP) has been used to calculate the sparse coefficient, and the classifications of the test samples is obtained.
1) Indian Pines data
The images of Indian pines in northwest Indiana were collected by AVIRIS sensor. The images consist of 145 × 145 pixels and 224 spectral reflection bands with a wavelength range of 0.4~2.5 nm, including 16 types of feature elements. The false color map and real ground object distribution are shown in Figure 4.
2) Salinas Scene data
The AVIRIS spectrometer collects images of the Salinas Valley in California, USA, with a size of 512 × 217 pixels and a total of 224 bands. After the bands covering the water absorption area are removed, 204 bands are used, including 16 types of ground feature elements. The false color map and the real ground object distribution are shown in Figure 5.
3) Pavia University data
Images of the Italian University of Pavia campus taken by the Rosis Spectrometer. It is 610 × 340 pixels in size and has a total of 115 wavebands. The 103 wavebands after the wavebands covering the water-absorbing region contain a total of 9 types of features are removed. The false color map and the real ground object distribution are shown in Figure 6.
In the experiment, 10% of each type of ground object of the three kinds of data is randomly selected as the training samples, and the rest is the test samples.
Confusion Matrix (CM) is usually used in the classification and evaluation of hyperspectral images. A confusion matrix is generally defined as follow.
P=[p11p21p12…p22…p1np2n…………pn1pn2…pnn] | (14) |
where, n denotes the number of objects in the category, pij represents the number of samples belonging to the class i that are assigned to the class j. The total amount of data in each row denotes the true number of objects in the category. The total amount of data in each column represents the total number of samples.
Based on the confusion matrix, three classification indexes can be obtained, which are Overall Accuracy (OA), Average Accuracy (AA) and Kappa coefficient.
OA=∑ni=1piiN | (15) |
where, N represents the total number of samples in the classification. pii represents the number of correctly classified samples of the class i. OA represents the probability that the classified result corresponds to its true label for each random sample.
CAi=piiNi | (16) |
where, Ni represents the total number of samples for the first category in the class i. CAi represents the probability that category i is correctly classified.
Kappa=(n(∑Ni=1pii)−∑Ni=1(∑Nj=1pij∑Nj=1pji))n2−∑Ni=1(∑Nj=1pij∑Nj=1pji) | (17) |
The Kappa coefficient comprehensively considers the number of correctly classified objects and the error of misclassified objects on the confusion matrix.
In the experiment, a large number of alternative values are tested, and some classical values are selected from literatures, these parameter values are experimentally modified until the most reasonable parameter values are determined. These selected parameter values have obtained the optimal solution, so that they can accurately and efficiently verify the effectiveness.
The quality of sample selection directly affects the efficiency of the experiment, and also affects the performance of classifier. In order to select the best samples, the experimental results of Information Entropy (IE), Min Error (ME), Breaking Ties (BT) and Least Confidence (LC) on three kinds of hyperspectral images are compared. The experiment is 10 initial samples for each class, and all remaining samples are test samples. Two hundred unlabeled samples are selected by four sample selection methods in each iteration. The label samples are used to expand the training samples, train the classifier and classify the test samples. The quality of the selection samples is determined according to the classification results after each iteration is executed. The classification accuracy of different sample selection methods on different data sets is shown in Table 1.
Data | Selection method | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | IE | 78.39 | 79.10 | 79.99 | 80.71 | 82.11 | 83.47 | 84.10 | 85.51 | 86.48 | 87.23 |
ME | 80.33 | 86.83 | 91.08 | 92.92 | 95.28 | 97.02 | 98.00 | 98.45 | 98.90 | 99.13 | |
BT | 91.47 | 95.47 | 98.22 | 98.36 | 98.64 | 98.59 | 98.66 | 98.71 | 99.34 | 99.29 | |
LC | 84.74 | 88.95 | 91.63 | 94.33 | 95.27 | 96.46 | 98.15 | 98.55 | 98.62 | 98.66 | |
Pavia University | IE | 69.87 | 71.31 | 71.75 | 71.74 | 72.04 | 72.40 | 73.25 | 73.36 | 73.76 | 74.45 |
ME | 73.01 | 75.23 | 78.74 | 82.78 | 89.14 | 92.82 | 95.14 | 96.10 | 97.13 | 97.82 | |
BT | 87.54 | 92.63 | 94.51 | 95.24 | 95.84 | 96.10 | 96.39 | 96.50 | 96.69 | 96.71 | |
LC | 74.91 | 76.33 | 80.76 | 84.45 | 87.23 | 89.89 | 90.27 | 90.67 | 90.56 | 91.47 | |
Salinas Scene | IE | 84.16 | 84.60 | 84.65 | 85.02 | 85.09 | 85.25 | 85.50 | 85.65 | 85.91 | 85.98 |
ME | 85.14 | 89.29 | 92.81 | 94.31 | 96.48 | 97.41 | 98.04 | 98.20 | 98.66 | 98.88 | |
BT | 95.30 | 96.98 | 98.26 | 98.71 | 98.95 | 98.90 | 99.03 | 99.21 | 99.25 | 99.24 | |
LC | 88.60 | 91.19 | 92.84 | 93.58 | 93.80 | 95.74 | 97.56 | 97.72 | 98.56 | 98.86 |
From Table 1, it can be seen that the classification accuracies of ME, BT and LC are greatly improved on the three data sets with the increase of the number of iterations. The results of the BT method are significantly better than those of the other methods. The accuracy can be improved in the first few iterations, which indicate that the BT method can select samples with greater classification improvement. Therefore, the BT method is chosen as the sample selection method in this paper.
In the labeling process, the samples are not completely labeled correctly. The more samples are screened, the more samples may be misclassified. This will make the training set noisier and affects the generalization ability of the classifier. If the number of samples is too small, the number of labeled samples will not improve the classification accuracy of the classifier or will reduce the classification efficiency. The results of classification accuracy under different sample sizes are shown in Table 2.
Data | Quantity | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | 200 | 77.30 | 77.57 | 78.32 | 77.97 | 78.59 | 78.97 | 78.96 | 79.28 | 79.51 | 79.23 |
400 | 77.54 | 78.60 | 79.80 | 80.80 | 82.13 | 82.20 | 83.03 | 83.77 | 83.83 | 83.85 | |
600 | 77.52 | 79.54 | 79.37 | 79.27 | 80.08 | 81.99 | 83.89 | 83.60 | 84.30 | 84.48 | |
800 | 77.75 | 79.84 | 80.87 | 80.22 | 82.11 | 83.91 | 84.41 | 84.57 | 85.12 | 85.94 | |
1000 | 77.85 | 80.49 | 80.28 | 82.74 | 81.35 | 82.60 | 84.18 | 85.88 | 86.77 | 87.84 | |
1200 | 77.85 | 79.95 | 79.79 | 80.38 | 81.59 | 84.41 | 84.72 | 85.74 | 87.48 | 88.85 | |
1400 | 78.18 | 80.20 | 80.09 | 83.96 | 85.00 | 85.78 | 87.93 | 89.90 | 91.07 | 91.20 | |
1600 | 78.55 | 80.56 | 80.59 | 84.12 | 86.90 | 87.82 | 89.02 | 90.87 | 91.49 | 91.83 | |
1800 | 78.34 | 79.76 | 79.23 | 82.64 | 85.39 | 87.32 | 88.81 | 89.97 | 90.69 | 91.46 | |
2000 | 78.01 | 79.16 | 80.24 | 82.55 | 86.02 | 87.48 | 88.66 | 89.63 | 90.02 | 90.46 | |
Pavia University | 200 | 68.75 | 73.93 | 76.73 | 78.18 | 79.23 | 80.92 | 81.85 | 82.40 | 82.74 | 83.60 |
400 | 66.41 | 73.11 | 75.45 | 78.20 | 81.18 | 82.13 | 82.57 | 83.39 | 84.07 | 83.98 | |
600 | 68.88 | 76.35 | 78.22 | 80.73 | 82.59 | 83.29 | 83.68 | 84.57 | 84.95 | 84.98 | |
800 | 69.89 | 77.50 | 80.31 | 81.91 | 83.22 | 84.85 | 84.99 | 84.79 | 85.16 | 85.31 | |
1000 | 70.28 | 76.35 | 79.92 | 82.68 | 83.83 | 84.48 | 84.84 | 85.21 | 85.37 | 85.04 | |
1200 | 70.24 | 75.18 | 80.32 | 83.13 | 84.14 | 85.04 | 85.30 | 85.55 | 85.04 | 84.90 | |
1400 | 70.34 | 76.23 | 80.57 | 82.46 | 83.92 | 84.71 | 85.59 | 85.77 | 85.87 | 86.47 | |
1600 | 70.40 | 75.87 | 80.64 | 83.06 | 83.93 | 84.69 | 85.28 | 86.02 | 86.19 | 85.83 | |
1800 | 69.77 | 76.12 | 80.18 | 82.99 | 85.19 | 85.04 | 84.89 | 85.26 | 85.68 | 85.68 | |
2000 | 69.71 | 75.90 | 82.29 | 83.40 | 84.41 | 85.23 | 85.59 | 85.77 | 85.86 | 85.87 | |
Salinas Scene | 200 | 85.09 | 87.26 | 89.04 | 89.35 | 89.24 | 89.22 | 88.85 | 88.47 | 88.14 | 88.03 |
400 | 84.94 | 88.34 | 89.88 | 89.26 | 89.12 | 88.88 | 88.26 | 87.68 | 87.40 | 87.25 | |
600 | 85.69 | 90.85 | 90.80 | 90.42 | 89.71 | 89.04 | 88.63 | 87.84 | 87.12 | 86.60 | |
800 | 85.36 | 89.17 | 88.87 | 87.96 | 87.46 | 86.85 | 86.42 | 85.69 | 85.13 | 84.58 | |
1000 | 85.35 | 88.93 | 89.88 | 89.04 | 88.34 | 87.58 | 87.03 | 85.49 | 85.08 | 85.08 | |
1200 | 85.04 | 89.66 | 90.08 | 88.98 | 87.67 | 87.14 | 86.13 | 85.30 | 85.09 | 84.85 | |
1400 | 85.07 | 88.46 | 89.39 | 88.63 | 88.10 | 88.06 | 87.34 | 86.81 | 86.69 | 86.58 | |
1600 | 85.47 | 89.54 | 89.88 | 88.87 | 87.68 | 86.21 | 85.00 | 84.34 | 83.89 | 83.70 | |
1800 | 85.50 | 90.16 | 90.15 | 89.45 | 88.58 | 87.71 | 86.79 | 86.52 | 86.31 | 85.93 | |
2000 | 85.48 | 89.40 | 89.38 | 88.60 | 87.73 | 85.98 | 85.67 | 85.12 | 85.39 | 85.34 |
It can be seen from Table 2 that the sample selection screening quantity in different data sets presents different rules. Indian Pines datasets have the highest classification accuracy after 10 iterations are finished. Pavia University datasets have the highest classification accuracy after 8–10 iterations are finished. Salinas Scene datasets have the highest classification accuracy after 2–4 iterations are finished. The sample screening quantity with the highest accuracy is regarded as the experiment parameter, which are 1600 for Indian Pines, 1400 for Pavia University and 600 for Salinas Scene.
The size of the block window determines the neighborhood information of the samples, which directly affect the accuracy of the pseudo-tagging method. Due to the different size of data, the optimal block window size is also determined by a large number of experiments. The classification accuracy under different block window sizes is shown in Table 3.
Data | Block window size | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | 3 | 77.62 | 77.86 | 78.35 | 78.65 | 78.91 | 79.14 | 79.09 | 79.12 | 79.56 | 79.56 |
4 | 77.73 | 78.24 | 78.65 | 78.75 | 78.70 | 78.66 | 79.14 | 79.18 | 79.70 | 79.74 | |
5 | 78.41 | 79.61 | 79.41 | 81.28 | 81.11 | 81.73 | 82.67 | 82.84 | 84.18 | 84.68 | |
6 | 77.85 | 80.14 | 80.41 | 80.91 | 80.90 | 83.54 | 84.91 | 85.09 | 86.42 | 88.30 | |
7 | 78.86 | 80.58 | 82.58 | 84.90 | 86.55 | 86.99 | 88.47 | 88.87 | 89.58 | 90.98 | |
8 | 78.33 | 78.95 | 81.00 | 84.24 | 85.44 | 86.37 | 87.60 | 88.13 | 88.57 | 89.19 | |
9 | 79.61 | 81.00 | 83.39 | 85.51 | 85.71 | 86.34 | 86.78 | 86.86 | 87.07 | 87.86 | |
10 | 78.61 | 79.32 | 83.45 | 85.45 | 85.59 | 85.69 | 86.12 | 87.00 | 87.17 | 87.32 | |
Pavia University | 5 | 68.39 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 |
10 | 68.20 | 67.67 | 69.05 | 69.06 | 68.30 | 68.13 | 68.06 | 67.98 | 67.89 | 67.62 | |
15 | 70.91 | 70.59 | 70.96 | 70.50 | 71.89 | 73.54 | 73.80 | 74.58 | 75.25 | 75.32 | |
20 | 71.47 | 71.63 | 73.60 | 76.12 | 75.99 | 75.61 | 77.18 | 77.51 | 77.81 | 78.14 | |
25 | 71.25 | 74.57 | 78.52 | 79.15 | 81.95 | 82.32 | 83.66 | 84.88 | 85.45 | 85.51 | |
30 | 70.28 | 76.35 | 79.92 | 82.68 | 83.83 | 84.48 | 84.84 | 85.21 | 85.37 | 85.04 | |
35 | 71.48 | 77.43 | 80.15 | 81.75 | 83.18 | 83.62 | 84.11 | 83.65 | 83.23 | 83.17 | |
40 | 73.40 | 77.50 | 80.36 | 81.92 | 83.15 | 82.68 | 82.23 | 82.38 | 82.18 | 82.04 | |
Salinas Scene | 5 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 |
10 | 83.61 | 84.16 | 84.95 | 85.03 | 84.77 | 84.68 | 84.59 | 84.82 | 84.61 | 85.48 | |
15 | 83.86 | 83.17 | 85.51 | 87.04 | 87.70 | 88.44 | 89.77 | 89.60 | 91.01 | 91.09 | |
20 | 84.16 | 86.84 | 89.41 | 90.56 | 90.22 | 90.69 | 91.02 | 91.13 | 90.87 | 90.68 | |
25 | 84.11 | 88.22 | 89.46 | 89.68 | 89.35 | 88.77 | 87.86 | 87.49 | 86.71 | 86.78 | |
30 | 85.35 | 88.93 | 89.88 | 89.04 | 88.34 | 87.58 | 87.03 | 85.49 | 85.08 | 85.08 | |
35 | 86.17 | 88.80 | 88.67 | 87.67 | 86.63 | 85.84 | 85.26 | 84.87 | 83.95 | 83.06 | |
40 | 86.86 | 89.25 | 89.02 | 87.78 | 86.37 | 84.74 | 83.60 | 82.83 | 81.69 | 80.92 |
As can be seen from Table 3, there have different changes in classification accuracy for different datasets. Compared with the other two data sets, the size of Indian Pines is the smallest, so its experimental block window side length values are from 3 to 10. With the increase of the number of iterations, the classification accuracy showed a trend of gradual increase, and the optimal accuracy is obtained when the side length is 7. When the side length of the block window for Pavia and Salinas datasets is too small, the classification accuracy will not improve with the increase of iterations. This indicates that the neighborhood information cannot distinguish the category at this time.
With the increase of the side length of the block window, the number of iterations to achieve the optimal classification accuracy is advanced, but the optimal accuracy decreases. The block window size is larger, the more noise information will be introduced, and it will affect the sample labeling accuracy. Therefore, the block window size of the Indian Pines dataset is 7 × 7, and the block window sizes of the Pavia and Salinas datasets are 25 × 25 and 20 × 20, respectively.
In fact, the determination of pseudo-tags of samples mainly depends on the determination of classifiers. K-Nearest Neighbor (KNN), Sparse Representation-Based Classifier (SRC), Neighbor Rough Set (NRS) and mixed logistic regression (MLR) are employed to determine the pseudo-tags. The experimental results of single classifier and different combination classifiers on different datasets are shown in Tables 4–6.
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 314 | 673 | 1139 | 1630 | 2178 | 2680 | 3324 | 4063 | 4820 | 5474 |
OA (%) | 78.69 | 79.93 | 81.20 | 82.05 | 82.49 | 84.38 | 85.70 | 86.52 | 87.33 | 87.77 | |
SRC | NUM | 311 | 648 | 1068 | 1525 | 2032 | 2606 | 3248 | 3899 | 4668 | 5522 |
OA (%) | 78.64 | 80.42 | 80.10 | 81.56 | 82.87 | 84.39 | 85.83 | 87.04 | 87.92 | 88.19 | |
NRS | NUM | 315 | 673 | 1116 | 1597 | 2136 | 2697 | 3265 | 3815 | 4437 | 5016 |
OA (%) | 78.82 | 81.02 | 80.96 | 83.71 | 84.94 | 85.57 | 86.87 | 88.01 | 89.25 | 89.42 | |
MLR | NUM | 133 | 295 | 580 | 936 | 1296 | 1790 | 2396 | 3077 | 3858 | 4648 |
OA (%) | 77.55 | 79.31 | 82.21 | 83.91 | 83.99 | 85.75 | 87.36 | 87.68 | 88.27 | 88.41 | |
KNN + SRC | NUM | 317 | 706 | 1120 | 1702 | 2294 | 2967 | 3728 | 4494 | 5233 | 5950 |
OA (%) | 78.71 | 80.96 | 81.99 | 83.97 | 84.96 | 85.82 | 86.99 | 87.75 | 87.93 | 88.10 | |
KNN + NRS | NUM | 317 | 707 | 1198 | 1691 | 2308 | 2954 | 3625 | 4450 | 5374 | 6305 |
OA (%) | 78.78 | 80.27 | 80.73 | 82.51 | 83.82 | 85.45 | 87.56 | 89.06 | 89.95 | 90.19 | |
KNN + MLR | NUM | 318 | 712 | 1206 | 1794 | 2555 | 3398 | 4292 | 5072 | 5783 | 6678 |
OA (%) | 78.79 | 80.56 | 82.57 | 86.08 | 87.10 | 87.87 | 88.50 | 88.88 | 89.18 | 89.31 | |
SRC + KNN | NUM | 317 | 706 | 1134 | 1673 | 2223 | 2875 | 3658 | 4317 | 5011 | 5748 |
OA (%) | 78.71 | 81.09 | 81.51 | 83.17 | 84.27 | 85.84 | 86.94 | 87.54 | 88.26 | 88.64 | |
SRC + NRS | NUM | 318 | 730 | 1205 | 1778 | 2372 | 3091 | 3969 | 4813 | 5787 | 6641 |
OA (%) | 78.81 | 80.79 | 82.37 | 85.03 | 86.81 | 88.45 | 88.93 | 89.86 | 90.49 | 90.78 | |
SRC + MLR | NUM | 315 | 708 | 1153 | 1744 | 2457 | 3322 | 4231 | 5042 | 5800 | 6735 |
OA (%) | 78.80 | 80.92 | 82.16 | 85.16 | 86.24 | 87.61 | 88.39 | 88.71 | 89.07 | 89.19 | |
NRS + KNN | NUM | 317 | 707 | 1202 | 1712 | 2385 | 3021 | 3700 | 4538 | 5399 | 6333 |
OA (%) | 78.74 | 80.67 | 81.33 | 83.36 | 84.46 | 86.35 | 87.93 | 89.23 | 90.03 | 90.43 | |
NRS + SRC | NUM | 318 | 734 | 1207 | 1728 | 2378 | 3061 | 3959 | 4799 | 5691 | 6739 |
OA (%) | 78.77 | 81.01 | 82.56 | 84.52 | 86.37 | 87.43 | 88.95 | 90.11 | 90.61 | 90.90 | |
NRS + MLR | NUM | 318 | 694 | 1148 | 1690 | 2446 | 3194 | 4102 | 4950 | 5768 | 6792 |
OA (%) | 78.91 | 81.39 | 81.62 | 85.51 | 87.21 | 89.63 | 90.28 | 90.92 | 91.28 | 91.88 | |
MLR + KNN | NUM | 318 | 677 | 1166 | 1689 | 2429 | 3246 | 4018 | 4737 | 5677 | 6672 |
OA (%) | 78.30 | 81.13 | 81.85 | 85.94 | 87.22 | 88.28 | 88.77 | 88.98 | 89.20 | 89.29 | |
MLR + SRC | NUM | 315 | 704 | 1258 | 1741 | 2439 | 3286 | 4097 | 4867 | 5775 | 6760 |
OA (%) | 78.31 | 81.81 | 82.55 | 85.09 | 86.86 | 88.30 | 89.01 | 89.44 | 90.21 | 90.71 | |
MLR + NRS | NUM | 318 | 683 | 1154 | 1701 | 2458 | 3301 | 4219 | 5057 | 5997 | 6889 |
OA (%) | 78.46 | 81.49 | 82.11 | 86.14 | 87.86 | 89.70 | 90.68 | 91.58 | 92.15 | 92.42 |
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 236 | 444 | 661 | 965 | 1210 | 1466 | 1812 | 2254 | 2762 | 3251 |
OA (%) | 71.19 | 73.36 | 74.70 | 75.69 | 76.10 | 76.16 | 77.28 | 78.48 | 77.60 | 77.69 | |
SRC | NUM | 225 | 477 | 733 | 1080 | 1492 | 2047 | 2793 | 3654 | 4605 | 5631 |
OA (%) | 72.21 | 72.19 | 76.97 | 79.12 | 80.08 | 81.77 | 83.09 | 84.07 | 84.95 | 85.27 | |
NRS | NUM | 225 | 438 | 763 | 1007 | 1256 | 1487 | 1726 | 1893 | 1991 | 2179 |
OA (%) | 71.22 | 73.74 | 75.14 | 76.59 | 76.15 | 76.47 | 76.41 | 76.20 | 75.68 | 75.40 | |
MLR | NUM | 100 | 244 | 396 | 605 | 829 | 1035 | 1304 | 1587 | 1872 | 2183 |
OA (%) | 72.36 | 76.37 | 77.18 | 79.04 | 79.40 | 81.16 | 82.21 | 82.18 | 83.36 | 85.85 | |
KNN + SRC | NUM | 248 | 473 | 787 | 1118 | 1509 | 1996 | 2552 | 3111 | 3975 | 4837 |
OA (%) | 71.48 | 72.82 | 74.61 | 76.59 | 78.38 | 80.08 | 80.82 | 81.62 | 82.40 | 83.99 | |
KNN + NRS | NUM | 240 | 491 | 775 | 1067 | 1382 | 1850 | 2403 | 3040 | 3794 | 4415 |
OA (%) | 71.11 | 73.83 | 77.44 | 78.83 | 79.50 | 79.58 | 79.39 | 80.11 | 80.74 | 81.48 | |
KNN + MLR | NUM | 244 | 515 | 822 | 1176 | 1616 | 2162 | 2905 | 3698 | 4745 | 5682 |
OA (%) | 71.37 | 75.03 | 77.84 | 78.08 | 79.73 | 79.59 | 81.30 | 83.88 | 84.90 | 85.08 | |
SRC + KNN | NUM | 248 | 476 | 795 | 1215 | 1725 | 2298 | 3055 | 3967 | 4977 | 5913 |
OA (%) | 71.38 | 72.54 | 75.26 | 78.09 | 79.19 | 80.21 | 81.88 | 83.31 | 83.35 | 83.51 | |
SRC + NRS | NUM | 242 | 511 | 867 | 1385 | 1889 | 2513 | 3201 | 3988 | 4910 | 5794 |
OA (%) | 71.40 | 74.12 | 77.56 | 80.80 | 82.19 | 82.55 | 83.02 | 83.80 | 84.34 | 85.09 | |
SRC + MLR | NUM | 236 | 507 | 841 | 1289 | 1731 | 2261 | 3016 | 3928 | 4939 | 6015 |
OA (%) | 71.52 | 73.62 | 76.47 | 78.45 | 79.51 | 81.54 | 83.10 | 84.37 | 84.68 | 84.66 | |
NRS + KNN | NUM | 240 | 486 | 803 | 1119 | 1541 | 1992 | 2557 | 3190 | 3939 | 4611 |
OA (%) | 71.05 | 74.37 | 76.57 | 77.77 | 78.06 | 77.21 | 76.75 | 77.35 | 78.88 | 79.28 | |
NRS + SRC | NUM | 242 | 501 | 803 | 1296 | 1792 | 2433 | 3089 | 3839 | 4776 | 5798 |
OA (%) | 71.44 | 74.50 | 76.72 | 79.93 | 81.44 | 81.90 | 82.85 | 83.81 | 85.00 | 85.48 | |
NRS + MLR | NUM | 234 | 517 | 828 | 1237 | 1796 | 2446 | 3354 | 4220 | 5061 | 5857 |
OA (%) | 71.49 | 75.47 | 77.89 | 80.86 | 83.68 | 84.43 | 84.93 | 85.12 | 85.57 | 85.46 | |
MLR + KNN | NUM | 244 | 486 | 746 | 1170 | 1658 | 2300 | 3205 | 4208 | 5336 | 6514 |
OA (%) | 71.47 | 75.30 | 76.67 | 78.76 | 80.35 | 82.97 | 85.05 | 86.05 | 86.88 | 87.02 | |
MLR + SRC | NUM | 236 | 504 | 903 | 1310 | 1708 | 2207 | 3009 | 4024 | 5098 | 6234 |
OA (%) | 71.71 | 74.01 | 79.30 | 79.81 | 80.40 | 83.03 | 86.27 | 86.93 | 87.97 | 88.53 | |
MLR + NRS | NUM | 234 | 524 | 787 | 1205 | 1738 | 2374 | 3116 | 3953 | 4754 | 5586 |
OA (%) | 71.63 | 75.94 | 79.83 | 80.66 | 82.75 | 84.64 | 85.98 | 86.19 | 86.37 | 86.87 |
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 133 | 251 | 443 | 684 | 963 | 1304 | 1672 | 2047 | 2425 | 2857 |
OA (%) | 83.38 | 86.46 | 86.46 | 86.65 | 87.26 | 87.31 | 87.79 | 87.92 | 87.66 | 87.15 | |
SRC | NUM | 144 | 275 | 441 | 666 | 937 | 1255 | 1606 | 1968 | 2391 | 2811 |
OA (%) | 83.10 | 83.56 | 85.11 | 85.83 | 85.81 | 86.63 | 86.50 | 86.64 | 86.95 | 86.96 | |
NRS | NUM | 148 | 271 | 423 | 668 | 952 | 1263 | 1569 | 1940 | 2351 | 2799 |
OA (%) | 84.04 | 86.06 | 87.62 | 87.63 | 87.10 | 87.48 | 88.44 | 88.43 | 88.32 | 87.92 | |
MLR | NUM | 102 | 177 | 302 | 451 | 621 | 867 | 1176 | 1518 | 1848 | 2217 |
OA (%) | 82.88 | 85.41 | 87.11 | 88.36 | 88.73 | 90.68 | 91.52 | 92.20 | 91.70 | 92.02 | |
KNN + SRC | NUM | 146 | 294 | 479 | 694 | 976 | 1330 | 1691 | 2106 | 2520 | 2985 |
OA (%) | 83.60 | 84.65 | 85.38 | 86.76 | 87.39 | 87.39 | 88.00 | 88.11 | 87.53 | 87.23 | |
KNN + NRS | NUM | 150 | 297 | 500 | 761 | 1108 | 1466 | 1891 | 2354 | 2834 | 3316 |
OA (%) | 83.53 | 86.43 | 87.14 | 87.49 | 87.64 | 87.37 | 88.06 | 88.03 | 87.95 | 88.03 | |
KNN + MLR | NUM | 143 | 285 | 508 | 768 | 1132 | 1546 | 2026 | 2526 | 3049 | 3590 |
OA (%) | 83.38 | 85.50 | 86.34 | 88.35 | 88.80 | 90.07 | 90.24 | 90.08 | 89.86 | 89.79 | |
SRC + KNN | NUM | 146 | 287 | 472 | 686 | 957 | 1281 | 1673 | 2061 | 2485 | 2892 |
OA (%) | 83.25 | 84.70 | 85.79 | 85.88 | 86.70 | 86.92 | 86.89 | 86.93 | 86.85 | 86.90 | |
SRC + NRS | NUM | 150 | 280 | 493 | 755 | 1017 | 1328 | 1708 | 2114 | 2539 | 2982 |
OA (%) | 83.07 | 85.55 | 86.28 | 85.47 | 86.66 | 86.97 | 87.52 | 87.05 | 87.19 | 87.16 | |
SRC + MLR | NUM | 150 | 271 | 483 | 720 | 1108 | 1499 | 1958 | 2451 | 2969 | 3487 |
OA (%) | 83.15 | 84.27 | 87.25 | 87.88 | 88.88 | 89.29 | 89.36 | 89.85 | 89.70 | 89.85 | |
NRS + KNN | NUM | 150 | 298 | 519 | 814 | 1148 | 1556 | 2037 | 2538 | 3027 | 3522 |
OA (%) | 84.01 | 87.12 | 87.79 | 87.30 | 87.54 | 88.38 | 88.09 | 88.25 | 87.95 | 87.88 | |
NRS + SRC | NUM | 150 | 284 | 488 | 803 | 1158 | 1539 | 1988 | 2482 | 2955 | 3423 |
OA (%) | 83.91 | 87.23 | 86.98 | 87.73 | 88.30 | 88.97 | 89.57 | 89.57 | 89.60 | 89.43 | |
NRS + MLR | NUM | 153 | 293 | 509 | 762 | 1104 | 1509 | 1940 | 2441 | 2934 | 3452 |
OA (%) | 83.87 | 85.82 | 88.25 | 89.93 | 90.40 | 90.51 | 90.75 | 90.48 | 90.12 | 89.61 | |
MLR + KNN | NUM | 143 | 299 | 521 | 825 | 1187 | 1602 | 2046 | 2514 | 2993 | 3407 |
OA (%) | 82.80 | 85.42 | 87.67 | 88.79 | 89.68 | 90.06 | 90.54 | 90.92 | 91.27 | 91.32 | |
MLR + SRC | NUM | 150 | 292 | 521 | 799 | 1123 | 1537 | 2007 | 2448 | 2929 | 3367 |
OA (%) | 82.91 | 85.45 | 88.89 | 90.12 | 90.21 | 90.93 | 90.99 | 91.83 | 92.08 | 92.64 | |
MLR + NRS | NUM | 153 | 315 | 564 | 841 | 1197 | 1605 | 2064 | 2561 | 3060 | 3500 |
OA (%) | 82.81 | 84.87 | 88.16 | 89.34 | 89.69 | 90.03 | 90.42 | 90.52 | 91.19 | 91.43 |
As can be seen in Table 4, with the increase of iterations, the classification accuracy of each category samples increased gradually. Compared with the results of the single classifier, the SRC has the largest number of samples after 10 iterations are finished, but the classification effect is not the best value. The classifier with the best classification effect is NRS. The number of samples and classification accuracy of the combination classifiers are mostly better than that of single classifier. The experiment results are different for combination classifiers with different priority. The classifier with NRS can achieve more than 90% classification effect after 10 iterations are finished. The number of combinations with MLR is more than 6600 after 10 iterations are finished, and the best combination is MLR + NRS after 10 iterations are finished.
As can be seen in Table 5, compared with the experimental results by single classifier, the number of labeled samples with SRC is the largest value after 10 iterations are finished, which is higher than the other three methods. However, the MLR obtained best classification results. After 10 iterations are finished, the number of labeled samples of two classifiers is more than that of the single classifier. KNN, SRC and NRS are considered as the first priority classifiers, the classification results of the sample set after 10 iterations are not as good as those of MLR. The combination of MLR as the first priority classifier has better classification effect than single MLR after 10 iterations are finished.
For Salinas Scene data, the number of labeled samples by MLR after 10 iterations is the smallest result, but the classification accuracy of the labeled samples is the highest result. After 10 iterations are finished, the number of iterations of the two classifiers is also higher than that of the single classifier. However, from the perspective of the performance of labeled samples in classification, MLR+SRC has higher classification accuracy than MLR, which indicates that the addition of classifiers improves the classification accuracy.
From three experiment results, it can be seen that the method with the largest number of labeled samples does not necessarily achieve the best classification results. The labeled samples are needed to improve the classification accuracy of the classifier, so the obtained labeled samples after 10 iterations are taken as the evaluation criteria. The Indian Pines dataset uses a combination of classifiers MLR + NRS. The Pavia University and Salinas Scene data sets use MLR + SRC.
Based on the analysis, the settings of the related parameters are shown in Table 7.
Data set | Indian Pines | Pavia University | Salinas Scene |
Selection policy | BT | BT | BT |
Number of selections | 1600 | 1400 | 600 |
Window size | 7 × 7 | 25 × 25 | 20 × 20 |
Combination of classifiers | MLR + NRS | MLR + SRC | MLR + SRC |
Number of labeled samples | 6889 | 6234 | 3367 |
Firstly, the LBP is used to extract the features of spatial texture information of hyperspectral remote sensing images. Secondly, the sample labeling method based on neighborhood information and priority classifier discrimination is used to obtain the learned pseudo-labeled samples. The SRC classifier is trained with the labeled samples, and the test samples are predicted. The obtained classification results are compared with those of the SRC classifier on the training data, and the classification results of training models with different training dataset are shown in Table 8.
Training samples | Index | Initial samples | Labeling samples |
Indian Pines | AA | 67.93% | 84.70% |
OA | 77.38% | 92.42% | |
KAPPA | 0.746 | 0.914 | |
Pavia University | AA | 60.53% | 81.87% |
OA | 69.00% | 88.53% | |
KAPPA | 0.609 | 0.848 | |
Salinas Scene | AA | 82.59% | 87.76% |
OA | 84.00% | 92.64% | |
KAPPA | 0.823 | 0.918 |
It can be seen from Table 8, for Indian Pines dataset, the classification results of AA, OA and KAPPA are 84.7%, 94.42% and 0.914, respectively. For Pavia University dataset, the classification results of AA, OA and KAPPA are 81.87%, 88.53% and 0.848, respectively. For Salinas Scene dataset, the classification results of AA, OA and KAPPA are 87.76%, 92.64% and 0.918, respectively. Therefore, the proposed classification method obtains higher classification accuracy.
The classification visualizations of the proposed classification method for the initial samples and labeled samples are shown in Figures 7 and 8.
By comparing with the results of the experiments, it can find that the classification results of the trained classifier with expanded samples on the three datasets are better than those of the trained classifier with initial samples. Moreover, from the classification visualization, it can also see that the obtained classification results by the classifier and the labeled samples is smoother and has fewer discrete points, which indicates that the generalization ability of the classifier is improved by labeling the samples.
For the processing and analysis difficulties of hyperspectral images, a new sample labeling method based on neighborhood information and priority classifier discrimination is developed to implement a new classification method of hyperspectral remote sensing images based on texture features and semi-supervised learning by introducing LBP, sparse representation and mixed logistic regression. The LBP is employed to extract the texture features of the hyperspectral remote sensing images. The multivariate logistic regression model is used to select the unlabeled samples with the largest amount of information, and the unlabeled samples with neighborhood information and priority classifier tags are selected to obtain the pseudo-labeled samples after learning. The problem of limited labeled samples of hyperspectral images is solved. The Indian Pines dataset, Salinas scene dataset and Pavia University dataset are selected in here. The experiment results show that the block window of Indian Pines dataset is 7 × 7, and the block windows of Pavia University and Salinas scene are 25 × 25 and 20 × 20, respectively. The combination of MLR and SRC can obtain better classification results. The obtained classification results are smoother and have fewer discrete points, which indicate that the generalization ability of the classifier is improved by labeling the samples. For Indian Pines dataset, the classification results of AA, OA and KAPPA are 84.7%, 94.42% and 0.914, respectively. For Pavia University dataset, the classification results of AA, OA and KAPPA are 81.87%, 88.53% and 0.848, respectively. For Salinas Scene dataset, the classification results of AA, OA and KAPPA are 87.76%, 92.64% and 0.918, respectively. Therefore, the proposed classification method obtains higher classification accuracy by comparing with the other methods.
However, the proposed classification method has the more computing time, so the next step should be more in-depth research to reduce the time complexity.
This research was funded by the Sichuan Science and Technology Program, grant number 2021YFS0407, 2022YFS0593, 2023YFG0028; the Sichuan Provincial Transfer Payment Program, Chian under Grant R21ZYZF0006; the A Ba Achievements Transformation Program under Grant R21CGZH0001, R22CGZH0006, R22CGZH0007; the Chengdu Science and technology planning project, grant number 2021-YF05-00933-SN, the Research Foundation for Civil Aviation University of China grant number 2020KYQD123.
The authors declare no conflict of interest.
[1] |
H. Chen, F. Miao, Y. Chen, A hyperspectral image classification method using multifeature vectors and optimized KELM, IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 14 (2021), 2781–2795. https://doi.org/10.1109/JSTARS.2021.3059451 doi: 10.1109/JSTARS.2021.3059451
![]() |
[2] |
G. Y. Chen, Multiscale filter-based hyperspectral image classification with PCA and SVM, J. Electr. Eng., 72 (2021), 40–45. https://doi.org/10.2478/jee-2021-0006 doi: 10.2478/jee-2021-0006
![]() |
[3] |
Z. Dou, K. Gao, X. Zhang, H. Wang, L. Han, Band selection of hyperspectral images using attention-based autoencoders, IEEE Geosci. Remote Sens. Lett., 18 (2020), 147–151. https://doi.org/10.1109/LGRS.2020.2967815 doi: 10.1109/LGRS.2020.2967815
![]() |
[4] |
M. Z. Chen, H. D. Shao, H. X. Dou, W. Li, B. Liu, Data augmentation and intelligent fault diagnosis of planetary gearbox using ILoFGAN under extremely limited samples, IEEE Trans. Reliab., 2022. https://doi.org/10.1109/TR.2022.3215243 doi: 10.1109/TR.2022.3215243
![]() |
[5] |
C. Chen, Y. Ma, G. Ren, Hyperspectral classification using deep belief networks based on conjugate gradient update and pixel-centric spectral block features, IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 13 (2021), 4060–4069. https://doi.org/10.1109/JSTARS.2020.3008825 doi: 10.1109/JSTARS.2020.3008825
![]() |
[6] |
X. Zhang, H. Wang, C. Du, X. Y. Fan, L. Cui, H. M. Chen, et al., Custom-molded offloading footwear effectively prevents recurrence and amputation, and lowers mortality rates in high-risk diabetic foot patients, a multicenter, prospective observational study, Diabetes, Metab. Syndr. Obesity, Targets Ther., 15 (2022), 103–109. https://doi.org/10.2147/DMSO.S341364 doi: 10.2147/DMSO.S341364
![]() |
[7] |
T. Jin, Y. Zhu, Y. Shu, J. Cao, H. Y. Yan, D. P. Jiang, Uncertain optimal control problem with the first hitting time objective and application to a portfolio selection model, J. Intell. Fuzzy Syst., 44 (2022), 1585–1599. https://doi.org/10.3233/JIFS-222041. doi: 10.3233/JIFS-222041
![]() |
[8] |
X. B. Zhou, H. J. Ma, J. G. Gu, H. L. Chen, W. Deng, Parameter adaptation-based ant colony optimization with dynamic hybrid mechanism, Eng. Appl. Artif. Intell., 114 (2022), 105139. https://doi.org/10.1016/j.engappai.2022.105139 doi: 10.1016/j.engappai.2022.105139
![]() |
[9] |
C. Yu, B. Gong, M. Song, E. Y. Zhao, C. I. Chang, Multiview calibrated prototype learning for few-shot hyperspectral image classification, IEEE Trans. Geosci. Remote Sens., 60 (2022), 5544713. https://doi.org/10.1109/TGRS.2022.3225947. doi: 10.1109/TGRS.2022.3225947
![]() |
[10] |
H. D. Shao, W. Li, B. P. Cai, J. F. Wan, Y. M. Xiao, S. Yan, Dual-threshold attention-guided GAN and limited infrared thermal images for rotating machinery fault diagnosis under speed fluctuation, IEEE Trans. Ind. Inf., 2022. https://doi.org/10.1109/TⅡ.2022.3232766. doi: 10.1109/TⅡ.2022.3232766
![]() |
[11] |
H. Y. Chen, M. Fang, S. Xu, Hyperspectral remote sensing image classification with CNN based on quantum genetic-optimized sparse representation, IEEE Access, 8 (2020), 99900–99909. https://doi.org/10.1109/ACCESS.2020.2997912 doi: 10.1109/ACCESS.2020.2997912
![]() |
[12] |
Y. M. Xiao, H. D. Shao, S. Y. Han, Z. Q. Huo, J. F. Wan, Novel joint transfer network for unsupervised bearing fault diagnosis from simulation domain to experimental domain, IEEE/ASME Trans. Mechatron., 27 (2022), 5254–5263. https://doi.org/10.1109/TMECH.2022.3177174 doi: 10.1109/TMECH.2022.3177174
![]() |
[13] |
I. Dumke, M. Ludvigsen, S. L. Ellefmo, F. Soreide, G. Johnsen, B. Murton, Underwater hyperspectral imaging using a stationary platform in the transatlantic geotraverse hydrothermal field, IEEE Trans. Geosci. Remote Sens., 57 (2019), 2947–2962. https://doi.org/10.1109/TGRS.2018.2878923 doi: 10.1109/TGRS.2018.2878923
![]() |
[14] |
W. Huang, Y. Huang, H. Wang, Y. Liu, H. J. Shim, Local binary patterns and superpixel-based multiple kernels for hyperspectral image classification, IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 13 (2020), 4550–4563. https://doi.org/10.1109/JSTARS.2020.3014492 doi: 10.1109/JSTARS.2020.3014492
![]() |
[15] |
T. Jin, S. Gao, H. Xia, Reliability analysis for the fractional-order circuit system subject to the uncertain random fractional-order model with Caputo type, J. Adv. Res., 32 (2021), 15–26. https://doi.org/10.1016/j.jare.2021.04.008 doi: 10.1016/j.jare.2021.04.008
![]() |
[16] |
X. Jiang, W. Liu, Y. Zhang, Spectral-spatial hyperspectral image classification using dual-channel capsule networks, IEEE Geosci. Remote Sens. Lett., 18 (2020), 1094–1098. https://doi.org/10.1109/LGRS.2020.2991405 doi: 10.1109/LGRS.2020.2991405
![]() |
[17] |
M. Seifi, H. Ghassemian, A probabilistic SVM approach for hyperspectral image classification using spectral and texture features, Int. J. Remote Sens., 38 (2017), 4265–4284. https://doi.org/10.1080/01431161.2017.1317941 doi: 10.1080/01431161.2017.1317941
![]() |
[18] |
Y. Song, X. Cai, X. Zhou, Dynamic hybrid mechanism-based differential evolution algorithm and its application, Expert Syst. Appl., 213 (2022), 118834. https://doi.org/10.1016/j.eswa.2022.118834 doi: 10.1016/j.eswa.2022.118834
![]() |
[19] |
X. Yang, W. Cao, Y. Lu, Hyperspectral image transformer classification networks, IEEE Trans. Geosci. Remote Sens., 60 (2022). https://doi.org/10.1109/TGRS.2022.3171551 doi: 10.1109/TGRS.2022.3171551
![]() |
[20] |
C. Yu, C. Liu, H. Yu, Unsupervised domain adaptation with dense-based compaction for hyperspectral imagery, IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 14 (2021), 12287–12299. https://doi.org/10.1109/JSTARS.2021.3128932 doi: 10.1109/JSTARS.2021.3128932
![]() |
[21] |
H. M. Zhao, X. X. Yang, B. J. Chen, H. Y. Chen, W. Deng, Bearing fault diagnosis using transfer learning and optimized deep belief network, Meas. Sci. Technol., 33 (2022), 065009. https://doi.org/10.1088/1361-6501/ac543a doi: 10.1088/1361-6501/ac543a
![]() |
[22] |
X. B. Zhou, X. Cai, H. Zhang, Multi-strategy competitive-cooperative co-evolutionary algorithm and its application, Inf. Sci., 2023. https://doi.org/10.1016/j.ins.2023.03.142. doi: 10.1016/j.ins.2023.03.142
![]() |
[23] |
C. Xie, L. Zhou, S. Ding, Experimental and numerical investigation on self-propulsion performance of polar merchant ship in brash ice channel, Ocean Eng., 269 (2023), 113424. https://doi.org/10.1016/j.oceaneng.2022.113424 doi: 10.1016/j.oceaneng.2022.113424
![]() |
[24] |
X. Shang, M. Song, C. I. Chang, An iterative random training sample selection approach to constrained energy minimization for hyperspectral image classification, IEEE Geosci. Remote Sens. Lett., 18 (2020), 1625–1629. https://doi.org/10.1109/LGRS.2020.3005078 doi: 10.1109/LGRS.2020.3005078
![]() |
[25] |
M. Li, J. Zhang, J. Song, A clinical-oriented non severe depression diagnosis method based on cognitive behavior of emotional conflict, IEEE Trans. Comput. Social Syst., 2022. http://dx.doi.org/10.1109/TCSS.2022.3152091 doi: 10.1109/TCSS.2022.3152091
![]() |
[26] |
M. Li, W. Zhang, B. Hu, Automatic assessment of depression and anxiety through encoding pupil-wave from HCI in VR scenes, ACM Trans. Multimedia Comput. Commun. Appl., 2022. http://dx.doi.org/10.1145/3513263 doi: 10.1145/3513263
![]() |
[27] |
C. Shi, C. M. Pun, Multiscale superpixel-based hyperspectral image classification using recurrent neural networks with stacked autoencoders, IEEE Trans. Multimedia, 22 (2019), 487–501. https://doi.org/10.1109/TMM.2019.2928491 doi: 10.1109/TMM.2019.2928491
![]() |
[28] |
X. Ye, J. Ma, H. Xiong, Local affine preservation with motion consistency for feature matching of remote sensing images, IEEE Trans. Geosci. Remote Sens., 60 (2021), 1–12. https://doi.org/10.1109/TGRS.2021.3128292 doi: 10.1109/TGRS.2021.3128292
![]() |
[29] |
J. Yin, C. Qi, Q. Chen, Spatial-spectral network for hyperspectral image classification, A 3-D CNN and Bi-LSTM framework, Remote Sens., 13 (2021), 353. https://doi.org/10.3390/rs13122353 doi: 10.3390/rs13122353
![]() |
[30] |
C. Yu, S. Zhou, M. Song, Semisupervised hyperspectral band selection based on dual-constrained low-rank representation, IEEE Geosci. Remote Sens. Lett., 19 (2021), 1–5. https://doi.org/10.1109/lgrs.2021.3049267 doi: 10.1109/lgrs.2021.3049267
![]() |
[31] |
H. L. Chen, C. Y. Li, M. Mafarja, Slime mould algorithm, A comprehensive review of recent variants and applications, Int. J. Syst. Sci., 54 (2023), 204–235. https://doi.org/10.1080/00207721.2022.2153635 doi: 10.1080/00207721.2022.2153635
![]() |
[32] |
G. Camps-Valls, T. Bandos, D. Zhou, Semi-supervised graph-based hyperspectral image classification, IEEE Trans. Geosci. Remote Sens., 45 (2007), 3044–3054. https://doi.org/10.1109/TGRS.2007.895416 doi: 10.1109/TGRS.2007.895416
![]() |
[33] | C. Yang, S. C. Liu, L. Bruzzone, A semisupervised feature metric-based band selection method for hyperspectral image classification, in 2012 4th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), IEEE, 2012. https://doi.org/10.1109/WHISPERS.2012.6874326 |
[34] |
K. Tan, E. Li, D. Qian, An efficient semi-supervised classification approach for hyperspectral imagery, ISPRS J. Photogramm. Remote Sens., 97 (2014), 36–45. https://doi.org/10.1016/j.isprsjprs.2014.08.003 doi: 10.1016/j.isprsjprs.2014.08.003
![]() |
[35] | S. Samiappan, R. J. Moorhead, Semi-supervised co-training and active learning framework for hyperspectral image classification, in 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), IEEE, (2015), 401–404. https://doi.org/10.1109/IGARSS.2015.7325785 |
[36] |
J. Zhang, Z. Meng, F. Zhao, Convolution transformer mixer for hyperspectral image classification, IEEE Geosci. Remote Sens. Lett., 2022. https://doi.org/10.1109/LGRS.2022.3208935. doi: 10.1109/LGRS.2022.3208935
![]() |
[37] |
H. M. Zhao, P. P. Zhang, R. C. Zhang, A novel performance trend prediction approach using ENBLS with GWO, Meas. Sci. Technol., 34 (2023), 025018. https://doi.org/10.1088/1361-6501/ac9a61 doi: 10.1088/1361-6501/ac9a61
![]() |
[38] |
S. Zhou, Z. Xue, P. Du, Semisupervised stacked autoencoder with cotraining for hyperspectral image classification, IEEE Trans. Geosci. Remote Sensi., 57 (2019), 3813–3826. https://doi.org/10.1109/TGRS.2018.2888485 doi: 10.1109/TGRS.2018.2888485
![]() |
[39] |
K. Zhong, G. Zhou, W. Deng, MOMPA, Multi-objective marine predator algorithm, Comput. Methods Appl. Mech. Eng., 385 (2021), 114029. https://doi.org/10.1016/j.cma.2021.114029 doi: 10.1016/j.cma.2021.114029
![]() |
[40] |
C. Huang, X. Zhou, X. Ran, Co-evolutionary competitive swarm optimizer with three-phase for large-scale complex optimization problem, Inf. Sci., 619 (2023), 2–18. https://doi.org/10.1016/j.ins.2022.11.019 doi: 10.1016/j.ins.2022.11.019
![]() |
[41] |
Z. Duan, P. Song, C. Yang, The impact of hyperglycaemic crisis episodes on long-term outcomes for inpatients presenting with acute organ injury: A prospective, multicentre follow-up study, Front. Endocrinol., 13 (2022). https://doi.org/10.3389/fendo.2022.1057089 doi: 10.3389/fendo.2022.1057089
![]() |
[42] |
J. Xu, Y. Zhao, H. Chen, ABC-GSPBFT, PBFT with grouping score mechanism and optimized consensus process for flight operation data-sharing, Inf. Sci., 624 (2023), 110–127. https://doi.org/10.1016/j.ins.2022.12.068 doi: 10.1016/j.ins.2022.12.068
![]() |
[43] |
W. Deng, J. Xu, X. Gao, An enhanced MSIQDE algorithm with novel multiple strategies for global optimization problems, IEEE Trans. Syst. Man Cybern. Syst., 52 (2022), 1578–1587. https://doi.org/10.1109/TSMC.2020.3030792 doi: 10.1109/TSMC.2020.3030792
![]() |
[44] |
J. Cai, S. Ding, Q. Zhang, Broken ice circumferential crack estimation via image techniques, Ocean Eng., 259 (2022), 111735. https://doi.org/10.1016/j.oceaneng.2022.111735 doi: 10.1016/j.oceaneng.2022.111735
![]() |
[45] |
X. Zhang, H. Wang, C. Du, Custom-molded offloading footwear effectively prevents recurrence and amputation, and lowers mortality rates in high-risk diabetic foot patients: A multicenter, prospective observational study, Diabetes Metab. Syndr. Obes. Targets Ther., 15 (2022), 103–109. https://doi.org/10.2147/DMSO.S341364 doi: 10.2147/DMSO.S341364
![]() |
[46] |
X. D. Zhao, M. M. Zhang, R. Tao, Fractional Fourier image transformer for multimodal remote sensing data classification, IEEE Trans. Neural Networks Learn. Syst., 2022. https://doi.org/10.1109/TNNLS.2022.3189994 doi: 10.1109/TNNLS.2022.3189994
![]() |
[47] | Z. Zhang, M. Crawford, Semi-supervised multi-metric active learning for classification of hyperspectral images, in 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), IEEE, (2016), 1843–1847. https://doi.org/10.1109/IGARSS.2016.7729473 |
[48] |
F. Melgani, L. Bruzzone, Classification of hyperspectral remote sensing images with support vector machines, IEEE Trans. Geosci. Remote Sens., 42 (2004), 1778–1790. https://doi.org/10.1109/TGRS.2004.831865 doi: 10.1109/TGRS.2004.831865
![]() |
[49] | T. Z. Ratle, R. Kanevski, Learning manifolds in forensic data, in Artificial Neural Networks–ICANN 2006: 16th International Conference, Springer, Berlin, Heidelberg, (2006), 894–903. |
[50] |
Y. Chen, M. N. Nasser, T. D. Tran, Hyperspectral image classification using dictionary-based sparse representation, IEEE Trans. Geosci. Remote Sens., 49 (2011), 3973–3985. https://doi.org/10.1109/TGRS.2011.2129595 doi: 10.1109/TGRS.2011.2129595
![]() |
[51] |
Y. Chen, M. N. Nasser, T. D. Tran, Hyperspectral image classification via kernel sparse representation, IEEE Trans. Geosci. Remote Sens., 51 (2013), 217–231. https://doi.org/10.1109/TGRS.2012.2201730 doi: 10.1109/TGRS.2012.2201730
![]() |
[52] | M. Cui, S. Prasad, Multiscale sparse representation classification for robust hyperspectral image analysis, in 2013 IEEE Global Conference on Signal and Information Processing, (2013), 969–972. https://doi.org/10.1109/GlobalSIP.2013.6737054 |
[53] |
Y. Y. Tang, H. Yuan, L. Li, Manifold-based sparse representation for hyperspectral image classification, IEEE Trans. Geosci. Remote Sens., 52 (2014), 7606–7618. https://doi.org/10.1109/TGRS.2014.2315209 doi: 10.1109/TGRS.2014.2315209
![]() |
[54] | C. Wang, H. Wang, B. Hu, A novel spatial-spectral sparse representation for hyperspectral image classification based on neighborhood segmentation, Spectrosc. Spectral Anal., 36 (2016), 2919–2924. |
[55] |
H. R. Wang, T. Celik, Sparse representation-based hyperspectral image classification, Signal Image Video Process., 12 (2018), 1009–1017. https://doi.org/10.1007/s11760-018-1249-1 doi: 10.1007/s11760-018-1249-1
![]() |
[56] |
S. Hu, C. Xu, J. Peng, Weighted Kernel joint sparse representation for hyperspectral image classification, IET Image Process., 13 (2019), 254–260. https://doi.org/10.1049/iet-ipr.2018.0124 doi: 10.1049/iet-ipr.2018.0124
![]() |
[57] |
Z. H. Xue, P. J. Du, J. Li, Sparse graph regularization for hyperspectral remote sensing image classification, IEEE Trans. Geosci. Remote Sens., 55 (2017), 2351–2366. https://doi.org/10.1109/TGRS.2016.2641985 doi: 10.1109/TGRS.2016.2641985
![]() |
[58] |
M. Yang, C. H. Li, J. Guan, A supervised-learning p-norm distance metric for hyperspectral remote sensing image classification, IEEE Geosci. Remote Sens. Lett., 15 (2018), 1432–1436. https://doi.org/10.1109/LGRS.2018.2841023 doi: 10.1109/LGRS.2018.2841023
![]() |
[59] |
C. J. Zhang, G. D. Li, S. H. Du, Multi-scale dense networks for hyperspectral remote sensing image classification, IEEE Trans. Geosci. Remote Sens., 57 (2019), 9201–9222. https://doi.org/10.1109/TGRS.2019.2925615 doi: 10.1109/TGRS.2019.2925615
![]() |
[60] |
Z. X. Liu, L. Ma, Q. Du, Class-wise distribution adaptation for unsupervised classification of hyperspectral remote sensing images, IEEE Trans. Geosci. Remote Sens., 59 (2021), 508–521. https://doi.org/10.1109/TGRS.2020.2997863 doi: 10.1109/TGRS.2020.2997863
![]() |
[61] |
Q. Y. Wang, Q. Zhang, J. P. Zhang, Graph-based semisupervised learning with weighted features for hyperspectral remote sensing image classification, IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 15 (2022), 6356–6370. https://doi.org/10.1109/JSTARS.2022.3195639 doi: 10.1109/JSTARS.2022.3195639
![]() |
[62] |
J. Bi, G. Zhou, Y. Zhou, Artificial electric field algorithm with greedy state transition strategy for spherical multiple traveling salesmen problem, Int. J. Comput. Intell. Syst., 15 (2022), 5. https://doi.org/10.1007/s44196-021-00059-0 doi: 10.1007/s44196-021-00059-0
![]() |
[63] |
C. Chang, Y. Kuo, S. Chen, Self-Mutual information-based band selection for hyperspectral image classification, IEEE Trans. Geosci. Remote Sens., 59 (2021), 5979–5997. https://doi.org/10.1109/TGRS.2020.3024602 doi: 10.1109/TGRS.2020.3024602
![]() |
[64] |
W. Deng, L. Zhang, X. Zhou, Multi-strategy particle swarm and ant colony hybrid optimization for airport taxiway planning problem, Inf. Sci., 612 (2022), 576–593. https://doi.org/10.1016/j.ins.2022.08.115 doi: 10.1016/j.ins.2022.08.115
![]() |
[65] |
Y. Gu, L. Zhou, S. Ding, Numerical simulation of ship maneuverability in level ice considering ice crushing failure, Ocean Eng., 251 (2022), 111110. https://doi.org/10.1016/j.oceaneng.2022.111110 doi: 10.1016/j.oceaneng.2022.111110
![]() |
[66] |
C. Huang, X. Zhou, X. Ran, Adaptive cylinder vector particle swarm optimization with differential evolution for UAV path planning, Eng. Appl. Artif. Intell., 121 (2023), 105942. https://doi.org/10.1016/j.engappai.2023.105942 doi: 10.1016/j.engappai.2023.105942
![]() |
[67] |
T. Jin, H. Xia, Lookback option pricing models based on the uncertain fractional-order differential equation with Caputo type, J. Ambient Intell. Hum. Comput., 2021 (2021), 1–14. https://doi.org/10.1007/s12652-021-03516-y3 doi: 10.1007/s12652-021-03516-y3
![]() |
[68] |
Z. Ren, X. Zhen, Z. Jiang, Underactuated control and analysis of single blade installation using a jackup installation vessel and active tugger line force control, Mar. struct., 88 (2023), 103338. https://doi.org/10.1016/j.marstruc.2022.103338 doi: 10.1016/j.marstruc.2022.103338
![]() |
[69] |
Y. Song, G. Zhao, B. Zhang, An enhanced distributed differential evolution algorithm for portfolio optimization problems, Eng. Appl. Artif. Intell., 121 (2023), 106004. https://doi.org/10.1016/j.engappai.2023.106004 doi: 10.1016/j.engappai.2023.106004
![]() |
[70] |
H. Zhao, C. Wang, H. Chen, A hybrid classification method with dual-channel CNN and KELM for hyperspectral remote sensing images, Int. J. Remote Sens., 44 (2023), 289–310. https://doi.org/10.1080/01431161.2022.2162352 doi: 10.1080/01431161.2022.2162352
![]() |
[71] |
X. Zhao, M. Zhang, R. Tao, Fractional Fourier image transformer for multimodal remote sensing data classification, IEEE Trans. Neural Networks Learn. Syst., (2022), 1–13. https://doi.org/10.1109/TNNLS.2022.3189994 doi: 10.1109/TNNLS.2022.3189994
![]() |
[72] |
T. Ojala, I. Harwood, A comparative study of texture measures with classification based on feature distributions, Pattern Recognit., 29 (1996), 51–59. https://doi.org/10.1016/0031-3203(95)00067-4 doi: 10.1016/0031-3203(95)00067-4
![]() |
1. | Songyang Lyu, Ray C. C. Cheung, Efficient and Automatic Breast Cancer Early Diagnosis System Based on the Hierarchical Extreme Learning Machine, 2023, 23, 1424-8220, 7772, 10.3390/s23187772 | |
2. | Yansong Du, 2023, Research on Photographic Image Classification Based on Multi-model Fusion and Data Augmentation, 979-8-3503-1545-5, 1, 10.1109/ICIICS59993.2023.10420882 |
Data | Selection method | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | IE | 78.39 | 79.10 | 79.99 | 80.71 | 82.11 | 83.47 | 84.10 | 85.51 | 86.48 | 87.23 |
ME | 80.33 | 86.83 | 91.08 | 92.92 | 95.28 | 97.02 | 98.00 | 98.45 | 98.90 | 99.13 | |
BT | 91.47 | 95.47 | 98.22 | 98.36 | 98.64 | 98.59 | 98.66 | 98.71 | 99.34 | 99.29 | |
LC | 84.74 | 88.95 | 91.63 | 94.33 | 95.27 | 96.46 | 98.15 | 98.55 | 98.62 | 98.66 | |
Pavia University | IE | 69.87 | 71.31 | 71.75 | 71.74 | 72.04 | 72.40 | 73.25 | 73.36 | 73.76 | 74.45 |
ME | 73.01 | 75.23 | 78.74 | 82.78 | 89.14 | 92.82 | 95.14 | 96.10 | 97.13 | 97.82 | |
BT | 87.54 | 92.63 | 94.51 | 95.24 | 95.84 | 96.10 | 96.39 | 96.50 | 96.69 | 96.71 | |
LC | 74.91 | 76.33 | 80.76 | 84.45 | 87.23 | 89.89 | 90.27 | 90.67 | 90.56 | 91.47 | |
Salinas Scene | IE | 84.16 | 84.60 | 84.65 | 85.02 | 85.09 | 85.25 | 85.50 | 85.65 | 85.91 | 85.98 |
ME | 85.14 | 89.29 | 92.81 | 94.31 | 96.48 | 97.41 | 98.04 | 98.20 | 98.66 | 98.88 | |
BT | 95.30 | 96.98 | 98.26 | 98.71 | 98.95 | 98.90 | 99.03 | 99.21 | 99.25 | 99.24 | |
LC | 88.60 | 91.19 | 92.84 | 93.58 | 93.80 | 95.74 | 97.56 | 97.72 | 98.56 | 98.86 |
Data | Quantity | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | 200 | 77.30 | 77.57 | 78.32 | 77.97 | 78.59 | 78.97 | 78.96 | 79.28 | 79.51 | 79.23 |
400 | 77.54 | 78.60 | 79.80 | 80.80 | 82.13 | 82.20 | 83.03 | 83.77 | 83.83 | 83.85 | |
600 | 77.52 | 79.54 | 79.37 | 79.27 | 80.08 | 81.99 | 83.89 | 83.60 | 84.30 | 84.48 | |
800 | 77.75 | 79.84 | 80.87 | 80.22 | 82.11 | 83.91 | 84.41 | 84.57 | 85.12 | 85.94 | |
1000 | 77.85 | 80.49 | 80.28 | 82.74 | 81.35 | 82.60 | 84.18 | 85.88 | 86.77 | 87.84 | |
1200 | 77.85 | 79.95 | 79.79 | 80.38 | 81.59 | 84.41 | 84.72 | 85.74 | 87.48 | 88.85 | |
1400 | 78.18 | 80.20 | 80.09 | 83.96 | 85.00 | 85.78 | 87.93 | 89.90 | 91.07 | 91.20 | |
1600 | 78.55 | 80.56 | 80.59 | 84.12 | 86.90 | 87.82 | 89.02 | 90.87 | 91.49 | 91.83 | |
1800 | 78.34 | 79.76 | 79.23 | 82.64 | 85.39 | 87.32 | 88.81 | 89.97 | 90.69 | 91.46 | |
2000 | 78.01 | 79.16 | 80.24 | 82.55 | 86.02 | 87.48 | 88.66 | 89.63 | 90.02 | 90.46 | |
Pavia University | 200 | 68.75 | 73.93 | 76.73 | 78.18 | 79.23 | 80.92 | 81.85 | 82.40 | 82.74 | 83.60 |
400 | 66.41 | 73.11 | 75.45 | 78.20 | 81.18 | 82.13 | 82.57 | 83.39 | 84.07 | 83.98 | |
600 | 68.88 | 76.35 | 78.22 | 80.73 | 82.59 | 83.29 | 83.68 | 84.57 | 84.95 | 84.98 | |
800 | 69.89 | 77.50 | 80.31 | 81.91 | 83.22 | 84.85 | 84.99 | 84.79 | 85.16 | 85.31 | |
1000 | 70.28 | 76.35 | 79.92 | 82.68 | 83.83 | 84.48 | 84.84 | 85.21 | 85.37 | 85.04 | |
1200 | 70.24 | 75.18 | 80.32 | 83.13 | 84.14 | 85.04 | 85.30 | 85.55 | 85.04 | 84.90 | |
1400 | 70.34 | 76.23 | 80.57 | 82.46 | 83.92 | 84.71 | 85.59 | 85.77 | 85.87 | 86.47 | |
1600 | 70.40 | 75.87 | 80.64 | 83.06 | 83.93 | 84.69 | 85.28 | 86.02 | 86.19 | 85.83 | |
1800 | 69.77 | 76.12 | 80.18 | 82.99 | 85.19 | 85.04 | 84.89 | 85.26 | 85.68 | 85.68 | |
2000 | 69.71 | 75.90 | 82.29 | 83.40 | 84.41 | 85.23 | 85.59 | 85.77 | 85.86 | 85.87 | |
Salinas Scene | 200 | 85.09 | 87.26 | 89.04 | 89.35 | 89.24 | 89.22 | 88.85 | 88.47 | 88.14 | 88.03 |
400 | 84.94 | 88.34 | 89.88 | 89.26 | 89.12 | 88.88 | 88.26 | 87.68 | 87.40 | 87.25 | |
600 | 85.69 | 90.85 | 90.80 | 90.42 | 89.71 | 89.04 | 88.63 | 87.84 | 87.12 | 86.60 | |
800 | 85.36 | 89.17 | 88.87 | 87.96 | 87.46 | 86.85 | 86.42 | 85.69 | 85.13 | 84.58 | |
1000 | 85.35 | 88.93 | 89.88 | 89.04 | 88.34 | 87.58 | 87.03 | 85.49 | 85.08 | 85.08 | |
1200 | 85.04 | 89.66 | 90.08 | 88.98 | 87.67 | 87.14 | 86.13 | 85.30 | 85.09 | 84.85 | |
1400 | 85.07 | 88.46 | 89.39 | 88.63 | 88.10 | 88.06 | 87.34 | 86.81 | 86.69 | 86.58 | |
1600 | 85.47 | 89.54 | 89.88 | 88.87 | 87.68 | 86.21 | 85.00 | 84.34 | 83.89 | 83.70 | |
1800 | 85.50 | 90.16 | 90.15 | 89.45 | 88.58 | 87.71 | 86.79 | 86.52 | 86.31 | 85.93 | |
2000 | 85.48 | 89.40 | 89.38 | 88.60 | 87.73 | 85.98 | 85.67 | 85.12 | 85.39 | 85.34 |
Data | Block window size | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | 3 | 77.62 | 77.86 | 78.35 | 78.65 | 78.91 | 79.14 | 79.09 | 79.12 | 79.56 | 79.56 |
4 | 77.73 | 78.24 | 78.65 | 78.75 | 78.70 | 78.66 | 79.14 | 79.18 | 79.70 | 79.74 | |
5 | 78.41 | 79.61 | 79.41 | 81.28 | 81.11 | 81.73 | 82.67 | 82.84 | 84.18 | 84.68 | |
6 | 77.85 | 80.14 | 80.41 | 80.91 | 80.90 | 83.54 | 84.91 | 85.09 | 86.42 | 88.30 | |
7 | 78.86 | 80.58 | 82.58 | 84.90 | 86.55 | 86.99 | 88.47 | 88.87 | 89.58 | 90.98 | |
8 | 78.33 | 78.95 | 81.00 | 84.24 | 85.44 | 86.37 | 87.60 | 88.13 | 88.57 | 89.19 | |
9 | 79.61 | 81.00 | 83.39 | 85.51 | 85.71 | 86.34 | 86.78 | 86.86 | 87.07 | 87.86 | |
10 | 78.61 | 79.32 | 83.45 | 85.45 | 85.59 | 85.69 | 86.12 | 87.00 | 87.17 | 87.32 | |
Pavia University | 5 | 68.39 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 |
10 | 68.20 | 67.67 | 69.05 | 69.06 | 68.30 | 68.13 | 68.06 | 67.98 | 67.89 | 67.62 | |
15 | 70.91 | 70.59 | 70.96 | 70.50 | 71.89 | 73.54 | 73.80 | 74.58 | 75.25 | 75.32 | |
20 | 71.47 | 71.63 | 73.60 | 76.12 | 75.99 | 75.61 | 77.18 | 77.51 | 77.81 | 78.14 | |
25 | 71.25 | 74.57 | 78.52 | 79.15 | 81.95 | 82.32 | 83.66 | 84.88 | 85.45 | 85.51 | |
30 | 70.28 | 76.35 | 79.92 | 82.68 | 83.83 | 84.48 | 84.84 | 85.21 | 85.37 | 85.04 | |
35 | 71.48 | 77.43 | 80.15 | 81.75 | 83.18 | 83.62 | 84.11 | 83.65 | 83.23 | 83.17 | |
40 | 73.40 | 77.50 | 80.36 | 81.92 | 83.15 | 82.68 | 82.23 | 82.38 | 82.18 | 82.04 | |
Salinas Scene | 5 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 |
10 | 83.61 | 84.16 | 84.95 | 85.03 | 84.77 | 84.68 | 84.59 | 84.82 | 84.61 | 85.48 | |
15 | 83.86 | 83.17 | 85.51 | 87.04 | 87.70 | 88.44 | 89.77 | 89.60 | 91.01 | 91.09 | |
20 | 84.16 | 86.84 | 89.41 | 90.56 | 90.22 | 90.69 | 91.02 | 91.13 | 90.87 | 90.68 | |
25 | 84.11 | 88.22 | 89.46 | 89.68 | 89.35 | 88.77 | 87.86 | 87.49 | 86.71 | 86.78 | |
30 | 85.35 | 88.93 | 89.88 | 89.04 | 88.34 | 87.58 | 87.03 | 85.49 | 85.08 | 85.08 | |
35 | 86.17 | 88.80 | 88.67 | 87.67 | 86.63 | 85.84 | 85.26 | 84.87 | 83.95 | 83.06 | |
40 | 86.86 | 89.25 | 89.02 | 87.78 | 86.37 | 84.74 | 83.60 | 82.83 | 81.69 | 80.92 |
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 314 | 673 | 1139 | 1630 | 2178 | 2680 | 3324 | 4063 | 4820 | 5474 |
OA (%) | 78.69 | 79.93 | 81.20 | 82.05 | 82.49 | 84.38 | 85.70 | 86.52 | 87.33 | 87.77 | |
SRC | NUM | 311 | 648 | 1068 | 1525 | 2032 | 2606 | 3248 | 3899 | 4668 | 5522 |
OA (%) | 78.64 | 80.42 | 80.10 | 81.56 | 82.87 | 84.39 | 85.83 | 87.04 | 87.92 | 88.19 | |
NRS | NUM | 315 | 673 | 1116 | 1597 | 2136 | 2697 | 3265 | 3815 | 4437 | 5016 |
OA (%) | 78.82 | 81.02 | 80.96 | 83.71 | 84.94 | 85.57 | 86.87 | 88.01 | 89.25 | 89.42 | |
MLR | NUM | 133 | 295 | 580 | 936 | 1296 | 1790 | 2396 | 3077 | 3858 | 4648 |
OA (%) | 77.55 | 79.31 | 82.21 | 83.91 | 83.99 | 85.75 | 87.36 | 87.68 | 88.27 | 88.41 | |
KNN + SRC | NUM | 317 | 706 | 1120 | 1702 | 2294 | 2967 | 3728 | 4494 | 5233 | 5950 |
OA (%) | 78.71 | 80.96 | 81.99 | 83.97 | 84.96 | 85.82 | 86.99 | 87.75 | 87.93 | 88.10 | |
KNN + NRS | NUM | 317 | 707 | 1198 | 1691 | 2308 | 2954 | 3625 | 4450 | 5374 | 6305 |
OA (%) | 78.78 | 80.27 | 80.73 | 82.51 | 83.82 | 85.45 | 87.56 | 89.06 | 89.95 | 90.19 | |
KNN + MLR | NUM | 318 | 712 | 1206 | 1794 | 2555 | 3398 | 4292 | 5072 | 5783 | 6678 |
OA (%) | 78.79 | 80.56 | 82.57 | 86.08 | 87.10 | 87.87 | 88.50 | 88.88 | 89.18 | 89.31 | |
SRC + KNN | NUM | 317 | 706 | 1134 | 1673 | 2223 | 2875 | 3658 | 4317 | 5011 | 5748 |
OA (%) | 78.71 | 81.09 | 81.51 | 83.17 | 84.27 | 85.84 | 86.94 | 87.54 | 88.26 | 88.64 | |
SRC + NRS | NUM | 318 | 730 | 1205 | 1778 | 2372 | 3091 | 3969 | 4813 | 5787 | 6641 |
OA (%) | 78.81 | 80.79 | 82.37 | 85.03 | 86.81 | 88.45 | 88.93 | 89.86 | 90.49 | 90.78 | |
SRC + MLR | NUM | 315 | 708 | 1153 | 1744 | 2457 | 3322 | 4231 | 5042 | 5800 | 6735 |
OA (%) | 78.80 | 80.92 | 82.16 | 85.16 | 86.24 | 87.61 | 88.39 | 88.71 | 89.07 | 89.19 | |
NRS + KNN | NUM | 317 | 707 | 1202 | 1712 | 2385 | 3021 | 3700 | 4538 | 5399 | 6333 |
OA (%) | 78.74 | 80.67 | 81.33 | 83.36 | 84.46 | 86.35 | 87.93 | 89.23 | 90.03 | 90.43 | |
NRS + SRC | NUM | 318 | 734 | 1207 | 1728 | 2378 | 3061 | 3959 | 4799 | 5691 | 6739 |
OA (%) | 78.77 | 81.01 | 82.56 | 84.52 | 86.37 | 87.43 | 88.95 | 90.11 | 90.61 | 90.90 | |
NRS + MLR | NUM | 318 | 694 | 1148 | 1690 | 2446 | 3194 | 4102 | 4950 | 5768 | 6792 |
OA (%) | 78.91 | 81.39 | 81.62 | 85.51 | 87.21 | 89.63 | 90.28 | 90.92 | 91.28 | 91.88 | |
MLR + KNN | NUM | 318 | 677 | 1166 | 1689 | 2429 | 3246 | 4018 | 4737 | 5677 | 6672 |
OA (%) | 78.30 | 81.13 | 81.85 | 85.94 | 87.22 | 88.28 | 88.77 | 88.98 | 89.20 | 89.29 | |
MLR + SRC | NUM | 315 | 704 | 1258 | 1741 | 2439 | 3286 | 4097 | 4867 | 5775 | 6760 |
OA (%) | 78.31 | 81.81 | 82.55 | 85.09 | 86.86 | 88.30 | 89.01 | 89.44 | 90.21 | 90.71 | |
MLR + NRS | NUM | 318 | 683 | 1154 | 1701 | 2458 | 3301 | 4219 | 5057 | 5997 | 6889 |
OA (%) | 78.46 | 81.49 | 82.11 | 86.14 | 87.86 | 89.70 | 90.68 | 91.58 | 92.15 | 92.42 |
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 236 | 444 | 661 | 965 | 1210 | 1466 | 1812 | 2254 | 2762 | 3251 |
OA (%) | 71.19 | 73.36 | 74.70 | 75.69 | 76.10 | 76.16 | 77.28 | 78.48 | 77.60 | 77.69 | |
SRC | NUM | 225 | 477 | 733 | 1080 | 1492 | 2047 | 2793 | 3654 | 4605 | 5631 |
OA (%) | 72.21 | 72.19 | 76.97 | 79.12 | 80.08 | 81.77 | 83.09 | 84.07 | 84.95 | 85.27 | |
NRS | NUM | 225 | 438 | 763 | 1007 | 1256 | 1487 | 1726 | 1893 | 1991 | 2179 |
OA (%) | 71.22 | 73.74 | 75.14 | 76.59 | 76.15 | 76.47 | 76.41 | 76.20 | 75.68 | 75.40 | |
MLR | NUM | 100 | 244 | 396 | 605 | 829 | 1035 | 1304 | 1587 | 1872 | 2183 |
OA (%) | 72.36 | 76.37 | 77.18 | 79.04 | 79.40 | 81.16 | 82.21 | 82.18 | 83.36 | 85.85 | |
KNN + SRC | NUM | 248 | 473 | 787 | 1118 | 1509 | 1996 | 2552 | 3111 | 3975 | 4837 |
OA (%) | 71.48 | 72.82 | 74.61 | 76.59 | 78.38 | 80.08 | 80.82 | 81.62 | 82.40 | 83.99 | |
KNN + NRS | NUM | 240 | 491 | 775 | 1067 | 1382 | 1850 | 2403 | 3040 | 3794 | 4415 |
OA (%) | 71.11 | 73.83 | 77.44 | 78.83 | 79.50 | 79.58 | 79.39 | 80.11 | 80.74 | 81.48 | |
KNN + MLR | NUM | 244 | 515 | 822 | 1176 | 1616 | 2162 | 2905 | 3698 | 4745 | 5682 |
OA (%) | 71.37 | 75.03 | 77.84 | 78.08 | 79.73 | 79.59 | 81.30 | 83.88 | 84.90 | 85.08 | |
SRC + KNN | NUM | 248 | 476 | 795 | 1215 | 1725 | 2298 | 3055 | 3967 | 4977 | 5913 |
OA (%) | 71.38 | 72.54 | 75.26 | 78.09 | 79.19 | 80.21 | 81.88 | 83.31 | 83.35 | 83.51 | |
SRC + NRS | NUM | 242 | 511 | 867 | 1385 | 1889 | 2513 | 3201 | 3988 | 4910 | 5794 |
OA (%) | 71.40 | 74.12 | 77.56 | 80.80 | 82.19 | 82.55 | 83.02 | 83.80 | 84.34 | 85.09 | |
SRC + MLR | NUM | 236 | 507 | 841 | 1289 | 1731 | 2261 | 3016 | 3928 | 4939 | 6015 |
OA (%) | 71.52 | 73.62 | 76.47 | 78.45 | 79.51 | 81.54 | 83.10 | 84.37 | 84.68 | 84.66 | |
NRS + KNN | NUM | 240 | 486 | 803 | 1119 | 1541 | 1992 | 2557 | 3190 | 3939 | 4611 |
OA (%) | 71.05 | 74.37 | 76.57 | 77.77 | 78.06 | 77.21 | 76.75 | 77.35 | 78.88 | 79.28 | |
NRS + SRC | NUM | 242 | 501 | 803 | 1296 | 1792 | 2433 | 3089 | 3839 | 4776 | 5798 |
OA (%) | 71.44 | 74.50 | 76.72 | 79.93 | 81.44 | 81.90 | 82.85 | 83.81 | 85.00 | 85.48 | |
NRS + MLR | NUM | 234 | 517 | 828 | 1237 | 1796 | 2446 | 3354 | 4220 | 5061 | 5857 |
OA (%) | 71.49 | 75.47 | 77.89 | 80.86 | 83.68 | 84.43 | 84.93 | 85.12 | 85.57 | 85.46 | |
MLR + KNN | NUM | 244 | 486 | 746 | 1170 | 1658 | 2300 | 3205 | 4208 | 5336 | 6514 |
OA (%) | 71.47 | 75.30 | 76.67 | 78.76 | 80.35 | 82.97 | 85.05 | 86.05 | 86.88 | 87.02 | |
MLR + SRC | NUM | 236 | 504 | 903 | 1310 | 1708 | 2207 | 3009 | 4024 | 5098 | 6234 |
OA (%) | 71.71 | 74.01 | 79.30 | 79.81 | 80.40 | 83.03 | 86.27 | 86.93 | 87.97 | 88.53 | |
MLR + NRS | NUM | 234 | 524 | 787 | 1205 | 1738 | 2374 | 3116 | 3953 | 4754 | 5586 |
OA (%) | 71.63 | 75.94 | 79.83 | 80.66 | 82.75 | 84.64 | 85.98 | 86.19 | 86.37 | 86.87 |
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 133 | 251 | 443 | 684 | 963 | 1304 | 1672 | 2047 | 2425 | 2857 |
OA (%) | 83.38 | 86.46 | 86.46 | 86.65 | 87.26 | 87.31 | 87.79 | 87.92 | 87.66 | 87.15 | |
SRC | NUM | 144 | 275 | 441 | 666 | 937 | 1255 | 1606 | 1968 | 2391 | 2811 |
OA (%) | 83.10 | 83.56 | 85.11 | 85.83 | 85.81 | 86.63 | 86.50 | 86.64 | 86.95 | 86.96 | |
NRS | NUM | 148 | 271 | 423 | 668 | 952 | 1263 | 1569 | 1940 | 2351 | 2799 |
OA (%) | 84.04 | 86.06 | 87.62 | 87.63 | 87.10 | 87.48 | 88.44 | 88.43 | 88.32 | 87.92 | |
MLR | NUM | 102 | 177 | 302 | 451 | 621 | 867 | 1176 | 1518 | 1848 | 2217 |
OA (%) | 82.88 | 85.41 | 87.11 | 88.36 | 88.73 | 90.68 | 91.52 | 92.20 | 91.70 | 92.02 | |
KNN + SRC | NUM | 146 | 294 | 479 | 694 | 976 | 1330 | 1691 | 2106 | 2520 | 2985 |
OA (%) | 83.60 | 84.65 | 85.38 | 86.76 | 87.39 | 87.39 | 88.00 | 88.11 | 87.53 | 87.23 | |
KNN + NRS | NUM | 150 | 297 | 500 | 761 | 1108 | 1466 | 1891 | 2354 | 2834 | 3316 |
OA (%) | 83.53 | 86.43 | 87.14 | 87.49 | 87.64 | 87.37 | 88.06 | 88.03 | 87.95 | 88.03 | |
KNN + MLR | NUM | 143 | 285 | 508 | 768 | 1132 | 1546 | 2026 | 2526 | 3049 | 3590 |
OA (%) | 83.38 | 85.50 | 86.34 | 88.35 | 88.80 | 90.07 | 90.24 | 90.08 | 89.86 | 89.79 | |
SRC + KNN | NUM | 146 | 287 | 472 | 686 | 957 | 1281 | 1673 | 2061 | 2485 | 2892 |
OA (%) | 83.25 | 84.70 | 85.79 | 85.88 | 86.70 | 86.92 | 86.89 | 86.93 | 86.85 | 86.90 | |
SRC + NRS | NUM | 150 | 280 | 493 | 755 | 1017 | 1328 | 1708 | 2114 | 2539 | 2982 |
OA (%) | 83.07 | 85.55 | 86.28 | 85.47 | 86.66 | 86.97 | 87.52 | 87.05 | 87.19 | 87.16 | |
SRC + MLR | NUM | 150 | 271 | 483 | 720 | 1108 | 1499 | 1958 | 2451 | 2969 | 3487 |
OA (%) | 83.15 | 84.27 | 87.25 | 87.88 | 88.88 | 89.29 | 89.36 | 89.85 | 89.70 | 89.85 | |
NRS + KNN | NUM | 150 | 298 | 519 | 814 | 1148 | 1556 | 2037 | 2538 | 3027 | 3522 |
OA (%) | 84.01 | 87.12 | 87.79 | 87.30 | 87.54 | 88.38 | 88.09 | 88.25 | 87.95 | 87.88 | |
NRS + SRC | NUM | 150 | 284 | 488 | 803 | 1158 | 1539 | 1988 | 2482 | 2955 | 3423 |
OA (%) | 83.91 | 87.23 | 86.98 | 87.73 | 88.30 | 88.97 | 89.57 | 89.57 | 89.60 | 89.43 | |
NRS + MLR | NUM | 153 | 293 | 509 | 762 | 1104 | 1509 | 1940 | 2441 | 2934 | 3452 |
OA (%) | 83.87 | 85.82 | 88.25 | 89.93 | 90.40 | 90.51 | 90.75 | 90.48 | 90.12 | 89.61 | |
MLR + KNN | NUM | 143 | 299 | 521 | 825 | 1187 | 1602 | 2046 | 2514 | 2993 | 3407 |
OA (%) | 82.80 | 85.42 | 87.67 | 88.79 | 89.68 | 90.06 | 90.54 | 90.92 | 91.27 | 91.32 | |
MLR + SRC | NUM | 150 | 292 | 521 | 799 | 1123 | 1537 | 2007 | 2448 | 2929 | 3367 |
OA (%) | 82.91 | 85.45 | 88.89 | 90.12 | 90.21 | 90.93 | 90.99 | 91.83 | 92.08 | 92.64 | |
MLR + NRS | NUM | 153 | 315 | 564 | 841 | 1197 | 1605 | 2064 | 2561 | 3060 | 3500 |
OA (%) | 82.81 | 84.87 | 88.16 | 89.34 | 89.69 | 90.03 | 90.42 | 90.52 | 91.19 | 91.43 |
Data set | Indian Pines | Pavia University | Salinas Scene |
Selection policy | BT | BT | BT |
Number of selections | 1600 | 1400 | 600 |
Window size | 7 × 7 | 25 × 25 | 20 × 20 |
Combination of classifiers | MLR + NRS | MLR + SRC | MLR + SRC |
Number of labeled samples | 6889 | 6234 | 3367 |
Training samples | Index | Initial samples | Labeling samples |
Indian Pines | AA | 67.93% | 84.70% |
OA | 77.38% | 92.42% | |
KAPPA | 0.746 | 0.914 | |
Pavia University | AA | 60.53% | 81.87% |
OA | 69.00% | 88.53% | |
KAPPA | 0.609 | 0.848 | |
Salinas Scene | AA | 82.59% | 87.76% |
OA | 84.00% | 92.64% | |
KAPPA | 0.823 | 0.918 |
Data | Selection method | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | IE | 78.39 | 79.10 | 79.99 | 80.71 | 82.11 | 83.47 | 84.10 | 85.51 | 86.48 | 87.23 |
ME | 80.33 | 86.83 | 91.08 | 92.92 | 95.28 | 97.02 | 98.00 | 98.45 | 98.90 | 99.13 | |
BT | 91.47 | 95.47 | 98.22 | 98.36 | 98.64 | 98.59 | 98.66 | 98.71 | 99.34 | 99.29 | |
LC | 84.74 | 88.95 | 91.63 | 94.33 | 95.27 | 96.46 | 98.15 | 98.55 | 98.62 | 98.66 | |
Pavia University | IE | 69.87 | 71.31 | 71.75 | 71.74 | 72.04 | 72.40 | 73.25 | 73.36 | 73.76 | 74.45 |
ME | 73.01 | 75.23 | 78.74 | 82.78 | 89.14 | 92.82 | 95.14 | 96.10 | 97.13 | 97.82 | |
BT | 87.54 | 92.63 | 94.51 | 95.24 | 95.84 | 96.10 | 96.39 | 96.50 | 96.69 | 96.71 | |
LC | 74.91 | 76.33 | 80.76 | 84.45 | 87.23 | 89.89 | 90.27 | 90.67 | 90.56 | 91.47 | |
Salinas Scene | IE | 84.16 | 84.60 | 84.65 | 85.02 | 85.09 | 85.25 | 85.50 | 85.65 | 85.91 | 85.98 |
ME | 85.14 | 89.29 | 92.81 | 94.31 | 96.48 | 97.41 | 98.04 | 98.20 | 98.66 | 98.88 | |
BT | 95.30 | 96.98 | 98.26 | 98.71 | 98.95 | 98.90 | 99.03 | 99.21 | 99.25 | 99.24 | |
LC | 88.60 | 91.19 | 92.84 | 93.58 | 93.80 | 95.74 | 97.56 | 97.72 | 98.56 | 98.86 |
Data | Quantity | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | 200 | 77.30 | 77.57 | 78.32 | 77.97 | 78.59 | 78.97 | 78.96 | 79.28 | 79.51 | 79.23 |
400 | 77.54 | 78.60 | 79.80 | 80.80 | 82.13 | 82.20 | 83.03 | 83.77 | 83.83 | 83.85 | |
600 | 77.52 | 79.54 | 79.37 | 79.27 | 80.08 | 81.99 | 83.89 | 83.60 | 84.30 | 84.48 | |
800 | 77.75 | 79.84 | 80.87 | 80.22 | 82.11 | 83.91 | 84.41 | 84.57 | 85.12 | 85.94 | |
1000 | 77.85 | 80.49 | 80.28 | 82.74 | 81.35 | 82.60 | 84.18 | 85.88 | 86.77 | 87.84 | |
1200 | 77.85 | 79.95 | 79.79 | 80.38 | 81.59 | 84.41 | 84.72 | 85.74 | 87.48 | 88.85 | |
1400 | 78.18 | 80.20 | 80.09 | 83.96 | 85.00 | 85.78 | 87.93 | 89.90 | 91.07 | 91.20 | |
1600 | 78.55 | 80.56 | 80.59 | 84.12 | 86.90 | 87.82 | 89.02 | 90.87 | 91.49 | 91.83 | |
1800 | 78.34 | 79.76 | 79.23 | 82.64 | 85.39 | 87.32 | 88.81 | 89.97 | 90.69 | 91.46 | |
2000 | 78.01 | 79.16 | 80.24 | 82.55 | 86.02 | 87.48 | 88.66 | 89.63 | 90.02 | 90.46 | |
Pavia University | 200 | 68.75 | 73.93 | 76.73 | 78.18 | 79.23 | 80.92 | 81.85 | 82.40 | 82.74 | 83.60 |
400 | 66.41 | 73.11 | 75.45 | 78.20 | 81.18 | 82.13 | 82.57 | 83.39 | 84.07 | 83.98 | |
600 | 68.88 | 76.35 | 78.22 | 80.73 | 82.59 | 83.29 | 83.68 | 84.57 | 84.95 | 84.98 | |
800 | 69.89 | 77.50 | 80.31 | 81.91 | 83.22 | 84.85 | 84.99 | 84.79 | 85.16 | 85.31 | |
1000 | 70.28 | 76.35 | 79.92 | 82.68 | 83.83 | 84.48 | 84.84 | 85.21 | 85.37 | 85.04 | |
1200 | 70.24 | 75.18 | 80.32 | 83.13 | 84.14 | 85.04 | 85.30 | 85.55 | 85.04 | 84.90 | |
1400 | 70.34 | 76.23 | 80.57 | 82.46 | 83.92 | 84.71 | 85.59 | 85.77 | 85.87 | 86.47 | |
1600 | 70.40 | 75.87 | 80.64 | 83.06 | 83.93 | 84.69 | 85.28 | 86.02 | 86.19 | 85.83 | |
1800 | 69.77 | 76.12 | 80.18 | 82.99 | 85.19 | 85.04 | 84.89 | 85.26 | 85.68 | 85.68 | |
2000 | 69.71 | 75.90 | 82.29 | 83.40 | 84.41 | 85.23 | 85.59 | 85.77 | 85.86 | 85.87 | |
Salinas Scene | 200 | 85.09 | 87.26 | 89.04 | 89.35 | 89.24 | 89.22 | 88.85 | 88.47 | 88.14 | 88.03 |
400 | 84.94 | 88.34 | 89.88 | 89.26 | 89.12 | 88.88 | 88.26 | 87.68 | 87.40 | 87.25 | |
600 | 85.69 | 90.85 | 90.80 | 90.42 | 89.71 | 89.04 | 88.63 | 87.84 | 87.12 | 86.60 | |
800 | 85.36 | 89.17 | 88.87 | 87.96 | 87.46 | 86.85 | 86.42 | 85.69 | 85.13 | 84.58 | |
1000 | 85.35 | 88.93 | 89.88 | 89.04 | 88.34 | 87.58 | 87.03 | 85.49 | 85.08 | 85.08 | |
1200 | 85.04 | 89.66 | 90.08 | 88.98 | 87.67 | 87.14 | 86.13 | 85.30 | 85.09 | 84.85 | |
1400 | 85.07 | 88.46 | 89.39 | 88.63 | 88.10 | 88.06 | 87.34 | 86.81 | 86.69 | 86.58 | |
1600 | 85.47 | 89.54 | 89.88 | 88.87 | 87.68 | 86.21 | 85.00 | 84.34 | 83.89 | 83.70 | |
1800 | 85.50 | 90.16 | 90.15 | 89.45 | 88.58 | 87.71 | 86.79 | 86.52 | 86.31 | 85.93 | |
2000 | 85.48 | 89.40 | 89.38 | 88.60 | 87.73 | 85.98 | 85.67 | 85.12 | 85.39 | 85.34 |
Data | Block window size | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
Indian Pines | 3 | 77.62 | 77.86 | 78.35 | 78.65 | 78.91 | 79.14 | 79.09 | 79.12 | 79.56 | 79.56 |
4 | 77.73 | 78.24 | 78.65 | 78.75 | 78.70 | 78.66 | 79.14 | 79.18 | 79.70 | 79.74 | |
5 | 78.41 | 79.61 | 79.41 | 81.28 | 81.11 | 81.73 | 82.67 | 82.84 | 84.18 | 84.68 | |
6 | 77.85 | 80.14 | 80.41 | 80.91 | 80.90 | 83.54 | 84.91 | 85.09 | 86.42 | 88.30 | |
7 | 78.86 | 80.58 | 82.58 | 84.90 | 86.55 | 86.99 | 88.47 | 88.87 | 89.58 | 90.98 | |
8 | 78.33 | 78.95 | 81.00 | 84.24 | 85.44 | 86.37 | 87.60 | 88.13 | 88.57 | 89.19 | |
9 | 79.61 | 81.00 | 83.39 | 85.51 | 85.71 | 86.34 | 86.78 | 86.86 | 87.07 | 87.86 | |
10 | 78.61 | 79.32 | 83.45 | 85.45 | 85.59 | 85.69 | 86.12 | 87.00 | 87.17 | 87.32 | |
Pavia University | 5 | 68.39 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 | 68.38 |
10 | 68.20 | 67.67 | 69.05 | 69.06 | 68.30 | 68.13 | 68.06 | 67.98 | 67.89 | 67.62 | |
15 | 70.91 | 70.59 | 70.96 | 70.50 | 71.89 | 73.54 | 73.80 | 74.58 | 75.25 | 75.32 | |
20 | 71.47 | 71.63 | 73.60 | 76.12 | 75.99 | 75.61 | 77.18 | 77.51 | 77.81 | 78.14 | |
25 | 71.25 | 74.57 | 78.52 | 79.15 | 81.95 | 82.32 | 83.66 | 84.88 | 85.45 | 85.51 | |
30 | 70.28 | 76.35 | 79.92 | 82.68 | 83.83 | 84.48 | 84.84 | 85.21 | 85.37 | 85.04 | |
35 | 71.48 | 77.43 | 80.15 | 81.75 | 83.18 | 83.62 | 84.11 | 83.65 | 83.23 | 83.17 | |
40 | 73.40 | 77.50 | 80.36 | 81.92 | 83.15 | 82.68 | 82.23 | 82.38 | 82.18 | 82.04 | |
Salinas Scene | 5 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 | 83.94 |
10 | 83.61 | 84.16 | 84.95 | 85.03 | 84.77 | 84.68 | 84.59 | 84.82 | 84.61 | 85.48 | |
15 | 83.86 | 83.17 | 85.51 | 87.04 | 87.70 | 88.44 | 89.77 | 89.60 | 91.01 | 91.09 | |
20 | 84.16 | 86.84 | 89.41 | 90.56 | 90.22 | 90.69 | 91.02 | 91.13 | 90.87 | 90.68 | |
25 | 84.11 | 88.22 | 89.46 | 89.68 | 89.35 | 88.77 | 87.86 | 87.49 | 86.71 | 86.78 | |
30 | 85.35 | 88.93 | 89.88 | 89.04 | 88.34 | 87.58 | 87.03 | 85.49 | 85.08 | 85.08 | |
35 | 86.17 | 88.80 | 88.67 | 87.67 | 86.63 | 85.84 | 85.26 | 84.87 | 83.95 | 83.06 | |
40 | 86.86 | 89.25 | 89.02 | 87.78 | 86.37 | 84.74 | 83.60 | 82.83 | 81.69 | 80.92 |
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 314 | 673 | 1139 | 1630 | 2178 | 2680 | 3324 | 4063 | 4820 | 5474 |
OA (%) | 78.69 | 79.93 | 81.20 | 82.05 | 82.49 | 84.38 | 85.70 | 86.52 | 87.33 | 87.77 | |
SRC | NUM | 311 | 648 | 1068 | 1525 | 2032 | 2606 | 3248 | 3899 | 4668 | 5522 |
OA (%) | 78.64 | 80.42 | 80.10 | 81.56 | 82.87 | 84.39 | 85.83 | 87.04 | 87.92 | 88.19 | |
NRS | NUM | 315 | 673 | 1116 | 1597 | 2136 | 2697 | 3265 | 3815 | 4437 | 5016 |
OA (%) | 78.82 | 81.02 | 80.96 | 83.71 | 84.94 | 85.57 | 86.87 | 88.01 | 89.25 | 89.42 | |
MLR | NUM | 133 | 295 | 580 | 936 | 1296 | 1790 | 2396 | 3077 | 3858 | 4648 |
OA (%) | 77.55 | 79.31 | 82.21 | 83.91 | 83.99 | 85.75 | 87.36 | 87.68 | 88.27 | 88.41 | |
KNN + SRC | NUM | 317 | 706 | 1120 | 1702 | 2294 | 2967 | 3728 | 4494 | 5233 | 5950 |
OA (%) | 78.71 | 80.96 | 81.99 | 83.97 | 84.96 | 85.82 | 86.99 | 87.75 | 87.93 | 88.10 | |
KNN + NRS | NUM | 317 | 707 | 1198 | 1691 | 2308 | 2954 | 3625 | 4450 | 5374 | 6305 |
OA (%) | 78.78 | 80.27 | 80.73 | 82.51 | 83.82 | 85.45 | 87.56 | 89.06 | 89.95 | 90.19 | |
KNN + MLR | NUM | 318 | 712 | 1206 | 1794 | 2555 | 3398 | 4292 | 5072 | 5783 | 6678 |
OA (%) | 78.79 | 80.56 | 82.57 | 86.08 | 87.10 | 87.87 | 88.50 | 88.88 | 89.18 | 89.31 | |
SRC + KNN | NUM | 317 | 706 | 1134 | 1673 | 2223 | 2875 | 3658 | 4317 | 5011 | 5748 |
OA (%) | 78.71 | 81.09 | 81.51 | 83.17 | 84.27 | 85.84 | 86.94 | 87.54 | 88.26 | 88.64 | |
SRC + NRS | NUM | 318 | 730 | 1205 | 1778 | 2372 | 3091 | 3969 | 4813 | 5787 | 6641 |
OA (%) | 78.81 | 80.79 | 82.37 | 85.03 | 86.81 | 88.45 | 88.93 | 89.86 | 90.49 | 90.78 | |
SRC + MLR | NUM | 315 | 708 | 1153 | 1744 | 2457 | 3322 | 4231 | 5042 | 5800 | 6735 |
OA (%) | 78.80 | 80.92 | 82.16 | 85.16 | 86.24 | 87.61 | 88.39 | 88.71 | 89.07 | 89.19 | |
NRS + KNN | NUM | 317 | 707 | 1202 | 1712 | 2385 | 3021 | 3700 | 4538 | 5399 | 6333 |
OA (%) | 78.74 | 80.67 | 81.33 | 83.36 | 84.46 | 86.35 | 87.93 | 89.23 | 90.03 | 90.43 | |
NRS + SRC | NUM | 318 | 734 | 1207 | 1728 | 2378 | 3061 | 3959 | 4799 | 5691 | 6739 |
OA (%) | 78.77 | 81.01 | 82.56 | 84.52 | 86.37 | 87.43 | 88.95 | 90.11 | 90.61 | 90.90 | |
NRS + MLR | NUM | 318 | 694 | 1148 | 1690 | 2446 | 3194 | 4102 | 4950 | 5768 | 6792 |
OA (%) | 78.91 | 81.39 | 81.62 | 85.51 | 87.21 | 89.63 | 90.28 | 90.92 | 91.28 | 91.88 | |
MLR + KNN | NUM | 318 | 677 | 1166 | 1689 | 2429 | 3246 | 4018 | 4737 | 5677 | 6672 |
OA (%) | 78.30 | 81.13 | 81.85 | 85.94 | 87.22 | 88.28 | 88.77 | 88.98 | 89.20 | 89.29 | |
MLR + SRC | NUM | 315 | 704 | 1258 | 1741 | 2439 | 3286 | 4097 | 4867 | 5775 | 6760 |
OA (%) | 78.31 | 81.81 | 82.55 | 85.09 | 86.86 | 88.30 | 89.01 | 89.44 | 90.21 | 90.71 | |
MLR + NRS | NUM | 318 | 683 | 1154 | 1701 | 2458 | 3301 | 4219 | 5057 | 5997 | 6889 |
OA (%) | 78.46 | 81.49 | 82.11 | 86.14 | 87.86 | 89.70 | 90.68 | 91.58 | 92.15 | 92.42 |
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 236 | 444 | 661 | 965 | 1210 | 1466 | 1812 | 2254 | 2762 | 3251 |
OA (%) | 71.19 | 73.36 | 74.70 | 75.69 | 76.10 | 76.16 | 77.28 | 78.48 | 77.60 | 77.69 | |
SRC | NUM | 225 | 477 | 733 | 1080 | 1492 | 2047 | 2793 | 3654 | 4605 | 5631 |
OA (%) | 72.21 | 72.19 | 76.97 | 79.12 | 80.08 | 81.77 | 83.09 | 84.07 | 84.95 | 85.27 | |
NRS | NUM | 225 | 438 | 763 | 1007 | 1256 | 1487 | 1726 | 1893 | 1991 | 2179 |
OA (%) | 71.22 | 73.74 | 75.14 | 76.59 | 76.15 | 76.47 | 76.41 | 76.20 | 75.68 | 75.40 | |
MLR | NUM | 100 | 244 | 396 | 605 | 829 | 1035 | 1304 | 1587 | 1872 | 2183 |
OA (%) | 72.36 | 76.37 | 77.18 | 79.04 | 79.40 | 81.16 | 82.21 | 82.18 | 83.36 | 85.85 | |
KNN + SRC | NUM | 248 | 473 | 787 | 1118 | 1509 | 1996 | 2552 | 3111 | 3975 | 4837 |
OA (%) | 71.48 | 72.82 | 74.61 | 76.59 | 78.38 | 80.08 | 80.82 | 81.62 | 82.40 | 83.99 | |
KNN + NRS | NUM | 240 | 491 | 775 | 1067 | 1382 | 1850 | 2403 | 3040 | 3794 | 4415 |
OA (%) | 71.11 | 73.83 | 77.44 | 78.83 | 79.50 | 79.58 | 79.39 | 80.11 | 80.74 | 81.48 | |
KNN + MLR | NUM | 244 | 515 | 822 | 1176 | 1616 | 2162 | 2905 | 3698 | 4745 | 5682 |
OA (%) | 71.37 | 75.03 | 77.84 | 78.08 | 79.73 | 79.59 | 81.30 | 83.88 | 84.90 | 85.08 | |
SRC + KNN | NUM | 248 | 476 | 795 | 1215 | 1725 | 2298 | 3055 | 3967 | 4977 | 5913 |
OA (%) | 71.38 | 72.54 | 75.26 | 78.09 | 79.19 | 80.21 | 81.88 | 83.31 | 83.35 | 83.51 | |
SRC + NRS | NUM | 242 | 511 | 867 | 1385 | 1889 | 2513 | 3201 | 3988 | 4910 | 5794 |
OA (%) | 71.40 | 74.12 | 77.56 | 80.80 | 82.19 | 82.55 | 83.02 | 83.80 | 84.34 | 85.09 | |
SRC + MLR | NUM | 236 | 507 | 841 | 1289 | 1731 | 2261 | 3016 | 3928 | 4939 | 6015 |
OA (%) | 71.52 | 73.62 | 76.47 | 78.45 | 79.51 | 81.54 | 83.10 | 84.37 | 84.68 | 84.66 | |
NRS + KNN | NUM | 240 | 486 | 803 | 1119 | 1541 | 1992 | 2557 | 3190 | 3939 | 4611 |
OA (%) | 71.05 | 74.37 | 76.57 | 77.77 | 78.06 | 77.21 | 76.75 | 77.35 | 78.88 | 79.28 | |
NRS + SRC | NUM | 242 | 501 | 803 | 1296 | 1792 | 2433 | 3089 | 3839 | 4776 | 5798 |
OA (%) | 71.44 | 74.50 | 76.72 | 79.93 | 81.44 | 81.90 | 82.85 | 83.81 | 85.00 | 85.48 | |
NRS + MLR | NUM | 234 | 517 | 828 | 1237 | 1796 | 2446 | 3354 | 4220 | 5061 | 5857 |
OA (%) | 71.49 | 75.47 | 77.89 | 80.86 | 83.68 | 84.43 | 84.93 | 85.12 | 85.57 | 85.46 | |
MLR + KNN | NUM | 244 | 486 | 746 | 1170 | 1658 | 2300 | 3205 | 4208 | 5336 | 6514 |
OA (%) | 71.47 | 75.30 | 76.67 | 78.76 | 80.35 | 82.97 | 85.05 | 86.05 | 86.88 | 87.02 | |
MLR + SRC | NUM | 236 | 504 | 903 | 1310 | 1708 | 2207 | 3009 | 4024 | 5098 | 6234 |
OA (%) | 71.71 | 74.01 | 79.30 | 79.81 | 80.40 | 83.03 | 86.27 | 86.93 | 87.97 | 88.53 | |
MLR + NRS | NUM | 234 | 524 | 787 | 1205 | 1738 | 2374 | 3116 | 3953 | 4754 | 5586 |
OA (%) | 71.63 | 75.94 | 79.83 | 80.66 | 82.75 | 84.64 | 85.98 | 86.19 | 86.37 | 86.87 |
Classifier | Index | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
KNN | NUM | 133 | 251 | 443 | 684 | 963 | 1304 | 1672 | 2047 | 2425 | 2857 |
OA (%) | 83.38 | 86.46 | 86.46 | 86.65 | 87.26 | 87.31 | 87.79 | 87.92 | 87.66 | 87.15 | |
SRC | NUM | 144 | 275 | 441 | 666 | 937 | 1255 | 1606 | 1968 | 2391 | 2811 |
OA (%) | 83.10 | 83.56 | 85.11 | 85.83 | 85.81 | 86.63 | 86.50 | 86.64 | 86.95 | 86.96 | |
NRS | NUM | 148 | 271 | 423 | 668 | 952 | 1263 | 1569 | 1940 | 2351 | 2799 |
OA (%) | 84.04 | 86.06 | 87.62 | 87.63 | 87.10 | 87.48 | 88.44 | 88.43 | 88.32 | 87.92 | |
MLR | NUM | 102 | 177 | 302 | 451 | 621 | 867 | 1176 | 1518 | 1848 | 2217 |
OA (%) | 82.88 | 85.41 | 87.11 | 88.36 | 88.73 | 90.68 | 91.52 | 92.20 | 91.70 | 92.02 | |
KNN + SRC | NUM | 146 | 294 | 479 | 694 | 976 | 1330 | 1691 | 2106 | 2520 | 2985 |
OA (%) | 83.60 | 84.65 | 85.38 | 86.76 | 87.39 | 87.39 | 88.00 | 88.11 | 87.53 | 87.23 | |
KNN + NRS | NUM | 150 | 297 | 500 | 761 | 1108 | 1466 | 1891 | 2354 | 2834 | 3316 |
OA (%) | 83.53 | 86.43 | 87.14 | 87.49 | 87.64 | 87.37 | 88.06 | 88.03 | 87.95 | 88.03 | |
KNN + MLR | NUM | 143 | 285 | 508 | 768 | 1132 | 1546 | 2026 | 2526 | 3049 | 3590 |
OA (%) | 83.38 | 85.50 | 86.34 | 88.35 | 88.80 | 90.07 | 90.24 | 90.08 | 89.86 | 89.79 | |
SRC + KNN | NUM | 146 | 287 | 472 | 686 | 957 | 1281 | 1673 | 2061 | 2485 | 2892 |
OA (%) | 83.25 | 84.70 | 85.79 | 85.88 | 86.70 | 86.92 | 86.89 | 86.93 | 86.85 | 86.90 | |
SRC + NRS | NUM | 150 | 280 | 493 | 755 | 1017 | 1328 | 1708 | 2114 | 2539 | 2982 |
OA (%) | 83.07 | 85.55 | 86.28 | 85.47 | 86.66 | 86.97 | 87.52 | 87.05 | 87.19 | 87.16 | |
SRC + MLR | NUM | 150 | 271 | 483 | 720 | 1108 | 1499 | 1958 | 2451 | 2969 | 3487 |
OA (%) | 83.15 | 84.27 | 87.25 | 87.88 | 88.88 | 89.29 | 89.36 | 89.85 | 89.70 | 89.85 | |
NRS + KNN | NUM | 150 | 298 | 519 | 814 | 1148 | 1556 | 2037 | 2538 | 3027 | 3522 |
OA (%) | 84.01 | 87.12 | 87.79 | 87.30 | 87.54 | 88.38 | 88.09 | 88.25 | 87.95 | 87.88 | |
NRS + SRC | NUM | 150 | 284 | 488 | 803 | 1158 | 1539 | 1988 | 2482 | 2955 | 3423 |
OA (%) | 83.91 | 87.23 | 86.98 | 87.73 | 88.30 | 88.97 | 89.57 | 89.57 | 89.60 | 89.43 | |
NRS + MLR | NUM | 153 | 293 | 509 | 762 | 1104 | 1509 | 1940 | 2441 | 2934 | 3452 |
OA (%) | 83.87 | 85.82 | 88.25 | 89.93 | 90.40 | 90.51 | 90.75 | 90.48 | 90.12 | 89.61 | |
MLR + KNN | NUM | 143 | 299 | 521 | 825 | 1187 | 1602 | 2046 | 2514 | 2993 | 3407 |
OA (%) | 82.80 | 85.42 | 87.67 | 88.79 | 89.68 | 90.06 | 90.54 | 90.92 | 91.27 | 91.32 | |
MLR + SRC | NUM | 150 | 292 | 521 | 799 | 1123 | 1537 | 2007 | 2448 | 2929 | 3367 |
OA (%) | 82.91 | 85.45 | 88.89 | 90.12 | 90.21 | 90.93 | 90.99 | 91.83 | 92.08 | 92.64 | |
MLR + NRS | NUM | 153 | 315 | 564 | 841 | 1197 | 1605 | 2064 | 2561 | 3060 | 3500 |
OA (%) | 82.81 | 84.87 | 88.16 | 89.34 | 89.69 | 90.03 | 90.42 | 90.52 | 91.19 | 91.43 |
Data set | Indian Pines | Pavia University | Salinas Scene |
Selection policy | BT | BT | BT |
Number of selections | 1600 | 1400 | 600 |
Window size | 7 × 7 | 25 × 25 | 20 × 20 |
Combination of classifiers | MLR + NRS | MLR + SRC | MLR + SRC |
Number of labeled samples | 6889 | 6234 | 3367 |
Training samples | Index | Initial samples | Labeling samples |
Indian Pines | AA | 67.93% | 84.70% |
OA | 77.38% | 92.42% | |
KAPPA | 0.746 | 0.914 | |
Pavia University | AA | 60.53% | 81.87% |
OA | 69.00% | 88.53% | |
KAPPA | 0.609 | 0.848 | |
Salinas Scene | AA | 82.59% | 87.76% |
OA | 84.00% | 92.64% | |
KAPPA | 0.823 | 0.918 |