Abstract
In recent years, more and more deep learning frameworks are being applied to hyperspectral image classification tasks and have achieved great results. However, the existing network models have higher model complexity and require more time consumption. Traditional hyperspectral image classification methods tend to ignore the correlation between local spatial features. In this paper, a new hyperspectral image classification method is proposed, which combines twodimensional Gabor filter with random patch convolution (GRPC) feature extraction to obtain spatialspectral feature information. The method firstly performs dimensionality reduction through principal component analysis and linear discriminant analysis and extracts the edge texture and spatial information of the image using a Gabor filter for the reduceddimensional image. Next, the extracted information is convolved with random patches to extract spectral features. Finally, the spatial features and multilevel spectral features are fused to classify the images using the Support Vector Machine classifier. In order to verify the performance of this method, experiments were conducted on three widely used datasets of Indian Pines, Pavia University and Kennedy Space Center. The overall classification accuracy reached 98.09%, 99.64% and 96.53%, which are all higher than other comparison methods. The experimental results reveal the superiority of the proposed method in classification accuracy.
Introduction
In recent years, with the development of remote sensing technology, multisource remote sensing imaging means such as hyperspectral, infrared, and radar play an increasingly important role in important fields such as precision agriculture, resource investigation, environmental monitoring, national defense and military^{1,2}. However, hyperspectral images (HSIs) is a threedimensional data cube with special high dimensionality, strong correlation between adjacent bands and highly nonlinear data structure and few training samples, and it makes the classification of HSI with many difficulties^{3}.
How to effectively overcome the above problems, extract multilevel nonlinear discriminant features from HSI data, and improve the classification accuracy of the model, researchers have proposed many methods^{4,5}. For example, methods based on deep learning include Stacked AutoEncoders (SAE)^{6} and Deep Belief Networks (DBN)^{7,8}, Convolutional Neural Networks (CNN)^{9} and Random Patches Networks (RPNet)^{10} and so on. This method learns multilevel nonlinear discriminant features unsupervised from HSI data, has better robustness and discriminant, and can improve the classification accuracy of the model. But deep learning methods require a large number of samples, while HSI training samples are limited. If only spectral or spatial features are used for classification, it is prone to the wellknown Hughes phenomenon. This paper proposes a new method based on a twodimensional Gabor filter and random patches convolution (GRPC), which combines spectralspatial features for HSI image classification.
The specific contributions of the method in this paper are as follows:

1.
For the first time, we introduce a GRPC method combining random patches convolution and covariance matrix representation into hyperspectral image classification. GRPC has a simple structure, and the experiments show that its performance can match the state of the art.

2.
GRPC not only uses the feature extraction capability of random patches convolution and the advantages of Gabor filters, but also realizes the stacking of spectral and spatial features, revealing the importance of spatial structure features that are usually ignored in HSI classification.

3.
GRPC is a competitive and robust method, which can overcome the pepper noise and oversmoothing phenomenon in HSI classification, and it can obtain higher classification accuracy even with limited training samples.
Finally, in order to verify the superiority of GRPC, we also compared with traditional deep learning methods and other classic classification methods. The main ideas for the work to be carried out include the following two aspects: one is to fully consider the benefits of HSI feature information; the other is to improve the classification accuracy of the model while ensuring time efficiency.
The remainder of the paper is organized as follows. In “Related works” section, a comprehensive overview of related work is given. In “Methodology” section, the related technologies, methods and algorithm flow used in this article are described. The information and experimental environment of the data set used are described in “Experimental and analysis” section, which also discusses the experimental parameters and results are discussed. In “Discussion” section, the method of this paper is discussed. Then “Conclusions” section summarizes the paper.
Related works
HSI classification is a very popular research field in recent years. This section discusses existing methods based on deep learning.
Deep learning algorithms can learn representative and discriminative features hierarchically from data. Through the excellent information representation ability brought by deep structure, the automatic extraction and representation of features can be realized^{11}. Designing a reasonable deep network structure can greatly improve the recognition accuracy in classification and target detection applications. For HSI data, the feature extraction of hyperspectral data is extremely difficult due to the high latitude and particularity of hyperspectral images^{12}. In order to solve these problems, deep learning algorithms are more and more widely used in HSI feature extraction, classification and target detection. The early deep learning methods used for HSI feature extraction and classification are SAE and DBN. However, the input of SAE and DBN models needs to stretch the threedimensional structure of the HSI into a onedimensional feature vector, which will lose the spatial information of the HSI. Different from SAE and DBN input, it is based on the CNN method, which can directly process the input HSI unsupervised and output the classification accuracy^{13}. Aiming at the problem of hyperspectral image classification, Hu et al.^{14} first proposed a CNN with a 5layer network depth to extract the spectral features of HSI, and achieved better classification results when using spectral feature classification. Yue et al.^{15} proposed a method based on 2DCNN. First, Principal Component Analysis (PCA) was used to reduce the dimensionality of HSI, and the first 3 principal components were retained. The neighborhood extracts spatial features, but the 2DCNN model alone ignores the spectral information. In order to extract highorder features, Ghasrodashti et al.^{16} constructed a deep autoencoder with multilayer stacking and spatial update capabilities in an unsupervised manner, and extracted the spectral and spatial by adding similarity angle map criteria, fuzzy patterns and multiscale feature weights are used to extract spatial and spectral features. Moreover, Roy et al.^{17} proposed a mixed spectral CNN model, using 2DCNN to extract spectral features and 3DCNN to extract spatial features, which learns more abstract feature information, and fusing the two types of feature information for HSI classification. This model almost fully utilizes the spectral information and spatial information in the image. However, the CNN model contains a large number of parameters and requires a large number of training samples and training time. Actually, there are fewer training samples in the hyperspectral image, which is very easy to cause overfitting, resulting in poor generalization of the classification model^{18}. Therefore, Ghasrodashti et al.^{19} proposed a sparse classification of hyperspectral images based on extended hidden Markov random fields. The objective function of the sparse classifier was improved by constructing a dictionary containing the minimum spatialspectral correlation and sparse coding, then satisfactory results were obtained. In order to improve the computational efficiency of the model, Xu et al.^{10} proposed a new RPNet. It uses random projection to determine that the convolution kernel is different from traditional CNN and does not require training. It has excellent feature extraction capabilities to avoid overfitting problems. And the computational burden of the network is relatively low. It can be seen that this work is flawed. The spatial information in the hyperspectral image is ignored or not used, while the spatial information is the best value information that can be obtained from the hyperspectral image.
With the development of image processing technology, spatial features play an increasingly critical role in HSI classification^{20}. Among the existing spatial information extraction technologies, Gabor filters have attracted much attention because of their ability to provide distinctive feature information. Gabor filter is an effective unsupervised feature information extraction method. It has a stronger ability to describe the texture and spatial features, can extract HSI texture and spatial structure information, and reduce the model's limited dependence on training samples^{21}. Many researchers have also proved that Gabor filters can achieve better results when applied to HSI classification. For example, Feng Xiao et al.^{22} combined a threedimensional Gabor filter with a support vector machine (SVM) for hyperspectral image classification, which can effectively improve the classification accuracy and efficiency; Wang Liguo et al.^{23} based on the empirical mode decomposition of the spectral data, the Gabor filtering operation is performed to better mine the texture features of the image; Chen et al.^{24} extracted twodimensional Gabor features from the hyperspectral data after PCA dimensionality reduction, and input them into 2DCNN for classification, which improved the classification accuracy and reduced the model’s dependence on training samples and excessive smoothing.
In the field of image processing, it is difficult for a single feature and a single model to achieve the desired performance. An important and effective method is integration^{25}. Therefore, in order to solve the abovementioned problems in the process of hyperspectral image classification, it is inspired by the early fusion of literature^{26,27}. This paper proposes a joint spectralspatial feature method GRPC for HSI classification. This method combines all Gabor spatial features and multiscale convolution features, and has a high degree of discrimination for HSI classification. In order to avoid the "dimensionality catastrophe" problem^{28}, GRPC uses PCA and Linear Discriminant Analysis (LDA) to preprocess the HSI dimensionality reduction, and project the HSI data into a lowdimensional feature space; the Gabor filter is used to extract the texture and internal spatial structure information of the image after dimensionality reduction. At the same time, the Gabor features extracted by the model are used as input, and the deep spectral features of the image are extracted using random patches convolution features. For the extracted spectral features, in order to increase the feature sparsity, a modified linear unit activation function is used for feature activation; finally, the feature stacking method is used to stack spatial features and multilevel convolution features for classification. GRPC not only uses the feature extraction capability of random patches convolution and the advantages of Gabor filters, but also realizes the multilayer fusion of feature maps, making the network have multiscale advantages. What needs to be explained here is that the Gabor filter based on frequency domain and direction is similar to the filter in the human visual system. Random patches convolution is similar to CNN and is a network inspired by neuroscience to a large extent. Therefore, the inherent relationship between random patches convolution and Gabor makes the combination of the two technologies have potential advantages over other methods.
Methodology
HSI preprocessing
PCA is a common HSI dimensionality reduction method, but often fails to show better category differentiation when dealing with highdimensional data. The Linear Discriminant Analysis (LDA) algorithm can increase the interclass to intraclass distance ratio, but the algorithm requires the intraclass distance matrix to be nonsingular, and the data after PCA dimensionality reduction just meets this requirement^{29}. Therefore, in this paper, we first apply the PCA algorithm to reduce the dimensionality of the original data and eliminate the redundancy of the data, letting the hyperspectral original input data be \({\text{X}} \in R^{{\left( {rc \times n} \right)}}\), r, c and n, which are the row number, column number and spectral band number respectively, and use the following way to represent the reduced data \(X_{P} \in R^{rc \times P}\); then we apply the LDA algorithm to reduce the dimensionality of the data processed by the PCA algorithm using the second projection of \(X_{LDA} \in R^{rc \times P}\). This process reduces extra operations and ensures that the intraclass matrix is not singular. The data processed by the twodimensional reduction process retains the main information and increases the interclass to intraclass distance ratio, which improves the sample differentiation^{30}.
Gabor filter
The twodimensional Gabor filter can be used for feature extraction of images, which can effectively capture the asymmetric relationship between the spatial and frequency domains of images and extract intrinsic information such as spatial features, which is an important factor in discriminating different features of HIS^{31}. The twodimensional Gabor filter is composed of an imaginary part and a real part, which is a sinusoidal curve function modulated by a Gaussian envelope, defined as follows.
where \({\text{x}^{\prime}} = {\text{xcos}}\uptheta + {\text{ysin}}\uptheta\) and \({\text{y}^{\prime}} =  {\text{sin}}\uptheta + {\text{ycos}}\uptheta\); \({\uplambda }\) is the wavelength, the wavelength value is specified in pixels, usually greater than or equal to 2, but not greater than onefifth of the input image size; \({\uptheta }\) is the angle of the Gabor kernel direction, this parameter specifies the direction of the Gabor function parallel stripes, taking a value range of \(\left[ {0,{\uppi }} \right]\); \({\uppsi }\) is the phase offset, the default is \({\uppi }/2\); \({\upgamma }\) is the spatial aspect ratio, which determines the ellipticity of the Gabor function shape, when \({\upgamma } = 1\) is, the Gabor kernel shape is circular, when \({\upgamma } < 1\), the shape elongates with the stripe direction, and the default value is 0.5; \({\upsigma }\) is the standard deviation of the Gaussian envelope, determined by the wavelength \({\uplambda }\) and the spatial frequency bandwidth \({\text{b}}_{{\text{w}}}\), defined as follows.
where the value of the bandwidth \({\text{b}}_{{\text{w}}}\) is a positive real number, the smaller the bandwidth, the larger the standard deviation, and the larger the shape of the Gabor kernel; when its value is taken in \(\left[ {1,5} \right]\), it does not affect the results of the algorithm in this paper much, and its default value is 1.
Random patches convolution
Unlike traditional convolutional feature extraction, the convolutional kernel of the random patches convolution part is determined using random projection, the basic principle of which is derived from the JL Lemma and is an effective method for dimensionality reduction^{32}. Arriaga et al.^{33} achieved classification by projecting data into a random lowdimensional space. It is proved that the random projection can well preserve the boundary between different categories, and only a small number of samples are needed to train the classifier in the lowdimensional space. This feature is conducive to image classification tasks, especially in the case of very limited training samples. The random patches convolution part used in this paper mainly consists of data PCA whitening, random projection and convolutional feature extraction, as described below.
PCA data whitening: PCA whitening refers to the standardization of the features of each dimension after PCA dimensionality reduction. In this way, the variances of different wavebands are similar, and the correlation between different wavebands is reduced, which is conducive to the completion of the image classification task.
Random projection and convolutional feature extraction: The random patches used in this paper are randomly selected from the PCAwhitened data \({\text{X}}_{{{\text{whiten}}}}\) using a random function with k random pixels \({\text{X}}_{{{\text{whiten}}}} \in {\text{R}}^{{{\text{rc}} \times {\text{P}}}}\), and around each pixel, a \({\text{w}} \times {\text{w}} \times {\text{P}}\) patches is taken to obtain k random patches \({\text{P}}_{1} ,{\text{P}}_{2} , \ldots ,{\text{P}}_{{\text{k}}} \in {\text{R}}^{{{\text{w}} \times {\text{w}} \times {\text{P}}}}\). For those pixels distributed at the edge of the image, a mirror image is used to fill in the neighboring blank pixels. Then, all k random patches are used as convolution kernels to convolve the whitened data with the random patches to obtain k feature maps
\({\text{I}}_{{\text{i}}} \in {\text{R}}^{{\left( {{\text{r}} \times {\text{c}}} \right)}}\) is the idimensional feature map, \({\text{X}}_{{{\text{whiten}}}}^{{\text{j}}} \in {\text{R}}^{{{\text{r}} \times {\text{c}}}}\) is the jth dimension of the downscaled data, and \({\text{P}}_{{\text{i}}}^{{\left( {\text{j}} \right)}} \in {\text{R}}^{{{\text{w}} \times {\text{w}}}}\) is the jth dimension of the ith random patches. The step size of the convolution operation is set to 1. For blank pixels at the edges, they are filled by mirroring the image.
Hyperspectral images classification method based on GRPC
In order to make full use of the effectiveness of HSI spectral and spatial features and improve the classification accuracy of HSI, this paper combines a twodimensional Gabor filter with random patches convolution feature extraction, and proposes a new joint spatial and spectral feature information HSI classification method GRPC. The steps of this method are shown in Fig. 1, which mainly includes four steps: First, use PCA and LDA algorithms to reduce the dimensionality of the input original HSI, and retain the main information of the HSI, eliminating the redundancy of the data, and increase the interclass and the distance ratio within the class; secondly, the twodimensional Gabor filter is used to extract the spatial structure feature information of the image after the dimensionality is reduced; again, the extracted Gabor features are used as input, and the multilevel spectral feature information of the image is extracted using random patches convolution features; finally, connect spatial and spectral features to perform feature stacking, then use SVM classifier to predict category labels, and output classification accuracy.
Among them, the random patches convolution feature extraction part contains multiple layers, and each layer contains a PCA, whitening, random projection and convolution feature extraction. The method of this paper uses a modified linear unit activation for the convolved feature map. The function performs feature activation, the activation function can be expressed as \(f\left( I \right) = \max \left( {0,I  S} \right)\), \({\text{S}} = \left[ {S_{2} \cdots S_{2} } \right]\) is an average matrix composed of k \(S_{2}\), and \(S_{2}\) is the mean vector of I in the second dimension, K is the number of feature maps, if I is the convolution feature of the first layer, then \(Z_{1} = f\left( I \right)\). Feature stacking can better meet the precise classification requirements of different targets at different scales by combining spectral and spatial features in classification^{34}. The feature stacking of GRPC is to transform the acquired spectral and spatial feature information and the original spectral data into the same dimension for splicing. If Z is regarded as the final feature vector used in classification, the final feature vector in each pixel \(\left( {{\text{i}},{\text{j}}} \right)\) is as follows, \({\text{M}}_{{\left( {{\text{i}},{\text{j}}} \right)}} = \left\{ {{\text{G}}_{1} , \ldots ,{\text{G}}_{{\text{R}}} ,{\text{Z}}_{1} , \ldots ,{\text{Z}}_{{\text{L}}} ,{\text{X}}} \right\}\); \({\text{G}}_{{\text{R}}}\) are different Gabor features, \({\text{Z}}_{{\text{L}}}\) is the convolution feature of different layers, and L is the number of network layers extracted by the convolution feature. It is worth noting that these features still need to be standardized, and the standardized formula is shown in formula (4).
\({\text{M}}_{{\left( {{\text{i}},{\text{j}}} \right)}}^{{{\text{norm}}}}\), \({\text{var}}\left( {{\text{M}}_{{\left( {{\text{i}},{\text{j}}} \right)}} } \right)\) and \({\text{mean}}\left( {{\text{M}}_{{\left( {{\text{i}},{\text{j}}} \right)}} } \right)\) are the standardized, \({\text{M}}_{{\left( {{\text{i}},{\text{j}}} \right)}}\) variance and mean, respectively. Finally, the standardized final features are used to classify the HSI using the SVM classifier. The SVM classifier has the advantages of simple calculation and good classification performance. The implementation details of the GRPC method are shown in Algorithm 1.
Experimental and analysis
Data and experimental environment description
In order to verify the effectiveness of the method proposed in this paper, experiments were conducted on the three most widely used public data sets of Indian Pines^{35}, Pavia University^{35} and Kennedy Space Center^{35}. The three data sets are all images with the absorption band removed from water mist. The Indian Pines image was obtained from an AVIRIS spectrometer in a forest area in northwestern Indiana, USA. The image size is 145 × 145 pixels, the wavelength is 0.40–2.50 μm, and the number of bands is 200, the spatial resolution is 20 m; the Pavia University image is obtained by ROSIS spectrometer at the University of Pavia, Italy. The image size is 610 × 340 pixels, the wavelength is 0.43–0.86 μm, the number of bands is 103, and the spatial resolution is 1.3 m; the Kennedy Space Center image was obtained by an AVIRIS spectrometer near the Kennedy Space Center in Florida. The image size is 512 × 614 pixels, the wavelength is 0.40–2.50 μm, the number of bands is 176, and the spatial resolution is 18.0 m. The three data sets are divided into 16 categories, 9 categories and 13 categories. Their falsecolor images and groundtruth images are shown in Figs. 2, 3, and 4, respectively. The numbers of training samples and test samples are shown in Tables 1, 2 and 3.
Regarding the Falsecolor map and Groundtruth map in Figs. 2, 3, 4, 8, 9 and 10, we used Python to draw them ourselves. The software version is Python 3.7.6^{36}. The website link is https://www.python.org/downloads/release/python376/.
The whole experiment process was carried out on a local PC. The operating system was Windows10 Enterprise Edition, and the hardware configuration was Intel Core i54570S @ 2.90 GHz, NVIDIA GeForce GT 750 M and 8 GB RAM. Code will be available at https://github.com/ShenmingQu/GRPC.
Parameter analysis
Gabor Filter: In general, a Gabor filter is considered to be a multidirectional and multiscale filter^{20}. In this paper, in order to reduce the computational complexity of the network while maintaining a robust feature representation capability, a multidirectional and singlescale Gabor filter is used. By testing a combination of three center frequencies (0.1, 0.2 and 0.4 Hz) and two sets of filter directions (four directions θ = 0, π/4, π/2, 3π/4 and eight directions θ = 0, π/8, π/4, 3π/8, π/2, 5π/8, 3π/4, 7π/8), this article uses a filter bank of 0.2 Hz and four directions. In addition, the influence of Gabor window size on classification accuracy is evaluated. As shown in Fig. 5, with the window size changes, the classification accuracy of Indian Pines and Pavia University increases first and then stabilizes. The classification accuracy of Kennedy Space Center has increased after it has decreased. The reason is that if the window is too small, it will highlight the details in the image, but it is susceptible to noise interference. If the window is too large, it is not conducive to express the local characteristics of the image and lose the details of the image. Finally, the window size is set to 3 × 3.
Number of PCA Principal Components and Number of Network Layers: This paper analyzes the influence of the principal component parameters and the number of network layers L on the classification accuracy through experiments. As shown in (a) of Fig. 6, with the gradual increase of the parameter P, the OA of this method in Kennedy Space Center gradually stabilizes, and the OA of the two data sets of Indian Pines and Pavia University has a significant downward trend. As the value of P increases, the experimental time of the three data sets also becomes significantly longer. Considering the balance between accuracy and time consumption, this article sets P to 3. As shown in Fig. 6b, the classification accuracy increases as the number of network layers increases. When the number of network layers is less than 6, the classification accuracy gradually increases and tends to be stable. When the number of network layers is greater than 6, Pavia University and Kennedy Space Center drop significantly. This shows that although too deep network layers can extract more abstract feature information, it will also cause loss of information, which will affect the classification accuracy. Finally, the parameter L is set to 6.
Number and Size of Random Patches: The number k and size w of random patches are two important parameters of GRPC, and the size of these two parameters will affect the classification accuracy. In this paper, we experimentally evaluate the impact of different k and w on the method in this paper. As shown in Fig. 7, a smaller k cannot achieve higher classification accuracy, and in this experiment, the parameter k is set to 23. As for the parameter w, in general, a larger w can help to improve the classification accuracy, but too large a w can increase the phenomenon of transition smoothing, and in this experiment, the parameter w is set to 24.
Evaluation metrics
In order to better evaluate the performance of this method in the experiment, two indicators, Overall Accuracy (OA) and Kappa Coefficient (Kappa) are used to evaluate the classification performance of the model.
OA is equal to the sum of correctly classified pixels divided by the total number of pixels. The number of correctly classified pixels is distributed along the diagonal of the confusion matrix, and the total number of pixels is equal to the total number of pixels of all real reference sources. The formula is as follows, where h_{ii} is the number of correctly classified pixels distributed along the diagonal of the confusion matrix, N is the total number of samples, and n is the number of categories.
Kappa coefficient is an indicator of consistency test. It is done by multiplying the total number (N) of all real reference pixels by the sum of the diagonal (h_{kk}) of the confusion matrix, and then subtracting the number of real reference pixels in each category and the classified pixels (h_{ik}, h_{kj}) after the product of the total number, divided by the square of the total number of pixels minus the product of the total number of true reference pixels in each category and the total number of classified pixels in the category, and then sum the result. The kappa coefficient comprehensively considers the various factors in the confusion matrix and can more comprehensively reflect the accuracy of the overall classification. The larger the value of the Kappa coefficient, the higher the accuracy of the corresponding classification algorithm. The general formula is as follows.
Classification results
In order to verify the advantages of this method in the classification of hyperspectral remote sensing images, this method is compared with the current mainstream PCA (using RBFSVM to classify the first 20 principal components), EMP^{37}, SAE^{38}, 3DCNN^{39}, GaborBased^{21} and RPNet^{10} and other methods are compared. Among them, PCA and EMP are two classic and representative machine learning methods; SAE is a classic machine learning method; 3DCNN achieves HSI classification only by extracting local spectralspatial information of pixels; GaborBased has the same data input as the method in this paper; RPNet uses spectral feature information for classification, and has the same model structure as the convolution feature extraction module in the method in this paper, but the difference lies in the input data. The number of training samples and the number of test samples for each method remain the same. The number of detailed training samples and test samples is shown in Tables 1, 2 and 3. The classification accuracy of different methods on the three data sets is shown in Tables 4, 5 and 6. The bold font in the table represents the highest value in the same industry. The following comparative analysis can draw conclusions.
The OA and Kappa coefficients of the method in this paper are significantly higher than those of the compared methods. On the Indian Pines dataset, the method achieved an OA of 98.09% and a Kappa coefficient of 0.9778, an improvement of 2% and 0.0232 compared to RPNet, and 2.31% and 0.0267 compared to GaborBased; for the Pavia University dataset, the OA reached 99.64% and the Kappa coefficient reached 0.9951, an improvement of 0.30% and 0.0041 compared to RPNet, and 0.94% and 0.0129 compared to GaborBased; for the Kennedy Space Center dataset, OA reached 96.53% and the Kappa coefficient reached 0.9612, an improvement of 2.38% and 0.38% compared to RPNet improved by 2.38% and 0.0263, compared to GaborBased by 3.86% and 0.0449, etc.
The Gabor filter and random patches convolution included in the model in this paper have good feature extraction capabilities for HSI. Moreover, the method in this paper achieves higher classification accuracy than RPNet and GaborBased, and also demonstrates the advantage of using the combination of Gabor filters and RPNet to extract spatial information including edges and textures.
This method is also able to improve the classification accuracy of a variety of features, such as Alfalfa, Grasspasturemowed and Soybeanmintill in the Indian Pines dataset; Asphalt, Gravel and Bitumen in the Pavia University dataset; and Scrub, CP/Oak and Salt marsh in the Kennedy Space Center dataset. Features such as Scrub, CP/Oak and Salt marsh in the Kennedy Space Center dataset. It also demonstrates the advantage of using multilevel features to bring different feature information at different scales to the classification accuracy.
Figures 8, 9 and 10 show the ground truth sample plots of the three data sets and the classification result plots obtained by comparing the methods and the method of this paper respectively. From the classification result plots in Figs. 8, 9 and 10, it can be seen that the method in this paper does not have large blocks of noise points in the classification result plots, and the classification results of all types of features have high classification accuracy, which is consistent with the classification accuracy results in Tables 4, 5 and 6.
Due to the high cost of acquiring hyperspectral datasets, the number of samples available is very limited. In order to further investigate the robustness of the algorithm of this paper's method under different numbers of training samples, another 10%, 20%, 30%, 40% and 50% of the training samples of each type of feature were selected as training samples on three datasets of Indian Pines, Pavia University and Kennedy Space Center, respectively, and compared with 3DCNN, GaborBased and RPNet methods for comparison experiments. Ten experiments were conducted for each method with different training sample ratios, and the average of OA accuracy was taken after the experiments for comparison. Figure 11 shows the accuracy rates under different training samples. It can be seen from Fig. 11 that the OA accuracy rises gradually as the number of training samples increases, and the method in this paper achieves higher classification accuracy with different numbers of training samples and has obvious advantages. It can be seen that this method has better robustness under the limited number of samples.
Discussion
In Figs. 7, 8, 9, 10, 11 and Tables 4, 5, 6, the comparison of GRPC in the three data sets with several other comparison methods shows that the GRPC method can achieve better visual effects and higher accuracy. This proves the effectiveness of the spectralspatial feature extraction mode in the GRPC method. There are three reasons for this. First, the twodimensional Gabor filter is used to extract the spectralspatial feature information of the image, so as to obtain the local spatial correlation information of all regions of the entire image. Second, random patches convolution can extract shallow and deep features, allowing the combination of multiscale and multilayer features. Third, random patches convolution and Gabor localized spatial features in GRPC have great potential for overcoming the pepper noise and oversmoothing phenomenon in HSI processing, and a good classification effect is still achieved with limited training samples. Through the above classification results and quantitative evaluation, the method in this paper can become a novel and effective spectralspatial classification framework.
In order to verify the superiority of GRPC spatialspectral joint feature classification and the influence of LDA during data preprocessing. Three different methods were also designed as variants of GRPC in the experiment. One is that NGRPC that does not use Gabor features in the feature stacking part, and the parameter settings are the same as GRPC. Second, OGRPC, which only stacks Gabor features and random patches convolution deep features, is different from GRPC in that only the last layer of features of random patches convolution is used in the feature stacking part. Third, in the data preprocessing part, only PCA dimensionality reduction is used to preprocess the input image PGRPC. Table 7 shows the classification accuracy of the four methods. It can be seen from Table 7. The two simple feature fusion methods, NGRPC and OGRPC, did not achieve better classification accuracy. The reason is that NGRPC, which uses only spectral features, ignores the importance of spatial structure information; OGRPC, which uses only the last layer of convolution features and spatial features, ignores the more powerful spectral features of the shallow layer. PGRPC has not achieved better classification accuracy. The reason is that PCA often fails to show good class discrimination when processing highdimensional data.
The time cost of the algorithm is one of the important factors affecting its application in remote sensing. This section evaluates the time consumption of the proposed GRPC and existing HSI classification methods such as SAE, 3DCNN, GaborBased and RPNet. As shown in Table 8, the methods based on SAE and 3DCNN are the slowest, because these two methods need to be pretrained first, and then finetuned to achieve the convergence strategy; compared with SAE and 3DCNN, GaborBased runs relatively faster without pretraining operations, but GaborBased reduces its efficiency when extracting each highdimensional pixel feature; For GRPC and RPNet, the RPNet method is obviously faster, because GRPC is affected by the extraction of spatial features and the fusion of spectral and spatial features. This process occupies one third of the total running time of the method. However, GRPC classification accuracy is higher, and the overall time consumption is acceptable. All the experimental environments here remain unchanged. For the computing environment and parameters used in this section, please refer to the beginning of this section.
In summary, the comparison of experiments and algorithms on the three data sets shows that the GRPC method can extract highly distinguished features by combining multiscale and multilayer convolution information, and fuse feature information. For HSI classification, GRPC is a competitive and robust method. Specifically, the experiments in this article show that both random patches convolution and twodimensional Gabor filters are reliable technologies, and GRPC is still robust even with limited training samples.
Conclusions
Aiming at the problem of low HSI training samples and low classification accuracy, this paper proposes a HSI classification method GRPC that combines spatial feature information and spectral feature information. This method first performs PCA and LDA dimensionality reduction processing on HSI, which retains useful information and increases computational efficiency. Secondly, in order to effectively use the spectral and spatial information, Gabor filter and random patches convolution are used to extract the spatial and spectral features of HSI respectively. Finally, the feature stacking method is used to fuse spatial spectral information, and SVM is used to classify the image. Through experiments, several important parameters that affect the accuracy of the network are discussed separately, and through comparative experiments on the three data sets of Indian Pine, Pavia University and Kennedy Space Center, the superiority of the network is verified. Compared with the method, the OA and Kappa coefficients are improved to different degrees, and this network still has better classification accuracy even with limited training samples. In the subsequent research work, it is planned to further optimize the operating efficiency of the algorithm while ensuring the classification accuracy.
Data availability
The three datasets of Indian Pines, Pavia University, and Kennedy Space Center used in this study are publicly available at http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes.
References
Bruzzone, L., Chi, M. & Marconcini, M. Hyperspectral Data Exploitation: Theory and Applications[J]. Photogramm. Eng. Remote Sens. 81(6), 441–442 (2007).
Chen, W., Zhang, L. & Bo, D. Kernel slow feature analysis for scene change detection. IEEE Trans. Geosci. Remote Sens. 55(4), 1–18 (2017).
Zhou, C., Tu, B., Ren, Q. & Chen, S. Spatial peakaware collaborative representation for hyperspectral imagery classification. IEEE Geosci. Remote Sens. Lett. 19, 1–5. https://doi.org/10.1109/LGRS.2021.3083416 (2021).
Kordi Ghasrodashti, E., Helfroush, M. S. & Danyali, H. Spectralspatial classification of hyperspectral images using wavelet transform and hidden Markov random fields. Geocarto Int. 33, 771–790 (2017).
Fang, L. et al. Classification of hyperspectral images by exploiting spectralspatial information of superpixel via multiple kernels. IEEE Trans. Geosci. Remote Sens. 53(12), 6663–6674 (2015).
Huang, H. et al. Dimensionality reduction of hyperspectral imagery based on spatialspectral manifold learning. IEEE Trans. Cybern. 50, 2604–2616 (2019).
Zhao, X. et al. Spectralspatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 8(6), 2381–2392 (2015).
Tao, C. et al. Unsupervised spectralspatial feature learning with stacked sparse autoencoder for hyperspectral imagery classification. IEEE Geosci. Remote Sens. Lett. 12(12), 2438–2442 (2015).
Chen, Y. et al. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 54(10), 6232–6251 (2016).
Xu, Y. et al. Hyperspectral image classification via a random patches network. ISPRS J. Photogramm. Remote Sens. 142(AUG.), 344–357 (2018).
Jiang, J., Ma, J. & Liu, X. Multilayer spectralspatial graphs for label noisy robust hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. https://doi.org/10.1109/TNNLS.2020.3029523 (2020).
Tu, B., Zhou, C., Liao, X., Zhang, G. & Peng, Y. Spectral–spatial hyperspectral classification via structuralkernel collaborative representation. IEEE Geosci. Remote Sens. Lett. 18(5), 861–865. https://doi.org/10.1109/LGRS.2020.2988124 (2021).
Yin, J. et al. Spatialspectral network for hyperspectral image classification: A 3D CNN and BiLSTM framework. Remote Sens. 13(12), 2353 (2021).
Hu, W. et al. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 1–12 (2015).
Yue, J. et al. Spectral–spatial classification of hyperspectral images using deep convolutional neural networks. Remote Sens. Lett. 6(4–6), 468–477 (2015).
Ghasrodashti, E. K. & Sharma, N. Hyperspectral image classification using an extended autoencoder method. Signal Process. Image Commun. 92(3), 116111 (2020).
Roy, S. K. et al. HybridSN: Exploring 3D–2D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 17(2), 277–281 (2020).
Liu, B. et al. A semisupervised convolutional neural network for hyperspectral image classification. Remote Sens. Lett. 8(9), 839–848 (2017).
Kordi Ghasrodashti, E., Helfroush, M. S. & Danyali, H. Sparsebased classification of hyperspectral images using extended hidden Markov random fields. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 11, 4101–4112 (2018).
Huang, K. K. et al. Hyperspectral image classification via discriminant Gabor ensemble filter. IEEE Trans. Cybern. https://doi.org/10.1109/TCYB.2021.3051141 (2021).
He, L. et al. Discriminative lowrank Gabor filtering for spectralspatial hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 55(3), 1381–1395 (2017).
Feng, X. et al. Hyperspectral image classification based on 3D Gabor filter and support vector machines. Spectrosc. Spectr. Anal. 34(8), 2218 (2014).
Wang, L. et al. Hyperspectral image classification by combining empirical mode decomposition with Gabor filtering. J. Harbin Eng. Univ. 37(2), 284–290 (2016) (in Chinese).
Chen, Y. et al. Hyperspectral images classification with Gabor filtering and convolutional neural network. IEEE Geosci. Remote Sens. Lett. 14(12), 2355–2359 (2017).
Sun, Y., Fu, Z. & Fan, L. A novel hyperspectral image classification pattern using random patches convolution and local covariance. Remote Sens. 11(16), 1954 (2019).
Zhang, X. et al. Spectralspatial feature learning using clusterbased group sparse coding for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 9(9), 4142–4159 (2017).
Zhang, X. et al. Multifeature hyperspectral image classification with local and nonlocal spatial information via Markov random field in semantic space. IEEE Trans. Geosci. Remote Sens. 56, 1409–1424 (2017).
Du, P. et al. Multikernel adaptive collaborative representation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 56, 4664–4677 (2018).
Liu, J., Chen, S. & Tan, X. A study on three linear discriminant analysis based methods in small sample size problem. Pattern Recogn. 41(1), 102–116 (2008).
Park, B. C. et al. Terahertz single conductance quantum and topological phase transitions in topological insulator Bi_{2}Se_{3} ultrathin films. Nat. Commun. 6, 6552 (2015).
Li, H. C. et al. Gabor featurebased composite kernel method for hyperspectral image classification. Electron. Lett. 54(10), 628–630 (2018).
Johnson, W. B. & Lindenstrauss, J. Extensions of Lipschitz mappings into a Hilbert space. Proc. Conf. Modern Anal. Probab. 189–206 (1984).
Arriaga, R. I. & Vempala, S. An algorithmic theory of learning: Robust concepts and random projection. Mach. Learn. 63(2), 161–182 (2006).
Chen, Y. et al. Spectralspatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 8(6), 2381–2392 (2015).
Ehu.eus. Hyperspectral Remote Sensing Scenes (2017). [online] Available at: http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes. Accessed 10 Mar 2017.
Python.org. Python Release Python 3.7.6 [DB/OL]. [12182019]. https://www.python.org/downloads/release/python376/.
Benediktsson, J. A., Palmason, J. A. & Sveinsson, J. R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 43(3), 480–491 (2005).
Chen, Y. et al. Deep learningbased classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 7(6), 2094–2107 (2014).
He, M., Bo, L., Chen, H. Multiscale 3D deep convolutional neural network for hyperspectral image classification. In 2017 IEEE International Conference on Image Processing (ICIP) (IEEE, 2018).
Acknowledgements
This work was funded by the Science and Technology Development Program of Henan Province, China, the National College Student Innovation and Entrepreneurship Training Program, and the Graduate Education Innovation and Quality Improvement Program of Henan University. The project numbers are: 212102210538, 222102210101, 202110475135 and SYL20040121. The authors would like to thank Professor Junjun Jiang for his advice and guidance on the article.
Author information
Authors and Affiliations
Contributions
Supervision, Project administration, Investigation, Methodology, Resources, S.Q.; Software, Validation, Data Curation, Writing—Original Draft, Visualization, X.L.; Conceptualization, Formal analysis, Writing—Review and Editing, Funding acquisition, Z.G.; All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Shenming, Q., Xiang, L. & Zhihua, G. A new hyperspectral image classification method based on spatialspectral features. Sci Rep 12, 1541 (2022). https://doi.org/10.1038/s41598022054225
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598022054225
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.