|
1.IntroductionOver the past decade, hyperspectral imaging is of widespread interest among the remote sensing research community due to its ability to discriminate between the variety of ground objects.1 The hyperspectral data consist of rich information in both spectral and spatial domains, which has opened opportunities in numerous diverse field applications, such as land cover classification,2 target detection,3 tree species classification,4 food technology,5 and medical imaging.6 The hyperspectral data present a very difficult challenge caused by a large number of narrow spectral bands with a small number of available labeled training samples. This problem along with other difficulties, such as high variations of the spectral signature from identical material, high similarities of the spectral signatures between some different materials, and noise from the sensors and environment, will significantly decrease the classification accuracy. Therefore, feature extraction is an essential task in hyperspectral image processing to explore the hidden discriminant features of hyperspectral data that are useful for the classification and in turn increases the classification accuracy. Researchers have proposed various feature extraction techniques in the past few years for the extraction of features from the hyperspectral images. Feature extraction is the transformation of the original feature space into a new set of coordinates or features.7 The feature extraction process preserves the most informative contents of the original high-dimensional feature space. Principle component analysis (PCA) is one of the most commonly used feature extraction techniques.8,9 This is because the PCA is an invertible transformation, which makes easy to interpret the extracted features. PCA finds the projections with lower reconstruction error for the whole data. It works on the global features and ignores the local information. Hence, segmented PCA (SPCA) is proposed as an extended version of PCA where PCA is applied to the blocks that are composed of correlation between bands to use local information.10 As an extension to PCA, modified algorithms are proposed, such as maximum noise fraction (MNF)11 and kernel PCA (KPCA).12 Probabilistic PCA (PPCA) is a generative latent variable model in connection with maximum likelihood function, which is also used to extract features.13 Independent component analysis14 is then proposed to extract class discriminant features. Another best-known feature extraction approach is linear discriminant analysis (LDA).15 LDA finds the projections that preserve the most discriminative information. Many other extensions to the above-mentioned two approaches have been developed, such as regularized LDA,15 nonparametric weighted feature extraction (NWFE),16 and kernel NWFE.17 The PCA- and LDA-based methods assume that the distribution of the samples in a class is Gaussian; however, sample distribution is not always Gaussian, and sometimes it may have complex multimodal structure. Therefore, locality preserving feature extraction methods have emerged, which includes local Fisher’s discriminant analysis (LFDA) and locality-preserving projection (LPP). In Refs. 18 and 19, random feature selection (RFS)-based methods are developed to explore the diverse feature set that leads to the higher classification performance. Clustering-based feature extraction techniques are also widely used for feature extraction, which removes the redundancies and the correlated features;20,21 however, most of the clustering methods focus on only spectral features rather than exploring the hidden discriminant features. The above-mentioned approaches are nearly matrix-based approaches or vector-based approaches. However, the original hyperspectral data are represented in a three-dimensional (3-D) volumetric array, which includes two spatial dimensions and one spectral dimension. Therefore, it is more obvious to represent the hyperspectral data as a 3-D cube or tensor22 to preserve the higher-order statistical structure. The transform-domain method, 3-D discrete wavelet transform (DWT), is used to extract the texture feature at different scales and frequencies and has achieved the significant classification performance.23,24 Recently, deep learning techniques have emerged too as the most powerful methods for feature extraction of the hyperspectral data. Deep learning techniques that are extensively used for feature extraction includes deep belief network,25 stacked autoencoder,26 convolutional neural network (CNN),27 and recurrent neural network.28 A 3-D convolutional neural network (3-D CNN) framework is proposed in Ref. 29 to extract the deep spectral–spatial features. It is observed that the tensor-based or 3-D methods provide the significant performance as the joint spectral–spatial structure is adequately preserved. Although the deep learning methods provide the significant deep feature representation of the high-dimensional data that can improve the classification performance of the system, it also increases the computation time and the complexity of the algorithm. From the study of various existing feature extraction techniques, we found the following challenges such as:
In this work, three-dimensional discrete cosine transform (3-D DCT) for classification of hyperspectral images is proposed. DCT exhibits excellent energy compaction properties, and large DCT coefficients are located in the low-frequency region. Therefore, DCT is chosen for feature extraction in hyperspectral image classification. DCT extracts highly discriminative and informative features from the hyperspectral images. The proposed method transforms the hyperspectral image into a DCT coefficient matrix and looks for a signature pattern in the DCT domain for classifying different land cover classes. Further, support vector machine (SVM) classifier is used to obtain labels of unknown samples of the hyperspectral images. To the best our knowledge, this is the first time where DCT is used for feature extraction in hyperspectral image classification. This technique has shown very distinct features that are more suitable for hyperspectral classification, including high classification accuracy and computational efficiency. The main contribution of this paper can be summarized as follows:
The rest of the paper is arranged as follows: overview of feature extraction using 3-D DCT is given in Sec. 2. Section 3 deals with experimentation on standard benchmark datasets and discusses the findings of the experiments, and finally, Sec. 4 presents the conclusion and future directions. 2.Proposed Three-Dimensional Discrete Cosine Transform-Based Feature Extraction FrameworkIn this section, the proposed feature extraction framework for hyperspectral image classification is explained in detail. As shown in Fig. 1, the proposed approach consists of two stages such as feature extraction and classification. The following subsection deals with the detailed explanation of the various stages present in the proposed system. Fig. 1DownloadConsider the hyperspectral image dataset that is represented as, X∈RH×W×N, where H and W be the height and width of hyperspectral image and N is a total number of spectral bands or the feature dimension. Assume the training samples of hyperspectral image data as x=[x1,x1,…,xM], y=[y1,y1,…,yM], are the labels of training samples that belong to the k classes in the data that are denoted as Ω=[Ω1,Ω2,…,Ωk]. 2.1.Principal Component AnalysisPrincipal component analysis (PCA) is widely used in the image preprocessing step to reduce the dimension and redundancy. PCA reduces the dimension of the image by choosing information only from the significant bands. It uses a vector space transform, which reduces the dimensionality of the original dataset and can be interpreted as a dataset with fewer variables, called principal components (PCs).32 Let us consider the hyperspectral image X∈RH×W×N, where H and W be the height and width of the hyperspectral image and N is the total number of spectral bands or the feature dimension, respectively. The pixel vector of a hyperspectral image is represented as follows: The mean μ of all image pixel vectors can be written as follows: where M=H×W denotes the total number of pixels in a spectral band.The covariance matrix can be given as follows: The eigenvalue decomposition of the covariance matrix is follows: where D is the diagonal matrix composed of eigenvalues λ1,λ2,…,λN of covariance matrix CM and B is the orthonormal matrix comprised of N eigenvectors and is given as follows:Then linear transformation can be calculated as follows: which is the PCA pixel vector and j is the number of principal components with the highest eigenvalues. Each pixel vector is mapped similarly, and the new low-dimensional image is obtained.2.2.Three-Dimensional Discrete Cosine TransformDiscrete cosine transform (DCT) is one of the most widely used techniques in numerous areas of image processing including the denoising and compression.33 Due to energy compaction property of DCT, the image information is represented using a few DCT coefficients. Thus, making DCT more suitable for image compression applications. As DCT is a linear and invertible transformation, it can provide easy separation of the transformation coefficients. The extracted independent transformation coefficients give a meaningful data structure that allows extracting information at a finer level of precision. The favorable outcome of such transformation is the removal of the interpixel redundancy as well as interband redundancy. In this paper, 3-D DCT is applied to the hyperspectral cube, which encodes the information in the form of DCT coefficients. It should also be noted that 3-D DCT can be achieved by applying two-dimensional (2-D) DCT to a pixel vector. The 2-D DCT f(m,n) of size M×N is given as follows: where 0≤u≤M−1, 0≤v≤N−1 and To extract the features from the low-dimensional hyperspectral image, 2-D DCT is applied to each pixel vector of low-dimensional image and coefficients of DCT are computed. One of the main characteristics of DCT is its ability to convert the energy of the image into a few coefficients. Thus, in the field of pattern recognition, DCT coefficients are widely used as features.The DCT coefficients of each pixel at position (m,n) of the low-dimensional hyperspectral image can be directly concatenated to form its feature vector: where E(·) is the expectation operator.Let ˆf∈RH×W×j be the final concatenated cube of the 3-D DCT-based feature vector is given as 2.3.Support Vector Machine ClassifierSVM has been widely used in the classification of hyperspectral image because of its particular advantages in solving problems about small-sized samples training, nonlinear, and high dimensions.34,35 In this paper, the SVM classifier is employed to get the final classification map. Assume the training samples of hyperspectral image data as x=[x1,x1,…,xM], y=[y1,y1,…,yM], are the labels of training samples that belong to the k classes in the data that are denoted as, Ω=[Ω1,Ω2,…,Ωk] and a nonlinear kernel mapping (·), xi is a pixel vector with j-dimensional spectrum. The SVM technique solves Constrained to where w is normal to the decision hyperplane, b represents the closest distance to the origin of the coordinate system, and l denotes the number of samples. Parameter C is the regularization parameter, which controls the generalization capabilities of the classifier, and ξi are the positive slack variables.Figure 1 shows a flowchart of the proposed technique, and the entire process is summarized in Algorithm 1. Algorithm 13-D DCT-based hyperspectral image classification.
3.ExperimentationIn this section, to assess the effectiveness of the proposed method, a series of experiments on three standard datasets were conducted, namely Indian pines, Pavia University, and Salinas dataset.36 All the experiments are conducted using MATLAB 2018a on PC with 16 GB RAM and 2.70 GHz CPU. To verify the efficacy of the proposed method, few traditional feature extraction methods were considered for comparison. Widely studied methods such as SVM,34 SVM-PCA,9 ICDA,37 and LDA38 were compared. The 3-D DWT, a transform-based feature extraction method, is also considered.23 For the SVM method, the original hyperspectral image is directly used for the classification without any feature extraction step. 3.1.Dataset Description
3.2.Performance MetricsThe performance of the proposed method is compared with other competing methods using three widely used quality metrics, i.e., overall accuracy, average accuracy, class-wise accuracy, and kappa coefficient. Overall accuracy (OA) is the percentage of correctly classified pixels in the whole scene. Average accuracy (AA) is the mean of percentage of correctly labeled pixels for each class. Classwise accuracy is also known as producer’s accuracy. Kappa coefficient is a robust measure of the degree of agreement, which integrates diagonal and off-diagonal entries of the confusion matrix. 3.3.Parameter SettingIn the beginning, to evaluate the effectiveness of the proposed method with less amount of labeled data, 20% samples for each class from the reference data of Indian pine dataset, Pavia University dataset, and Salinas dataset are randomly chosen as training samples, and the remaining samples in each class are used for testing purpose. This experiment is repeated for 10 times to evaluate an average of OA, AA, and κ. The training and testing samples used for conducting tests are shown in Table 1Table 2–3. Also, some parameters need to be tuned for the conduction of tests. For all the SVM-based methods, the penalty parameter C and the radial basis function (RBF) parameter γ are tuned through fivefold cross-validation (γ=2−3,2−2,…,22,23, C=21,22,…,28). Also, few other parameters of these methods need to be tuned. For the proposed technique, the RBF parameter γ and the penalty parameter C are tuned as the methods above. Table 1Details of Indian pines dataset including some classes, class name, training, testing, and the total number of samples.
Table 2Details of Pavia University dataset including number of classes, class name, training, testing, and the total number of samples.
Table 3Details of Salinas dataset including number of classes, class name, training, testing, and the total number of samples.
For SVM-PCA, 25 principal components (PCs) have obtained best classification accuracy. So, in this experiment, the number of PCs is set to 25.10 The number of independent components (ICs) are selected such that it could give a better result and have lesser computation burden. Also, it has been observed that a lesser or greater number of ICs may have redundant information. So as per Ref. 37, the number of ICs is set to 18. 3.4.Classification ResultsThis section discusses the classification results obtained for Indian pines dataset, Pavia University dataset, and Salinas dataset, the impact of the different proportions of training samples on overall accuracy and execution time taken by all competing method. First, we illustrate how DCT influenced the original spectra of the real hyperspectral datasets. Here, the original spectra of hyperspectral datasets, such as Indian Pines, Pavia University, and Salinas dataset, are shown in Figs. 2(a), 2(d), and 2(g), respectively. The transformed output spectra after PCA transformation on hyperspectral datasets, such as Indian Pines, Pavia University, and Salinas dataset, are shown in Figs. 2(b), 2(e), and 2(h), respectively. The transformed output spectra after applying DCT on hyperspectral datasets, such as Indian Pines, Pavia University, and Salinas dataset, are shown in Figs. 2(c), 2(f), and 2(i), respectively. The spectral curves shown in Figs. 2(b), 2(e), and 2(h) indicate the high correlation between various classes of hyperspectral images, which influences the discrimination among the classes. Figures 2(a), 2(d), and 2(g) indicates the original spectral curves of datasets. It shows slightly more separation between the land cover classes. However, these curves are obtained by considering all available spectral bands, which lead to heavy computations. However, in Figs. 2(c), 2(f), and 2(i), the spectral responses of land cover classes look more separated from each other in DCT domain, which directly influences the performance of the hyperspectral image classification. Fig. 2Download3.4.1.Result analysis by comparing the proposed method with different classification methods on Indian Pines datasetThe information required for experimentation, such as ground-truth data, training sample map, and testing sample map of Indian pines dataset is shown in Fig. 3. The classification map of all competing techniques on Indian pines dataset is shown Fig. 4, and the classification results (i.e., OA, classwise accuracy, AA, and κ) of all competing methods and the proposed method are shown in Table 4. Fig. 3DownloadFig. 4DownloadTable 4Comparison of classification accuracies (%) obtained by proposed method with competing methods for Indian pines dataset.
The performance of the proposed method is compared with traditional methods such as SVM, SVM-PCA, LDA, ICDA, and transform-based method such as 3-D DWT. From Table 4, it can be shown that the proposed technique attains greatest performance in terms of overall accuracy, average accuracy, classwise accuracy as well as the κ. For the PCA-based classification algorithm, the original image was reduced into few principal components that are then used for classification using SVM classifier. The PCA-based classification technique decreases the dimensionality of hyperspectral images in the spectral domain. However, it increases the discrepancy in the spatial domain (i.e., texture or shape variation). Therefore, the classification accuracies of the SVM-PCA-based method are not solely better for Indian pines dataset. By exploiting spectral–spatial features, 3-D DWT has achieved better performance in terms of OA, AA, and κ over all other competing methods, such as SVM, SVM-PCA, ICDA, and LDA. The proposed approach shows excellent and comparable classification performance due to the application of 3-D DCT features. The classification map of SVM-, SVM-PCA-, ICDA-, and LDA-based approaches have shown some salt and pepper noise that is less visible in the DWT method and proposed 3-D DCT method. This noise will disappear if the spatial information is considered for classification along with spectral information. When compared with other competing approaches, the proposed approach improves the classification accuracy significantly as shown in Table 4 (boldface). For instance, the classification accuracy of classes “Corn-no till” and “Corn-min till” increases from 46.50% to 75.83%, 22.29% to 70.63%, respectively. However, it is observed that the proposed method is not performing well in terms of the classwise accuracy of individual classes such as “Alfalfa,” “Grass-pasture-moved,” and “Oat” as shown in Table 4. The reason behind the lesser accuracy is that, the classes, such as “Alfalfa,” “Grass-pasture-mowed,” and “Oats,” have a limited number of samples (also called small classes) as shown in Table 1. By selecting 20% samples per class as training samples, these classes are represented by only a few samples in the training set, which probably do not provide a fair-enough representation of the class. For pixel-wise classifier SVM, the training samples are too limited to learn an effective model. Moreover, for classes, such as, “Grass Pasture,” “Oat,” “Buildings-grass-trees-drives,” and “Stone-steel-towers,” 3-D DWT method outperforms 3-D DCT method due to the localization property of the 3-D DWT method. 3-D DCT considers only frequency content of the signal and ignores the localized information. 3.4.2.Result analysis by comparing the proposed method with different classification methods on Pavia University datasetThe information used for experimentation such as ground-truth data, training sample map, and testing sample map of Pavia University dataset is shown in Fig. 5. The classification map of all competing techniques on Pavia University dataset is shown Fig. 6, and the classification results (i.e., OA, classwise accuracy, AA, and κ) of all competing methods and the proposed method are presented in Table 5. From Fig. 6 and Table 5, it can be shown that the proposed technique attains most excellent performance in terms of OA, AA, class wise accuracy as well as κ. Also, it is noted that the traditional feature extraction methods, such as SVM, SVM-PCA, ICDA, and LDA, yield similar results. Fig. 5DownloadFig. 6DownloadTable 5Comparison of classification accuracies (%) obtained by proposed method with competing methods for Pavia University dataset.
Because of the inherent multiresolution approach to the complex data, the transform-based feature extraction method 3-D DWT shows remarkable performance in comparison with traditional feature extraction methods, SVM, SVM-PCA, ICDA, and LDA. However, the proposed methods have achieved better performance over the 3-D DWT method, which means the energy coefficients preserves more complementary information of original feature space. As shown in Fig. 6, the proposed approach can help to eliminate most of the noisy pixels generated by the other methods, and the overall classification accuracy increases by >2%. For example, misclassified pixels from other comparable methods were corrected in the green region at the center of Fig. 6, which is very close to the ground truth and also the overall classification map has become smoother. Compared with other competing approaches, the proposed approach has improved classification accuracy significantly as shown in Table 5 (boldface). For instance, the classification accuracy of class “Asphalt” increases from 88.71% to 95.12%, and the classification accuracy of class “Trees” increases from 79.63% to 95.63%. Moreover, class “Shadow” is identified with 100% accuracy. As shown in Fig. 6, for the proposed method, many pixels of the “Bare soil” class are misclassified as “Meadows” class because of the complex structure of the classes. Also, some of the pixels of “Gravel” class are misclassified as other classes, such as “Bitumen” and “Self-blocking bricks.” By visual inspection, it is observed that the proposed method produces a more smooth and accurate classification map. For the classes “Gravel,” “Bare soil,” “Bitumen,” and “Self-blocking bricks,” 3-D DWT produces better accuracy than 3-D DCT. It is due to the localization property of the 3-D DWT method, whereas 3-D DCT performs a transform of frequency contents only. 3.4.3.Result analysis by comparing the proposed method with different classification methods on Salinas datasetThe information used for experimentation, such as ground-truth data, training sample map, and testing sample map of Salinas dataset, is shown in Fig. 7. Figure 8 shows the classification map of all competing techniques on Salinas dataset and the statistical results (i.e., OA, class wise accuracy, AA, and κ) of all competing methods, and the proposed method is summarized in Table 6. It is clear that the classification map of the proposed method has less noise and is more accurate. From Fig. 8 and Table 6, it can be shown that the proposed technique attains the most significant performance in terms of OA, AA, class wise accuracy as well as the κ. Table 6 shows that ICDA and SVM-PCA methods perform better than the SVM method. Furthermore, the LDA method balances both interclass and intraclass criteria using a balancing parameter that outperforms the SVM, SVM-PCA, and ICDA. Fig. 7DownloadFig. 8DownloadTable 6Comparison of classification accuracies (%) obtained by proposed method with competing methods for Salinas dataset.
Due to the inherent multiresolution property, the transform-based feature extraction method 3-D DWT shows remarkable performance in comparison with the traditional feature extraction methods. However, the proposed methods have achieved better performance over the 3-D DWT method, which means the energy coefficients of the DCT preserves more complementary information of original feature space. As shown in Fig. 8, the proposed approach can help to eliminate most of the noisy pixels generated by the other methods, and the overall classification accuracy increases by >2%. As shown in Fig. 8, overall classification map is very close to the ground truth. Compared with other competing approaches, the proposed approach has improved the classification accuracy significantly as shown in Table 6 (boldface). It also is shown that the proposed method presents higher performances, especially in classes with a small number of training samples such as “Fallow-rough-plow,” “Lettuce-romaine-4wk,” “Lettuce-romaine-6wk,” and “Lettuce-romaine-7wk.” The best classwise accuracy is produced by the proposed method for most of the classes (11 out of the 16 classes). Also, the class “Lettuce-romaine” is correctly identified with 100% accuracy. However, it can be seen that the proposed approach produces slightly lesser classification accuracy for classes such as “Broccoli-green-weeds-2,” “Fallow-smooth,” “Celery,” “Soil-vineyard-develop,” and “Lettuce-romaine-5wk,” which is almost negligible. The reason is that the DCT coefficients do not consider the localized information about the data. 3.4.4.Influence of different proportions of training samples on overall accuracyTo verify the superiority of the proposed method as the number of training samples increases, additional tests are conducted by considering randomly chosen 10%, 20%, 30%, 40%, and 50% training samples39,40 from each class of all datasets. The remaining samples are used as testing samples. Figure 9 shows OA obtained by the proposed method for a different proportion of training samples, and it is observed that the proposed method achieves a better result as sample proportion increases. Thus, the proposed method obtains the sufficient information to divulge the discriminative features of the hyperspectral data. Fig. 9Download3.4.5.Computational timeFigure 10 shows the computational time or the execution time (in seconds) of all competing methods for all datasets, such as Indian Pines, Pavia University, and Salinas dataset. Fig. 10DownloadAs shown in Fig. 10, the proposed method takes more computation time than traditional feature extraction methods but less than the transform-based feature extraction method. The 3-D DWT approach requires much more computational time, which degrades the competency of the 3-D DWT method when applied for the high dimensionality of data. The reason behind the expensive computations of the DWT method is the recursive computation of approximation and detail coefficients. In contrast, the DCT involves computation of real data only, which reduces computation burden. Also, DCT captures local variation present in the hyperspectral data that increase discrimination between different classes. So, by taking into account the overall accuracy and computational time, the proposed 3-D DCT approach significantly outperforms the other competent methods. 4.Conclusion and Future WorkIn this paper, a 3-D DCT-based feature extraction technique for hyperspectral image classification is proposed. This study proved that DCT allows a more efficient representation of the hyperspectral data by removing the redundancy between the neighboring pixels and adjacent bands and provides excellent decorrelation for the hyperspectral images. This technique is beneficial to extract discriminative features from high-dimensional data and is computationally competent. The experimental results on three standard benchmark datasets demonstrate that the proposed technique is more useful in extracting informative features and removing the redundant ones. Experimental results also show that compared with popular feature extraction methods, the proposed technique has significant performance on hyperspectral image classification. The proposed method has achieved a maximum classification accuracy of 94.62% for Salinas dataset. Although the proposed method is competitive with other state-of-the-art methods, there are still two crucial research directions deserving future attention. First, the spectral information can be integrated with the spatial information, such as edge-preserving filtering,41 Markov random field,42 discriminative random field method,43 and morphological profiles44 to improve the classification performance further. Second, the computational efficiency of the proposed method will be increased by parallel processing and graphics processing unit programming. AcknowledgmentsWe would like to thank the Council of Scientific & Industrial Research (CSIR), New Delhi, India for the award of CSIR-SRF and Vellore Institute of Technology, Vellore, India for providing the infrastructure facility. ReferencesM. Khodadadzadeh et al.,
“A new framework for hyperspectral image classification using multiple spectral and spatial features,”
in IEEE Geoscience and Remote Sensing Symp.,
4628
–4631
(2014). https://doi.org/10.1109/IGARSS.2014.6947524 Google Scholar
B. B. Damodaran and R. R. Nidamanuri,
“Dynamic linear classifier system for hyperspectral image classification for land cover mapping,”
IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 7
(6), 2080
–2093
(2014). https://doi.org/10.1109/JSTARS.2013.2294857 Google Scholar
H. Ren and C. I. Chang,
“Automatic spectral target recognition in hyperspectral imagery,”
IEEE Trans. Aerosp. Electron. Syst., 39
(4), 1232
–1249
(2003). Google Scholar
M. Dalponte et al.,
“Semi-supervised SVM for individual tree crown species classification,”
ISPRS J. Photogramm. Remote Sens., 110 77
–87
(2015). https://doi.org/10.1016/j.isprsjprs.2015.10.010 IRSEE9 0924-2716 Google Scholar
S. Marshall et al.,
“Hyperspectral imaging for food applications,”
in 23rd European Signal Processing Conf. (EUSIPCO),
2854
–2858
(2015). https://doi.org/10.1109/EUSIPCO.2015.7362906 Google Scholar
M. A. Calin et al.,
“Hyperspectral imaging in the medical field: present and future,”
Appl. Spectrosc. Rev., 49
(6), 435
–447
(2014). https://doi.org/10.1080/05704928.2013.838678 APSRBB 0570-4928 Google Scholar
C. Burges,
“Dimension reduction: a guided tour,”
Found. Trends Mach. Learn., 2
(4), 275
–364
(2010). https://doi.org/10.1561/2200000002 Google Scholar
“Introduction to statistical pattern recognition,”
Academic Press Professional, Inc., San Diego, California
(1990). Google Scholar
P. Dong, J. Liu,
“Hyperspectral image classification using support vector machines with an efficient principal component analysis scheme,”
Foundations of Intelligent Systems, 122 Springer(2011). Google Scholar
F. Tsai, E. K. Lin and K. Yoshino,
“Spectrally segmented principal component analysis of hyperspectral imagery for mapping invasive plant species,”
Int. J. Remote Sens., 28
(5), 1023
–1039
(2007). https://doi.org/10.1080/01431160600887706 IJSEDK 0143-1161 Google Scholar
X. Liu et al.,
“A maximum noise fraction transform with improved noise estimation for hyperspectral images,”
Sci. China Ser. F-Inf. Sci., 52
(9), 1578
–1587
(2009). Google Scholar
M. Fauvel, J. Chanussot and J. A. Benediktsson,
“Kernel principal component analysis for the classification of hyperspectral remote sensing data over urban areas,”
EURASIP J. Adv. Signal Process., 2009
(2), 783194
(2009). https://doi.org/10.1155/2009/783194 Google Scholar
M. E. Tipping,
“Probabilistic principal component analysis,”
J. R. Stat. Soc. Ser. B, 61
(3), 611
–622
(1999). https://doi.org/10.1111/rssb.1999.61.issue-3 JSTBAJ 0035-9246 Google Scholar
A. Villa et al.,
“Hyperspectral image classification with independent component discriminant analysis,”
IEEE Trans. Geosci. Remote Sens., 49
(12), 4865
–4876
(2011). https://doi.org/10.1109/TGRS.2011.2153861 IGRSD2 0196-2892 Google Scholar
T. V. Bandos, L. Bruzzone and G. Camps-Valls,
“Classification of hyperspectral images with regularized linear discriminant analysis,”
IEEE Trans. Geosci. Remote Sens., 47
(3), 862
–873
(2009). https://doi.org/10.1109/TGRS.2008.2005729 IGRSD2 0196-2892 Google Scholar
B. C. Kuo and D. A. Landgrebe,
“Nonparametric weighted feature extraction for classification,”
IEEE Trans. Geosci. Remote Sens., 42
(5), 1096
–1105
(2004). https://doi.org/10.1109/TGRS.2004.825578 IGRSD2 0196-2892 Google Scholar
B. C. Kuo, C. H. Li and J. M. Yang,
“Kernel nonparametric weighted feature extraction for hyperspectral image classification,”
IEEE Trans. Geosci. Remote Sens., 47
(4), 1139
–1155
(2009). https://doi.org/10.1109/TGRS.2008.2008308 IGRSD2 0196-2892 Google Scholar
J. Xia et al.,
“Improving random forest with ensemble of features and semisupervised feature extraction,”
IEEE Geosci. Remote Sens. Lett., 12
(7), 1471
–1475
(2015). https://doi.org/10.1109/LGRS.2015.2409112 Google Scholar
S. Samiappan, S. Prasad and L. Bruce,
“Non-uniform random feature selection and kernel density scoring with SVM based ensemble classification for hyperspectral image analysis,”
IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 6
(2), 792
–800
(2013). https://doi.org/10.1109/JSTARS.2013.2237757 Google Scholar
M. Imani and H. Ghassemian,
“Band clustering-based feature extraction for classification of hyperspectral images using limited training samples,”
IEEE Geosci. Remote Sens. Lett., 11
(8), 1325
–1329
(2014). https://doi.org/10.1109/LGRS.2013.2292892 Google Scholar
S. Sawant and M. Prabukumar,
“Band fusion based hyper spectral image classification,”
Int. J. Pure Appl. Math., 117
(17), 71
–76
(2017). Google Scholar
L. Zhang et al.,
“Compression of hyperspectral remote sensing images by tensor approach,”
Neurocomputing, 147
(1), 358
–363
(2015). https://doi.org/10.1016/j.neucom.2014.06.052 NRCGEO 0925-2312 Google Scholar
X. Cao et al.,
“Integration of 3-dimensional discrete wavelet transform and Markov random field for hyperspectral image classification,”
Neurocomputing, 226 90
–100
(2017). https://doi.org/10.1016/j.neucom.2016.11.034 NRCGEO 0925-2312 Google Scholar
Z. Zhu et al.,
“Three-dimensional Gabor feature extraction for hyperspectral imagery classification using a memetic framework,”
Inf. Sci., 298 274
–287
(2015). https://doi.org/10.1016/j.ins.2014.11.045 Google Scholar
Y. Chen et al.,
“Spectral—spatial classification of hyperspectral data based on deep belief network,”
IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 8
(6), 2381
–2392
(2015). https://doi.org/10.1109/JSTARS.2015.2388577 Google Scholar
X. Wan et al.,
“Stacked sparse autoencoder in hyperspectral data classification using spectral-spatial, higher order statistics and multifractal spectrum features,”
Infrared Phys. Technol., 86 77
–89
(2017). https://doi.org/10.1016/j.infrared.2017.08.021 IPTEEY 1350-4495 Google Scholar
B. Liu et al.,
“Supervised deep feature extraction for hyperspectral image classification,”
IEEE Trans. Geosci. Remote Sens., 56 1909
–1921
(2018). https://doi.org/10.1109/TGRS.2017.2769673 IGRSD2 0196-2892 Google Scholar
C. Shi and C. M. Pun,
“Multi-scale hierarchical recurrent neural networks for hyperspectral image classification,”
Neurocomputing, 294 82
–93
(2018). https://doi.org/10.1016/j.neucom.2018.03.012 NRCGEO 0925-2312 Google Scholar
Y. Chen et al.,
“Deep feature extraction and classification of hyperspectral images based on convolutional neural networks,”
IEEE Trans. Geosci. Remote Sens., 54
(10), 6232
–6251
(2016). https://doi.org/10.1109/TGRS.2016.2584107 IGRSD2 0196-2892 Google Scholar
S. Prasad and L. M. Bruce,
“Limitations of principal components analysis for hyperspectral target recognition,”
IEEE Geosci. Remote Sens. Lett., 5
(4), 625
–629
(2008). https://doi.org/10.1109/LGRS.2008.2001282 Google Scholar
Y. Zhen et al.,
“Classification based on 3-D DWT and decision fusion for hyperspectral image analysis,”
IEEE Geosci. Remote Sens. Lett., 11
(1), 173
–177
(2014). https://doi.org/10.1109/LGRS.2013.2251316 Google Scholar
A. R. Webb, Statistical Pattern Recognition, 71
(8), John Wiley & Sons, Ltd., England
(2011). Google Scholar
S. A. Khayam,
“The discrete cosine transform (DCT): theory and application,”
(2003). Google Scholar
F. Melgani and L. Bruzzone,
“Classification of hyperspectral remote sensing,”
IEEE Trans. Geosci. Remote Sens., 42
(8), 1778
–1790
(2004). https://doi.org/10.1109/TGRS.2004.831865 IGRSD2 0196-2892 Google Scholar
T. A. Moughal,
“Hyperspectral image classification using support vector machine,”
J. Phys. Conf. Ser., 439 012042
(2017). https://doi.org/10.1088/1742-6596/439/1/012042 JPCSDZ 1742-6588 Google Scholar
“Hyperspectral Remote Sensing Scenes,”
(2007) http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes Google Scholar
N. Falco, J. A. Benediktsson and L. Bruzzone,
“A study on the effectiveness of different independent component analysis algorithms for hyperspectral image classification,”
IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 7
(6), 2183
–2199
(2014). https://doi.org/10.1109/JSTARS.2014.2329792 Google Scholar
C.-I. Chang and H. Ren,
“An experiment-based quantitative and comparative analysis of target detection and image classification algorithms for hyperspectral imagery,”
IEEE Trans. Geosci. Remote Sens., 38
(2), 1044
–1063
(2000). https://doi.org/10.1109/36.841984 IGRSD2 0196-2892 Google Scholar
X. Xu et al.,
“Multisource remote sensing data classification based on convolutional neural network,”
IEEE Trans. Geosci. Remote Sens., 56
(2), 937
–949
(2018). https://doi.org/10.1109/TGRS.2017.2756851 IGRSD2 0196-2892 Google Scholar
B. Kumar and O. Dikshit,
“Hyperspectral image classification based on morphological profiles and decision fusion,”
Int. J. Remote Sens., 38
(20), 5830
–5854
(2017). https://doi.org/10.1080/01431161.2017.1348636 IJSEDK 0143-1161 Google Scholar
X. Kang et al.,
“Spectral—spatial hyperspectral image classification with edge-preserving filtering,”
IEEE Trans. Geosci. Remote Sens., 52
(5), 2666
–2677
(2014). https://doi.org/10.1109/TGRS.2013.2264508 IGRSD2 0196-2892 Google Scholar
X. Cao et al.,
“Hyperspectral image classification with Markov random fields and a convolutional neural network,”
IEEE Trans. Image Process., 27
(5), 2354
–2367
(2018). https://doi.org/10.1109/TIP.2018.2799324 IIPRE4 1057-7149 Google Scholar
A. C. Karaca et al.,
“Comparative evaluation of vector machine based hyperspectral classification methods,”
in IEEE Int. Geoscience and Remote Sensing Symp. (IGARSS),
4970
–4973
(2012). https://doi.org/10.1109/IGARSS.2012.6352496 Google Scholar
M. Fauvel and J. Benediktsson,
“Spectral and spatial classification of hyperspectral data using SVMs and morphological profile,”
IEEE Trans. Geosci. Remote Sens., 46
(11), 3804
–3814
(2008). https://doi.org/10.1109/TGRS.2008.922034 IGRSD2 0196-2892 Google Scholar
BiographyManoharan Prabukumar received his BE degree in electronics and communication engineering from Periyar University, Tamilnadu, in 2002, his MTech degree in computer vision and image processing from Amrita school of Engineering, Coimbatore, in 2007, and his PhD in computer graphics from Vellore Institute of Technology (VIT), Tamilnadu, India, in 2014. Currently, he is working as an associate professor at the School of Information Technology and Engineering, VIT. His research interests include hyperspectral remote sensing, image processing, computer graphics, and machine learning. Shrutika Sawant received her BE and ME degrees in electronics and telecommunication engineering from Shivaji University, Maharashtra, India, in 2009 and 2012, respectively. Currently, she is pursuing her PhD in hyperspectral image processing from Vellore Institute of Technology (VIT), Vellore, Tamilnadu, India. She has been awarded with the senior research fellowship (SRF) from Council of Scientific and Industrial Research (CSIR), New Delhi. Her research interests include hyperspectral remote sensing, image processing, and machine learning. Sathishkumar Samiappan received his BEngg degree in electronics and communication from Bharathiar University, Coimbatore, in 2003, his MTech degree in computer science and engineering from Amrita University, Coimbatore, India, in 2006, and his PhD in electrical and computer engineering at Mississippi State University (MSU), Starkville, Mississippi. Currently, he is an assistant research professor with the Geosystems Research Institute at MSU. His research interests include low-altitude remote sensing, pattern recognition, image processing, machine learning, and hyperspectral image classification. Loganathan Agilandeeswari completed her PhD and is working as an associate professor at the School of Information Technology and Engineering, VIT, Vellore. She was awarded a best researcher award for year 2015 to 2016. She received her bachelor’s degree in information technology and her master’s degree in computer science and engineering from Anna University during 2005 and 2009, respectively. She has published 25+ papers in peer-reviewed reputed journals. She also the author of the various books, such as computer networks, mobile computing, and communication engineering. |