Learning Multi-Modality Features for Scene Classification of High-Resolution Remote Sensing Images

HTML  XML Download Download as PDF (Size: 1515KB)  PP. 185-193  
DOI: 10.4236/jcc.2018.611018    495 Downloads   1,083 Views  Citations

ABSTRACT

Scene classification of high-resolution remote sensing (HRRS) image is an important research topic and has been applied broadly in many fields. Deep learning method has shown its high potential to in this domain, owing to its powerful learning ability of characterizing complex patterns. However the deep learning methods omit some global and local information of the HRRS image. To this end, in this article we show efforts to adopt explicit global and local information to provide complementary information to deep models. Specifically, we use a patch based MS-CLBP method to acquire global and local representations, and then we consider a pretrained CNN model as a feature extractor and extract deep hierarchical features from full-connection layers. After fisher vector (FV) encoding, we obtain the holistic visual representation of the scene image. We view the scene classification as a reconstruction procedure and train several class-specific stack denoising autoencoders (SDAEs) of corresponding class, i.e., one SDAE per class, and classify the test image according to the reconstruction error. Experimental results show that our combination method outperforms the state-of-the-art deep learning classification methods without employing fine-tuning.

Share and Cite:

Zhao, F. , Zhang, X. , Mu, X. , Yi, Z. and Yang, Z. (2018) Learning Multi-Modality Features for Scene Classification of High-Resolution Remote Sensing Images. Journal of Computer and Communications, 6, 185-193. doi: 10.4236/jcc.2018.611018.

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.