scispace - formally typeset
Journal ArticleDOI

Mining Deep Semantic Representations for Scene Classification of High-Resolution Remote Sensing Imagery

Fan Hu, +3 more
- 01 Sep 2020 - 
- Vol. 6, Iss: 3, pp 522-536
TLDR
This paper proposes to build powerful semantic features using the probabilistic latent semantic analysis (pLSA) model, by employing the pre-trained deep convolutional neural networks (CNNs) as feature extractors rather than relying on the hand-crafted features.
Abstract
Scene classification is one of the most fundamental task in interpretation of high-resolution remote sensing (HRRS) images. Many recent works show that the probabilistic topic models which are capable of mining latent semantics of images can be effectively applied to HRRS scene classification. However, the existing approaches based on topic models simply utilize low-level hand-crafted features to form semantic features, which severely limit the representative capability of the semantic features derived from topic models. To alleviate this problem, this paper propose to build powerful semantic features using the probabilistic latent semantic analysis (pLSA) model, by employing the pre-trained deep convolutional neural networks (CNNs) as feature extractors rather than relying on the hand-crafted features. Specifically, we develop two methods to generate semantic features, called multi-scale deep semantic representation (MSDS) and multi-level deep semantic representation (MLDS), by extracting CNN features from different layers: (1) in MSDS, the final semantic features are learned by the pLSA with multi-scale features extracted from the convolutional layer of a pre-trained CNN; (2) in MLDS, we extract CNN features for densely sampled image patches at different size level from the fully-connected layer of a pre-trained CNN, and concatenate the sematic features learned by the pLSA at each level. We comprehensively evaluate the two methods on two public HRRS scene datasets, and achieve significant performance improvement over the state-of-the-art. The outstanding results demonstrate that the pLSA model is capable of discovering considerably discriminative semantic features from the deep CNN features.

read more

Citations
More filters
Journal ArticleDOI

Relation Network for Multilabel Aerial Image Classification

TL;DR: Zhang et al. as mentioned in this paper proposed an attention-aware label relational reasoning network, which consists of three elemental modules: a label-wise feature parcel learning module, an attentional region extraction module, and a label relational inference module.
Journal ArticleDOI

SCViT: A Spatial-Channel Feature Preserving Vision Transformer for Remote Sensing Image Scene Classification

TL;DR: In this paper , a spatial-channel feature preserving vision transformer model (SCViT) is proposed, which considers both the detailed geometric information of the high spatial resolution (HSR) imagery and the contribution of the different channels contained in the classification token.
Journal ArticleDOI

Deep Feature Aggregation Framework Driven by Graph Convolutional Network for Scene Classification in Remote Sensing

TL;DR: Zhang et al. as discussed by the authors developed a deep feature aggregation framework driven by graph convolutional network (DFAGCN) for high-spatial resolution (HSR) scene classification.
Journal ArticleDOI

Object-Scale Adaptive Convolutional Neural Networks for High-Spatial Resolution Remote Sensing Image Classification

TL;DR: A novel method called object-scale adaptive convolutional neural network (OSA-CNN), which combines OBIA with CNN, is proposed for HSR image classification, which effectively enhances the image classification accuracy.
Journal ArticleDOI

Two-stream feature aggregation deep neural network for scene classification of remote sensing images

TL;DR: A novel architecture termed two-stream feature aggregation deep neural network (TFADNN) is developed for HSR scene classification and results indicate that the TFADNN method achieves satisfactory classification performance compared with some state-of-the-art methods.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Proceedings ArticleDOI

ImageNet: A large-scale hierarchical image database

TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Journal ArticleDOI

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Related Papers (5)