4.3 Article

Large patch convolutional neural networks for the scene classification of high spatial resolution imagery

Journal

JOURNAL OF APPLIED REMOTE SENSING
Volume 10, Issue -, Pages -

Publisher

SPIE-SOC PHOTO-OPTICAL INSTRUMENTATION ENGINEERS
DOI: 10.1117/1.JRS.10.025006

Keywords

large patch sampling; deep convolutional neural networks; remote sensing; high spatial resolution image; scene classification

Funding

  1. National Natural Science Foundation of China [41371344]
  2. State Key Laboratory of Earth Surface Processes and Resource Ecology [2015-KF-02]
  3. Program for Changjiang Scholars and Innovative Research Team in University [IRT1278]
  4. Natural Science Foundation of Hubei Province [2015CFA002]
  5. Open Research Fund Program of Shenzhen Key Laboratory of Spatial Smart Sensing and Services (Shenzhen University)

Ask authors/readers for more resources

The increase of the spatial resolution of remote-sensing sensors helps to capture the abundant details related to the semantics of surface objects. However, it is difficult for the popular object-oriented classification approaches to acquire higher level semantics from the high spatial resolution remote-sensing (HSR-RS) images, which is often referred to as the semantic gap. Instead of designing sophisticated operators, convolutional neural networks (CNNs), a typical deep learning method, can automatically discover intrinsic feature descriptors from a large number of input images to bridge the semantic gap. Due to the small data volume of the available HSR-RS scene datasets, which is far away from that of the natural scene datasets, there have been few reports of CNN approaches for HSR-RS image scene classifications. We propose a practical CNN architecture for HSR-RS scene classification, named the large patch convolutional neural network (LPCNN). The large patch sampling is used to generate hundreds of possible scene patches for the feature learning, and a global average pooling layer is used to replace the fully connected network as the classifier, which can greatly reduce the total parameters. The experiments confirm that the proposed LPCNN can learn effective local features to form an effective representation for different land-use scenes, and can achieve a performance that is comparable to the state-of-the-art on public HSR-RS scene datasets. (C) 2016 Society of Photo-Optical Instrumentation Engineers (SPIE)

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.3
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available