EMMCNN: An ETPS-Based Multi-Scale and Multi-Feature Method Using CNN for High Spatial Resolution Image Land-Cover Classification
"> Figure 1
<p>DenseNet architecture with dense blocks and transition layers.</p> "> Figure 2
<p>Flowchart of the proposed extended topology-preserving segmentation (ETPS)-based multi-scale and multi-feature method using the convolutional neural network (CNN) for high spatial resolution (HSR) image land-cover classification.</p> "> Figure 3
<p>Comparison of superpixel segmentation using extended topology-preserving segmentation (ETPS)based on different color compositions. (<b>a</b>) ETPS segmentation based on true color. (<b>b</b>) ETPS segmentation based on standard false-color.</p> "> Figure 4
<p>1-D CNN-MLP hybrid network with attention-based weighting design for comprehensive land-cover classification. (<b>a</b>) Multi-scale CNN feature fusion and encoding. (<b>b</b>) Hand-delineated feature encoding. (<b>c</b>) Multi-scale and multi-feature fusion and classification. CNN: convolutional neural network; MLP: multi-layer perception.</p> "> Figure 5
<p>Study areas (<b>a</b>) and GaoFen-2 images of Beilun urban scenes (<b>c</b>) and Cixi rural scenes (<b>b</b>) in true color.</p> "> Figure 6
<p>Study areas (<b>a</b>), GaoFen-2 images of Xiaoshan for training and testing (<b>b</b>) and validation (<b>c</b>), and GaoFen-2 images of Yuecheng for training and testing (<b>d</b>) and validation (<b>e</b>) in true color.</p> "> Figure 7
<p>Examples of different land-cover classes in false-color for Beilun and Cixi datasets. (<b>a</b>) Images of 10 land-cover classes in the Beilun dataset. (<b>b</b>) Images of 8 land-cover classes in the Cixi dataset.</p> "> Figure 8
<p>Confusion matrices of testing samples in the Beilun dataset for the ESSCNN, EMMCNN, OCNN, PCNN, SMCNN, SMMCNN, and ORF methods. The numbers 1 to 10 in horizontal and vertical axis denote the farmland, woodland, grassland, dense building, sparse building, road, impervious surface, facility land, vacant, and water classes, respectively. ESSCNN: extended topology-preserving segmentation (ETPS)-based single-scale and single-feature convolutional neural network (CNN); EMMCNN: ETPS-based multi-scale and multi-feature CNN; OCNN: object-based CNN; PCNN: patch-based CNN; SMCNN: simple linear iterative clustering (SLIC)-based multi-scale CNN; SMMCNN: SLIC-based multi-scale and multi-feature CNN; ORF: object-based random forest.</p> "> Figure 9
<p>Three typical image subsets (<b>a</b>, <b>b</b>, and <b>c</b>) in the Beilun dataset and their classification results using ESSCNN, EMMCNN, OCNN, PCNN, SMCNN, SMMCNN, and ORF methods. The white and black circles denote the correct and incorrect classification, respectively. ESSCNN: extended topology-preserving segmentation (ETPS)-based single-scale and single-feature convolutional neural network (CNN); EMMCNN: ETPS-based multi-scale and multi-feature CNN; OCNN: object-based CNN; PCNN: patch-based CNN; SMCNN: simple linear iterative clustering (SLIC)-based multi-scale CNN; SMMCNN: SLIC-based multi-scale and multi-feature CNN; ORF: object-based random forest.</p> "> Figure 10
<p>Confusion matrices of testing samples in the Cixi dataset for the ESSCNN, EMMCNN, OCNN, PCNN, SMCNN, SMMCNN, and ORF methods. The numbers 1 to 8 in horizontal and vertical axis denote the water, farmland, woodland, grassland, building, road, impervious surface, and facility land classes, respectively. ESSCNN: extended topology-preserving segmentation (ETPS)-based single-scale and single-feature convolutional neural network (CNN); EMMCNN: ETPS-based multi-scale and multi-feature CNN; OCNN: object-based CNN; PCNN: patch-based CNN; SMCNN: simple linear iterative clustering (SLIC)-based multi-scale CNN; SMMCNN: SLIC-based multi-scale and multi-feature CNN; ORF: object-based random forest.</p> "> Figure 11
<p>Three typical image subsets (<b>a</b>, <b>b</b>, and <b>c</b>) in the Cixi dataset and their classification results using ESSCNN, EMMCNN, OCNN, PCNN, SMCNN, SMMCNN, and ORF methods. The white and black circles denote the correct and incorrect classification, respectively. ESSCNN: extended topology-preserving segmentation (ETPS)-based single-scale and single-feature convolutional neural network (CNN); EMMCNN: ETPS-based multi-scale and multi-feature CNN; OCNN: object-based CNN; PCNN: patch-based CNN; SMCNN: simple linear iterative clustering (SLIC)-based multi-scale CNN; SMMCNN: SLIC-based multi-scale and multi-feature CNN; ORF: object-based random forest.</p> "> Figure 12
<p>Confusion matrices of testing (test) and validating (val) samples in the Xiaoshan dataset for the ESSCNN, EMMCNN, OCNN, PCNN, SMCNN, SMMCNN, and ORF methods. The numbers 1 to 7 in horizontal and vertical axis denote the farmland, woodland, building, road, impervious surface, vacant, and water classes, respectively. ESSCNN: extended topology-preserving segmentation (ETPS)-based single-scale and single-feature convolutional neural network (CNN); EMMCNN: ETPS-based multi-scale and multi-feature CNN; OCNN: object-based CNN; PCNN: patch-based CNN; SMCNN: simple linear iterative clustering (SLIC)-based multi-scale CNN; SMMCNN: SLIC-based multi-scale and multi-feature CNN; ORF: object-based random forest.</p> "> Figure 13
<p>Confusion matrices of testing (test) and validating (val) samples in the Yuecheng dataset for the ESSCNN, EMMCNN, OCNN, PCNN, SMCNN, SMMCNN, and ORF methods. The numbers 1 to 7 in horizontal and vertical axis denote the farmland, woodland, building, road, impervious surface, vacant, and water classes, respectively. ESSCNN: extended topology-preserving segmentation (ETPS)-based single-scale and single-feature convolutional neural network (CNN); EMMCNN: ETPS-based multi-scale and multi-feature CNN; OCNN: object-based CNN; PCNN: patch-based CNN; SMCNN: simple linear iterative clustering (SLIC)-based multi-scale CNN; SMMCNN: SLIC-based multi-scale and multi-feature CNN; ORF: object-based random forest.</p> "> Figure 14
<p>Classification accuracy comparison among single-feature, multi-feature, and multi-feature with attention-based weighting methods upon multi-scale combinations. The 1st to 36th columns of the horizontal axis represent the COMB1 to COMB36 combining solutions, as shown in <a href="#remotesensing-12-00066-t002" class="html-table">Table 2</a>. OA: overall accuracy; KC: Kappa coefficient.</p> ">
Abstract
:1. Introduction
2. Related Work
2.1. Superpixel Segmentation
2.2. CNN-Based Classification
3. Methodology
3.1. ETPS Superpixel Segmentation
3.2. Multi-Scale CNN Feature Extraction
3.3. Multi-Scale and Multi-Feature Combination
3.4. 1-D CNN-MLP Comprehensive Classification
4. Experiments
4.1. Datasets
4.2. Parameter Settings
4.3. Comparison Methods
4.4. Evaluation Criteria
4.5. Experimental Analysis
4.5.1. Beilun Dataset
4.5.2. Cixi Dataset
4.5.3. Xiaoshan Dataset
4.5.4. Yuecheng Dataset
5. Discussion
5.1. Evaluation of Single Spatial Scales
5.2. Evaluation of Multi-Scale Combinations
6. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Ma, L.; Li, M.; Ma, X.; Cheng, L.; Du, P.; Liu, Y. A review of supervised object-based land-cover image classification. ISPRS J. Photogramm. Remote Sens. 2017, 130, 277–293. [Google Scholar] [CrossRef]
- Zhao, W.; Du, S.; Wang, Q.; Emery, W.J. Contextually guided very-high-resolution imagery classification with semantic segments. ISPRS J. Photogramm. Remote Sens. 2017, 132, 48–60. [Google Scholar] [CrossRef]
- Lv, X.; Ming, D.; Chen, Y.; Wang, M. Very high resolution remote sensing image classification with SEEDS-CNN and scale effect analysis for superpixel CNN classification. Int. J. Remote Sens. 2019, 40, 506–531. [Google Scholar] [CrossRef]
- Kurtz, C.; Passat, N.; Gancarski, P.; Puissant, A. Extraction of complex patterns from multiresolution remote sensing images: A hierarchical top-down methodology. Pattern Recognit. 2012, 45, 685–706. [Google Scholar] [CrossRef] [Green Version]
- Zhao, W.; Du, S.; Emery, W.J. Object-based convolutional neural network for high-resolution imagery classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3386–3396. [Google Scholar] [CrossRef]
- Zhang, C.; Pan, X.; Li, H.; Gardiner, A.; Sargent, I.; Hare, J.; Atkinson, P.M. A hybrid MLP-CNN classifier for very fine resolution remotely sensed image classification. ISPRS J. Photogramm. Remote Sens. 2018, 140, 133–144. [Google Scholar] [CrossRef] [Green Version]
- Kelly, M.; Blanchard, S.D.; Kersten, E.; Koy, K. Terrestrial remotely sensed imagery in support of public health: New avenues of research using object-based image analysis. Remote Sens. 2011, 3, 2321–2345. [Google Scholar] [CrossRef] [Green Version]
- Myint, S.W.; Gober, P.; Brazel, A.; Grossman-Clarke, S.; Weng, Q. Per-pixel vs. object-based classification of urban land cover extraction using high spatial resolution imagery. Remote Sens. Environ. 2011, 115, 1145–1161. [Google Scholar] [CrossRef]
- Blaschke, T.; Hay, G.J.; Kelly, M.; Lang, S.; Hofmann, P.; Addink, E.; Feitosa, R.Q.; der Meer, F.; der Werff, H.; Van Coillie, F.; et al. Geographic object-based image analysis-towards a new paradigm. ISPRS J. Photogramm. Remote Sens. 2014, 87, 180–191. [Google Scholar] [CrossRef] [Green Version]
- Li, M.; Ma, L.; Blaschke, T.; Cheng, L.; Tiede, D. A systematic comparison of different object-based classification techniques using high spatial resolution imagery in agricultural environments. Int. J. Appl. Earth Obs. Geoinf. 2016, 49, 87–98. [Google Scholar] [CrossRef]
- Chen, G.; Weng, Q.; Hay, G.J.; He, Y. Geographic object-based image analysis (GEOBIA): Emerging trends and future opportunities. GIScience Remote Sens. 2018, 55, 159–182. [Google Scholar] [CrossRef]
- Huang, B.; Zhao, B.; Song, Y. Urban land-use mapping using a deep convolutional neural network with high spatial resolution multispectral remote sensing imagery. Remote Sens. Environ. 2018, 214, 73–86. [Google Scholar] [CrossRef]
- Zhang, C.; Sargent, I.; Pan, X.; Li, H.; Gardiner, A.; Hare, J.; Atkinson, P.M. An object-based convolutional neural network (OCNN) for urban land use classification. Remote Sens. Environ. 2018, 216, 57–70. [Google Scholar] [CrossRef] [Green Version]
- Zhang, G.; Jia, X.; Hu, J. Superpixel-based graphical model for remote sensing image mapping. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5861–5871. [Google Scholar] [CrossRef]
- Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of hyperspectral images by exploiting spectral--spatial information of superpixel via multiple kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef] [Green Version]
- Fang, L.; Li, S.; Kang, X.; Benediktsson, J.A. Spectral-spatial classification of hyperspectral images with a superpixel-based discriminative sparse model. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4186–4201. [Google Scholar] [CrossRef]
- Zhao, W.; Jiao, L.; Ma, W.; Zhao, J.; Zhao, J.; Liu, H.; Cao, X.; Yang, S. Superpixel-based multiple local CNN for panchromatic and multispectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4141–4156. [Google Scholar] [CrossRef]
- Feng, W.; Sui, H.; Huang, W.; Xu, C.; An, K. Water body extraction from very high-resolution remote sensing imagery using deep U-Net and a superpixel-based conditional random field model. IEEE Geosci. Remote Sens. Lett. 2018, 16, 618–622. [Google Scholar] [CrossRef]
- Bahmanyar, R.; Cui, S.; Datcu, M. A comparative study of bag-of-words and bag-of-topics models of EO image patches. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1357–1361. [Google Scholar] [CrossRef] [Green Version]
- Zhao, B.; Zhong, Y.; Zhang, L. A spectral-structural bag-of-features scene classifier for very high spatial resolution remote sensing imagery. ISPRS J. Photogramm. Remote Sens. 2016, 116, 73–85. [Google Scholar] [CrossRef]
- Tong, X.-Y.; Xia, G.-S.; Lu, Q.; Shen, H.; Li, S.; You, S.; Zhang, L. Learning transferable deep models for land-use classification with high-resolution remote sensing images. arXiv 2018, arXiv:1807.05713. [Google Scholar]
- Sun, Y.; Zhang, X.; Xin, Q.; Huang, J. Developing a multi-filter convolutional neural network for semantic segmentation using high-resolution aerial imagery and LiDAR data. ISPRS J. Photogramm. Remote Sens. 2018, 143, 3–14. [Google Scholar] [CrossRef]
- Wang, Q.; Yuan, Z.; Du, Q.; Li, X. GETNET: A general end-to-end 2-D CNN framework for hyperspectral image change detection. IEEE Trans. Geosci. Remote Sens. 2018, 57, 3–13. [Google Scholar] [CrossRef] [Green Version]
- Zhang, F.; Du, B.; Zhang, L. Scene classification via a gradient boosting random convolutional network framework. IEEE Trans. Geosci. Remote Sens. 2015, 54, 1793–1802. [Google Scholar] [CrossRef]
- Cheng, G.; Wang, Y.; Xu, S.; Wang, H.; Xiang, S.; Pan, C. Automatic road detection and centerline extraction via cascaded end-to-end convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3322–3337. [Google Scholar] [CrossRef]
- Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
- Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.-S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef] [Green Version]
- Ma, L.; Liu, Y.; Zhang, X.; Ye, Y.; Yin, G.; Johnson, B.A. Deep learning in remote sensing applications: A meta-analysis and review. ISPRS J. Photogramm. Remote Sens. 2019, 152, 166–177. [Google Scholar] [CrossRef]
- Zhou, T.; Miao, Z.; Zhang, J. Combining CNN with hand-crafted features for image classification. In Proceedings of the 2018 14th IEEE International Conference on Signal Processing (ICSP), Beijing, China, 12–16 August 2018; pp. 554–557. [Google Scholar]
- Cao, Q.; Zhong, Y.; Ma, A.; Zhang, L. Urban land use/land cover classification based on feature fusion fusing hyperspectral image and lidar data. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 8869–8872. [Google Scholar]
- Zhao, W.; Du, S. Learning multiscale and deep representations for classifying remotely sensed imagery. ISPRS J. Photogramm. Remote Sens. 2016, 113, 155–165. [Google Scholar] [CrossRef]
- Zhao, W.; Guo, Z.; Yue, J.; Zhang, X.; Luo, L. On combining multiscale deep learning features for the classification of hyperspectral remote sensing imagery. Int. J. Remote Sens. 2015, 36, 3368–3379. [Google Scholar] [CrossRef]
- Längkvist, M.; Kiselev, A.; Alirezaie, M.; Loutfi, A. Classification and segmentation of satellite orthoimagery using convolutional neural networks. Remote Sens. 2016, 8, 329. [Google Scholar] [CrossRef] [Green Version]
- Liang, M.; Jiao, L.; Yang, S.; Liu, F.; Hou, B.; Chen, H. Deep multiscale spectral-spatial feature fusion for hyperspectral images classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2911–2924. [Google Scholar] [CrossRef]
- Tian, Q.; Wan, S.; Jin, P.; Xu, J.; Zou, C.; Li, X. A novel feature fusion with self-adaptive weight method based on deep learning for image classification. In Proceedings of the Pacific Rim Conference on Multimedia, Hefei, China, 21–22 September 2018; pp. 426–436. [Google Scholar]
- Zhang, Y.; Huynh, C.P.; Ngan, K.N. Feature fusion with predictive weighting for spectral image classification and segmentation. IEEE Trans. Geosci. Remote Sens. 2019, 6792–6807. [Google Scholar] [CrossRef]
- Cheng, G.; Li, Z.; Han, J.; Yao, X.; Guo, L. Exploring hierarchical convolutional features for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6712–6722. [Google Scholar] [CrossRef]
- Liu, Y.; Cao, G.; Sun, Q.; Siegel, M. Hyperspectral classification via deep networks and superpixel segmentation. Int. J. Remote Sens. 2015, 36, 3459–3482. [Google Scholar] [CrossRef]
- Stutz, D.; Hermans, A.; Leibe, B. Superpixels: An evaluation of the state-of-the-art. Comput. Vis. Image Underst. 2018, 166, 1–27. [Google Scholar] [CrossRef] [Green Version]
- Yamaguchi, K.; McAllester, D.; Urtasun, R. Efficient joint segmentation, occlusion labeling, stereo and flow estimation. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 756–771. [Google Scholar]
- Yao, J.; Boben, M.; Fidler, S.; Urtasun, R. Real-time coarse-to-fine topologically preserving segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, America, 8–10 June 2015; pp. 2947–2955. [Google Scholar]
- LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc.: Lake Tahoe, America, 2012; pp. 1097–1105. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, America, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, America, 22–25 July 2017; pp. 4700–4708. [Google Scholar]
- Han, X.; Zhong, Y.; Cao, L.; Zhang, L. Pre-trained Alexnet architecture with pyramid pooling and supervision for high spatial resolution remote sensing image scene classification. Remote Sens. 2017, 9, 848. [Google Scholar] [CrossRef] [Green Version]
- Anwer, R.M.; Khan, F.S.; van de Weijer, J.; Molinier, M.; Laaksonen, J. Binary patterns encoded convolutional neural networks for texture recognition and remote sensing scene classification. ISPRS J. Photogramm. Remote Sens. 2018, 138, 74–85. [Google Scholar] [CrossRef] [Green Version]
- Xu, Y.; Xie, Z.; Feng, Y.; Chen, Z. Road extraction from high-resolution remote sensing imagery using deep learning. Remote Sens. 2018, 10, 1461. [Google Scholar] [CrossRef] [Green Version]
- Tao, Y.; Xu, M.; Lu, Z.; Zhong, Y. DenseNet-based depth-width double reinforced deep learning neural network for high-resolution remote sensing image per-pixel classification. Remote Sens. 2018, 10, 779. [Google Scholar] [CrossRef] [Green Version]
- Mahdianpari, M.; Salehi, B.; Rezaee, M.; Mohammadimanesh, F.; Zhang, Y. Very deep convolutional neural networks for complex land cover mapping using multispectral remote sensing imagery. Remote Sens. 2018, 10, 1119. [Google Scholar] [CrossRef] [Green Version]
- Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
- Benz, U.C.; Hofmann, P.; Willhauck, G.; Lingenfelder, I.; Heynen, M. Multi-resolution, object-oriented fuzzy analysis of remote sensing data for GIS-ready information. ISPRS J. Photogramm. Remote Sens. 2004, 58, 239–258. [Google Scholar] [CrossRef]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Paisitkriangkrai, S.; Sherrah, J.; Janney, P.; Hengel, V.-D. Others Effective semantic pixel labelling with convolutional networks and conditional random fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, America, 11–12 June 2015; pp. 36–43. [Google Scholar]
Layers | DCNN-5 | Output Size | DCNN-6 | Output Size |
---|---|---|---|---|
Convolution | 3 × 3 conv | 80 × 80, 24 | 3 × 3 conv | 80 × 80, 24 |
Dense block (1) | 80 × 80, 84 | 80 × 80, 96 | ||
Transition layer (1) | 1 × 1 conv | 80 × 80, 42 | 1 × 1 conv | 80 × 80, 48 |
2 × 2 avg pool, stride 2 | 40 × 40, 42 | 2 × 2 avg pool, stride 2 | 40 × 40, 48 | |
Dense block (2) | 40 × 40, 102 | 40 × 40, 120 | ||
Transition layer (2) | 1 × 1 conv | 40 × 40, 51 | 1 × 1 conv | 40 × 40, 60 |
2 × 2 avg pool, stride 2 | 20 × 20, 51 | 2 × 2 avg pool, stride 2 | 20 × 20, 60 | |
Dense block (3) | 20 × 20, 111 | 20 × 20, 132 | ||
Transition layer (3) | 1 × 1 conv | 20 × 20, 55 | 1 × 1 conv | 20 × 20, 66 |
2 × 2 avg pool, stride 2 | 10 × 10, 55 | 2 × 2 avg pool, stride 2 | 10 × 10, 66 | |
Dense block (4) | 10 × 10, 115 | 10 × 10, 138 | ||
GAP layer | 10 × 10 GAP | 1 × 1, 115 | 10 × 10 GAP | 1 × 1, 138 |
Classification layer | softmax | softmax |
Scale | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 | 21 | 22 | 23 | 24 | 25 | 26 | 27 | 28 | 29 | 30 | 31 | 32 | 33 | 34 | 35 | 36 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
24×24 | * | * | * | * | * | * | * | * | ||||||||||||||||||||||||||||
32×32 | * | * | * | * | * | * | * | * | * | * | * | * | ||||||||||||||||||||||||
40×40 | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | |||||||||||||||||||
48×48 | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | ||||||||||||||||
56×56 | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | |||||||||||||
64×64 | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | ||||||||||||
72×72 | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | ||||||||||||
80×80 | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | |||||||||||||
88×88 | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | ||||||||||||||||
96×96 | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | * | |||||||||||||||||||
104×104 | * | * | * | * | * | * | * | * | * | * | * | * | ||||||||||||||||||||||||
112×112 | * | * | * | * | * | * | * | * |
Class | Proportion | Train | Test | Subclass |
---|---|---|---|---|
Farmland | 7.09% | 4262 | 2841 | Paddy fields, dry lands, nurseries, orchards |
Woodland | 6.74% | 4056 | 2703 | Timber forest, shrub forest, planted forest |
Grassland | 10.37% | 6236 | 4156 | Native grassland, planted grassland |
Dense building | 14.65% | 8811 | 5873 | Dense high-rise and low-rise buildings |
Sparse building | 4.67% | 2811 | 1874 | Sparse high-rise and low-rise buildings |
Road | 11.88% | 7143 | 4761 | Highways, overpasses, streets |
Impervious surface | 17.91% | 10769 | 7179 | Squares, stadiums, parking lots, storage fields, rolled surface |
Facility land | 8.18% | 4923 | 3281 | Oil drums, container fields, docks, industrial facilities |
Vacant | 5.15% | 3096 | 2063 | Digging lands, bare surface |
Water | 13.37% | 8040 | 5360 | Rivers, rivulets, ponds, lakes |
Class | Proportion | Train | Test | Subclass |
---|---|---|---|---|
Water | 9.93% | 4627 | 3085 | Rivers, rivulets, ponds, lakes |
Farmland | 27.62% | 12864 | 8576 | Paddy field, dry land, nursery, orchard |
Woodland | 16.56% | 7714 | 5143 | Timber forest, shrub forest, planted forest |
Grassland | 4.41% | 2056 | 1371 | Native grassland, planted grassland |
Building | 20.39% | 9499 | 6333 | Low-rise and mid-rise buildings |
Road | 3.54% | 1648 | 1099 | Streets, country roads |
Impervious surface | 1.95% | 909 | 607 | Threshing ground, rolled surface |
Facility land | 15.59% | 7260 | 4840 | Greenhouses, agricultural facilities |
Class | Proportion | Train | Test | Subclass |
---|---|---|---|---|
Farmland | 22.01% | 11200 | 7500 | Paddy field, dry land, nursery, orchard |
Woodland | 21.88% | 11152 | 7441 | Timber forest, shrub forest, planted forest, grassland |
Building | 25.48% | 13047 | 8602 | Low-rise and mid-rise buildings |
Road | 5.92% | 3013 | 2019 | Streets, country roads |
Impervious surface | 5.58% | 2836 | 1902 | Threshing ground, rolled surface, facility land |
Vacant | 6.29% | 3208 | 2134 | Digging lands, bare surface |
Water | 12.85% | 6527 | 4391 | Rivers, rivulets, ponds, lakes |
Class | Proportion | Train | Test | Subclass |
---|---|---|---|---|
Farmland | 20.10% | 10029 | 6537 | Paddy field, dry land, nursery, orchard |
Woodland | 13.33% | 6619 | 4367 | Timber forest, shrub forest, planted forest, grassland |
Building | 27.03% | 13345 | 8935 | Low-rise and mid-rise buildings |
Road | 6.72% | 3331 | 2205 | Streets, country roads |
Impervious surface | 6.34% | 3176 | 2046 | Threshing ground, rolled surface, facility land |
Vacant | 5.12% | 2494 | 1722 | Digging lands, bare surface |
Water | 21.37% | 10459 | 7157 | Rivers, rivulets, ponds, lakes |
Class | Xiaoshan Dataset | Yuecheng Dataset | ||
---|---|---|---|---|
Proportion | Validate | Proportion | Validate | |
Farmland | 27.13% | 13581 | 14.44% | 5592 |
Woodland | 17.95% | 8986 | 19.89% | 7704 |
Building | 27.52% | 13773 | 29.18% | 11303 |
Road | 4.81% | 2408 | 7.66% | 2966 |
Impervious surface | 7.57% | 3787 | 14.34% | 5556 |
Vacant | 2.85% | 1428 | 4.55% | 1763 |
Water | 12.17% | 6091 | 9.95% | 3855 |
Land-Cover | ESSCNN | EMMCNN | OCNN | PCNN | SMCNN | SMMCNN | ORF | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | |
Farmland | 92.05% | 91.51% | 92.68% | 93.81% | 76.93% | 82.16% | 85.23% | 86.86% | 89.09% | 87.49% | 91.57% | 89.70% | 73.04% | 66.67% |
Woodland | 81.55% | 84.97% | 85.69% | 85.53% | 68.86% | 71.91% | 64.12% | 85.60% | 79.91% | 79.83% | 79.14% | 81.71% | 72.75% | 66.66% |
Grassland | 86.32% | 90.72% | 90.41% | 90.27% | 73.79% | 69.71% | 85.77% | 77.96% | 73.72% | 91.70% | 82.79% | 88.21% | 64.19% | 67.44% |
Dense building | 88.76% | 87.39% | 89.85% | 90.01% | 76.90% | 72.65% | 88.94% | 76.81% | 82.41% | 87.20% | 86.61% | 85.19% | 75.19% | 56.54% |
Sparse building | 83.15% | 82.12% | 86.37% | 86.52% | 57.21% | 60.16% | 75.80% | 75.09% | 74.73% | 75.83% | 77.35% | 80.22% | 23.02% | 48.62% |
Road | 73.61% | 75.61% | 77.12% | 77.86% | 57.99% | 56.16% | 55.22% | 68.00% | 72.31% | 65.60% | 71.78% | 68.61% | 52.76% | 58.03% |
Impervious surface | 86.89% | 82.40% | 87.56% | 87.45% | 66.29% | 65.76% | 76.57% | 76.13% | 82.09% | 77.86% | 80.94% | 83.54% | 54.37% | 53.66% |
Facility land | 89.73% | 89.93% | 92.51% | 90.49% | 66.99% | 71.61% | 85.34% | 82.24% | 88.67% | 80.37% | 89.86% | 82.53% | 33.85% | 51.16% |
Vacant | 92.09% | 89.21% | 92.05% | 93.33% | 65.96% | 64.85% | 89.38% | 74.38% | 79.50% | 79.41% | 82.41% | 87.72% | 32.66% | 52.54% |
Water | 90.10% | 90.40% | 91.69% | 90.57% | 81.39% | 85.57% | 82.37% | 88.81% | 91.02% | 84.56% | 90.82% | 87.46% | 84.72% | 83.47% |
AA | 86.43% | 88.59% | 69.23% | 78.87% | 81.35% | 83.33% | 56.65% | |||||||
OA | 86.45% | 88.56% | 70.57% | 79.08% | 81.55% | 83.67% | 61.72% | |||||||
KC | 0.847 | 0.871 | 0.667 | 0.763 | 0.792 | 0.816 | 0.564 |
Land-Cover | ESSCNN | EMMCNN | OCNN | PCNN | SMCNN | SMMCNN | ORF | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | |
Water | 83.59% | 81.99% | 84.46% | 82.89% | 69.11% | 61.27% | 64.94% | 82.52% | 61.03% | 80.14% | 59.67% | 86.25% | 81.49% | 83.82% |
Farmland | 90.19% | 88.02% | 90.29% | 90.02% | 75.94% | 83.21% | 91.29% | 78.63% | 80.28% | 83.41% | 86.62% | 83.87% | 87.26% | 82.35% |
Woodland | 91.89% | 90.77% | 92.63% | 92.13% | 85.48% | 78.19% | 91.14% | 80.52% | 88.22% | 85.88% | 88.66% | 89.67% | 87.22% | 83.49% |
Grassland | 59.47% | 75.45% | 70.39% | 79.01% | 20.79% | 48.83% | 19.33% | 78.41% | 47.83% | 71.71% | 49.26% | 79.02% | 19.12% | 77.63% |
Building | 92.33% | 89.25% | 92.89% | 89.65% | 91.51% | 79.16% | 92.04% | 86.08% | 92.72% | 82.10% | 94.19% | 82.28% | 91.55% | 79.53% |
Road | 57.68% | 67.39% | 59.25% | 70.54% | 13.64% | 58.41% | 24.31% | 65.00% | 35.74% | 52.48% | 33.16% | 62.42% | 45.98% | 72.21% |
Impervious surface | 65.27% | 72.87% | 74.00% | 75.22% | 32.28% | 54.80% | 39.37% | 67.93% | 48.24% | 66.60% | 41.21% | 79.99% | 18.16% | 75.67% |
Facility land | 89.39% | 89.92% | 90.32% | 90.20% | 84.93% | 71.92% | 83.59% | 83.51% | 90.11% | 74.34% | 93.04% | 76.43% | 82.40% | 76.98% |
AA | 78.73% | 81.78% | 59.21% | 63.25% | 68.02% | 68.23% | 64.15% | |||||||
OA | 87.13% | 88.35% | 75.78% | 81.39% | 80.27% | 82.63% | 80.77% | |||||||
KC | 0.841 | 0.857 | 0.701 | 0.767 | 0.756 | 0.784 | 0.761 |
Land-Cover | ESSCNN | EMMCNN | OCNN | PCNN | SMCNN | SMMCNN | ORF | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | |
Farmland | 89.99% | 89.45% | 91.91% | 90.31% | 75.68% | 81.60% | 90.21% | 83.22% | 85.77% | 88.77% | 87.21% | 89.57% | 78.15% | 64.58% |
Woodland | 86.97% | 86.06% | 88.85% | 88.01% | 70.62% | 74.71% | 83.21% | 82.74% | 81.77% | 84.42% | 83.58% | 85.52% | 65.62% | 71.60% |
Building | 91.73% | 87.85% | 92.31% | 89.89% | 88.74% | 78.39% | 86.92% | 88.06% | 89.03% | 85.43% | 91.74% | 84.17% | 90.14% | 68.60% |
Road | 67.82% | 72.71% | 69.98% | 75.08% | 56.46% | 47.71% | 56.36% | 68.36% | 59.34% | 65.16% | 63.96% | 63.60% | 54.99% | 70.69% |
Impervious surface | 62.42% | 73.74% | 69.97% | 76.43% | 29.03% | 49.65% | 65.11% | 60.17% | 62.93% | 62.19% | 60.27% | 70.26% | 5.78% | 52.67% |
Vacant | 91.06% | 89.20% | 92.10% | 93.01% | 70.20% | 70.41% | 85.12% | 89.44% | 83.32% | 83.31% | 83.42% | 87.58% | 37.78% | 75.28% |
Water | 89.72% | 92.00% | 90.66% | 92.93% | 86.29% | 82.23% | 85.33% | 90.40% | 89.43% | 84.03% | 87.40% | 88.15% | 81.86% | 90.76% |
AA | 82.81% | 85.11% | 68.14% | 78.89% | 78.80% | 79.65% | 59.19% | |||||||
OA | 86.91% | 88.63% | 75.22% | 83.48% | 83.16% | 84.42% | 70.98% | |||||||
KC | 0.839 | 0.860 | 0.694 | 0.796 | 0.793 | 0.808 | 0.634 |
Land-Cover | ESSCNN | EMMCNN | OCNN | PCNN | SMCNN | SMMCNN | ORF | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | |
Farmland | 65.38% | 75.09% | 67.04% | 77.68% | 51.37% | 77.40% | 69.60% | 71.87% | 58.24% | 76.97% | 57.86% | 77.02% | 63.14% | 60.72% |
Woodland | 66.31% | 58.39% | 68.02% | 60.02% | 61.75% | 55.10% | 65.49% | 58.95% | 63.40% | 56.27% | 66.16% | 56.13% | 62.40% | 52.81% |
Building | 83.23% | 72.05% | 83.64% | 72.87% | 83.06% | 67.03% | 76.50% | 75.48% | 82.74% | 69.16% | 85.90% | 69.38% | 76.93% | 66.49% |
Road | 32.64% | 43.55% | 36.50% | 42.52% | 33.75% | 27.69% | 34.64% | 43.95% | 31.36% | 40.48% | 34.41% | 39.57% | 29.95% | 34.94% |
Impervious surface | 19.65% | 35.59% | 24.48% | 41.94% | 12.29% | 32.88% | 30.15% | 36.84% | 22.22% | 34.06% | 19.42% | 42.02% | 2.62% | 31.93% |
Vacant | 43.43% | 26.81% | 43.70% | 30.74% | 43.76% | 26.39% | 34.70% | 25.08% | 43.26% | 24.85% | 40.70% | 23.43% | 31.43% | 23.29% |
Water | 76.03% | 76.41% | 76.99% | 76.47% | 75.60% | 64.67% | 75.71% | 75.89% | 76.61% | 71.92% | 75.21% | 76.41% | 69.72% | 80.97% |
AA | 55.24% | 57.20% | 51.65% | 55.25% | 53.98% | 54.24% | 48.03% | |||||||
OA | 65.62% | 67.20% | 60.46% | 65.43% | 63.24% | 64.20% | 60.06% | |||||||
KC | 0.568 | 0.588 | 0.507 | 0.567 | 0.541 | 0.551 | 0.493 |
Land-Cover | ESSCNN | EMMCNN | OCNN | PCNN | SMCNN | SMMCNN | ORF | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | |
Farmland | 91.60% | 90.34% | 92.96% | 91.90% | 84.36% | 73.73% | 92.93% | 80.02% | 86.99% | 88.92% | 91.68% | 85.05% | 79.13% | 64.13% |
Woodland | 75.83% | 75.84% | 78.33% | 79.85% | 52.95% | 48.59% | 63.10% | 68.79% | 68.33% | 70.80% | 66.30% | 76.94% | 36.86% | 46.92% |
Building | 89.61% | 87.56% | 90.69% | 89.29% | 83.56% | 75.09% | 85.15% | 83.28% | 87.35% | 85.73% | 91.02% | 83.25% | 88.16% | 62.81% |
Road | 73.53% | 76.41% | 76.94% | 76.74% | 41.35% | 63.60% | 61.71% | 68.67% | 76.12% | 60.25% | 73.34% | 64.66% | 46.75% | 56.49% |
Impervious surface | 59.33% | 65.74% | 64.76% | 70.46% | 25.09% | 42.96% | 40.65% | 61.32% | 42.82% | 63.17% | 42.84% | 70.82% | 8.17% | 44.71% |
Vacant | 83.31% | 88.51% | 89.48% | 89.38% | 45.20% | 48.88% | 79.95% | 78.87% | 73.68% | 82.97% | 70.92% | 89.21% | 3.91% | 61.03% |
Water | 92.96% | 91.83% | 93.54% | 93.18% | 83.47% | 88.93% | 90.51% | 89.99% | 92.38% | 87.88% | 91.07% | 89.88% | 86.78% | 91.45% |
AA | 80.88% | 83.81% | 59.43% | 73.43% | 75.38% | 75.31% | 49.97% | |||||||
OA | 85.62% | 87.52% | 71.11% | 80.30% | 81.68% | 82.73% | 67.19% | |||||||
KC | 0.822 | 0.846 | 0.641 | 0.756 | 0.774 | 0.785 | 0.584 |
Land-Cover | ESSCNN | EMMCNN | OCNN | PCNN | SMCNN | SMMCNN | ORF | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | UA | PA | |
Farmland | 67.56% | 52.05% | 69.16% | 49.80% | 63.37% | 40.72% | 81.61% | 40.81% | 60.64% | 52.52% | 75.94% | 46.60% | 74.37% | 36.91% |
Woodland | 49.17% | 56.08% | 46.44% | 56.74% | 35.34% | 42.60% | 32.81% | 55.47% | 47.36% | 55.38% | 34.49% | 57.70% | 26.81% | 50.96% |
Building | 80.55% | 73.13% | 81.30% | 76.35% | 76.55% | 71.34% | 77.94% | 74.80% | 76.72% | 75.41% | 83.90% | 68.63% | 86.11% | 60.85% |
Road | 46.83% | 56.67% | 53.33% | 53.97% | 43.44% | 36.14% | 42.74% | 54.93% | 59.24% | 34.08% | 53.19% | 43.92% | 37.92% | 44.56% |
Impervious surface | 33.89% | 44.27% | 34.38% | 45.27% | 21.27% | 38.60% | 17.49% | 38.38% | 32.17% | 46.70% | 26.22% | 47.58% | 3.04% | 33.73% |
Vacant | 19.48% | 20.90% | 16.80% | 21.73% | 14.84% | 10.90% | 25.73% | 16.74% | 13.14% | 18.23% | 8.77% | 23.88% | 1.58% | 15.38% |
Water | 80.71% | 78.61% | 82.52% | 76.96% | 50.49% | 76.65% | 77.75% | 79.92% | 78.44% | 77.75% | 80.06% | 78.55% | 79.91% | 71.91% |
AA | 54.03% | 54.85% | 43.61% | 50.87% | 52.53% | 51.80% | 44.25% | |||||||
OA | 60.39% | 60.93% | 50.51% | 55.69% | 58.14% | 58.48% | 52.51% | |||||||
KC | 0.513 | 0.521 | 0.394 | 0.460 | 0.490 | 0.488 | 0.407 |
Single-Scale | Beilun Dataset | Cixi Dataset | ||||||
---|---|---|---|---|---|---|---|---|
Single-Feature | Multi-Feature | Single-Feature | Multi-Feature | |||||
OA | KC | OA | KC | OA | KC | OA | KC | |
24 × 24 | 75.24% | 0.720 | 75.56% | 0.724 | 76.02% | 0.703 | 81.73% | 0.774 |
32 × 32 | 78.71% | 0.760 | 79.01% | 0.763 | 78.72% | 0.737 | 82.59% | 0.784 |
40 × 40 | 81.41% | 0.790 | 81.50% | 0.791 | 80.80% | 0.763 | 83.98% | 0.802 |
48 × 48 | 83.36% | 0.812 | 83.39% | 0.812 | 82.43% | 0.783 | 84.60% | 0.810 |
56 × 56 | 84.82% | 0.829 | 84.97% | 0.830 | 83.73% | 0.800 | 85.52% | 0.822 |
64 × 64 | 85.25% | 0.834 | 85.43% | 0.835 | 85.31% | 0.819 | 86.33% | 0.831 |
72 × 72 | 85.68% | 0.838 | 85.82% | 0.840 | 85.83% | 0.825 | 86.69% | 0.836 |
80 × 80 | 86.04% | 0.843 | 86.22% | 0.844 | 86.61% | 0.835 | 87.12% | 0.841 |
88 × 88 | 86.20% | 0.844 | 86.39% | 0.847 | 86.57% | 0.834 | 87.08% | 0.841 |
96 × 96 | 86.40% | 0.846 | 86.45% | 0.847 | 86.89% | 0.838 | 87.29% | 0.843 |
104 × 104 | 86.45% | 0.847 | 86.52% | 0.848 | 86.97% | 0.840 | 87.38% | 0.845 |
112 × 112 | 86.42% | 0.847 | 86.55% | 0.848 | 87.13% | 0.841 | 87.54% | 0.846 |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, S.; Li, C.; Qiu, S.; Gao, C.; Zhang, F.; Du, Z.; Liu, R. EMMCNN: An ETPS-Based Multi-Scale and Multi-Feature Method Using CNN for High Spatial Resolution Image Land-Cover Classification. Remote Sens. 2020, 12, 66. https://doi.org/10.3390/rs12010066
Zhang S, Li C, Qiu S, Gao C, Zhang F, Du Z, Liu R. EMMCNN: An ETPS-Based Multi-Scale and Multi-Feature Method Using CNN for High Spatial Resolution Image Land-Cover Classification. Remote Sensing. 2020; 12(1):66. https://doi.org/10.3390/rs12010066
Chicago/Turabian StyleZhang, Shuyu, Chuanrong Li, Shi Qiu, Caixia Gao, Feng Zhang, Zhenhong Du, and Renyi Liu. 2020. "EMMCNN: An ETPS-Based Multi-Scale and Multi-Feature Method Using CNN for High Spatial Resolution Image Land-Cover Classification" Remote Sensing 12, no. 1: 66. https://doi.org/10.3390/rs12010066
APA StyleZhang, S., Li, C., Qiu, S., Gao, C., Zhang, F., Du, Z., & Liu, R. (2020). EMMCNN: An ETPS-Based Multi-Scale and Multi-Feature Method Using CNN for High Spatial Resolution Image Land-Cover Classification. Remote Sensing, 12(1), 66. https://doi.org/10.3390/rs12010066