An End-to-End Conditional Random Fields and Skip-Connected Generative Adversarial Segmentation Network for Remote Sensing Images
"> Figure 1
<p>Framework of several segmentation methods.</p> "> Figure 2
<p>Structure of generative model.</p> "> Figure 3
<p>Improvement in encoder-decoder model.</p> "> Figure 4
<p>The iteration in CRF layer.</p> "> Figure 5
<p>Structure of discriminative model.</p> "> Figure 6
<p>Flowchart of training and testing process.</p> "> Figure 7
<p>The experimental data 1. (<b>a</b>) Pauli SAR image; (<b>b</b>) The ground truth.</p> "> Figure 8
<p>The experimental data 2. (<b>a</b>) Remote sensing image; (<b>b</b>) The ground truth.</p> "> Figure 9
<p>Segmentation results 1.</p> "> Figure 10
<p>Segmentation results 2.</p> "> Figure 11
<p>Segmentation results 3.</p> "> Figure 12
<p>Segmentation results 4.</p> "> Figure 13
<p>Statistics on the evaluation indexes of Data Set 2.</p> ">
Abstract
:1. Introduction
1.1. Background
1.2. Problems and Motivations
1.3. Contribution
- (1)
- CRFAS proposes an end-to-end generative adversarial segmentation network based on Bayesian framework. In this algorithm, the joint training of prior network, likelihood network and posterior network realizes the estimation from prior to posterior. The prior network provides the prior information for the likelihood network and the posterior network. The prior information is combined with the discriminator of the posterior network to improve the likelihood network.
- (2)
- Based on the structure of GAN, CRFAS combines pixel-level and regional-level evaluation methods to calculate loss function. Since the pixel-level evaluation methods cannot discriminate spatial structure, CRFAS utilizes a region-level evaluation method, which can judge the similarity between the predicted label graph and the ground truth, to introduce spatial structure as a constraint. Through the discriminative model, CRFAS can obtain the adversarial loss by region-level evaluation method. Then, a new loss function considering the pixel information and spatial structure information is defined. The new loss function including cross-entropy loss and adversarial loss, guides the training of the segmentation network and improves the accuracy.
- (3)
- The integrated training of skip connected encoder-decoder network structure and CRF layer combines the advantages of both. The existing segmentation networks mainly use convolutional neural networks and post-optimized CRF. CRF only refines the segmentation results after training the convolutional neural networks, rather than participating in the training process of neural network parameters. CRFAS changes the way of training convolutional neural networks and conditional random fields separately before. By combining skip connected encoder-decoder network structure and CRF layer, the results of CRF can guide the training of CRF, and the result is improved by taking more information into account.
2. Methodology
2.1. Framework
2.2. Generative Model
2.2.1. Conditional Random Fields
2.2.2. Unitary Potential
2.2.3. CRF Layer
2.3. Discriminative Model
2.4. Loss Function
2.5. Model Training
3. Flowchart
4. Experiment
4.1. Experiment Data
4.2. Experiment Result
4.2.1. The Confusion Matrix and Overall Accuracy
4.2.2. F1 Score and mIOU
4.3. Discussion
- (1)
- A conditional generative adversarial segmentation network is proposed based on the Bayesian framework. The networks joint training and adversarial learning make the segmentation results as close as possible to the ground truth. While FCN is a segmentation network, and Deeplab adds conditional random fields as post-processing after segmentation network. We directly observe the segmented results and can see that at the lower part of Figure 12, FCN and Deeplab labeled farmland as water, but CRFAS and pixel2pixel did not.
- (2)
- To obtain details and consider global information, the skip-connected encoder-decoder architecture is integrated with CRF layer to form an end-to-end generative model, so as to improve the accuracy of segmentation. FCN8 and Pixel2pixel use skip connections between different layers to get detailed information, but the global information is not taken into account. Observing the segmentation results directly, it can be seen that in Figure 11 and Figure 12, CRFAS and Deeplab succeeded in segmenting rivers (blue), while FCN and pixel2pixel failed.
- (3)
- A new loss function including the cross-entropy loss and the adversarial loss is defined to guide the training of the whole segmentation network. In contrast, FCN and Deeplab only consider the difference of each pixel to calculate the cross-entropy loss, while pixel2pixel does not take the pixel-level difference into account. Combined with the above, for these two data sets, CRFAS completes the segmentation task and has highest accuracy and mIoU among the four semantic segmentation methods.
5. Conclusions
Author Contributions
Acknowledgments
Conflicts of Interest
References
- Li, D.; Zhang, G.; Wu, Z.; Yi, L. An edge embedded marker-based watershed algorithm for high spatial resolution remote sensing image segmentation. IEEE Trans. Image Process. 2010, 19, 2781–2787. [Google Scholar] [PubMed]
- Zhang, X.; Xiao, P.; Feng, X.; Wang, J.; Zuo, W. Hybrid region merging method for segmentation of high-resolution remote sensing images. ISPRS J. Photogramm. Remote Sens. 2014, 98, 19–28. [Google Scholar] [CrossRef]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Region-based convolutional networks for accurate object detection and segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 142–158. [Google Scholar] [CrossRef] [PubMed]
- Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.S.; Fraundorfer, F. Deep Learning in Remote Sensing: A Comprehensive Review and List of Resources. IEEE Geosci. Remote Sens. Mag. 2018, 5, 8–36. [Google Scholar] [CrossRef]
- Maggiori, E.; Tarabalka, Y.; Charpiat, G.; Alliez, P. Convolutional Neural Networks for Large-Scale Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2016, 55, 645–657. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. Comput. Sci. 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Noh, H.; Hong, S.; Han, B. Learning deconvolution network for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1520–1528. [Google Scholar]
- Peng, C.; Zhang, X.; Yu, G.; Luo, G.; Sun, J. Large Kernel Matters–Improve Semantic Segmentation by Global Convolutional Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4353–4361. [Google Scholar]
- Nogueira, K.; Miranda, W.O.; Dos Santos, J.A. Improving spatial feature representation from aerial scenes by using convolutional networks. In Proceedings of the 2015 28th SIBGRAPI Conference on Graphics, Patterns and Images, Salvador, Brazil, 26–29 August 2015; pp. 289–296. [Google Scholar]
- Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional architecture for fast feature embedding. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 675–678. [Google Scholar]
- Sermanet, P.; Eigen, D.; Zhang, X.; Mathieu, M.; Fergus, R.; LeCun, Y. Overfeat: Integrated recognition, localization and detection using convolutional networks. arXiv Preprint, 2013; arXiv:1312.6229. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
- Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
- Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
- Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef]
- Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. arXiv, 2016; arXiv:1612.01105. [Google Scholar]
- Krähenbühl, P.; Koltun, V. Efficient inference in fully connected crfs with gaussian edge potentials. In Proceedings of the Advances in Neural Information Processing Systems, Granada, Spain, 12–15 December 2011; pp. 109–117. [Google Scholar]
- Liu, F.; Lin, G.; Shen, C. CRF learning with CNN features for image segmentation. Pattern Recognit. 2015, 48, 2983–2992. [Google Scholar] [CrossRef] [Green Version]
- Liu, F.; Lin, G.; Shen, C. Discriminative training of deep fully connected continuous CRFs with task-specific loss. IEEE Trans. Image Process. 2017, 26, 2127–2136. [Google Scholar] [CrossRef] [PubMed]
- Zheng, S.; Jayasumana, S.; Romera-Paredes, B.; Vineet, V.; Su, Z.; Du, D.; Huang, C.; Torr, P.H. Conditional random fields as recurrent neural networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1529–1537. [Google Scholar]
- Luc, P.; Couprie, C.; Chintala, S.; Verbeek, J. Semantic segmentation using adversarial networks. arXiv Preprint, 2016; arXiv:1611.08408. [Google Scholar]
- Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
- Zhu, W.; Xiang, X.; Tran, T.D.; Xie, X. Adversarial deep structural networks for mammographic mass segmentation. arXiv Preprint, 2016; arXiv:1612.05970. [Google Scholar]
- Ma, F.; Gao, F.; Sun, J.; Zhou, H.; Hussain, A. Weakly Supervised Segmentation of SAR Imagery Using Superpixel and Hierarchically Adversarial CRF. Remote Sens. 2019, 11, 512. [Google Scholar] [CrossRef]
- Souly, N.; Spampinato, C.; Shah, M. Semi supervised semantic segmentation using generative adversarial network. In Proceedings of the IEEE International Conference on Computer Vision, Honolulu, HI, USA, 21–26 July 2017; pp. 5688–5696. [Google Scholar]
- Fu, C.; Lee, S.; Joon Ho, D.; Han, S.; Salama, P.; Dunn, K.W.; Delp, E.J. Three dimensional fluorescence microscopy image synthesis and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2221–2229. [Google Scholar]
- Huo, Y.; Xu, Z.; Bao, S.; Bermudez, C.; Plassard, A.J.; Liu, J.; Yao, Y.; Assad, A.; Abramson, R.G.; Landman, B.A. Splenomegaly segmentation using global convolutional kernels and conditional generative adversarial networks. SPIE Proc. 2018, 10574, 1057409. [Google Scholar]
- Hariharan, B.; Arbelaez, P.; Girshick, R.; Malik, J. Hypercolumns for Object Segmentation and Fine-Grained Localization. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
- Kluckner, S.; Bischof, H. Semantic classification by covariance descriptors within a randomized forest. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 665–672. [Google Scholar]
- Li, W.; He, C.; Fang, J.; Zheng, J.; Fu, H.; Yu, L. Semantic Segmentation-Based Building Footprint Extraction Using Very High-Resolution Satellite Images and Multi-Source GIS Data. Remote Sens. 2019, 11, 403. [Google Scholar] [CrossRef]
- Zhang, Z.; Wang, Y. JointNet: A Common Neural Network for Road and Building Extraction. Remote Sens. 2019, 11, 696. [Google Scholar] [CrossRef]
- Papadomanolaki, M.; Vakalopoulou, M.; Karantzalos, K. A Novel Object-Based Deep Learning Framework for Semantic Segmentation of Very High-Resolution Remote Sensing Data: Comparison with Convolutional and Fully Convolutional Networks. Remote Sens. 2019, 11, 684. [Google Scholar] [CrossRef]
- Panboonyuen, T.; Jitkajornwanich, K.; Lawawirojwong, S.; Srestasathiern, P.; Vateekul, P. Semantic Segmentation on Remotely Sensed Images Using an Enhanced Global Convolutional Network with Channel Attention and Domain Specific Transfer Learning. Remote Sens. 2019, 11, 83. [Google Scholar] [CrossRef]
- Pan, X.; Gao, L.; Marinoni, A.; Bing, Z.; Gamba, P. Semantic Labeling of High Resolution Aerial Imagery and LiDAR Data with Fine Segmentation Network. Remote Sens. 2018, 10, 743. [Google Scholar] [CrossRef]
- Mao, X.; Shen, C.; Yang, Y.B. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2802–2810. [Google Scholar]
- Kumar, P.; Nagar, P.; Arora, C.; Gupta, A. U-Segnet: Fully Convolutional Neural Network Based Automated Brain Tissue Segmentation Tool. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 3503–3507. [Google Scholar]
- Jiang, J.; Zheng, L.; Luo, F.; Zhang, Z. Rednet: Residual encoder-decoder network for indoor rgb-d semantic segmentation. arXiv Preprint, 2018; arXiv:1806.01054. [Google Scholar]
- Son, J.; Park, S.J.; Jung, K.H. Retinal vessel segmentation in fundoscopic images with generative adversarial networks. arXiv Preprint, 2017; arXiv:1706.09318. [Google Scholar]
- Liu, Y.; Fan, B.; Wang, L.; Bai, J.; Xiang, S.; Pan, C. Semantic labeling in very high resolution images via a self-cascaded convolutional neural network. ISPRS J. Photogramm. Remote Sens. 2018, 145, 78–95. [Google Scholar] [CrossRef] [Green Version]
- Wei, X.; Guo, Y.; Gao, X.; Yan, M.; Sun, X. A new semantic segmentation model for remote sensing images. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 1776–1779. [Google Scholar]
- Cheng, W.; Yang, W.; Wang, M.; Wang, G.; Chen, J. Context Aggregation Network for Semantic Labeling in Aerial Images. Remote Sens. 2019, 11, 1158. [Google Scholar] [CrossRef]
- Tong, X.; Xia, G.; Lu, Q.; Shen, H.; Li, S.; You, S.; Zhang, L. Learning Transferable Deep Models for Land-Use Classification with High-Resolution Remote Sensing Images. arXiv Preprint, 2018; arXiv:1807.05713. [Google Scholar]
Method | Category | Farmland | Forest | Others | Road | Building | Overall Acc. |
---|---|---|---|---|---|---|---|
Farmland | 46.17 | 8.97 | 38.63 | 2.16 | 4.07 | ||
Forest | 1.67 | 87.56 | 3.90 | 1.25 | 5.62 | ||
FCN | Others | 4.33 | 1.35 | 87.51 | 3.87 | 2.95 | 81.83 |
Road | 4.90 | 2.08 | 28.58 | 59.34 | 5.09 | ||
Building | 0.04 | 2.28 | 4.61 | 1.08 | 91.99 | ||
Farmland | 45.03 | 21.91 | 31.43 | 0.39 | 1.24 | ||
Forest | 0.74 | 89.01 | 5.14 | 1.61 | 3.50 | ||
Deeplab | Others | 1.64 | 1.16 | 92.26 | 3.27 | 1.66 | 83.87 |
Road | 0.87 | 4.19 | 34.30 | 55.09 | 5.56 | ||
Building | 0 | 0.76 | 4.13 | 0.60 | 94.48 | ||
Farmland | 35.88 | 6.23 | 23.77 | 2.55 | 31.56 | ||
Forest | 2.06 | 84.30 | 5.05 | 2.21 | 6.38 | ||
Pix2pix | Others | 2.18 | 0.61 | 88.02 | 4.79 | 4.40 | 82.43 |
Road | 0.28 | 0.87 | 20.02 | 75.00 | 3.83 | ||
Building | 0.42 | 2.75 | 5.41 | 1.81 | 89.62 | ||
Farmland | 47.81 | 9.70 | 23.06 | 2.77 | 15.15 | ||
Forest | 1.27 | 83.51 | 5.06 | 2.29 | 7.86 | ||
Pro. Aproach | Others | 2.22 | 0.93 | 93.72 | 3.12 | 0.01 | 84.74 |
Road | 0.67 | 2.45 | 23.42 | 68.79 | 4.67 | ||
Building | 0.47 | 3.71 | 4.89 | 2.22 | 88.70 |
Method | Category | Farmland | Water | Building | Grassland | Forest | Overall Acc. |
---|---|---|---|---|---|---|---|
FCN | Farmland | 98.39 | 0.08 | 0.01 | 1.27 | 0.25 | 76.31 |
Water | 24.99 | 75.01 | 0 | 0 | 0 | ||
Building | 6.40 | 0.11 | 93.49 | 0 | 0 | ||
Grassland | 98.71 | 0.06 | 0 | 0.64 | 0.59 | ||
Forest | 31.37 | 2.24 | 0 | 3.49 | 62.90 | ||
Deeplab | Farmland | 93.10 | 5.02 | 0 | 0.77 | 1.11 | 77.99 |
Water | 0.48 | 99.47 | 0 | 0 | 0.04 | ||
Building | 2.13 | 0 | 97.87 | 0 | 0 | ||
Grassland | 67.42 | 0.34 | 0 | 3.29 | 28.95 | ||
Forest | 0.42 | 0 | 0 | 0 | 99.58 | ||
Pix2pix | Farmland | 91.53 | 0 | 0 | 0 | 8.47 | 71.78 |
Water | 100 | 0 | 0 | 0 | 0 | ||
Building | 13.52 | 0 | 85.46 | 1.01 | 0 | ||
Grassland | 98.31 | 0 | 0 | 0 | 1.69 | ||
Forest | 32.65 | 0 | 0 | 0 | 67.35 | ||
Pro. Aproach | Farmland | 95.69 | 0.71 | 0 | 2.61 | 0.99 | 78.27 |
Water | 0.82 | 28.57 | 0 | 0 | 70.61 | ||
Building | 1.65 | 0 | 97.10 | 1.25 | 0 | ||
Grassland | 85.44 | 0.04 | 0 | 11.21 | 3.32 | ||
Forest | 22.00 | 0.81 | 0 | 0.06 | 77.13 |
Method | Category | Farmland | Water | Building | Grassland | Forest | Overall Acc. |
---|---|---|---|---|---|---|---|
FCN | Farmland | 99.69 | 0.06 | 0.09 | 0.12 | 0.04 | 92.45 |
Water | 50.07 | 49.22 | 0.69 | 0.02 | 0.02 | ||
Building | 33.20 | 0.10 | 66.67 | 0 | 0.03 | ||
Grassland | 0 | 0 | 0 | 0 | 0 | ||
Forest | 13.73 | 0.06 | 0.15 | 0.71 | 85.35 | ||
Deeplab | Farmland | 98.91 | 1.01 | 0.02 | 0 | 0.06 | 98.03 |
Water | 1.38 | 98.57 | 0.04 | 0 | 0 | ||
Building | 9.18 | 0.62 | 90.20 | 0 | 0 | ||
Grassland | 0 | 0 | 0 | 0 | 0 | ||
Forest | 0.14 | 0.08 | 0 | 0 | 99.78 | ||
Pix2pix | Farmland | 98.99 | 0.03 | 0.70 | 0.04 | 0.25 | 71.78 |
Water | 73.70 | 20.45 | 0.62 | 0.03 | 5.19 | ||
Building | 29.00 | 0 | 69.96 | 2.04 | 0 | ||
Grassland | 0 | 0 | 0 | 0 | 0 | ||
Forest | 96.56 | 0 | 0 | 0 | 3.44 | ||
Pro. Aproach | Farmland | 99.26 | 0.33 | 0.08 | 0.14 | 0.19 | 98.17 |
Water | 3.36 | 96.16 | 0.04 | 0 | 0.44 | ||
Building | 3.83 | 0 | 95.96 | 0.21 | 0 | ||
Grassland | 0 | 0 | 0 | 0 | 0 | ||
Forest | 7.27 | 18.28 | 0 | 0 | 74.45 |
Method | Category | Farmland | Water | Building | Grassland | Forest | Overall Acc. |
---|---|---|---|---|---|---|---|
FCN | Farmland | 98.90 | 0.50 | 0.56 | 0 | 0.04 | 93.16 |
Water | 68.99 | 30.99 | 0 | 0 | 0.01 | ||
Building | 29.41 | 0.40 | 70.19 | 0 | 0 | ||
Grassland | 0 | 0 | 0 | 0 | 0 | ||
Forest | 0 | 0 | 0 | 0 | 0 | ||
Deeplab | Farmland | 98.08 | 1.14 | 0.63 | 0 | 0.15 | 97.54 |
Water | 7.98 | 88.35 | 0 | 0 | 3.66 | ||
Building | 3.49 | 0.53 | 95.98 | 0 | 0 | ||
Grassland | 0 | 0 | 0 | 0 | 0 | ||
Forest | 0 | 0 | 0 | 0 | 0 | ||
Pixel2pixel | Farmland | 97.83 | 0.02 | 2.04 | 0.02 | 0.09 | 95.59 |
Water | 78.15 | 14.36 | 0.47 | 0 | 7.01 | ||
Building | 3.86 | 0 | 96.09 | 0.05 | 0 | ||
Grassland | 0 | 0 | 0 | 0 | 0 | ||
Forest | 0 | 0 | 0 | 0 | 0 | ||
Pro. Aproach | Farmland | 99.32 | 0 | 0.58 | 0.08 | 0.01 | 98.81 |
Water | 9.50 | 87.85 | 0.09 | 0 | 2.56 | ||
Building | 2.14 | 0.03 | 97.65 | 0.18 | 0 | ||
Grassland | 0 | 0 | 0 | 0 | 0 | ||
Forest | 0 | 0 | 0 | 0 | 0 |
Method | Category | Farmland | Water | Building | Grassland | Forest | Overall Acc. |
---|---|---|---|---|---|---|---|
FCN | Farmland | 93.99 | 0.42 | 4.93 | 0.22 | 0.44 | 87.11 |
Water | 25.68 | 72.38 | 1.16 | 0.07 | 0.71 | ||
Building | 19.05 | 0.65 | 79.95 | 0.02 | 0.33 | ||
Grassland | 98.71 | 0.06 | 0 | 0.64 | 0.59 | ||
Forest | 50.22 | 0.88 | 0.11 | 4.34 | 44.45 | ||
Deeplab | Farmland | 95.34 | 1.80 | 2.32 | 0.04 | 0.49 | 92.38 |
Water | 11.98 | 87.74 | 0.13 | 0 | 0.15 | ||
Building | 4.65 | 0.26 | 95.02 | 0 | 0.07 | ||
Grassland | 67.41 | 0.34 | 0 | 3.30 | 28.94 | ||
Forest | 31.31 | 6.29 | 0.08 | 3.70 | 58.62 | ||
Pixel2pixel | Farmland | 94.43 | 0.17 | 3.18 | 0.34 | 1.87 | 87.04 |
Water | 21.89 | 74.40 | 0.39 | 0.01 | 3.31 | ||
Building | 18.51 | 0.13 | 80.44 | 0.91 | 0.01 | ||
Grassland | 98.31 | 0 | 0 | 0 | 1.69 | ||
Forest | 82.50 | 0 | 0 | 0 | 17.50 | ||
Pro. Aproach | Farmland | 97.73 | 0.30 | 1.03 | 0.43 | 0.50 | 94.14 |
Water | 6.30 | 92.75 | 0.21 | 0.04 | 0.70 | ||
Building | 4.43 | 0.07 | 95.22 | 0.27 | 0 | ||
Grassland | 85.43 | 0.04 | 0 | 11.21 | 3.32 | ||
Forest | 64.64 | 1.31 | 0.12 | 0.11 | 33.83 |
Data | Method | F1 Score | mIOU | ||||
---|---|---|---|---|---|---|---|
Farmland | Forest | Others | Road | Building | |||
Data Set 1 | FCN | 0.4871 | 0.8840 | 0.8558 | 0.6698 | 0.8697 | 0.6270 |
Deeplab | 0.5634 | 0.8723 | 0.8781 | 0.6489 | 0.9105 | 0.6529 | |
Pix2pix | 0.4550 | 0.8807 | 0.8752 | 0.7555 | 0.8001 | 0.6266 | |
CRFAS | 0.5717 | 0.8570 | 0.9032 | 0.7349 | 0.8580 | 0.6612 |
Data | Method | F1 Score | mIOU | ||||
---|---|---|---|---|---|---|---|
Farmland | Water | Building | Grassland | Forest | |||
Image1 | FCN | 0.8484 | 0.3413 | 0.9656 | 0.0118 | 0.7617 | 0.4994 |
Deeplab | 0.8785 | 0.0713 | 0.9892 | 0.0621 | 0.7998 | 0.4995 | |
Pix2pix | 0.8104 | 0 | 0.9216 | 0 | 0.6239 | 0.3978 | |
CRFAS | 0.8565 | 0.1003 | 0.9849 | 0.1852 | 0.8195 | 0.5137 | |
Image2 | FCN | 0.9556 | 0.6560 | 0.7940 | 0 | 0.9130 | 0.5803 |
Deeplab | 0.9882 | 0.9355 | 0.9472 | 0 | 0.9882 | 0.7464 | |
Pix2pix | 0.9349 | 0.3385 | 0.7880 | 0 | 0.0530 | 0.3518 | |
CRFAS | 0.9915 | 0.9364 | 0.9760 | 0 | 0.8132 | 0.7004 | |
Image3 | FCN | 0.9604 | 0.4106 | 0.8094 | 0 | 0 | 0.3724 |
Deeplab | 0.9862 | 0.7635 | 0.9615 | 0 | 0 | 0.5032 | |
Pix2pix | 0.9747 | 0.2496 | 0.9240 | 0 | 0 | 0.3904 | |
CRFAS | 0.9934 | 0.9338 | 0.9715 | 0 | 0 | 0.5615 | |
Data set 2 | FCN | 0.9148 | 0.8247 | 0.7681 | 0.0099 | 0.5521 | 0.5109 |
Deeplab | 0.9508 | 0.8803 | 0.9141 | 0.0569 | 0.6398 | 0.6068 | |
Pix2pix | 0.9156 | 0.8479 | 0.8111 | 0.0000 | 0.1925 | 0.4738 | |
CRFAS | 0.9609 | 0.9523 | 0.9471 | 0.1543 | 0.4468 | 0.6209 |
Method | FCN | Deeplab | Pixel2pixel | CRFAS |
---|---|---|---|---|
Time (h) | 16.9 | 25.3 | 26.7 | 81.2 |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
He, C.; Fang, P.; Zhang, Z.; Xiong, D.; Liao, M. An End-to-End Conditional Random Fields and Skip-Connected Generative Adversarial Segmentation Network for Remote Sensing Images. Remote Sens. 2019, 11, 1604. https://doi.org/10.3390/rs11131604
He C, Fang P, Zhang Z, Xiong D, Liao M. An End-to-End Conditional Random Fields and Skip-Connected Generative Adversarial Segmentation Network for Remote Sensing Images. Remote Sensing. 2019; 11(13):1604. https://doi.org/10.3390/rs11131604
Chicago/Turabian StyleHe, Chu, Peizhang Fang, Zhi Zhang, Dehui Xiong, and Mingsheng Liao. 2019. "An End-to-End Conditional Random Fields and Skip-Connected Generative Adversarial Segmentation Network for Remote Sensing Images" Remote Sensing 11, no. 13: 1604. https://doi.org/10.3390/rs11131604
APA StyleHe, C., Fang, P., Zhang, Z., Xiong, D., & Liao, M. (2019). An End-to-End Conditional Random Fields and Skip-Connected Generative Adversarial Segmentation Network for Remote Sensing Images. Remote Sensing, 11(13), 1604. https://doi.org/10.3390/rs11131604