[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Deep Texton-Coherence Network for Camouflaged Object Detection

Published: 04 July 2022 Publication History

Abstract

Camouflaged object detection is a challenging visual task since the appearance and morphology of foreground objects and background regions are highly similar in nature. Recent CNN-based studies gradually integrated the high-level semantic information and the low-level local features of images through hierarchical and progressive structures to achieve camouflaged object detection. However, these methods ignore the <bold>spatial statistical properties</bold> of the local context, which is a critical cue for distinguishing and describing camouflaged objects. To address this problem, we propose a novel Deep Texton-Coherence Network (DTC-Net) that leverages the spatial organization of textons in the foreground and background regions as discriminative cues for camouflaged object detection. Specifically, a Local Bilinear module (LB) is devised to obtain the robust representation of texton to trivial details and illumination changes, by replacing the classic first-order linearization operations with bilinear second-order statistical operations in the convolution process. Next, these texton representations are associated with a Spatial Coherence Organization module (SCO) to capture irregular spatial coherence via a deformable convolutional strategy, and then the descriptions of the textons extracted by the LB module are used as weights to suppress features that are spatially adjacent but have different representations. Finally, the texton-coherence representation is integrated with the original features at different levels to achieve camouflaged object detection. Evaluation on the three most challenging camouflaged object detection datasets demonstrats the superiority of the proposed model when compared to the state-of-the-art methods. Furthermore, our ablation studies and performance analyses demonstrate the effectiveness of the texton-coherence module.

References

[1]
I. C. Cuthill et al., “Disruptive coloration and background pattern matching,” Nature, vol. 434, no. 7029, pp. 72–74, 2005.
[2]
T. Troscianko, C. P. Benton, P. G. Lovell, D. J. Tolhurst, and Z. Pizlo, “Camouflage and visual perception,” Philos. Trans. Roy. Soc. B Biol. Sci., vol. 364, no. 1516, pp. 449–461, 2009.
[3]
D.-P. Fan et al., “Inf-Net: Automatic COVID-19 lung infection segmentation from CT images,” IEEE Trans. Med. Imag., vol. 39, no. 8, pp. 2626–2637, Aug. 2020.
[4]
W. Zhai, J. Zhu, Y. Cao, and Z. Wang, “A generative adversarial network based framework for unsupervised visual surface inspection,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process., 2018, pp. 1283–1287.
[5]
Y. Zheng et al., “Detection of people with camouflage pattern via dense deconvolution network,” IEEE Signal Process. Lett., vol. 26, no. 1, pp. 29–33, Jan. 2019.
[6]
A. Tankus and Y. Yeshurun, “Convexity-based visual camouflage breaking,” Comput. Vis. Image Understanding, vol. 82, pp. 208–237, 2001.
[7]
N. U. Bhajantri and P. Nagabhushan, “Camouflage defect identification: A novel approach,” in Proc. IEEE 9th Int. Conf. Inf. Technol., 2006, pp. 145–148.
[8]
X. Feng, C. Guoying, H. Richang, and G. Jing, “Camouflage texture evaluation using a saliency map,” Multimedia Syst., vol. 21, pp. 169–175, 2015.
[9]
X. Zhang, C. Zhu, S. Wang, Y. Liu, and M. Ye, “A bayesian approach to camouflaged moving object detection,” IEEE Trans. Circuits Syst. Video Technol., vol. 27, no. 9, pp. 2001–2013, Sep. 2017.
[10]
M. Sultana, A. Mahmood, and S. K. Jung, “Unsupervised moving object detection in complex scenes using adversarial regularizations,” IEEE Trans. Multimedia, vol. 23, pp. 2005–2018, 2020.
[11]
M. Nawaz and H. Yan, “Saliency detection using deep features and affinity-based robust background subtraction,” IEEE Trans. Multimedia, vol. 23, pp. 2902–2916, 2020.
[12]
Y. Zhou, A. Mao, S. Huo, J. Lei, and S.-Y. Kung, “Salient object detection via fuzzy theory and object-level enhancement,” IEEE Trans. Multimedia, vol. 21, no. 1, pp. 74–85, Jan. 2019.
[13]
J. Li, Z. Pan, Q. Liu, and Z. Wang, “Stacked u-shape network with channel-wise attention for salient object detection,” IEEE Trans. Multimedia, vol. 23, pp. 1397–1409, 2020.
[14]
D.-P. Fan et al., “Camouflaged object detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2020, pp. 2774–2784.
[15]
B. Julesz, “Textons, the elements of texture perception, and their interactions,” Nature, vol. 290, no. 5802, pp. 91–97, 1981.
[16]
M. S. Allili, N. Baaziz, and M. Mejri, “Texture modeling using contourlets and finite mixtures of generalized gaussian distributions and applications,” IEEE Trans. Multimedia, vol. 16, no. 3, pp. 772–784, Mar. 2014.
[17]
W. Zhang, W. Zhang, K. Liu, and J. Gu, “A feature descriptor based on local normalized difference for real-world texture classification,” IEEE Trans. Multimedia, vol. 20, no. 4, pp. 880–888, Apr. 2018.
[18]
K. Zhu, Y. Cao, W. Zhai, and Z.-J. Zha, “One-shot texture retrieval using global grouping metric,” IEEE Trans. Multimedia, vol. 23, pp. 3726–3737, 2021.
[19]
S. Xie and Z. Tu, “Holistically-nested edge detection,” Int. J. Comput. Vis., vol. 125, no. 1, pp. 3–18, 2017.
[20]
P. Skurowski et al., “Animal camouflage analysis: Chameleon database,” Unpublished Manuscript, 2018.
[21]
T.-N. Le, T. V. Nguyen, Z. Nie, M.-T. Tran, and A. Sugimoto, “Anabranch network for camouflaged object segmentation,” Comput. Vis. Image Understanding, vol. 184, pp. 45–56, 2019.
[22]
F. Xue et al., “Camouflage performance analysis and evaluation framework based on features fusion,” Multimedia Tools Appl., vol. 75, pp. 4065–4082, 2016.
[23]
L. Song and W. Geng, “A new camouflage texture evaluation method based on wssim and nature image features,” in Proc. Int. Conf. Multimedia Technol., 2010, pp. 1–4.
[24]
G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, “Densely connected convolutional networks,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2017, pp. 2261–2269.
[25]
T.-Y. Lin, A. RoyChowdhury, and S. Maji, “Bilinear CNN models for fine-grained visual recognition,” in Proc. IEEE Int. Conf. Comput. Vis., 2015, pp. 1449–1457.
[26]
Y. Gao, O. Beijbom, N. Zhang, and T. Darrell, “Compact bilinear pooling,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 317–326.
[27]
X. Dai, J. Y. Ng, and L. S. Davis, “FASON: First and second order information fusion network for texture recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2017, pp. 6100–6108.
[28]
J. Dai et al., “Deformable convolutional networks,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 764–773.
[29]
X. Zhu, H. Hu, S. Lin, and J. Dai, “Deformable ConvNets V2: More deformable, better results,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 9300–9308.
[30]
K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 770–778.
[31]
D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” 2014, arXiv:1412.6980.
[32]
D.-P. Fan, M.-M. Cheng, Y. Liu, T. Li, and A. Borji, “Structure-measure: A new way to evaluate foreground maps,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 4558–4567.
[33]
R. Margolin, L. Zelnik-Manor, and A. Tal, “How to evaluate foreground maps?,” in IEEE Conf. Comput. Vis. Pattern Recognit., 2014, pp. 248–255.
[34]
F. Perazzi, P. Krähenbühl, Y. Pritch, and A. Hornung, “Saliency filters: Contrast based filtering for salient region detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2012, pp. 733–740.
[35]
D.-P. Fan et al., “Enhanced-alignment measure for binary foreground map evaluation,” in Proc. Int. Joint Conf. Artif. Intell., 2018, pp. 698–704.
[36]
J.-X. Zhao et al., “EgNet: Edge guidance network for salient object detection,” in IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 8778–8787.
[37]
K. Chen et al., “Hybrid task cascade for instance segmentation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 4969–4978.
[38]
Z. Wu, L. Su, and Q. Huang, “Cascaded partial decoder for fast and accurate salient object detection,” in IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 3902–3911.
[39]
T. Zhao and X. Wu, “Pyramid feature attention network for saliency detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 3080–3089.
[40]
X. Qin et al., “BASNet: Boundary-aware salient object detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 7471–7481.
[41]
Z. Huang, L. Huang, Y. Gong, C. Huang, and X. Wang, “Mask scoring R-CNN,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 6402–6411.
[42]
N. Liu, J. Han, and M.-H. Yang, “PiCANet: Learning pixel-wise contextual attention for saliency detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2018, pp. 3089–3098.
[43]
Z. Zhou, M. M. R. Siddiquee, N. Tajbakhsh, and J. Liang, “UNet: A nested U-net architecture for medical image segmentation,” in Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. Berlin, Germany: Springer, 2018.
[44]
H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2017, pp. 6230–6239.
[45]
K. He, G. Gkioxari, P. Dollar, and R. Girshick, “Mask R-CNN,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 2980–2988.
[46]
T.-Y. Lin et al., “Feature pyramid networks for object detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2017, pp. 936–944.
[47]
L. v. d. Maaten and G. Hinton, “Visualizing data using T-SNE,” J. Mach. Learn. Res., vol. 9, pp. 2579–2605, 2008.
[48]
M. Cimpoi, S. Maji, I. Kokkinos, S. Mohamed, and A. Vedaldi, “Describing textures in the wild,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2014, pp. 3606–3613.
[49]
P. Mallikarjuna et al., “The kth-tips2 database,” Computational Vision and Active Perception Laboratory. Stockholm, Sweden: A. Bonniers Forlag, 2006.
[50]
L. Sharan, C. Liu, R. Rosenholtz, and E. H. Adelson, “Recognizing materials using perceptually inspired features,” Int. J. Comput. Vis., vol. 103, pp. 348–371, 2013.
[51]
T.-Y. Lin and S. Maji, “Visualizing and understanding deep texture representations,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 2791–2799.
[52]
L. Sharan, R. Rosenholtz, and E. Adelson, “Material perception: What can you see in a brief glance?,” J. Vis., vol. 9, pp. 784–784, 2009.
[53]
B. Caputo, E. Hayman, and P. Mallikarjuna, “Class-specific material categorisation,” in Proc. IEEE Int. Conf. Comput. Vis., 2005, pp. 1597–1604.

Cited By

View all

Index Terms

  1. Deep Texton-Coherence Network for Camouflaged Object Detection
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Please enable JavaScript to view thecomments powered by Disqus.

          Information & Contributors

          Information

          Published In

          cover image IEEE Transactions on Multimedia
          IEEE Transactions on Multimedia  Volume 25, Issue
          2023
          8932 pages

          Publisher

          IEEE Press

          Publication History

          Published: 04 July 2022

          Qualifiers

          • Research-article

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • Downloads (Last 12 months)0
          • Downloads (Last 6 weeks)0
          Reflects downloads up to 03 Jan 2025

          Other Metrics

          Citations

          Cited By

          View all
          • (2025)Improving underwater camouflage object segmentation with dual-decoder attention networkThe Journal of Supercomputing10.1007/s11227-024-06584-x81:1Online publication date: 1-Jan-2025
          • (2024)Decoupling and Integration Network for Camouflaged Object DetectionIEEE Transactions on Multimedia10.1109/TMM.2024.336071026(7114-7129)Online publication date: 31-Jan-2024
          • (2024)UEDG:Uncertainty-Edge Dual Guided Camouflage Object DetectionIEEE Transactions on Multimedia10.1109/TMM.2023.329509526(4050-4060)Online publication date: 1-Jan-2024
          • (2024)A Universal Multi-View Guided Network for Salient Object and Camouflaged Object DetectionIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2024.341760734:11_Part_1(11184-11197)Online publication date: 21-Jun-2024

          View Options

          View options

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media