[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Attention-guided Multi-modality Interaction Network for RGB-D Salient Object Detection

Published: 23 October 2023 Publication History

Abstract

The past decade has witnessed great progress in RGB-D salient object detection (SOD). However, there are two bottlenecks that limit its further development. The first one is low-quality depth maps. Most existing methods directly use raw depth maps to perform detection, but low-quality depth images can bring negative impacts to the detection performance. Hence, it is not desirable to utilize depth maps indiscriminately. The other one is how to effectively predict salient maps with clear boundary and complete salient region. To address these problems, an Attention-Guided Multi-Modality Interaction Network (AMINet) is proposed. First, we propose a new quality enhancement strategy for unreliable depth images, named Depth Enhancement Module (DEM). With respect to the second issue, we propose Cross-Modality Attention Module (CMAM) to rapidly locate salient region. The Boundary-Aware Module (BAM) is designed to utilize high-level feature to guide the low-level feature generation in a top-down way to make up for the dilution of the boundary. To further improve the accuracy, we propose Atrous Refined Block (ARB) to adaptively compensate for the shortcoming of atrous convolution. By integrating these interactive modules, features from depth and RGB streams can be refined efficiently, which consequently boosts the detection performance. Experimental results demonstrate the proposed AMINet exceeds state-of-the-art (SOTA) methods on several public RGB-D datasets.

References

[1]
Radhakrishna Achanta, Sheila Hemami, Francisco Estrada, and Sabine Susstrunk. 2009. Frequency-tuned salient region detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1597–1604. DOI:
[2]
Ali Borji, Ming-Ming Cheng, Qibin Hou, Huaizu Jiang, and Jia Li. 2019. Salient object detection: A survey. Comput. Vis. Media 5, 2 (2019), 117–150.
[3]
Ali Borji, Ming-Ming Cheng, Huaizu Jiang, and Jia Li. 2015. Salient object detection: A benchmark. IEEE Trans. Image Process. 24, 12 (2015), 5706–5722. DOI:
[4]
Baian Chen, Zhilei Chen, Xiaowei Hu, Jun Xu, Haoran Xie, Jing Qin, and Mingqiang Wei. 2023. Dynamic message propagation network for RGB-D and video salient object detection. ACM Trans. Multim. Comput., Commun. Applic. 20, 1 (2023), 21 pages.
[5]
Chenglizhao Chen, Jipeng Wei, Chong Peng, and Hong Qin. 2021. Depth-quality-aware salient object detection. IEEE Trans. Image Process. 30 (2021), 2350–2363.
[6]
Chenglizhao Chen, Jipeng Wei, Chong Peng, Weizhong Zhang, and Hong Qin. 2020. Improved saliency detection in RGB-D images using two-phase depth estimation and selective deep fusion. IEEE Trans. Image Process. 29 (2020), 4296–4307.
[7]
Gang Chen, Feng Shao, Xiongli Chai, Hangwei Chen, Qiuping Jiang, Xiangchao Meng, and Yo-Sung Ho. 2022. Modality-induced transfer-fusion network for RGB-D and RGB-T salient object detection. IEEE Trans. Circ. Syst. Vid. Technol. 33, 4 (2022), 1787–1801.
[8]
Qian Chen, Keren Fu, Ze Liu, Geng Chen, Hongwei Du, Bensheng Qiu, and Ling Shao. 2021. EF-Net: A novel enhancement and fusion network for RGB-D saliency detection. Pattern Recog. 112 (2021), 107740.
[9]
Yupeng Cheng, Huazhu Fu, Xingxing Wei, Jiangjian Xiao, and Xiaochun Cao. 2014. Depth enhanced saliency detection method. In Proceedings of the International Conference on Internet Multimedia Computing and Service. 23–27.
[10]
David Eigen and Rob Fergus. 2015. Predicting depth, surface normals and semantic labels with a common multi-scale convolutional architecture. In Proceedings of the IEEE International Conference on Computer Vision. 2650–2658.
[11]
David Eigen, Christian Puhrsch, and Rob Fergus. 2014. Depth map prediction from a single image using a multi-scale deep network. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Vol. 2. 2366–2374.
[12]
Deng-Ping Fan, Ming-Ming Cheng, Yun Liu, Tao Li, and Ali Borji. 2017. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision. 4548–4557.
[13]
Deng-Ping Fan, Cheng Gong, Yang Cao, Bo Ren, Ming-Ming Cheng, and Ali Borji. 2018. Enhanced-alignment measure for binary foreground map evaluation. arXiv preprint arXiv:1805.10421 (2018).
[14]
Deng-Ping Fan, Zheng Lin, Zhao Zhang, Menglong Zhu, and Ming-Ming Cheng. 2020. Rethinking RGB-D salient object detection: Models, data sets, and large-scale benchmarks. IEEE Trans. Neural Netw. Learn. Syst. 32, 5 (2020), 2075–2089. DOI:
[15]
Deng-Ping Fan, Yingjie Zhai, Ali Borji, Jufeng Yang, and Ling Shao. 2020. BBS-Net: RGB-D salient object detection with a bifurcated backbone strategy network. In Proceedings of the European Conference on Computer Vision. 275–292.
[16]
Xian Fang, Mingfeng Jiang, Jinchao Zhu, Xiuli Shao, and Hongpeng Wang. 2023. M2RNet: Multi-modal and multi-scale refined network for RGB-D salient object detection. Pattern Recog. 135 (2023), 109139.
[17]
Huan Fu, Mingming Gong, Chaohui Wang, Kayhan Batmanghelich, and Dacheng Tao. 2018. Deep ordinal regression network for monocular depth estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2002–2011.
[18]
Keren Fu, Deng-Ping Fan, Ge-Peng Ji, Qijun Zhao, Jianbing Shen, and Ce Zhu. 2021. Siamese network for RGB-D salient object detection and beyond. IEEE Trans. Pattern Anal. Mach. Intell. 44, 9 (2021), 5541–5559.
[19]
Shang-Hua Gao, Ming-Ming Cheng, Kai Zhao, Xin-Yu Zhang, Ming-Hsuan Yang, and Philip Torr. 2019. Res2Net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 43, 2 (2019), 652–662.
[20]
Wei Gao, Guibiao Liao, Siwei Ma, Ge Li, Yongsheng Liang, and Weisi Lin. 2021. Unified information fusion network for multi-modal RGB-D and RGB-T salient object detection. IEEE Trans. Circ. Syst. Vid. Technol. 32, 4 (2021), 2091–2106.
[21]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 770–778.
[22]
Jie Hu, Li Shen, and Gang Sun. 2018. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 7132–7141.
[23]
Xiaowei Hu, Chi-Wing Fu, Lei Zhu, Tianyu Wang, and Pheng-Ann Heng. 2020. SAC-Net: Spatial attenuation context for salient object detection. IEEE Trans. Circ. Syst. Vid. Technol. 31, 3 (2020), 1079–1090.
[24]
Shuaixiong Hui, Qiang Guo, Xiaoyu Geng, and Caiming Zhang. 2023. Multi-guidance CNNs for salient object detection. ACM Trans. Multim. Comput. Commun. Appl. 19, 3 (2023), 19 pages. DOI:. Just Accepted.
[25]
Wei Ji, Jingjing Li, Shuang Yu, Miao Zhang, Yongri Piao, Shunyu Yao, Qi Bi, Kai Ma, Yefeng Zheng, Huchuan Lu, and Li Cheng. 2021. Calibrated RGB-D salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 9471–9481.
[26]
Wei Ji, Jingjing Li, Miao Zhang, Yongri Piao, and Huchuan Lu. 2020. Accurate RGB-D salient object detection via collaborative learning. In Proceedings of the European Conference on Computer Vision. 52–69.
[27]
Wei Ji, Ge Yan, Jingjing Li, Yongri Piao, Shunyu Yao, Miao Zhang, Li Cheng, and Huchuan Lu. 2022. DMRA: Depth-induced multi-scale recurrent attention network for RGB-D saliency detection. IEEE Trans. Image Process. 31 (2022), 2321–2336.
[28]
Lai Jiang, Mai Xu, Xiaofei Wang, and Leonid Sigal. 2021. Saliency-guided image translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 16509–16518.
[29]
Xiao Jin, Kang Yi, and Jing Xu. 2022. MoADNet: Mobile asymmetric dual-stream networks for real-time and lightweight RGB-D salient object detection. IEEE Trans. Circ. Syst. Vid. Technol. 32, 11 (2022), 7632–7645.
[30]
Ran Ju, Ling Ge, Wenjing Geng, Tongwei Ren, and Gangshan Wu. 2014. Depth saliency based on anisotropic center-surround difference. In Proceedings of the IEEE International Conference on Image Processing. 1115–1119.
[31]
Diederik P. Kingma and Jimmy Ba. 2017. Adam: A Method for Stochastic Optimization. arxiv:1412.6980 [cs.LG]
[32]
Dominik A. Klein and Simone Frintrop. 2011. Center-surround divergence of feature statistics for salient object detection. In Proceedings of the International Conference on Computer Vision. 2214–2219.
[33]
John Lafferty, Andrew McCallum, and Fernando C. N. Pereira. 2001. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. Proceedings of the 18th International Conference on Machine Learning, 282–289.
[34]
Gongyang Li, Zhi Liu, Minyu Chen, Zhen Bai, Weisi Lin, and Haibin Ling. 2021. Hierarchical alternate interaction network for RGB-D salient object detection. IEEE Trans. Image Process. 30 (2021), 3528–3542.
[35]
Nianyi Li, Jinwei Ye, Yu Ji, Haibin Ling, and Jingyi Yu. 2014. Saliency detection on light field. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2806–2813.
[36]
Haipeng Liu, Yang Wang, Meng Wang, and Yong Rui. 2022. Delving globally into texture and structure for image inpainting. In Proceedings of the 30th ACM International Conference on Multimedia. 1270–1278.
[37]
Jiang-Jiang Liu, Qibin Hou, Zhi-Ang Liu, and Ming-Ming Cheng. 2023. PoolNet+: Exploring the potential of pooling for salient object detection. IEEE Trans. Pattern Anal. Mach. Intell. 45, 1 (2023), 887–904.
[38]
Nian Liu, Ni Zhang, and Junwei Han. 2020. Learning selective self-mutual attention for RGB-D saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 13756–13765.
[39]
Nian Liu, Ni Zhang, Kaiyuan Wan, Ling Shao, and Junwei Han. 2021. Visual saliency transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 4722–4732.
[40]
Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. 2021. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV’21). 10012–10022.
[41]
Zhengyi Liu, Yacheng Tan, Qian He, and Yun Xiao. 2021. SwinNet: Swin transformer drives edge-aware RGB-D and RGB-T salient object detection. IEEE Trans. Circ. Syst. Vid. Technol. 32, 7 (2021), 4486–4497.
[42]
Zhengyi Liu, Yuan Wang, Zhengzheng Tu, Yun Xiao, and Bin Tang. 2021. TriTransNet: RGB-D salient object detection with a triplet transformer embedding network. In Proceedings of the 29th ACM International Conference on Multimedia. 4481–4490.
[43]
Ran Margolin, Lihi Zelnik-Manor, and Ayellet Tal. 2014. How to evaluate foreground maps? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 248–255.
[44]
Arsalan Mousavian, Hamed Pirsiavash, and Jana Košecká. 2016. Joint semantic segmentation and depth estimation with deep convolutional networks. In Proceedings of the 4th International Conference on 3D Vision. 611–619.
[45]
Yuzhen Niu, Yujie Geng, Xueqing Li, and Feng Liu. 2012. Leveraging stereopsis for saliency analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 454–461.
[46]
Youwei Pang, Lihe Zhang, Xiaoqi Zhao, and Huchuan Lu. 2020. Hierarchical dynamic filtering network for RGB-D salient object detection. In Proceedings of the European Conference on Computer Vision. 235–252.
[47]
Houwen Peng, Bing Li, Weihua Xiong, Weiming Hu, and Rongrong Ji. 2014. RGBD salient object detection: A benchmark and algorithms. In Proceedings of the European Conference on Computer Vision. 92–109.
[48]
Yongri Piao, Wei Ji, Jingjing Li, Miao Zhang, and Huchuan Lu. 2019. Depth-induced multi-scale recurrent attention network for saliency detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 7254–7263.
[49]
Yongri Piao, Zhengkun Rong, Miao Zhang, Weisong Ren, and Huchuan Lu. 2020. A2dele: Adaptive and attentive depth distiller for efficient RGB-D salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 9060–9069.
[50]
Biao Qian, Yang Wang, Richang Hong, and Meng Wang. 2023. Adaptive data-free quantization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 7960–7968.
[51]
Biao Qian, Yang Wang, Hongzhi Yin, Richang Hong, and Meng Wang. 2022. Switchable online knowledge distillation. In Proceedings of the European Conference on Computer Vision. Springer, 449–466.
[52]
Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014).
[53]
Peng Sun, Wenhu Zhang, Huanyu Wang, Songyuan Li, and Xi Li. 2021. Deep RGB-D saliency detection with depth-sensitive attention and automatic multi-modal fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1407–1417.
[54]
Fengyun Wang, Jinshan Pan, Shoukun Xu, and Jinhui Tang. 2022. Learning discriminative cross-modality features for RGB-D saliency detection. IEEE Trans. Image Process. 31 (2022), 1285–1297.
[55]
Fasheng Wang, Ruimin Wang, and Fuming Sun. 2023. DCMNet: Discriminant and cross-modality network for RGB-D salient object detection. Expert Syst. Applic. 214 (2023), 119047.
[56]
Liansheng Wang, Rongzhen Chen, Lei Zhu, Haoran Xie, and Xiaomeng Li. 2020. Deep sub-region network for salient object detection. IEEE Trans. Circ. Syst. Vid. Technol. 31, 2 (2020), 728–741.
[57]
Rui Wang, Stephen M. Pizer, and Jan-Michael Frahm. 2019. Recurrent neural network for (un-) supervised learning of monocular video visual odometry and depth. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5555–5564.
[58]
Tiantian Wang, Lihe Zhang, Shuo Wang, Huchuan Lu, Gang Yang, Xiang Ruan, and Ali Borji. 2018. Detect globally, refine locally: A novel approach to saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3127–3135.
[59]
Wenguan Wang, Jianbing Shen, Ming-Ming Cheng, and Ling Shao. 2019. An iterative and cooperative top-down and bottom-up inference network for salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5968–5977.
[60]
Wenguan Wang, Jianbing Shen, Xingping Dong, Ali Borji, and Ruigang Yang. 2019. Inferring salient objects from human fixations. IEEE Trans. Pattern Anal. Mach. Intell. 42, 8 (2019), 1913–1927.
[61]
Wenguan Wang, Jianbing Shen, and Ling Shao. 2017. Video salient object detection via fully convolutional networks. IEEE Trans. Image Process. 27, 1 (2017), 38–49.
[62]
Wenguan Wang, Jianbing Shen, Jianwen Xie, Ming-Ming Cheng, Haibin Ling, and Ali Borji. 2019. Revisiting video saliency prediction in the deep learning era. IEEE Trans. Pattern Anal. Mach. Intell. 43, 1 (2019), 220–237.
[63]
Wenguan Wang, Guolei Sun, and Luc Van Gool. 2022. Looking beyond single images for weakly supervised semantic segmentation learning. IEEE Trans. Pattern Anal. Mach. Intell. (2022), 15 pages. DOI:
[64]
Wenguan Wang, Shuyang Zhao, Jianbing Shen, Steven C. H. Hoi, and Ali Borji. 2019. Salient object detection with pyramid attention and salient edges. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1448–1457.
[65]
Yang Wang. 2021. Survey on deep multi-modal data analytics: Collaboration, rivalry, and fusion. ACM Trans. Multim. Comput., Commun. Applic. 17, 1s (2021), 1–25.
[66]
Yue Wang, Yuke Li, James H. Elder, Runmin Wu, Huchuan Lu, and Lu Zhang. 2020. Synergistic saliency and depth prediction for RGB-D saliency detection. In Proceedings of the Asian Conference on Computer Vision. 336–352.
[67]
Yang Wang, Jinjia Peng, Huibing Wang, and Meng Wang. 2022. Progressive learning with multi-scale attention network for cross-domain vehicle re-identification. Sci. China Inf. Sci. 65, 6 (2022), 160103.
[68]
Yanbo Wang, Fasheng Wang, Chang Wang, Jianjun He, and Fuming Sun. 2022. Learning saliency aware correlation filter for visual tracking. Comput. J. 65, 7 (2022), 1846–1859.
[69]
Sanghyun Woo, Jongchan Park, Joon-Young Lee, and In So Kweon. 2018. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV’18). 3–19.
[70]
Yu-Huan Wu, Yun Liu, Jun Xu, Jia-Wang Bian, Yu-Chao Gu, and Ming-Ming Cheng. 2021. MobileSal: Extremely efficient RGB-D salient object detection. IEEE Trans. Pattern Anal. Mach. Intell. 44, 12 (2021), 10261–10269. DOI:
[71]
Zhe Wu, Li Su, and Qingming Huang. 2019. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 3907–3916.
[72]
Hui-Chu Xiao, Wan-Lei Zhao, Jie Lin, Yi-Geng Hong, and Chong-Wah Ngo. 2022. Deeply activated salient region for instance search. ACM Trans. Multim. Comput. Commun. Appl. 18, 3s, Article 147 (Nov.2022), 19 pages. DOI:
[73]
Qiong Yan, Li Xu, Jianping Shi, and Jiaya Jia. 2013. Hierarchical saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1155–1162.
[74]
Yang Yang, Qi Qin, Yongjiang Luo, Yi Liu, Qiang Zhang, and Jungong Han. 2022. Bi-directional progressive guidance network for RGB-D salient object detection. IEEE Trans. Circ. Syst. Vid. Technol. 32, 8 (2022), 5346–5360.
[75]
Wei Yin, Yifan Liu, Chunhua Shen, and Youliang Yan. 2019. Enforcing geometric constraints of virtual normal for depth prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 5684–5693.
[76]
Chen Zhang, Runmin Cong, Qinwei Lin, Lin Ma, Feng Li, Yao Zhao, and Sam Kwong. 2021. Cross-modality discrepant interaction network for RGB-D salient object detection. In Proceedings of the 29th ACM International Conference on Multimedia. 2094–2102.
[77]
Jing Zhang, Jianwen Xie, Nick Barnes, and Ping Li. 2021. Learning generative vision transformer with energy-based latent space for saliency prediction. Adv. Neural Inf. Process. Syst. 34 (2021), 15448–15463.
[78]
Miao Zhang, Sun Xiao Fei, Jie Liu, Shuang Xu, Yongri Piao, and Huchuan Lu. 2020. Asymmetric two-stream architecture for accurate RGB-D saliency detection. In Proceedings of the European Conference on Computer Vision. 374–390.
[79]
Miao Zhang, Weisong Ren, Yongri Piao, Zhengkun Rong, and Huchuan Lu. 2020. Select, supplement and focus for RGB-D saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 3472–3481.
[80]
Miao Zhang, Shunyu Yao, Beiqi Hu, Yongri Piao, and Wei Ji. 2022. C2DFNet: Criss-cross dynamic filter network for RGB-D salient object detection. IEEE Trans. Multim. (2022), 1–13. DOI:
[81]
Wenbo Zhang, Ge-Peng Ji, Zhuo Wang, Keren Fu, and Qijun Zhao. 2021. Depth quality-inspired feature manipulation for efficient RGB-D salient object detection. In Proceedings of the 29th ACM International Conference on Multimedia. 731–740.
[82]
Xi Zhang and Xiaolin Wu. 2021. Attention-guided image compression by deep reconstruction of compressive sensed saliency skeleton. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 13354–13364.
[83]
Jiawei Zhao, Yifan Zhao, Jia Li, and Xiaowu Chen. 2020. Is depth really necessary for salient object detection? In Proceedings of the 28th ACM International Conference on Multimedia. 1745–1754.
[84]
Jia-Xing Zhao, Yang Cao, Deng-Ping Fan, Ming-Ming Cheng, Xuan-Yi Li, and Le Zhang. 2019. Contrast prior and fluid pyramid integration for RGBD salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 3927–3936.
[85]
Xiaoqi Zhao, Youwei Pang, Lihe Zhang, Huchuan Lu, and Lei Zhang. 2020. Suppress and balance: A simple gated network for salient object detection. In Proceedings of the European Conference on Computer Vision. 35–51.
[86]
Xiaoqi Zhao, Lihe Zhang, Youwei Pang, Huchuan Lu, and Lei Zhang. 2020. A single stream network for robust and real-time RGB-D salient object detection. In Proceedings of the European Conference on Computer Vision. 646–662.
[87]
Jiayuan Zhou, Lijun Wang, Huchuan Lu, Kaining Huang, Xinchu Shi, and Bocong Liu. 2022. MVSalNet: Multi-view augmentation for RGB-D salient object detection. In Proceedings of the European Conference on Computer Vision. Springer, 270–287.
[88]
Tao Zhou, Deng-Ping Fan, Ming-Ming Cheng, Jianbing Shen, and Ling Shao. 2021. RGB-D salient object detection: A survey. Comput. Vis. Media 7 (2021), 37–69.
[89]
Tao Zhou, Huazhu Fu, Geng Chen, Yi Zhou, Deng-Ping Fan, and Ling Shao. 2021. Specificity-preserving RGB-D saliency detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 4681–4691.

Cited By

View all
  • (2025)Asymmetric deep interaction network for RGB-D salient object detectionExpert Systems with Applications10.1016/j.eswa.2024.126083266(126083)Online publication date: Mar-2025
  • (2024)Network Information Security Monitoring Under Artificial Intelligence EnvironmentInternational Journal of Information Security and Privacy10.4018/IJISP.34503818:1(1-25)Online publication date: 21-Jun-2024
  • (2024)Transparent Depth Completion Using Segmentation FeaturesACM Transactions on Multimedia Computing, Communications, and Applications10.1145/369497820:12(1-19)Online publication date: 9-Sep-2024
  • Show More Cited By

Index Terms

  1. Attention-guided Multi-modality Interaction Network for RGB-D Salient Object Detection

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Multimedia Computing, Communications, and Applications
    ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 20, Issue 3
    March 2024
    665 pages
    EISSN:1551-6865
    DOI:10.1145/3613614
    • Editor:
    • Abdulmotaleb El Saddik
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 23 October 2023
    Online AM: 16 September 2023
    Accepted: 12 September 2023
    Revised: 29 August 2023
    Received: 03 February 2023
    Published in TOMM Volume 20, Issue 3

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Salient object detection
    2. boundary aware
    3. multi-modality
    4. depth enhancement

    Qualifiers

    • Research-article

    Funding Sources

    • National Natural Science Foundation of China
    • LiaoNing Revitalization Talents Program
    • Liaoning Baiqianwan Talents Program
    • Innovative Talents Program for Liaoning Universities

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)405
    • Downloads (Last 6 weeks)48
    Reflects downloads up to 14 Dec 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2025)Asymmetric deep interaction network for RGB-D salient object detectionExpert Systems with Applications10.1016/j.eswa.2024.126083266(126083)Online publication date: Mar-2025
    • (2024)Network Information Security Monitoring Under Artificial Intelligence EnvironmentInternational Journal of Information Security and Privacy10.4018/IJISP.34503818:1(1-25)Online publication date: 21-Jun-2024
    • (2024)Transparent Depth Completion Using Segmentation FeaturesACM Transactions on Multimedia Computing, Communications, and Applications10.1145/369497820:12(1-19)Online publication date: 9-Sep-2024
    • (2024)Rethinking Feature Mining for Light Field Salient Object DetectionACM Transactions on Multimedia Computing, Communications, and Applications10.1145/3676967Online publication date: 8-Jul-2024
    • (2024)Heterogeneous Fusion and Integrity Learning Network for RGB-D Salient Object DetectionACM Transactions on Multimedia Computing, Communications, and Applications10.1145/365647620:7(1-24)Online publication date: 15-May-2024
    • (2024)MultiRider: Enabling Multi-Tag Concurrent OFDM Backscatter by Taming In-band InterferenceProceedings of the 22nd Annual International Conference on Mobile Systems, Applications and Services10.1145/3643832.3661862(292-303)Online publication date: 3-Jun-2024
    • (2024)Cross-Modal and Cross-Level Attention Interaction Network for Salient Object DetectionIEEE Transactions on Artificial Intelligence10.1109/TAI.2023.33338275:6(2907-2920)Online publication date: Jun-2024
    • (2024)RGB-D Salient Object Detection Based on Cross-Modal and Cross-Level Feature FusionIEEE Access10.1109/ACCESS.2024.338152412(45134-45146)Online publication date: 2024
    • (2024)A multiple‐attention refinement network for RGB‐D salient object detectionIET Image Processing10.1049/ipr2.1326718:14(4551-4562)Online publication date: 21-Nov-2024
    • (2024)Leveraging modality‐specific and shared features for RGB‐T salient object detectionIET Computer Vision10.1049/cvi2.12307Online publication date: 25-Sep-2024
    • Show More Cited By

    View Options

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media