Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleDecember 2024
RGB-D Data Compression via Bi-Directional Cross-Modal Prior Transfer and Enhanced Entropy Modeling
ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM), Volume 21, Issue 2Article No.: 58, Pages 1–17https://doi.org/10.1145/3702997RGB-D data, being homogeneous cross-modal data, demonstrates significant correlations among data elements. However, current research focuses only on a uni-directional pattern of cross-modal contextual information, neglecting the exploration of bi-...
- research-articleNovember 2024
PGGNet: Pyramid gradual-guidance network for RGB-D indoor scene semantic segmentation
Highlights- We propose a pyramid network with progressive guidance structures.
- Global and local information of features is supplemented by the pyramid structure.
- The MEFM can reduce redundant information of depth maps and aggregate multiscale ...
In RGB-D (red–green–blue and depth) scene semantic segmentation, depth maps provide rich spatial information to RGB images to achieve high performance. However, properly aggregating depth information and reducing noise and information loss during ...
- research-articleSeptember 2024
Direct RGB-D visual odometry with point features
Intelligent Service Robotics (SPISR), Volume 17, Issue 5Pages 1077–1089https://doi.org/10.1007/s11370-024-00559-wAbstractIn this paper, we propose a traditional semi-dense direct visual odometry (VO) based on our preliminary study using low-order Gaussian derivative functions for solving a VO problem with pure frame-by-frame point tracking. With the off-line fitting ...
- ArticleJune 2024
Taking Advantage of Depth Information for Semantic Segmentation in Field-Measured Vineyards
AbstractRGB-D cameras mounted on moving agricultural robotic platforms provide detailed information about both appearance and volume of plants. Those images can be analysed by means of deep segmentation models; however, such methods usually dismiss depth ...
-
- research-articleMay 2024
CMDCF: an effective cross-modal dense cooperative fusion network for RGB-D SOD
Neural Computing and Applications (NCAA), Volume 36, Issue 23Pages 14361–14378https://doi.org/10.1007/s00521-024-09692-0AbstractThe success of vision transformer demonstrates that the transformer structure is also suitable for various vision tasks, including high-level classification tasks and low-level dense prediction tasks. Salient object detection (SOD) is a pixel-...
- research-articleNovember 2024
3D pose estimation using joint-based calibration in distributed RGB-D camera system
AbstractThis paper proposes a new approach that acquires camera parameters and generates an integrated 3D joint using an RGB-D camera network distributed in an arbitrary location in space. The proposed technique consists of three steps. In the first step,...
Graphical abstractDisplay Omitted
Highlights- Performing camera calibration using incomplete or partial joints obtained from images output from cameras installed in irregular positions.
- Temporal improvement of the precision of transformation parameters obtained through camera ...
- research-articleJuly 2024
Depth awakens: A depth-perceptual attention fusion network for RGB-D camouflaged object detection
AbstractCamouflaged object detection (COD) presents a persistent challenge in accurately identifying objects that seamlessly blend into their surroundings. However, most existing COD models overlook the fact that visual systems operate within a genuine ...
Highlights- An advanced RGB-D camouflaged object detection model is proposed.
- Knowledge about depth can greatly assist in recognizing camouflaged objects.
- Presentation of a fusion method that adaptively incorporates the generated depth.
- ArticleJanuary 2024
Bi-directional Interaction and Dense Aggregation Network for RGB-D Salient Object Detection
AbstractRGB-D salient object detection (SOD) which aims to detect the prominent regions in figures has attracted much attention recently. It jointly models the RGB and depth information. However, existing methods explore cross-modality information from ...
- research-articleJanuary 2024
Deep learning based computer vision under the prism of 3D point clouds: a systematic review
The Visual Computer: International Journal of Computer Graphics (VISC), Volume 40, Issue 11Pages 8287–8329https://doi.org/10.1007/s00371-023-03237-7AbstractPoint clouds consist of 3D data points and are among the most considerable data formats for 3D representations. Their popularity is due to their broad application areas, such as robotics and autonomous driving, and their employment in basic 3D ...
- research-articleMay 2024
Holographic traffic network characteristics: Prototype implementation
ICFNDS '23: Proceedings of the 7th International Conference on Future Networks and Distributed SystemsPages 657–662https://doi.org/10.1145/3644713.3644811One of the most important tasks of the current decade is the mass introduction of telepresence services, including the use of holographic copies of a person, robotic avatars, manipulator devices, as well as the implementation and deployment of their ...
- research-articleJanuary 2024
RGB-D Tracking via Hierarchical Modality Aggregation and Distribution Network
MMAsia '23: Proceedings of the 5th ACM International Conference on Multimedia in AsiaArticle No.: 67, Pages 1–7https://doi.org/10.1145/3595916.3626441The integration of dual-modal features has been pivotal in advancing RGB-Depth (RGB-D) tracking. However, current trackers are less efficient and focus solely on single-level features, resulting in weaker robustness in fusion and slower speeds that fail ...
- research-articleDecember 2023
Cross-Modal Transformer for RGB-D semantic segmentation of production workshop objects
Highlights- We propose Cross-Modal Transformer (CMFormer), a Transformer-based cross-modal semantic segmentation model, which achieves better cross-modal information interaction by capturing long-range contextual dependencies. The CMFormer includes ...
Scene understanding in a production workshop is an important technology to improve its intelligence level, semantic segmentation of production workshop objects is an effective method for realizing scene understanding. Since the varieties of ...
- research-articleDecember 2023
Hardening RGB-D object recognition systems against adversarial patch attacks
- Yang Zheng,
- Luca Demetrio,
- Antonio Emanuele Cinà,
- Xiaoyi Feng,
- Zhaoqiang Xia,
- Xiaoyue Jiang,
- Ambra Demontis,
- Battista Biggio,
- Fabio Roli
Information Sciences: an International Journal (ISCI), Volume 651, Issue Chttps://doi.org/10.1016/j.ins.2023.119701AbstractRGB-D object recognition systems improve their predictive performances by fusing color and depth information, outperforming neural network architectures that rely solely on colors. While RGB-D systems are expected to be more robust to adversarial ...
Highlights- We assess the performance of a state-of-art system for object detection based on color and depth features.
- We explain why RGB features are more vulnerable to attacks than depth features.
- We develop a defense against adversarial ...
- research-articleOctober 2023
Grape yield estimation with a smartphone’s colour and depth cameras using machine learning and computer vision techniques
Computers and Electronics in Agriculture (COEA), Volume 213, Issue Chttps://doi.org/10.1016/j.compag.2023.108174AbstractA smartphone with both colour and time of flight depth cameras is used for automated grape yield estimation of Chardonnay grapes. A new technique is developed to automatically identify grape berries in the smartphone’s depth maps. This utilises ...
Highlights- Grape yield estimation using a smartphone’s colour and time of flight depth cameras.
- Peaks in the depth map occur at berry locations due to diffuse scattering of light.
- Grape berries 3D locations are automatically detected using ...
- ArticleSeptember 2023
CarPatch: A Synthetic Benchmark for Radiance Field Evaluation on Vehicle Components
AbstractNeural Radiance Fields (NeRFs) have gained widespread recognition as a highly effective technique for representing 3D reconstructions of objects and scenes derived from sets of images. Despite their efficiency, NeRF models can pose challenges in ...
- research-articleSeptember 2023
Cross-modal attention fusion network for RGB-D semantic segmentation
AbstractRGB-D semantic segmentation is crucial for robots to understand scenes. Most existing methods take depth information as an additional input, leading to cross-modal semantic segmentation networks that cannot achieve the purpose of multi-scale ...
- ArticleJanuary 2024
Integrate Depth Information to Enhance the Robustness of Object Level SLAM
AbstractvSLAM (Visual Simultaneous Localization and Mapping) is a fundamental function in various robot applications. With the development of downstream applications, there is an increasing challenge for scene semantic understanding and stable operation ...
- ArticleAugust 2023
MBDNet: Mitigating the “Under-Training Issue” in Dual-Encoder Model for RGB-d Salient Object Detection
Advanced Intelligent Computing Technology and ApplicationsPages 99–111https://doi.org/10.1007/978-981-99-4761-4_9AbstractExisting RGB-D salient object detection methods generally rely on the dual-encoder structure for RGB and depth feature extraction. However, we observe that the encoders in such models are often not adequately trained to obtain superior feature ...
- ArticleMay 2024
Research on Improved Algorithm of Significance Object Detection Based on ATSA Model
Advances in Brain Inspired Cognitive SystemsPages 154–165https://doi.org/10.1007/978-981-97-1417-9_15AbstractSaliency detection refers to accurately positioning and extracting significant objects or regions in the image. Most effective object detection methods are based on RGB-D and adopt the dual-flow architecture with RGB and depth symmetry. At the ...