[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Concentration vs. Optical Density of ESKAPEE Bacteria: A Method to Determine the Optimum Measurement Wavelength
Previous Article in Journal
MambaPose: A Human Pose Estimation Based on Gated Feedforward Network and Mamba
Previous Article in Special Issue
CFRNet: Cross-Attention-Based Fusion and Refinement Network for Enhanced RGB-T Salient Object Detection
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wavelet-Driven Multi-Band Feature Fusion for RGB-T Salient Object Detection

1
School of Software Engineering, Shenyang University of Technology, Shenyang 110870, China
2
School of Mechanical Engineering & Automation, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 8159; https://doi.org/10.3390/s24248159
Submission received: 18 November 2024 / Revised: 15 December 2024 / Accepted: 19 December 2024 / Published: 20 December 2024
(This article belongs to the Special Issue Multi-Modal Image Processing Methods, Systems, and Applications)

Abstract

:
RGB-T salient object detection (SOD) has received considerable attention in the field of computer vision. Although existing methods have achieved notable detection performance in certain scenarios, challenges remain. Many methods fail to fully utilize high-frequency and low-frequency features during information interaction among different scale features, limiting detection performance. To address this issue, we propose a method for RGB-T salient object detection that enhances performance through wavelet transform and channel-wise attention fusion. Through feature differentiation, we effectively extract spatial characteristics of the target, enhancing the detection capability for global context and fine-grained details. First, input features are passed through the channel-wise criss-cross module (CCM) for cross-modal information fusion, adaptively adjusting the importance of features to generate rich fusion information. Subsequently, the multi-scale fusion information is input into the feature selection wavelet transforme module (FSW), which selects beneficial low-frequency and high-frequency features to improve feature aggregation performance and achieves higher segmentation accuracy through long-distance connections. Extensive experiments demonstrate that our method outperforms 22 state-of-the-art methods.

1. Introduction

Saliency object detection has numerous applications in the field of computer vision, primarily detecting the most prominent objects in a scene. In image segmentation, saliency detection can prioritize salient regions, improving segmentation efficiency and accuracy [1,2]. In object detection, it helps algorithms quickly locate key areas, thereby enhancing detection speed and accuracy [3]. In the medical field, saliency object detection can assist doctors in quickly identifying abnormal regions in organs [4,5]. Additionally, saliency detection finds wide applications in robotics and assisted driving.
However, these methods are primarily designed for monocular RGB cameras and perform poorly in low-light, overexposed, or noisy environments. Even with image enhancement algorithms, the results are often unsatisfactory. With the advancement of deep learning and information technology, thermal images have been introduced into saliency object detection. These images are less sensitive to lighting, compensating for the limitations of RGB images in information representation. However, insufficient fusion can result in the loss of valuable information, while overly complex fusion strategies may introduce excessive noise, affecting detection results. Therefore, effectively integrating the two different modalities and selecting information beneficial to the task is the primary challenge in RGB-T salient object detection.
Traditional methods primarily depend on manually designed saliency features and prior knowledge. For example, ref. [6] proposed a multitask manifold ranking with cross-modality consistency algorithm, ref. [7] proposed a multi-scale manifold ranking algorithm, and ref. [8] introduced a low-rank tensor model to learn an optimized graph affinity matrix, which can effectively suppress noisy values in the features. Although traditional methods achieve good detection results in some simple scenarios, their reliance on ranking algorithms and a large amount of prior knowledge limits their detection capability in complex scenes.
Deep learning has shown outstanding performance in visual tasks and can overcome some detection challenges in complex scenes. In recent years, many deep learning-based methods have emerged in the RGB-T SOD field, significantly improving detection accuracy. For example, ref. [9] proposed a dual decoder architecture that introduces global context and fully fuses bimodal features through cascaded multi-interaction modules to suppress bimodal bias. Ref. [10] introduced a parallel symmetric network (PSNet), which enhances bimodal feature representation through a cascaded aggregation module and integrates adjacent layer saliency cues using a parallel symmetric fusion module (PSF). Ref. [11] designed a dedicated MS module to capture cross-modal and cross-scale complementary information from RGB and thermal images, and by utilizing cross-level complementary information from different fusion levels, it generates accurate saliency maps with clear boundaries. Ref. [12] proposed a method that regulates cross-modal interaction by obtaining global illumination scores and uses RGB features to supplement thermal semantics during the encoding phase while enhancing RGB localization with thermal features during the decoding phase. Ref. [13] adopted an early fusion strategy to extract complementary information through simple concatenation, addition, and multiplication operations before inputting into the network. Additionally, it designed an SGM module to progressively refine low-level features with high-level semantics. Ref. [14] proposed a secondary fusion network called the global contextually guided lightweight network, which employs secondary cross-modal integration to remove redundant information while fusing and propagating effective modal information.
Although the above methods have achieved good results, some issues still remain. Most methods focus only on global features without distinguishing between low-frequency and high-frequency information, which may lead to an inability to fully separate noise from valuable information. The downside of this approach is that noise in the high-frequency information may be excessively amplified, resulting in detection outcomes that are less stable and accurate.
Motivated by the aforementioned two challenges, we divide the features of the two modalities into channels, processing the information in different channels in parallel. We enhance the contextual information within the feature maps of each channel and establish global dependencies between feature maps to improve the perception of fine-grained features, thereby achieving thorough fusion. Our solution is illustrated in Figure 1. Additionally, by selectively utilizing low-frequency and high-frequency information, we aim to more accurately highlight salient information. The advantage of low-frequency information is that it retains more robust and stable global structural features, avoiding the noise interference caused by excessive attention to high-frequency details. Unlike [15,16,17], our method emphasizes low-frequency information during the fusion process, allowing us to capture the overall contours of salient objects more accurately. For high-frequency information, we use 3D convolution, enabling the network to better understand the relationships and contextual information between objects. This enhances the reliability and consistency of detection, thereby improving the model’s generalization capability and performance in handling complex scenes.
In summary, our contributions are as follows:
  • We propose a novel RGB-T SOD method, which consists of three components: a feature encoder, a channel-wise criss-cross Module, and a feature selection wavelet transformer.
  • We propose a channel-wise criss-cross module (CCM) that performs channel decomposition and parallel computation on features from both RGB and thermal modalities, effectively preventing information loss caused by direct fusion. This module employs attention mechanism to adaptively fuse complementary information from both modalities, and enhances the capture of global contextual information and fine-grained features through dynamic weight allocation, thereby achieving more comprehensive and robust feature fusion.
  • We propose a contextual feature selection wavelet transformer (FSW) module that uses wavelet transform to decompose fused feature information into high-frequency and low-frequency components. The high-frequency features capture fine-grained edge details for accurate target localization, while the low-frequency features provide background and structural context. This design enables the model to remain sensitive to object edges while leveraging global context to improve robustness, particularly in complex scenes. By effectively integrating both frequency bands during feature aggregation, the FSW module enhances segmentation accuracy and object localization.
  • Extensive experimental validation shows that our method achieves outstanding performance on three datasets.

2. Related Work

2.1. RGB Salient Object Detection

Siris et al. [18] proposed a context-aware learning method for salient object detection in complex scenes. Their approach includes a semantic scene context refinement and a context instance translator, which are used to enhance the contextual features learned from salient objects and to capture the relationship between objects and the scene context. Wang et al. [19] introduced the hybrid feature alignment network (HFANet) for salient object detection in optical remote sensing images (RSI-SOD), aiming to address issues such as complex backgrounds, multi-scale variations, and irregular edge topologies. They designed a hybrid encoder that combines the strengths of CNNs and transformers for local and global context modeling, respectively. Wu et al. [20] proposed a new method called dynamic pyramid convolution (DPConv), which dynamically selects the most appropriate kernel size according to the input image’s needs, enhancing feature representation at different scales. Wu et al. [21] also introduced the extreme downsampling network (EDN), which aims to enhance SOD performance by improving high-level features. They incorporated an extreme downsampling technique to effectively learn a global view, enabling precise salient object localization. Additionally, they designed a Scale-dependent pyramid convolution (SCPC) decoder to recover object details. Li et al. [22] proposed a lightweight framework using a Complementary Tri-decoder (CTD), which decouples the U-shaped structure into a semantic path, spatial path, and boundary path to address issues of semantic context dilution, spatial detail loss, and boundary refinement. The framework gradually optimizes segmentation results through a cross aggregation module (CAM) and a boundary refinement module (BRM), while the scale-adaptive pooling module (SAP) enhances the handling of multi-scale features.

2.2. RGB-D Salient Object Detection

Jin et al. [23] proposed the complementary depth network (CDNet), which selects depth maps containing saliency information as the training target and uses RGB features to estimate meaningful auxiliary depth maps, thus providing more saliency features. Chen et al. [24] introduced a novel model, RD3D, which for the first time attempts to use 3D convolutional neural networks to enhance cross-modal fusion capability. By employing pre-fusion in the encoder stage and deep fusion in the decoder stage, the model fully integrates RGB and depth information. Song et al. [25] proposed the modality-aware decoder (MaD), which focuses more on the relationships between modalities. Through modality-aware embedding and reasoning, MaD improves RGB-D fusion without the need for complex designs. Sun et al. [26] introduced the cascade aggregation transformer network (CATNet), which progressively enhances feature representation and fusion through multiple stages. The network also uses a cascade correction mechanism to address deficiencies in multi-scale feature fusion, thereby improving detection performance. Wu et al. [27] proposed the hierarchical depth-aware network (HiDAnet), which enhances the discriminative ability of RGB and depth features through a granularity-based attention mechanism. The network achieves coarse-to-fine multi-modal and multi-level fusion through a cross-dual attention module, while also introducing a multi-scale loss function to fully utilize hierarchical information. Zhang et al. [28] proposed a two-stage method: first, a generative network creates high-quality pseudo-depth images to calibrate the original depth data; then, cross-modal fusion is performed using a feature calibration and fusion network.

2.3. RGB-T Salient Object Detection

Gao et al. [29] proposed the multi-level multi-scale fusion network (MMNet), which explores the importance of feature response stages through self-attention, interactive attention, and adversarial combination, and integrates them into cross-modal features during the adversarial combination stage. Zhou et al. [14] utilized channel attention and spatial attention to fully fuse features of corresponding sizes from RGB and thermal modalities, enhancing the effectiveness and complementarity of feature representation. Tu et al. [9] introduced a multi-interactive dual decoder, which mines the complementarity of different modalities and the multi-type clues of image content, enabling the method to perform well even in the presence of invalid modalities. In another work, Tu et al. [30] designed a modality alignment module that combines spatial affine transformations, feature affine transformations, and dynamic convolutions to simulate the strong correlations between RGB and thermal (T) modalities. Song et al. [31] proposed a novel graph model-based weighted network, which differs from previous methods that directly fuse bimodal information at the convolutional feature layer. Instead, their method uses a graph model to more accurately represent the correlation between the two modalities. Wang et al. [32] introduced a T-aware early fusion network to improve the limitations of existing RGB-T datasets in extremely low-light scenarios. This network fully leverages the advantages of thermal images, using normal-light data during both training and testing, and validating its effectiveness with low-light and extremely low-light data. Zhou et al. [33] proposed the position-aware relational learning network (PRLNet), which addresses boundary blur and intra-class and inter-class variation issues by exploring pixel-wise distance and directional relationships through the signed distance map auxiliary module (SDMAM) and the feature refinement method with a direction field (FRDF).

3. Method

In this section, we will first provide an overview of our proposed network architecture. Then, in Section 3.2, we will detail our proposed CCM module, followed by the introduction of the FSW module in Section 3.2. Finally, in Section 3.4, we will discuss the loss functions used for network supervision.

3.1. Architecture Overview

Before the data are input into the network, the input data are first preprocessed, and the data from both modalities are cropped to the same size. The input to the network consists of two images of size (3 × 384 × 384), and the network output is a single-channel black-and-white image, where the black regions represent the background and the white regions represent the detected object, as shown in Figure 2. Benefiting from the excellent performance of the Swin Transformer in image tasks, we employ two Swin Transformer networks as the backbone networks. First, the images from the two modalities are fed into two separate backbone networks for feature encoding, resulting in multi-level features for both modalities: the feature information for the RGB image is denoted as F R 1 , F R 2 , F R 3 , F R 4 , and the feature information for the thermal image is denoted as F T 1 , F T 2 , F T 3 , F T 4 . Additionally, the number of channels for the four levels of features is 128 , 256 , 512 , 1024 . Next, we design a cross-modal fusion module (CCM) to comprehensively facilitate information exchange between the two modalities, resulting in the fused information F i . Subsequently, we design an FSW module based on the dual-tree complex wavelet transform, which fully utilizes the low-frequency and high-frequency information in the fused data to assist in object localization. Finally, the final saliency prediction map is obtained through a decoder.

3.2. Channel-Wise Criss-Cross Module

The fusion of information from both modalities is a key step in RGB-T SOD tasks. Generally, RGB images capture rich detail information, while thermal images capture thermal characteristics of object surfaces. The two modalities complement each other, and in extreme environments, their interaction can compensate for the shortcomings of each, providing more enriched feature information for subsequent tasks. To fully leverage the feature information from both modalities and reduce redundant features, we propose a modality interaction method, as shown in Figure 3, to fuse the information from the two modalities. This module simultaneously focuses on both global and local features, enabling more thorough fusion of the two modalities’ information and utilizing the complementarity between the different modal features.
Compared to traditional fusion methods such as simple concatenation, early fusion, and late fusion, our proposed CCM offers several advantages. Simple concatenation directly combines features but lacks adaptive interaction between modalities. Early fusion may lose modality-specific information too early, while late fusion might miss important cross-modal correlations in the early stages. In contrast, our CCM enables both global and local feature interaction while preserving modality-specific characteristics through its unique structure.
First, we process the input features of the RGB and T modalities separately, using depthwise separable convolution (DW) to extract features at two different scales. The advantage of DW convolution is that it reduces the number of parameters and enhances computational efficiency while retaining the local information within each modality. This corresponds to the split operation shown in the figure. Two sets of features M i 1 , M i 2 and N i 1 , N i 2 are generated, respectively.
To better fuse the cross-modal information, we introduce the criss-cross attention mechanism [34]. Unlike simple fusion methods, this mechanism enables interactive computation of features along the spatial dimension, allowing the RGB and T modality features to focus on each other within local regions. Through the criss-cross attention mechanism, we independently process each set of features for the RGB and T modalities, obtaining features that are adaptively adjusted along the spatial dimension. These feature maps are enhanced within their respective modalities to capture local spatial dependencies. The specific process is as follows:
M i 1 , M i 2 = A T ( S p l i t ( F i R ) )
N i 1 , N i 2 = A T ( S p l i t ( F i T ) )
Here, M i 1 , M i 2 and N i 1 , N i 2 represent the modality features enhanced by criss-cross attention, where A T denotes the criss-cross attention mechanism.
To adaptively adjust the importance of features at both global and fine-grained levels, we apply a global average pooling (GAP) operation to the original input features from the RGB and thermal T modalities, generating global feature representations. These representations are then processed through a Sigmoid activation function to produce weight maps, which effectively capture global contextual information and are used to regulate the relative importance of the modality-specific features. Subsequently, we further capture local feature details at the split feature level. This multi-scale feature processing strategy distinguishes our method from traditional early or late fusion approaches, as it maintains both global context and local details throughout the fusion process. We employ local convolution operations to conduct a more fine-grained analysis of different regions of the feature maps, ensuring that key salient target features are highlighted and details are better emphasized.
R i g a t e = G A P ( F i R )
T i g a t e = G A P ( F i T )
F m = C o n c a t ( R i g a t e M i 1 , R i g a t e M i 2 , T i g a t e N i 1 , T i g a t e N i 2 )
R g a t e and T g a t e denote the weight maps for the RGB and T modalities, respectively. F m represents the fused features after applying the corresponding weights.
F i f u s e = F m + C o n v ( F i R ) + C o n v ( F i T )
Finally, we concatenate the four sets of weighted features along the channel dimension. Then, we perform element-wise addition of the original RGB features and T modality features with the weighted and fused features to form the final fused features F i f u s e . This approach preserves important information from the original modalities while further enhancing the complementarity between the modalities, thus providing richer feature representations for salient object detection.

3.3. Feature Selection Wavelet Transformer

In general, the low-frequency information in an image contains the overall structure, such as gradients and large-scale components, while the high-frequency information contains more edge details. To more effectively utilize both the high-frequency and low-frequency information in images, we designed a feature selection wavelet transformer module, which aims to simultaneously focus on and filter the beneficial features in both frequency bands to improve the performance during the feature aggregation process. We specifically choose dual-tree complex wavelet transform (DT-CWT) [35] for its unique advantages in our task: (1) it provides approximate shift invariance, which helps maintain consistent feature representations; (2) it offers enhanced directional selectivity in high-frequency bands, which is crucial for capturing edge and texture information in RGB-T images; (3) it ensures perfect reconstruction with limited redundancy, making it computationally efficient while preserving important feature information. Moreover, wavelet transform is particularly suitable for RGB-T SOD tasks as it enables effective multi-scale feature decomposition and maintains complete spatial structure information during feature fusion, which is essential for accurate object localization. The frequency domain analysis it provides also helps in better capturing complementary information between RGB and thermal modalities, especially in challenging scenarios with poor illumination or complex backgrounds. The module’s design adopts long-range connections, directly linking the output of the previous layer to the next layer. These connections allow our method to better capture contextual information and achieve higher segmentation accuracy. The structure is shown in Figure 4.
The input features for the method are the fused features from two adjacent scales, namely F i f u s e and F i + 1 f u s e . First, F i + 1 f u s e undergoes an upsampling operation to restore it to the same spatial dimensions as F i f u s e . Subsequently, the features at adjacent scales are processed with the DT-CWT, which decomposes each feature into four sub-bands of different frequencies: a low-frequency sub-band (LL) and three high-frequency sub-bands (LH, HL, HH). Specifically, the LL sub-band captures the approximate coefficients and overall object structure, while the high-frequency sub-bands capture different directional details: HL for horizontal edges, LH for vertical edges, and HH for diagonal features. This allows for separate processing of high-frequency and low-frequency information. The specific process is as follows:
( F n + 1 L L , F n + 1 H L , F n + 1 L H , F n + 1 H H ) = D T C W T ( U p ( F n + 1 f u s e ) )
( F n L L , F n H L , F n L H , F n H H ) = D T C W T ( F n f u s e )
Here, F n + 1 and F n represent the feature information at adjacent levels, while D T C W T denotes the dual-tree complex wavelet transform operation.
After obtaining the low-frequency and high-frequency sub-bands, we apply different processing methods to these types of information to enhance the internal correlations between features. Given the different roles of these sub-bands, we prioritize their processing accordingly: Specifically, the low-frequency sub-band mainly contains global structural information, which is crucial for identifying the general location and rough boundaries of salient objects, thus receiving focused processing through a self-attention mechanism. The three high-frequency sub-bands (LH, HL, HH) collectively provide complementary directional information about edges and textures, and are therefore processed using 3D convolution operations, enabling the model to more effectively capture the global shape and local details of salient objects. Specifically, we transform the low-frequency sub-band features into query (Q), key (K), and value (V) matrices through linear mapping:
Q = X L L W Q , K = X L L W K , V = X L L W V
Attention ( Q , K , V ) = softmax Q K T d k V
where ( W Q , W K , W V ) are the learned weight matrices, and the similarity between the Query and Key is calculated using the dot product, with 1 d k serving as a scaling factor to prevent excessively large values. The result is then normalized using the softmax function.
Subsequently, the low-frequency sub-bands of F n and F n + 1 , enhanced by the self-attention mechanism, are concatenated to obtain the fused low-frequency feature representation, denoted as c a t l l . The remaining high-frequency sub-bands undergo convolutional processing, where 3D convolution is employed to handle the increased dimensionality of the high-frequency information. This results in the corresponding fused high-frequency features. The fused low-frequency features are then concatenated with the fused high-frequency features. An inverse transform using the inverse dual-tree complex wavelet transform (IDTCWT) is applied to reconstruct the fused features, thereby integrating information from different frequencies at a multi-scale level. Finally, the two resulting feature maps are combined element-wise to yield the final output of multi-scale fused features. We reconstruct the fused low-frequency and high-frequency features back to the original space using the inverse wavelet transform, producing two restored feature maps. These feature maps are subsequently added element-wise to generate the final output of multi-scale fused features.
F r = I D T C W T ( c a t l l , F n H L , F n L H , F n H H )
F l = I D T C W T ( c a t l l , F n + 1 H L , F n + 1 L H , F n + 1 H H )
S = G e L U ( B N ( C o n v ( F r + F l ) ) )
c a t l l represents the concatenated low-frequency features, I D T C W T denotes the inverse wavelet transform, F r and F l represent the two aggregated sub-features, and S stands for the aggregated prediction value.
In summary, this module leverages multi-scale feature fusion in the frequency domain to not only preserve the global structure but also enhance fine-grained detection of salient targets, thereby improving overall performance.

3.4. Loss Function

In RGB-T salient object detection tasks, choosing an appropriate loss function is crucial for achieving accurate predictions. While several loss functions such as Focal Loss [36] and Dice Loss [37] are commonly used in segmentation tasks, we opt for a combination of binary cross entropy (BCE) [38] loss and intersection over union (IoU) [39] loss to supervise the network. This combination leverages the complementary advantages of both loss functions. The IoU loss primarily focuses on the spatial overlap between the prediction and the ground truth, while BCE loss emphasizes the predicted probability for each pixel. Compared to Focal Loss which mainly addresses class imbalance issues, our combination better handles the structural consistency requirements in RGB-T saliency detection. Additionally, unlike Dice Loss which may be sensitive to small objects and boundary regions, our approach provides better stability when dealing with thermal images where object boundaries might be less distinct.
Loss B C E = h = 1 H w = 1 W S ( h , w ) · log ( S ^ ( h , w ) ) + ( 1 S ( h , w ) ) · log ( 1 S ^ ( h , w ) )
S ^ is the predicted saliency probability value at position ( h , w ) , and S is the ground truth saliency label value at position ( h , w ) . The variables h and w represent the height and width of the input image, respectively, while L o s s B C E denotes the computed BCE loss value.
IoU = Intersection Union = h = 1 H w = 1 W ( S ^ ( h , w ) · S ( h , w ) ) h = 1 H w = 1 W ( S ^ ( h , w ) + S ( h , w ) ) h = 1 H w = 1 W ( S ^ ( h , w ) · S ( h , w ) )
IoULoss = 1 IoU
I n t e r s e c t i o n refers to the salient target areas that the model correctly predicts, while U n i o n represents the combined area of the predicted saliency region and the ground truth saliency region. S ^ ( h , w ) denotes the predicted probability of the salient target at position ( h , w ) , and S ( h , w ) represents the label value at that position. The total loss value can be expressed as
L o s s a l l = i = 1 3 BCE ( S i ^ , S ) + IoU ( S i ^ , S )
S i ^ represents the predicted values at different stages of the network, and S denotes the ground truth map.

4. Experiments and Results

4.1. Datasets

To comprehensively evaluate our method, we conducted experiments using three RGB-T datasets: VT821 [6], VT1000 [40], and VT5000 [41]. The VT821 dataset consists of 821 manually registered RGB-T image pairs, which include several challenging single-modality failure cases. The VT1000 dataset contains 1000 well-registered image pairs, and the VT5000 dataset consists of 5000 pairs, featuring more complex scenes and higher levels of difficulty. For the experiments, we used 2500 image pairs from the VT5000 dataset as the training set, while the remaining data, along with the other two datasets, were used for testing. The sample data used for the three RGB-T datasets are shown in Figure 5.

4.2. Implementation Details

We implemented the proposed method using PyTorch 2.3.0 and CUDA 12.1, and conducted training and testing on a computer equipped with an NVIDIA RTX 3090 24 GB GPU. Following the standard settings of Swin Transformer, we resized all input images to 384 × 384. Data augmentation techniques such as flipping and cropping were applied to enhance model robustness. The Adam optimizer was used for training, with a batch size of 16 (determined through experiments testing values between 8 and 32) for 100 epochs. The initial learning rate was set to 1 × 10 4 (selected via grid search in the range [ 1 × 10 5 , 1 × 10 2 ]) for optimal convergence stability.

4.3. Evaluation Metrics

In this study, we used four evaluation metrics to assess the performance of our method: Structure-measure ( S m ) [42], E-measure ( E m ) [43], F-measure ( F m ) [44], and mean absolute error (MAE) [45]. Specifically, S m is a spatial similarity metric used to evaluate the structural similarity between the predicted saliency map and the ground truth. E m is an enhanced alignment metric that considers both pixel-level and image-level statistical information. F m assesses the performance by combining precision and recall. MAE represents the average absolute error between the predicted values and the ground truth (GT), with smaller values indicating more accurate predictions. In addition to these four metrics, we also used the precision–recall (PR) curve and F-measure curve to evaluate the method’s performance.

4.4. Comparisons with State-of-the-Art Methods

4.4.1. Quantitative Comparisons

We compared our method with 22 state-of-the-art approaches, including MTMR [6], M3S-NIR [7], MIDD [9], SGDL [40], ADF [41], TNet [12], MGAI [31], DMRA [46], EGNet [47], BASNet [48], PoolNet [49], R3Net [50], PFA [51], DCNet [30], CPD [52], S2MA [53], GRNet [54], SSOD [55], CGFNet [56], ACMANet [57], CAVER [58], and LSNet [59]. All experimental data were obtained from resources published by the respective authors.
As shown in Table 1, our method achieved excellent results across all three datasets. On the VT5000 dataset, our method reached an S m value of 0.917, an MAE of 0.024, an F β of 0.909, and an E m of 0.958. Compared to the second-best approach, our method improved S m by 1.88%, reduced the MAE by 1.42%, and increased E m by 1.48%. Our approach also demonstrated considerable improvements on the other two datasets. Traditional methods, such as MTMR and M3S-NIR, often rely heavily on prior knowledge, making it difficult for them to effectively represent features when combining the two modalities, resulting in unsatisfactory detection outcomes. RGB-based methods, without the complement of thermal information, tend to perform poorly in extreme environments. Additionally, as illustrated in Figure 6, the PR and F-measure curves further reflect our method’s performance across different datasets.

4.4.2. Qualitative Comparison

Figure 7 presents a visual comparison between our method and several state-of-the-art approaches. The figure demonstrates that our method achieves more accurate detection results across various challenging scenarios, such as large targets, small targets, low contrast, overexposure, multiple targets, high noise levels, and target center offset. Compared to other methods, our approach consistently delivers better performance in these complex situations.
In the first row of Figure 7, which shows a noisy scene, with the advantage of our CCM module’s capability to adaptively fuse complementary information from RGB and thermal modalities, our method clearly detects the bottle’s shape with smooth edges and accurate detection while minimizing noise. In contrast, other methods often suffer from false positives and missed detections in this scenario, mainly due to their inability to effectively process information fusion between the two modalities. For example, while CAVER can detect the bottle, its result is somewhat distorted. Traditional methods like M3S-NIR and SGDL, due to their simple feature fusion strategies, produce incomplete contours, blurred shapes, and exhibit information loss.
In the second and sixth rows of Figure 7, leveraging the effective processing of high-frequency and low-frequency features by the FSW module, our method demonstrates superior performance in geometric shape detection. Specifically, the FSW module accurately captures target edge details through high-frequency components while maintaining structural integrity via low-frequency components, whereas other methods exhibit significant errors in edge and shape preservation. In the fourth row, where the target presents complex spatial structures, our approach, through the frequency decomposition mechanism of the FSW module, effectively preserves the intricate structural details and achieves precise edge localization, while other methods’ results show notable distortion. In the fifth row, with multiple salient targets positioned near image boundaries, the low-frequency features from the FSW module facilitate the preservation of global contextual information, resulting in more complete detection results, whereas most alternative methods produce incomplete results with higher noise levels. Finally, in the seventh and ninth rows, even under challenging conditions, such as RGB image overexposure or limited target feature information in thermal images, our method maintains accurate salient target extraction through the effective integration of different frequency components by the FSW module.

4.5. Ablation Study

The experimental results demonstrate the necessity and effectiveness of multi-modal fusion in our approach. As shown in Table 2, by comparing three different modal combinations (RGB-RGB, T-T, and RGB-T), we observe that using complementary modalities (RGB-T) consistently outperforms using a single modality twice (RGB-RGB or T-T) across all evaluation metrics. This validates that the performance improvement comes from the effective fusion of complementary information between RGB and thermal modalities, rather than the complexity of network architecture. The results strongly support our design choice of leveraging both RGB and thermal inputs for more robust saliency detection. To confirm the impact of essential components in our model, we performed experiments where we either removed or substituted these elements from the complete configuration. The resulting quantitative data are presented in Table 3. Furthermore, to gain deeper insights into how these components contribute to performance, we included visualizations of the ablation study outcomes, displayed in Figure 8.
In this work, we incorporated three main modules: CCM, FSW, and DTCWT. To validate the contribution of each, we conducted experiments by either omitting or substituting them. The results of these experiments are detailed in Table 3. When the CCM module was replaced with direct addition, the model’s performance on all three datasets significantly declined. Without CCM, the MAE increased by 0.01 and S m decreased by 0.031 on VT5000. This is due to insufficient interaction between the two modalities, resulting in excessive noise in the fused information. The visualized ablation results reveal noticeable distortions and blurred edges when CCM is removed. When the FSW module was removed, the MAE increased by 0.013. The visualized results show that without FSW, the detection maps contain substantial noise, and much of the spatial structure of the detected objects is lost. This happens because removing the FSW module reduces the model’s attention to both global and fine-grained features, making it less effective in utilizing spatial features. We also validated the effectiveness of the network’s supervision approach. When using only BCE or IoU separately, the results were inferior compared to the combined supervision approach, indicating that joint supervision improves performance.

4.6. Failure Cases

Despite the strong performance of our method in various challenging scenarios, we identified some key failure cases that require further investigation. As shown in Figure 9 (first row), our method struggles to detect occluded parts when the target is occluded and its color is similar to the occluder in the RGB image. This limitation arises from the difficulty in distinguishing the features of the target and the occluder, especially when thermal information is also compromised. To address this challenge, future work could introduce temporal consistency analysis and develop occlusion-aware feature aggregation mechanisms. Additionally, as shown in Figure 9 (second row), performance degrades when the thermal image provides limited contrast due to similar temperatures, and the target is small in the RGB image. In these cases, accurate localization becomes particularly challenging. To enhance performance in such scenarios, we propose developing multi-scale feature enhancement modules and introducing adaptive fusion strategies that dynamically adjust modality weights based on scene conditions. To systematically address these limitations, we could design a local–global detection mechanism in the future to enhance feature discrimination in complex scenes and integrate context-aware fusion strategies for challenging scenarios.

5. Conclusions

In this work, we present a wavelet-driven multi-band feature fusion approach for RGB-T salient object detection. To achieve effective multi-modal fusion, we design a feature fusion module based on gating and attention mechanisms, which assigns weights to different modal components and realizes cross-modal fusion between RGB and thermal features through channel splitting. Furthermore, we introduce a feature selection wavelet transformer module to enhance fine-grained multi-scale feature representation and global context modeling capabilities while preserving complete spatial structure information of objects. The effectiveness of our proposed method is demonstrated through comprehensive experimental evaluations.

Author Contributions

Conceptualization, X.W. and J.Z.; methodology, J.Z.; software, J.Z.; validation, J.Z.; formal analysis, J.Z. and K.S.; investigation, J.Z.; resources, X.W.; data curation, J.Z.; writing—original draft, J.Z.; writing—review and editing, J.Z., X.Y. and Y.H.; visualization, J.Z., Y.H. and X.Y.; supervision, X.W.; project administration, X.W.; funding acquisition, X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China (Grant No. 62306186) and the Scientific Research Project of Liaoning Education Department (Grant No. LJ212410142035).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this study are available at https://chenglongli.cn/code-dataset/ (accessed on 1 November 2024). The evaluation tool is available at https://github.com/lartpang/PySODEvalToolkit (accessed on 1 November 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Deng, B.; Liu, D.; Cao, Y.; Liu, H.; Yan, Z.; Chen, H. CFRNet: Cross-Attention-Based Fusion and Refinement Network for Enhanced RGB-T Salient Object Detection. Sensors 2024, 24, 7146. [Google Scholar] [CrossRef] [PubMed]
  2. Song, K.; Xue, X.; Wen, H.; Ji, Y.; Yan, Y.; Meng, Q. Misaligned Visible-Thermal Object Detection: A Drone-based Benchmark and Baseline. IEEE Trans. Intell. Veh. 2024, 1–12, in press. [Google Scholar] [CrossRef]
  3. Ramm, R.; de Dios Cruz, P.; Heist, S.; Kühmstedt, P.; Notni, G. Fusion of Multimodal Imaging and 3D Digitization Using Photogrammetry. Sensors 2024, 24, 2290. [Google Scholar] [CrossRef] [PubMed]
  4. Qureshi, I.; Yan, J.; Abbas, Q.; Shaheed, K.; Riaz, A.B.; Wahid, A.; Khan, M.W.J.; Szczuko, P. Medical image segmentation using deep semantic-based methods: A review of techniques, applications and emerging trends. Inf. Fusion 2023, 90, 316–352. [Google Scholar] [CrossRef]
  5. Song, K.; Zhao, Y.; Huang, L.; Yan, Y.; Meng, Q. RGB-T image analysis technology and application: A survey. Eng. Appl. Artif. Intell. 2023, 120, 105919. [Google Scholar] [CrossRef]
  6. Wang, G.; Li, C.; Ma, Y.; Zheng, A.; Tang, J.; Luo, B. RGB-T Saliency Detection Benchmark: Dataset, Baselines, Analysis and a Novel Approach. In Proceedings of the Image and Graphics Technologies and Applications; Wang, Y., Jiang, Z., Peng, Y., Eds.; Springer: Singapore, 2018; pp. 359–369. [Google Scholar]
  7. Tu, Z.; Xia, T.; Li, C.; Lu, Y.; Tang, J. M3S-NIR: Multi-modal Multi-scale Noise-Insensitive Ranking for RGB-T Saliency Detection. In Proceedings of the 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), San Jose, CA, USA, 28–30 March 2019; pp. 141–146. [Google Scholar] [CrossRef]
  8. Huang, L.; Song, K.; Gong, A.; Liu, C.; Yan, Y. RGB-T Saliency Detection via Low-Rank Tensor Learning and Unified Collaborative Ranking. IEEE Signal Process. Lett. 2020, 27, 1585–1589. [Google Scholar] [CrossRef]
  9. Tu, Z.; Li, Z.; Li, C.; Lang, Y.; Tang, J. Multi-Interactive Dual-Decoder for RGB-Thermal Salient Object Detection. IEEE Trans. Image Process. 2021, 30, 5678–5691. [Google Scholar] [CrossRef]
  10. Bi, H.; Wu, R.; Liu, Z.; Zhang, J.; Zhang, C.; Xiang, T.Z.; Wang, X. PSNet: Parallel symmetric network for RGB-T salient object detection. Neurocomputing 2022, 511, 410–425. [Google Scholar] [CrossRef]
  11. Zhang, Q.; Xi, R.; Xiao, T.; Huang, N.; Luo, Y. Enabling modality interactions for RGB-T salient object detection. Comput. Vis. Image Underst. 2022, 222, 103514. [Google Scholar] [CrossRef]
  12. Cong, R.; Zhang, K.; Zhang, C.; Zheng, F.; Zhao, Y.; Huang, Q.; Kwong, S. Does Thermal Really Always Matter for RGB-T Salient Object Detection? IEEE Trans. Multimed. 2023, 25, 6971–6982. [Google Scholar] [CrossRef]
  13. Huo, F.; Zhu, X.; Zhang, Q.; Liu, Z.; Yu, W. Real-Time One-Stream Semantic-Guided Refinement Network for RGB-Thermal Salient Object Detection. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  14. Zhou, W.; Guo, Q.; Lei, J.; Yu, L.; Hwang, J.N. ECFFNet: Effective and Consistent Feature Fusion Network for RGB-T Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 1224–1235. [Google Scholar] [CrossRef]
  15. Song, K.; Wen, H.; Xue, X.; Huang, L.; Ji, Y.; Yan, Y. Modality Registration and Object Search Framework for UAV-Based Unregistered RGB-T Image Salient Object Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  16. Zhou, W.; Sun, F.; Jiang, Q.; Cong, R.; Hwang, J.N. WaveNet: Wavelet Network with Knowledge Distillation for RGB-T Salient Object Detection. IEEE Trans. Image Process. 2023, 32, 3027–3039. [Google Scholar] [CrossRef]
  17. Yue, H.; Guo, J.; Yin, X.; Zhang, Y.; Zheng, S.; Zhang, Z.; Li, C. Salient object detection in low-light images via functional optimization-inspired feature polishing. Knowl.-Based Syst. 2022, 257, 109938. [Google Scholar] [CrossRef]
  18. Siris, A.; Jiao, J.; Tam, G.K.; Xie, X.; Lau, R.W. Scene Context-Aware Salient Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 4156–4166. [Google Scholar]
  19. Wang, Q.; Liu, Y.; Xiong, Z.; Yuan, Y. Hybrid Feature Aligned Network for Salient Object Detection in Optical Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  20. Wu, Z.; Li, S.; Chen, C.; Qin, H.; Hao, A. Salient Object Detection via Dynamic Scale Routing. IEEE Trans. Image Process. 2022, 31, 6649–6663. [Google Scholar] [CrossRef]
  21. Wu, Y.H.; Liu, Y.; Zhang, L.; Cheng, M.M.; Ren, B. EDN: Salient Object Detection via Extremely-Downsampled Network. IEEE Trans. Image Process. 2022, 31, 3125–3136. [Google Scholar] [CrossRef]
  22. Li, J.; Qiao, S.; Zhao, Z.; Xie, C.; Chen, X.; Xia, C. Rethinking Lightweight Salient Object Detection via Network Depth-Width Tradeoff. IEEE Trans. Image Process. 2023, 32, 5664–5677. [Google Scholar] [CrossRef]
  23. Jin, W.D.; Xu, J.; Han, Q.; Zhang, Y.; Cheng, M.M. CDNet: Complementary Depth Network for RGB-D Salient Object Detection. IEEE Trans. Image Process. 2021, 30, 3376–3390. [Google Scholar] [CrossRef]
  24. Chen, Q.; Zhang, Z.; Lu, Y.; Fu, K.; Zhao, Q. 3-D Convolutional Neural Networks for RGB-D Salient Object Detection and Beyond. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 4309–4323. [Google Scholar] [CrossRef] [PubMed]
  25. Song, M.; Song, W.; Yang, G.; Chen, C. Improving RGB-D Salient Object Detection via Modality-Aware Decoder. IEEE Trans. Image Process. 2022, 31, 6124–6138. [Google Scholar] [CrossRef] [PubMed]
  26. Sun, F.; Ren, P.; Yin, B.; Wang, F.; Li, H. CATNet: A Cascaded and Aggregated Transformer Network for RGB-D Salient Object Detection. IEEE Trans. Multimed. 2024, 26, 2249–2262. [Google Scholar] [CrossRef]
  27. Wu, Z.; Allibert, G.; Meriaudeau, F.; Ma, C.; Demonceaux, C. HiDAnet: RGB-D Salient Object Detection via Hierarchical Depth Awareness. IEEE Trans. Image Process. 2023, 32, 2160–2173. [Google Scholar] [CrossRef]
  28. Zhang, Q.; Qin, Q.; Yang, Y.; Jiao, Q.; Han, J. Feature Calibrating and Fusing Network for RGB-D Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 1493–1507. [Google Scholar] [CrossRef]
  29. Gao, W.; Liao, G.; Ma, S.; Li, G.; Liang, Y.; Lin, W. Unified Information Fusion Network for Multi-Modal RGB-D and RGB-T Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 2091–2106. [Google Scholar] [CrossRef]
  30. Tu, Z.; Li, Z.; Li, C.; Tang, J. Weakly Alignment-Free RGBT Salient Object Detection with Deep Correlation Network. IEEE Trans. Image Process. 2022, 31, 3752–3764. [Google Scholar] [CrossRef]
  31. Song, K.; Huang, L.; Gong, A.; Yan, Y. Multiple Graph Affinity Interactive Network and a Variable Illumination Dataset for RGBT Image Salient Object Detection. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3104–3118. [Google Scholar] [CrossRef]
  32. Wang, H.; Song, K.; Huang, L.; Wen, H.; Yan, Y. Thermal images-aware guided early fusion network for cross-illumination RGB-T salient object detection. Eng. Appl. Artif. Intell. 2023, 118, 105640. [Google Scholar] [CrossRef]
  33. Zhou, H.; Tian, C.; Zhang, Z.; Li, C.; Ding, Y.; Xie, Y.; Li, Z. Position-Aware Relation Learning for RGB-Thermal Salient Object Detection. IEEE Trans. Image Process. 2023, 32, 2593–2607. [Google Scholar] [CrossRef]
  34. Huang, Z.; Wang, X.; Wei, Y.; Huang, L.; Shi, H.; Liu, W.; Huang, T.S. CCNet: Criss-Cross Attention for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 6896–6908. [Google Scholar] [CrossRef] [PubMed]
  35. Goyal, A.; Meenpal, T. Patch-Based Dual-Tree Complex Wavelet Transform for Kinship Recognition. IEEE Trans. Image Process. 2021, 30, 191–206. [Google Scholar] [CrossRef] [PubMed]
  36. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed]
  37. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  38. de Boer, P.T.; Kroese, D.P.; Mannor, S.; Rubinstein, R.Y. A Tutorial on the Cross-Entropy Method. Ann. Oper. Res. 2005, 134, 19–67. [Google Scholar] [CrossRef]
  39. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar] [CrossRef]
  40. Tu, Z.; Xia, T.; Li, C.; Wang, X.; Ma, Y.; Tang, J. RGB-T Image Saliency Detection via Collaborative Graph Learning. IEEE Trans. Multimed. 2020, 22, 160–173. [Google Scholar] [CrossRef]
  41. Tu, Z.; Ma, Y.; Li, Z.; Li, C.; Xu, J.; Liu, Y. RGBT Salient Object Detection: A Large-Scale Dataset and Benchmark. IEEE Trans. Multimed. 2023, 25, 4163–4176. [Google Scholar] [CrossRef]
  42. Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4548–4557. [Google Scholar]
  43. Fan, D.P.; Ji, G.P.; Qin, X.; Cheng, M.M. Cognitive vision inspired object segmentation metric and loss function. Sci. Sin. Informationis 2021, 6, 5. [Google Scholar]
  44. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  45. Perazzi, F.; Krähenbühl, P.; Pritch, Y.; Hornung, A. Saliency filters: Contrast based filtering for salient region detection. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 733–740. [Google Scholar]
  46. Piao, Y.; Ji, W.; Li, J.; Zhang, M.; Lu, H. Depth-induced multi-scale recurrent attention network for saliency detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7254–7263. [Google Scholar]
  47. Zhao, J.X.; Liu, J.J.; Fan, D.P.; Cao, Y.; Yang, J.; Cheng, M.M. EGNet: Edge guidance network for salient object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8779–8788. [Google Scholar]
  48. Qin, X.; Fan, D.P.; Huang, C.; Diagne, C.; Zhang, Z.; Sant’Anna, A.C.; Suarez, A.; Jagersand, M.; Shao, L. Boundary-aware segmentation network for mobile and web applications. arXiv 2021, arXiv:2101.04704. [Google Scholar]
  49. Liu, J.J.; Hou, Q.; Cheng, M.M.; Feng, J.; Jiang, J. A simple pooling-based design for real-time salient object detection. In Proceedings of theIEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3917–3926. [Google Scholar]
  50. Deng, Z.; Hu, X.; Zhu, L.; Xu, X.; Qin, J.; Han, G.; Heng, P.A. R3net: Recurrent residual refinement network for saliency detection. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; AAAI Press: Menlo Park, CA, USA, 2018; Volume 684690. [Google Scholar]
  51. Zhao, T.; Wu, X. Pyramid feature attention network for saliency detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3085–3094. [Google Scholar]
  52. Wu, Z.; Su, L.; Huang, Q. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3907–3916. [Google Scholar]
  53. Liu, N.; Zhang, N.; Han, J. Learning Selective Self-Mutual Attention for RGB-D Saliency Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13753–13762. [Google Scholar] [CrossRef]
  54. Wen, H.; Song, K.; Huang, L.; Wang, H.; Yan, Y. Cross-modality salient object detection network with universality and anti-interference. Knowl.-Based Syst. 2023, 264, 110322. [Google Scholar] [CrossRef]
  55. Liu, Z.; Huang, X.; Zhang, G.; Fang, X.; Wang, L.; Tang, B. Scribble-supervised rgb-t salient object detection. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; pp. 2369–2374. [Google Scholar]
  56. Wang, J.; Song, K.; Bao, Y.; Huang, L.; Yan, Y. CGFNet: Cross-guided fusion network for RGB-T salient object detection. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 2949–2961. [Google Scholar] [CrossRef]
  57. Xu, C.; Li, Q.; Zhou, Q.; Jiang, X.; Yu, D.; Zhou, Y. Asymmetric cross-modal activation network for RGB-T salient object detection. Knowl.-Based Syst. 2022, 258, 110047. [Google Scholar] [CrossRef]
  58. Pang, Y.; Zhao, X.; Zhang, L.; Lu, H. CAVER: Cross-Modal View-Mixed Transformer for Bi-Modal Salient Object Detection. IEEE Trans. Image Process. 2023, 32, 892–904. [Google Scholar] [CrossRef]
  59. Zhou, W.; Zhu, Y.; Lei, J.; Yang, R.; Yu, L. LSNet: Lightweight spatial boosting network for detecting salient objects in RGB-thermal images. IEEE Trans. Image Process. 2023, 32, 1329–1340. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The design concept of our method.
Figure 1. The design concept of our method.
Sensors 24 08159 g001
Figure 2. The overall architecture of our proposed method. Where U P represents upsampling.
Figure 2. The overall architecture of our proposed method. Where U P represents upsampling.
Sensors 24 08159 g002
Figure 3. The detailed architecture of CCM.
Figure 3. The detailed architecture of CCM.
Sensors 24 08159 g003
Figure 4. The structure of an FSW.
Figure 4. The structure of an FSW.
Sensors 24 08159 g004
Figure 5. The dataset used in this experiment: (a) VT821, (b) VT1000, (c) VT5000.
Figure 5. The dataset used in this experiment: (a) VT821, (b) VT1000, (c) VT5000.
Sensors 24 08159 g005
Figure 6. Quantitative comparison of our method with other state-of-the-art methods: (a) PR curve. (b) Fm curve.
Figure 6. Quantitative comparison of our method with other state-of-the-art methods: (a) PR curve. (b) Fm curve.
Sensors 24 08159 g006
Figure 7. Qualitative comparison of our model with eleven recent state-of-the-art models.
Figure 7. Qualitative comparison of our model with eleven recent state-of-the-art models.
Sensors 24 08159 g007
Figure 8. The visualization of ablation, where ’w/o’ stands for the absence of the corresponding module.
Figure 8. The visualization of ablation, where ’w/o’ stands for the absence of the corresponding module.
Sensors 24 08159 g008
Figure 9. Visualization of some typical failure cases in our method.
Figure 9. Visualization of some typical failure cases in our method.
Sensors 24 08159 g009
Table 1. Quantitative results of our method and 22 compared methods. The best two results are shown in red and green. ↑ indicates that higher values are better, while ↓ indicates that lower values are better.
Table 1. Quantitative results of our method and 22 compared methods. The best two results are shown in red and green. ↑ indicates that higher values are better, while ↓ indicates that lower values are better.
MethodsVT821VT1000VT5000
S m MAE F β max E m max S m MAE F β max E m max S m MAE F β max E m max
M3S-NIR0.7230.1400.7380.8370.7260.1450.7350.8280.6520.1680.5960.760
MTMR0.7250.1080.6900.8120.7060.1190.7150.8360.6800.1140.6130.792
SGDL0.7650.0850.7350.8400.7870.0900.7700.8590.7510.0890.6950.829
S2MA0.8290.0810.7790.8550.9210.0290.9130.9520.8550.0550.8120.895
PFA0.7610.0960.7110.8540.8130.0780.8050.8880.7480.0990.7190.857
DMRA0.6660.2160.6610.7660.7840.1240.7980.8630.6590.1840.6010.758
LSNet0.8790.0330.8450.9210.9260.0230.9220.9630.8770.0370.8500.924
BASNet0.8230.0670.7630.8580.9090.0300.9010.9440.8390.0540.7910.884
ADF0.8100.0770.7520.8390.9100.0340.9080.9500.8640.0480.8370.911
CPD0.8180.0790.7580.8620.9070.0310.8970.9470.8550.0460.8180.905
DCNet0.8770.0330.8510.9200.9230.0210.9190.9610.8720.0350.8530.925
EGNet0.8300.0630.7560.8570.9100.0330.8980.9450.8530.0500.8080.893
MIDD0.8710.0450.8510.9180.9070.0290.9060.9520.8560.0460.8390.913
PoolNet0.7880.0820.7070.8420.8490.0630.8260.9040.7880.0800.7270.852
R3Net0.7820.0810.7110.8190.8860.0370.8760.9390.8120.0590.7530.863
MGAI0.8910.0310.8730.9350.9290.0210.9260.9660.8830.0340.8620.931
CGFNet0.8800.0380.8660.9200.9230.0230.9230.9590.8830.0350.8690.927
SSOD0.8950.0270.8780.9420.9250.0200.9220.9640.8770.0330.8590.933
ACMANet0.8830.0350.8510.9260.9270.0210.9230.9640.8870.0330.8710.936
TNet0.8990.0300.8880.9380.9290.0210.9300.9660.8950.0330.8810.937
GRNet0.8930.0310.8660.9330.9310.0180.9270.9660.8880.0340.8700.931
CAVER0.8980.0260.8770.9340.9380.0160.9390.9730.9000.0280.8820.944
OURS0.9100.0250.8920.9430.9420.0150.9460.9790.9170.0240.9090.958
Table 2. Ablation study on different modal combinations.
Table 2. Ablation study on different modal combinations.
VT821VT1000VT5000
S m MAE F β max E m max S m MAE F β max E m max S m MAE F β max E m max
OnlyRGB0.8850.0340.8450.9120.9390.0160.9390.9740.9030.0270.8900.947
OnlyT0.8460.0430.8140.9070.9080.0270.9030.9560.8710.0380.8440.927
OURS0.9100.0250.8920.9430.9420.0150.9460.9790.9170.0240.9090.958
Table 3. Ablation experiment results. w/o means without the module or supervision.
Table 3. Ablation experiment results. w/o means without the module or supervision.
Setting TypeConfigurationVT821VT1000VT5000
S m MAE F β max S m MAE F β max S m MAE F β max
Modulew/o CCM0.8830.0300.8630.9250.0240.9170.8860.0340.876
w/o FSW0.8750.0340.8590.9220.0250.9130.8790.0370.868
w/o DT-CWT0.9050.0280.8850.9350.0180.9450.9100.0270.897
OURS0.9100.0250.8920.9420.0150.9460.9170.0240.909
Lossw/o BCE0.9070.0270.8890.9410.0160.9440.9110.0270.906
w/o IoU0.9080.0260.8910.9390.0160.9420.9100.0260.899
BCE+IoU0.9100.0250.8920.9420.0150.9460.9170.0240.909
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, J.; Wen, X.; He, Y.; Yang, X.; Song, K. Wavelet-Driven Multi-Band Feature Fusion for RGB-T Salient Object Detection. Sensors 2024, 24, 8159. https://doi.org/10.3390/s24248159

AMA Style

Zhao J, Wen X, He Y, Yang X, Song K. Wavelet-Driven Multi-Band Feature Fusion for RGB-T Salient Object Detection. Sensors. 2024; 24(24):8159. https://doi.org/10.3390/s24248159

Chicago/Turabian Style

Zhao, Jianxun, Xin Wen, Yu He, Xiaowei Yang, and Kechen Song. 2024. "Wavelet-Driven Multi-Band Feature Fusion for RGB-T Salient Object Detection" Sensors 24, no. 24: 8159. https://doi.org/10.3390/s24248159

APA Style

Zhao, J., Wen, X., He, Y., Yang, X., & Song, K. (2024). Wavelet-Driven Multi-Band Feature Fusion for RGB-T Salient Object Detection. Sensors, 24(24), 8159. https://doi.org/10.3390/s24248159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop