Literature DB >> 36015947

Dynamic Knowledge Distillation with Noise Elimination for RGB-D Salient Object Detection.

Guangyu Ren1, Yinxiao Yu2, Hengyan Liu1, Tania Stathaki1.   

Abstract

RGB-D salient object detection (SOD) demonstrates its superiority in detecting in complex environments due to the additional depth information introduced in the data. Inevitably, an independent stream is introduced to extract features from depth images, leading to extra computation and parameters. This methodology sacrifices the model size to improve the detection accuracy which may impede the practical application of SOD problems. To tackle this dilemma, we propose a dynamic knowledge distillation (DKD) method, along with a lightweight structure, which significantly reduces the computational burden while maintaining validity. This method considers the factors of both teacher and student performance within the training stage and dynamically assigns the distillation weight instead of applying a fixed weight on the student model. We also investigate the issue of RGB-D early fusion strategy in distillation and propose a simple noise elimination method to mitigate the impact of distorted training data caused by low quality depth maps. Extensive experiments are conducted on five public datasets to demonstrate that our method can achieve competitive performance with a fast inference speed (136FPS) compared to 12 prior methods.

Entities:  

Keywords:  RGB-D; dynamic knowledge distillation; salient object detection

Mesh:

Substances:

Year:  2022        PMID: 36015947      PMCID: PMC9416116          DOI: 10.3390/s22166188

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.847


1. Introduction

Salient object detection (SOD) aims at locating prominent objects in a given scenario under consideration. In recent years, SOD has attracted significant attention and substantial progress has been demonstrated in the field. The object detection task can be treated as a pre-processing methodology that can be subsequently used in diverse fields, such as image understanding [1], video detection and segmentation [2], semantic segmentation [3], object tracking [4], person re-identification [5] and others. However, due to the complicated real-life scenarios, RGB-based SOD still fails in generating satisfactory prediction maps. In order to overcome this issue and obtain better detection performance in complex scenarios, depth images along with an independent network have been introduced to provide supplementary information. Specifically, Figure 1 illustrates three fusion methods in RGB-D based SOD. Existing state-of-the-art methods mainly adopt late fusion or multi-scale fusion and focus on designing feature-enhanced modules and complicated feature-fusion modules, which indeed improve the overall detection results. However, due to the processing of a high volume of information, the models tend to become extremely complicated, leading to an issue that weakens the practicality of SOD using RGB-D data. In addition to the regular strategies, recent approaches [6,7] propose joint learning frameworks and treat the RGB-D based SOD as a multi-task learning problem. However, these frameworks employ extra network branches and supervision labels, which cause an analogous problem with the aforementioned frameworks. Different from them, the early fusion in Figure 1c integrates the separate inputs into a unified representation before the feature extraction process. It provides an alternative strategy to lighten the model but suffers from the noise issue caused by low-quality depth information. That motivates us to explore the potential of early fusion from a novel perspective and compress the model size for SOD while maintaining high detection accuracy.
Figure 1

Comparison of three fusion strategies.

Recently, knowledge distillation (KD) has been proposed [8] to transfer knowledge from a large model to a smaller one. The main idea is that a small student model mimics a cumbersome model, namely, a so-called teacher model, to achieve competitive performance. The cumbersome network has a larger knowledge capacity than smaller models, but this capacity may not be utilized for its full potential. In other words, a lightweight network can reach a similar performance to a cumbersome network by KD without increasing the number of parameters. Similar to human behaviours, this teacher–student learning process can be implemented by a simple and effective way, which forces the student model to directly learn the final prediction of the teacher model. KD has been applied in a range of machine learning applications. Zhang [9] utilizes the KD to RGB saliency detection and proposes an efficient model by reducing the number of channels. Piao [10] explores the cross-modal distillation on RGB-D data and uses an adaptive weight to distil the depth knowledge from the teacher model. Nevertheless, both adjust their student networks according to the teacher networks and distillation strategies. In addition, the adaptive distillation [10] is proposed for the cross-modal distillation and only considers the performance of the teacher model, which limits the utilization of this KD method. In order to tackle the above issues from a new perspective, we use a concise framework based on the early fusion strategy for RGB-D based SOD and propose a dynamic knowledge distillation (DKD) weight to help the model pay more attention on hard samples by considering both teacher and student performance. We also investigate the issue of RGB-D early fusion strategy in distillation and propose a simple noise elimination method to mitigate the impact of distorted training data caused by low quality depth maps. Combing these two methods can lead to a reasonable distillation strategy for RGB-D saliency detection. Our final model achieves a good balance between accuracy and model size on widely used benchmarks as shown in Figure 2. In a nutshell, our main contributions can be summarised as follows:
Figure 2

Weighted F-measure and model size on NLPR and SIP datasets. Our lightweight model with simple early fusion can achieve satisfactory detection results over different datasets. Ours and Ours* indicate simple VGG16-based and VGG19-based FPN, respectively.

We propose a novel dynamic distillation strategy, which can adaptively assign the distillation weight by simultaneously considering the detection performance of the teacher and student networks within the training stage. As a result, the final model can pay more attention on hard samples and improve the overall performance. We propose a noise elimination method by taking full merit of knowledge prior from the teacher network to alleviate the impact of depth maps with low quality. The student network can take benefit from this method without increasing extra parameters and computations. We adopt a single stream for RGB-D SOD in order to bypass the depth network and avoid designing a complicated model. This single stream achieves competitive performance by only using VGG16 (57.9 MB) and VGG19 (78.2 MB), which are more applicable for practical use. Extensive experimental results on five benchmarks demonstrate that our methods can achieve competing performance within a fast lightweight architecture.

2. Related Work

RGB-D Salient Object Detection. RGB-D based SOD has obtained increasing attention in order to handle object detection tasks in complicated environments. Depth information is firstly introduced by [11], where they model the distribution of depth-induced saliency by using Gaussian mixture models. Zhao [12] proposes a feature-enhanced module and a contrast-enhanced net, which augments the contrast between the foreground and background by fluid pyramid integration. Pang [13] adopts multi-scale fusion and proposes a dynamic dilated pyramid module with adaptive receptive fields, which is generated by densely integrating cross-modal features. Chen [14] constructs a lightweight depth stream and designs a refinement network, which is progressively stacked by guided residual blocks. This method can alternately alleviate the mutual degradation and refine predictions in a progressive way. Zhou [15] leverages a novel feature aggregation network, which utilizes the K-nearest neighbor graph neural networks and the non-local module to dig the geometric cues and global semantic features. Zhang [6] proposes a multi-stage cascaded learning framework and transfers the maximization of joint entropy problem in multi-modal learning tasks to the minimization of mutual information, which can explicitly model the complementary information between the RGB image and depth data. These previous works focus on alleviating the impact of depth maps and enhancing the feature integration through delicate modules and networks. Knowledge Distillation. Knowledge Distillation was formally publicised by [8] in a teacher–student learning framework. This method proposes an effective way to compress model size and attempts to imitate the human beings’ learning mechanism. Cheng [16] designs mathematical metrics to quantify and compare the methods of learning from the teacher model and learning from raw data. They explain the superiority of KD in three aspects. First, more reliable visual concepts can be learned through KD. Second, KD makes the model able to learn various concepts simultaneously. Furthermore, learning from KD can generate more stable optimization directions in the training phase. Zheng [17] proposes a novel divide-and-conquer distillation strategy for dense object detection. They transfer the semantic and localization knowledge separately and show that the student takes more benefit from the original logits distillation than feature imitation. Yang [18] explores the difference between the features of students and teachers and proposes a focal distillation to make the student focus on the teacher’s critical pixels and channels. Then they further design a global distillation to help the student learn the relation between pixels. Xu [19] follows the human learning process and proposes a teacher–student collaborative KD. This method combines the teacher–student KD and student self-distillation to enhance the performance. However, the student self-distillation model is built by extra multiple exit classifiers from deep to shallow. Recently, KD has been used in SOD tasks. Zhang [9] designs the student model by reducing the amount of channels and applies multi-scale KD on the corresponding scales between teacher and student models. Piao [10] applies cross-modal distillation on RGB-D based SOD and proposes an adaptive distiller to distil the depth information, which alleviates the impact of low-quality depth maps. Different from the aforementioned methods, our method takes both the performance of teacher and student models into consideration and generates a dynamic weight to control the regular teacher–student KD process. In addition, we analyse the depth issues in the specific RGB-D SOD task and optimize the training phase through a threshold.

3. Methodologies

3.1. Overview

Existing methodologies for RGB-D SOD tend to build two-stream networks in order to process RGB and depth features separately. This two-stream design could improve detection performance but meanwhile introduces a large amount of parameters, which increases the complexity and reduces the practicality of models. Feature pyramid network (FPN) [20] is an effective structure which utilizes multi-scale features in different resolutions to achieve detection tasks. Figure 3 illustrates the overall framework, we do not focus on designing networks and only adopt the classic FPN based on a VGG16 and a VGG19 [21] as the student model. In order to obtain a stronger teacher model, we employ four receptive field blocks [22] in multi-scale layers to boost the detection performance. Considering different cross-modal fusion strategies, we choose the simple early fusion way which directly concatenates RGB images and depth images to form four-channel inputs. Similar to normal KD, we transfer the probability distribution of the final layer from the teacher model to the student model by utilizing the so-called DKD.
Figure 3

Our framework consists of two stages. We adopt a cumbersome network as the teacher model and a feature pyramid network as the student model.

3.2. Dynamic Knowledge Distillation

As mentioned above, KD benefits the student model but the weight of knowledge transfer is still hand-designed. Piao [10] proposes an adaptive weight for cross-modal distillation. However, in [10] they only distill the depth information by considering the performance of teacher model. In our method, we consider both performances of teacher and student networks and combine these two factors as a dynamic weight for KD. Concretely, the accuracy of teacher model represents the detection performance which also indicates the confidence of knowledge. Inspired by IOU [23] used in SOD, we design a dynamic factor to modulate the correct knowledge which can be transferred from the teacher model as follows: where and G represent the prediction of teacher model and the ground truth, respectively. indicates the confidence of knowledge which can be transferred to the student model. Then, we propose another dynamic factor to show the degree of desired knowledge for the student model as follows: where represents the prediction of student model. This dynamic factor is error rate of the current training sample. In other words, KD should also consider the current performance of student model. is inversely related to the accuracy between the output of student model and the ground truth. This indicates that hard samples which have large error rates need to learn more from the teacher model. Therefore, we propose a simple and effective formulation to find a plausible distillation weight : here tanh is treated as a scale function: More specifically, we define the by the weighted geometric mean of the knowledge confidence from teacher and the knowledge demand from student. We define the hyper-parameter to balance the ratio between the teacher and student networks. It is worth noting that large variation of leads to convergence issue in training phase. In this case, we further use a tanh function to scale the . The overall loss function can be formulated as: where is the Kullback-Leibler divergence loss and represents the cross-entropy loss. In the final network, we set the distillation temperature to 5 in and .

3.3. Noises Elimination with the DKD

As mentioned above, we simplify the procedure of KD and the student network architecture in RGB-D task. Concretely, we only distill the final output distribution and abandon the depth stream by concatenating RGB and depth maps to form a four-channel input. However, this fusion strategy suffers from the noise issue caused by low-quality depth information. As illustrated in Figure 4, we investigate the reasons that cause the distortion of depth maps: (1) besides the salient object, other objects in depth image dominate salient features; (2) low contrast between salient object and background in depth; (3) depth distortion caused by camera. Intuitively, training loss is supposed to reduce drastically if the training data are distorted. Therefore we propose that these depth maps can be treated as noises when combining with RGB maps and further set an accuracy threshold during KD to control the impact of noises: where indicates a small weight which is set to 0.01 in this paper. provides a knowledge prior from teacher network and indicates whether the depth distortion happens. Here is set to 0.5. Under this circumstance, the student model is able to know the useless training data when receiving knowledge from the teacher model.
Figure 4

Low quality depths in training.

Compared to considering one aspect or enforcing a fixed weight to the student model, our dynamic weight considers both the correctness of teacher’s knowledge and the error of student network, which allows the student network to receive the knowledge according to the degree of difficulty of samples. varies little in the start of training phase. As for late stage of training, the student network is able to detect most simple scenarios except for some hard samples. Therefore, automatically assign to relatively bigger weights for hard samples which can be detect accurately in teacher network but student network. The noise elimination method takes full merit of the knowledge prior from teacher network and effectively reduce the negative impact of depth maps in low quality. Extensive experiments demonstrate in Section 4 that this DKD could boost the detection performance without increasing extra parameters and model size. The process of the proposed methods is illustrated as Algorithm 1. Stage 1: Training the teacher network; Stage 2: Training the student network; ; ; if    then else end if ;

4. Experiments

4.1. Datasets and Evaluation Metrics

Datasets. Extensive experiments are conducted on five widely used RGB-D datasets, namely, NLPR [24], NJUD [25], SIP [26], DES [27] and LFSD [28]. These datasets contain large-scale images with different resolutions and diverse scenarios. We adopt the same training dataset with [12], which contains 1500 samples from NJUD and 700 samples from NLPR. The rest images in these two datasets together with other three datasets are used for testing. Evaluation Metrics. We adopt five metrics to comprehensively evaluate SOD tasks. These metrics include the F-measure curves, the F-measure score (), the Mean Absolute Error (M), the S-measure () and the E-measure (). Specifically, measures the accuracy of the model as follows: where is set to 0.3 as default. M measures the error rate of the model as follows: where W denotes the width and H denotes the height of prediction. S is the prediction saliency map and G is the corresponding ground truth.

4.2. Implementation Details

Our model is implemented using Pytorch Toolbox and trained on a GTX TITAN X GPU for 40 epochs with mini-batch size 4. We use a VGG16- and VGG19-based FPN as our final student architecture. Both RGB and depth images are resized to 256 × 256. To avoid overfitting, simple flipping and rotating are adopted to augment the training dataset. The initial learning rate is set to 1 × 10 and we adopt a 0.0005 weight decay for the stochastic gradient descent (SGD) with a momentum of 0.9.

4.3. Comparisons with the State of Arts

We conduct our experiments on 12 different prevalent methods in recent years, including DF [29], CTMF [30], AFNet [31], MMCI [32], DMRA [33], CPFP [12], TANet [34], D3Net [26], A2delde [10], DANet [35], FANet [15] and CMINet [6]. For fair comparisons, we directly use the released evaluation results or generate the results by the public saliency maps under the same evaluation framework. Quantitative Evaluation. Table 1 shows the quantitative results over five datasets. It can be observed that our method achieves the best scores in most metrics, especially on the NJUD dataset which contains 500 testing image pairs, for which our method performs better as far as all metrics are concerned. As for LFSD and SIP, although higher results come from other methods, we still obtain competing results in smaller VGG16 and VGG19 based networks. Figure 5 shows the comparison results using one-dimensional curves. Our method is represented by the red line which demonstrates better overall performance in both lightweight student models. In addition, it is apparent in Table 2 that our final VGG16-based network only has 57.9 MB with a faster speed, which drastically improves the inference speed and reduces the number of parameters. The above results indicate that without designing complicated models, accurate detection results can be obtained by only using an FPN with the help of the proposed methods.
Table 1

Quantitative comparisons through the maximum of F-score , S-score , E-score , and error-score M, over five widely evaluated datasets. ↑ and ↓ indicate that larger and smaller scores are better. Ours and Ours * indicate simple VGG16-based and VGG19-based FPN repsectively. ite refers to the training iterations.

MetricDFCTMFAFNetMMCITANetDMRACPFPD3NetA2deleDANetFANetCMINetOursOurs *
NJUDFβ0.7890.8570.8040.8680.8880.8960.8900.9030.9050.8900.8920.9250.9280.934
Sα 0.7350.8490.7720.8590.8780.8850.8780.8950.8670.8970.8990.9390.9160.920
Eθ 0.8180.8660.8470.8820.9090.9200.9000.9010.9140.9260.9140.9560.9490.952
M 0.1510.0850.1000.0790.0610.0510.0530.0510.0520.0460.0440.0320.0320.030
NLPR Fβ 0.7520.8410.8160.8410.8760.8880.8840.9040.8910.9080.8850.9090.9220.930
Sα 0.7690.8600.7990.8560.8860.8980.8840.9060.8890.9080.9130.9410.9210.924
Eθ 0.8400.8690.8840.8720.9260.9420.9200.9340.9370.9450.9510.9640.9580.960
M 0.1100.0560.0580.0590.0410.0310.0380.0340.0310.0310.0260.0190.0220.021
DES Fβ 0.6250.8650.7750.8390.8530.9060.8820.9170.8970.9160.8740.9260.9260.928
Sα 0.6850.8630.7700.8480.8580.8990.8720.9040.8830.9050.8940.9530.9180.918
Eθ 0.8060.9110.8740.9040.9190.9440.9270.9560.9180.9610.9250.9700.9650.966
M 0.1310.0550.0680.0650.0460.0300.0380.0300.0300.0280.0260.0150.0220.023
LFSD Fβ 0.8540.8150.7800.8130.8270.8720.8500.8490.858-0.8550.8620.8650.862
Sα 0.7860.7960.7380.7870.8010.8470.8280.8320.833-0.8500.8770.8340.839
Eθ 0.8410.8510.8100.8400.8510.8990.8670.8600.875-0.8820.9110.8830.883
M 0.1420.1200.1330.1320.1110.0760.0880.0990.077-0.0760.0640.0800.078
SIP Fβ 0.7040.7200.7560.8400.8510.8470.8700.8820.8550.901-0.8870.8720.882
Sα 0.6530.7160.7200.8330.8350.8000.8500.8640.8280.878-0.8940.8550.865
Eθ 0.7940.8240.8150.8860.8940.8580.8990.9030.8900.914-0.9330.9080.914
M 0.1850.1390.1180.0860.0750.0880.0640.0630.0700.054-0.0440.0600.056
BackboneVGG16VGG16VGG16VGG16VGG16VGG19VGG16VGG16VGG16VGG16/19VGG16ResNet50VGG16VGG19
Epoch---30,000(ite)-5010,000(ite)305040401004040
Figure 5

F-measure curves across four benchmarks on VGG16 and VGG19-based FPN.

Table 2

The model size and inference speed of different methods.

MethodMMCITANetPCANetD3NetCPFPDMRADANetCMINetA2deleOurs
Model Size (MB)951.9929.7533.6519278238.8106.78457.357.9/78.2
FPS19-15-7103210120136
Qualitative Evaluation. Figure 6 exhibits the visual comparisons with prevalent methods in recent years. Images contain diverse objects and scenarios, which are picked from different testing datasets. It can be observed that the saliency maps generated by our method are closer to the ground truth. More specifically, row 1 shows the case where the depth image has low contrast especially on the bottom part and row 2 and 3 show complex backgrounds in the RGB images. Under these circumstances, our method generates better saliency maps with less distortion and irrelevant objects compared to other methods.
Figure 6

Visual comparisons with existing methods.

4.4. Ablation Studies

Dynamic Knowledge Distillation. As shown in Table 3, our baseline is an FPN with VGG19 backbone trained on a cross-entropy loss, which can achieve the basic detection task. RGB indicates that only using RGB maps in training and RGBD concatenates depth maps as input. It is worth noting that directly using early fusion strategy also shows potential in RGB-D saliency detection. Then, we employ KD on the baseline to compare the results in different weights on four datasets. It is observed that KD can improve the performance and our DKD achieves better results across four testing datasets.
Table 3

Ablation analysis on 4 datasets. RGB and RGBD indicate that the student network is trained without/with depth maps, respectively, by cross-entropy loss. s indicates the weight of KD. Here we use the mean value of F-score and to show the overall accuracy. ↑ and ↓ indicate that larger and smaller scores are better.

MetricRGBRGBDs = 0.3s = 0.5s = 0.7s = Dynamic+Threshold
SIPFm0.7040.7730.8320.8450.8430.8490.853
wFm 0.6540.7240.7960.8050.8090.8110.817
M 0.1080.0860.0630.0610.0590.0580.056
NJUD Fm 0.7760.8300.9020.9020.8980.9040.914
wFm 0.7390.7990.8950.8890.8800.8930.901
M 0.0800.0600.0300.0340.0370.0320.030
NLPR Fm 0.7800.8160.8730.8760.8700.8760.890
wFm 0.7460.7810.8770.8750.8650.8760.887
M 0.0460.0410.0220.0240.0260.0240.021
LFSD Fm 0.7130.7840.8250.8320.8300.8340.835
wFm 0.6560.7410.7800.7930.7900.7950.796
M 0.1420.1020.0860.0800.0800.0780.078
Furthermore, Figure 7 shows detection performance of the student network in different KD weights. It is observed that in the last 10,000 iterations of training stage, the proposed method has better overall accuracy, where the lowest accuracy is still above 0.4, which is even better than the teacher network. Specific examples in Figure 8 illustrates that compared with the DKD, conventional fixed weights suffer from more false positives and negatives. Consequently, it is demonstrated that our DKD adaptively controls the KD in an appropriate way, leading to the improvement of overall detection performance.
Figure 7

Teacher and student scatters with different distillation weights. We exhibit the last 10,000 training iterations and DKD shows better overall performance.

Figure 8

Heatmap visual comparisons of different distillation weights: heatmaps from left to right indicate weights 0.3, 0.5, 0.7 and dynamic, respectively.

Noise Elimination with DKD. We investigate the low accuracy issue in teacher network by visualising extremely hard samples which has been shown in Figure 4. It is observed in Table 3 that with the help of noise elimination, all evaluation metrics over four testing datasets approach better results. Red arrows and rectangles in Figure 9 label the details which are refined by the proposed methods, especially on the part of object in the low contrast background, further illustrating that the proposed noise elimination effectively mitigate the noise during distillation and make the student model be able to learn more semantic details in the useful training data.
Figure 9

Ablation studies of the proposed methods. Baseline represents that the student network is only trained on cross-entropy loss. DKD represents the proposed DKD and NE means noise elimination.

Further analysis. In order to show the generalization of the proposed DKD, we replace the teacher network to DANet. Experimental results in Table 4 indicate that our DKD can compress the model size of existing method and approach similar accuracy to the teacher model, especially on the DES, where the performance of student model even outperforms the teacher model. To this end, the proposed dynamic distillation strategy can be explored on different teacher models. We further conduct experiments on VGG16-based FPN with different KD hyper-parameters as shown in Table 5. Specifically, we set temperature to 10 and only use the RGB images in the distillation training phase. Experimental results demonstrate that the proposed DKD can be utilized on different networks with different training settings, proving the effectiveness and generalization of DKD and leading to the potential of achieving RGB-D SOD tasks through RGB data within a lightweight structure.
Table 4

Quantitative comparisons of applying DKD on DANet. We use the DANet as the teacher model and exploit the proposed DKD to transfer knowledge from DANet to the VGG-19 based FPN. ↑ and ↓ indicate that larger and smaller scores are better.

MetricDANetDANet + DKD
SIPFm0.8640.848
wFm 0.8290.811
M 0.0540.058
DES Fm 0.8910.892
wFm 0.8480.870
M 0.0280.025
Model Size (MB)106.778.2
Table 5

Distillation on VGG16 with RGB maps. ↑ and ↓ indicate that larger and smaller scores are better.

Metrics = 0.3s = 0.5s = 0.7s = Dynamic
NLPRFm0.8820.8760.8760.884
wFm 0.8690.8650.8660.879
M 0.0240.0250.0270.024
LFSD Fm 0.7900.7860.7880.798
wFm 0.7450.7350.7510.753
M 0.1040.1060.1010.097

5. Conclusions

In this paper, we propose a DKD strategy and a noise elimination method for RGB-D based SOD. The proposed dynamic strategy considers the performance of both teacher and student models to generate an adaptive weight for KD. In order to reduce the final model size, we adopt the early fusion strategy for features fusion from different domains and the simple FPN as the final student model without designing extra networks. In addition, we investigate the noise issue caused by depth maps and alleviate this problem by setting a threshold during KD. The propose methods can be exploited on different teacher models and provide a new perspective which avoids designing extra networks for RGB-D SOD. We conduct comprehensive experiments on five challenging benchmark datasets to demonstrate that our method achieves competitive performance by only using a simple FPN model, which significantly compresses the model size and increases the inference speed. We further apply this dynamic strategy on different distillation temperatures with diverse models to prove the effectiveness and generalization of our method.
  5 in total

1.  Unsupervised Object Class Discovery via Saliency-Guided Multiple Class Learning.

Authors:  Jun-Yan Zhu; Jiajun Wu; Yan Xu; Eric Chang; Zhuowen Tu
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2015-04       Impact factor: 6.226

2.  RGBD Salient Object Detection via Deep Fusion.

Authors:  Liangqiong Qu; Shengfeng He; Jiawei Zhang; Jiandong Tian; Yandong Tang; Qingxiong Yang
Journal:  IEEE Trans Image Process       Date:  2017-03-15       Impact factor: 10.856

3.  Three-stream Attention-aware Network for RGB-D Salient Object Detection.

Authors:  Hao Chen; Youfu Li
Journal:  IEEE Trans Image Process       Date:  2019-01-07       Impact factor: 10.856

4.  Person Re-Identification by Saliency Learning.

Authors:  Rui Zhao; Wanli Oyang; Xiaogang Wang
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-03-21       Impact factor: 6.226

5.  Rethinking RGB-D Salient Object Detection: Models, Data Sets, and Large-Scale Benchmarks.

Authors:  Deng-Ping Fan; Zheng Lin; Zhao Zhang; Menglong Zhu; Ming-Ming Cheng
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2021-05-03       Impact factor: 10.451

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.