Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

This paper proposes a novel depth-aware salient object detection and segmentation framework via multiscale discriminative saliency fusion (MDSF) and bootstrap learning for RGBD images (RGB color images with corresponding Depth maps) and stereoscopic images. By exploiting low-level feature contrasts, mid-level feature weighted factors and high-level location priors, various saliency measures on four classes of features are calculated based on multiscale region segmentation. A random forest regressor is learned to perform the discriminative saliency fusion (DSF) and generate the DSF saliency map at each scale, and DSF saliency maps across multiple scales are combined to produce the MDSF saliency map. Furthermore, we propose an effective bootstrap learning-based salient object segmentation method, which is bootstrapped with samples based on the MDSF saliency map and learns multiple kernel support vector machines. Experimental results on two large datasets show how various categories of features contribute to the saliency detection performance and demonstrate that the proposed framework achieves the better performance on both saliency detection and salient object segmentation.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TIP.2017.2711277DOI Listing

Publication Analysis

Top Keywords

salient object
16
discriminative saliency
12
saliency fusion
12
saliency map
12
saliency
10
depth-aware salient
8
object detection
8
detection segmentation
8
multiscale discriminative
8
bootstrap learning
8

Similar Publications

The impact of scene inversion on early scene-selective activity.

Biol Psychol

September 2025

Department of Psychology, Wright State University, Dayton OH. Electronic address:

Category-selectivity is a ubiquitous property of high-level visual cortex manifested in distinct cortical responses to faces, objects, and scenes. These signatures emerge early during visual processing, with each category sensitive to specific types of visual information at different time points. However, it is still not clear what information is extracted during early scene-selective processing, as scenes are rich, complex, and multidimensional stimuli.

View Article and Find Full Text PDF

Computational saliency map models have facilitated quantitative investigations into how bottom-up visual salience influences attention. Two primary approaches to modeling salience computation exist: one focuses on functional approximation, while the other explores neurobiological implementation. The former provides sufficient performance for applying saliency map models to eye-movement data analysis, whereas the latter offers hypotheses on how neuronal abnormalities affect visual salience.

View Article and Find Full Text PDF

When planning reach-to-grasp movements, individuals frequently face a tradeoff between biomechanical comfort (i.e., avoiding effortful actions) and "socio-emotional comfort" (i.

View Article and Find Full Text PDF

In blind individuals, language processing activates not only classic language networks, but also the "visual" cortex. What is represented in visual areas when blind individuals process language? Here, we show that area V5/MT in blind individuals, but not other visual areas, responds differently to spoken nouns and verbs. We further show that this effect is present for concrete nouns and verbs, but not abstract or pseudo nouns and verbs.

View Article and Find Full Text PDF

Due to the limited output categories, semi-supervised salient object detection faces challenges in adapting conventional semi-supervised strategies. To address this limitation, we propose a multi-branch architecture that extracts complementary features from labeled data. Specifically, we introduce TripleNet, a three-branch network architecture designed for contour, content, and holistic saliency prediction.

View Article and Find Full Text PDF