|Appears in Collections:||Computing Science and Mathematics Journal Articles|
|Peer Review Status:||Refereed|
|Title:||Robust Visual Saliency Optimization Based on Bidirectional Markov Chains|
Background and foreground possibility
|Citation:||Jiang F, Kong B, Li J, Dashtipour K & Gogate M (2020) Robust Visual Saliency Optimization Based on Bidirectional Markov Chains. Cognitive Computation. https://doi.org/10.1007/s12559-020-09724-6|
|Abstract:||Saliency detection aims to automatically highlight the most important area in an image. Traditional saliency detection methods based on absorbing Markov chain only take into account boundary nodes and often lead to incorrect saliency detection when the boundaries have salient objects. In order to address this limitation and enhance saliency detection performance, this paper proposes a novel task-independent saliency detection method based on the bidirectional absorbing Markov chains that jointly exploits not only the boundary information but also the foreground prior and background prior cues. More specifically, the input image is first segmented into number of superpixels, and the four boundary nodes (duplicated as virtual nodes) are selected. Subsequently, the absorption time upon transition node’s random walk to the absorbing state is calculated to obtain the foreground possibility. Simultaneously, foreground prior (as the virtual absorbing nodes) is used to calculate the absorption time and get the background possibility. In addition, the two aforementioned results are fused to form a combined saliency map which is further optimized by using a cost function. Finally, the superpixel-level saliency results are optimized by a regularized random walks ranking model at multi-scale. The comparative experimental results on four benchmark datasets reveal superior performance of our proposed method over state-of-the-art methods reported in the literature. The experiments show that the proposed method is efficient and can be applicable to the bottom-up image saliency detection and other visual processing tasks.|
|Rights:||This item has been embargoed for a period. During the embargo please use the Request a Copy feature at the foot of the Repository record to request a copy directly from the author. You can only request a copy if you wish to use this work for your own research or private study. This is a post-peer-review, pre-copyedit version of an article published in Cognitive Computation. The final authenticated version is available online at: https://doi.org/10.1007/s12559-020-09724-6|
|Notes:||Output Status: Forthcoming/Available Online|
|FnalSubmission.pdf||Fulltext - Accepted Version||1.96 MB||Adobe PDF||Under Embargo until 2021-05-30 Request a copy|
Note: If any of the files in this item are currently embargoed, you can request a copy directly from the author by clicking the padlock icon above. However, this facility is dependent on the depositor still being contactable at their original email address.
This item is protected by original copyright
Items in the Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
If you believe that any material held in STORRE infringes copyright, please contact email@example.com providing details and we will remove the Work from public display in STORRE and investigate your claim.