Dr Yijun Yan y.yan2@rgu.ac.uk
Research Fellow
Unsupervised image saliency detection with Gestalt-laws guided optimization and visual attention based refinement.
Yan, Yijun; Ren, Jinchang; Sun, Genyun; Zhao, Huimin; Han, Junwei; Li, Xuelong; Marshall, Stephen; Zhan, Jin
Authors
Professor Jinchang Ren j.ren@rgu.ac.uk
Professor of Computing Science
Genyun Sun
Huimin Zhao
Junwei Han
Xuelong Li
Stephen Marshall
Jin Zhan
Abstract
Visual attention is a kind of fundamental cognitive capability that allows human beings to focus on the region of interests (ROIs) under complex natural environments. What kind of ROIs that we pay attention to mainly depends on two distinct types of attentional mechanisms. The bottom-up mechanism can guide our detection of the salient objects and regions by externally driven factors, i.e. color and location, whilst the top-down mechanism controls our biasing attention based on prior knowledge and cognitive strategies being provided by visual cortex. However, how to practically use and fuse both attentional mechanisms for salient object detection has not been sufficiently explored. To the end, we propose in this paper an integrated framework consisting of bottom-up and top-down attention mechanisms that enable attention to be computed at the level of salient objects and/or regions. Within our framework, the model of a bottom-up mechanism is guided by the gestalt-laws of perception. We interpreted gestalt-laws of homogeneity, similarity, proximity and figure and ground in link with color, spatial contrast at the level of regions and objects to produce feature contrast map. The model of top-down mechanism aims to use a formal computational model to describe the background connectivity of the attention and produce the priority map. Integrating both mechanisms and applying to salient object detection, our results have demonstrated that the proposed method consistently outperforms a number of existing unsupervised approaches on five challenging and complicated datasets in terms of higher precision and recall rates, AP (average precision) and AUC (area under curve) values.
Citation
YAN, Y., REN, J., SUN, G., ZHAO, H., HAN, J., LI, X., MARSHALL, S. and ZHAN, J. 2018. Unsupervised image saliency detection with Gestalt-laws guided optimization and visual attention based refinement. Pattern recognition [online], 79, pages 65-78. Available from: https://doi.org/10.1016/j.patcog.2018.02.004
Journal Article Type | Article |
---|---|
Acceptance Date | Feb 2, 2018 |
Online Publication Date | Feb 5, 2018 |
Publication Date | Jul 31, 2018 |
Deposit Date | Oct 5, 2021 |
Publicly Available Date | Oct 5, 2021 |
Journal | Pattern recognition |
Print ISSN | 0031-3203 |
Electronic ISSN | 1873-5142 |
Publisher | Elsevier |
Peer Reviewed | Peer Reviewed |
Volume | 79 |
Pages | 65-78 |
DOI | https://doi.org/10.1016/j.patcog.2018.02.004 |
Keywords | Background connectivity; Gestalt laws guided optimization; Image saliency detection; Feature fusion; Human vision perception |
Public URL | https://rgu-repository.worktribe.com/output/1474899 |
Files
YAN 2018 Unsupervised image (AAM)
(1.7 Mb)
PDF
Publisher Licence URL
https://creativecommons.org/licenses/by-nc-nd/4.0/
You might also like
Hyperspectral imaging based corrosion detection in nuclear packages.
(2023)
Journal Article
Downloadable Citations
About OpenAIR@RGU
Administrator e-mail: publications@rgu.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search