共 50 条
Two-Stage Learning to Predict Human Eye Fixations via SDAEs
被引:114
|作者:
Han, Junwei
[1
]
Zhang, Dingwen
[1
]
Wen, Shifeng
[1
]
Guo, Lei
[1
]
Liu, Tianming
[2
]
Li, Xuelong
[3
]
机构:
[1] Northwestern Polytech Univ, Sch Automat, Xian 710072, Peoples R China
[2] Univ Georgia, Dept Comp Sci, Athens, GA 30602 USA
[3] Chinese Acad Sci, Xian Inst Opt & Precis Mech, State Key Lab Transient Opt & Photon, Ctr OPT IMagery Anal & Learning, Xian 710119, Peoples R China
基金:
国家教育部博士点专项基金资助;
美国国家科学基金会;
关键词:
Deep networks;
eye fixation prediction;
saliency detection;
stacked denoising autoencoders ( SDAEs);
VISUAL SALIENCY;
OBJECT DETECTION;
RETRIEVAL;
ATTENTION;
AUTOENCODERS;
FRAMEWORK;
MODEL;
D O I:
10.1109/TCYB.2015.2404432
中图分类号:
TP [自动化技术、计算机技术];
学科分类号:
0812 ;
摘要:
Saliency detection models aiming to quantitatively predict human eye-attended locations in the visual field have been receiving increasing research interest in recent years. Unlike traditional methods that rely on hand-designed features and contrast inference mechanisms, this paper proposes a novel framework to learn saliency detection models from raw image data using deep networks. The proposed framework mainly consists of two learning stages. At the first learning stage, we develop a stacked denoising autoencoder (SDAE) model to learn robust, representative features from raw image data under an unsupervised manner. The second learning stage aims to jointly learn optimal mechanisms to capture the intrinsic mutual patterns as the feature contrast and to integrate them for final saliency prediction. Given the input of pairs of a center patch and its surrounding patches represented by the features learned at the first stage, a SDAE network is trained under the supervision of eye fixation labels, which achieves both contrast inference and contrast integration simultaneously. Experiments on three publically available eye tracking benchmarks and the comparisons with 16 state-of-the-art approaches demonstrate the effectiveness of the proposed framework.
引用
收藏
页码:487 / 498
页数:12
相关论文