Abstract
Human vision system actively seeks salient regions and movements in video sequences to reduce the search effort. Modeling computational visual saliency map provides important information for semantic understanding in many real world applications. In this paper, we propose a novel video saliency detection model for detecting the attended regions that correspond to both interesting objects and dominant motions in video sequences. In spatial saliency map, we inherit the classical bottom-up spatial saliency map. In temporal saliency map, a novel optical flow model is proposed based on the dynamic consistency of motion. The spatial and the temporal saliency maps are constructed and further fused together to create a novel attention model. The proposed attention model is evaluated on three video datasets. Empirical validations demonstrate the salient regions detected by our dynamic consistent saliency map highlight the interesting objects effectively and efficiency. More importantly, the automatically video attended regions detected by proposed attention model are consistent with the ground truth saliency maps of eye movement data.
Original language | English |
---|---|
Title of host publication | Proceedings of the 27th AAAI Conference on Artificial Intelligence, AAAI 2013 |
Pages | 1063-1069 |
Number of pages | 7 |
Publication status | Published - 1 Dec 2013 |
Event | 27th AAAI Conference on Artificial Intelligence, AAAI 2013 - Bellevue, WA, United States Duration: 14 Jul 2013 → 18 Jul 2013 |
Conference
Conference | 27th AAAI Conference on Artificial Intelligence, AAAI 2013 |
---|---|
Country/Territory | United States |
City | Bellevue, WA |
Period | 14/07/13 → 18/07/13 |
ASJC Scopus subject areas
- Artificial Intelligence