Spatio-Temporal Saliency Networks for Dynamic Saliency Prediction
Journal Article
Abstract Computational saliency models for still images have gained significant popularity in recent years. Saliency prediction from videos, on the other hand, has received relatively little interest from the community. Motivated by this, in this work, we study the use of deep learning for dynamic saliency prediction and propose the so-called spatio-temporal saliency networks. The key to our models is the architecture of two-stream networks where we investigate different fusion mechanisms to integrate spatial and temporal information. We evaluate our models on the DIEM and UCF-Sports datasets and present highly competitive results against the existing state-of-the-art models. We also carry out some experiments on a number of still images from the MIT300 dataset by exploiting the optical flow maps predicted from these images. Our results show that considering inherent motion information in this way can be helpful for static saliency estimation.

BibTeX
@ARTICLE{Bak2018,
author={C. Bak and A. Kocak and E. Erdem and A. Erdem},
journal={IEEE Transactions on Multimedia},
title={Spatio-Temporal Saliency Networks
for Dynamic Saliency Prediction},
volume={20},
number={7},
pages={1688--1698},
year={2018}}