Scanpath and saliency prediction on 360 degree images

View/Open
Cita com:
hdl:2117/119346
Document typeArticle
Defense date2018-06-23
Rights accessOpen Access
This work is protected by the corresponding intellectual and industrial property rights.
Except where otherwise noted, its contents are licensed under a Creative Commons license
:
Attribution-NonCommercial-NoDerivs 3.0 Spain
Abstract
We introduce deep neural networks for scanpath and saliency prediction trained on 360-degree images. The scanpath prediction model called SaltiNet is based on a temporal-aware novel representation of saliency information named the saliency volume. The first part of the network consists of a model trained to generate saliency volumes, whose parameters are fit by back-propagation using a binary cross entropy (BCE) loss over downsampled versions of the saliency volumes. Sampling strategies over these volumes are used to generate scanpaths over the 360-degree images. Our experiments show the advantages of using saliency volumes, and how they can be used for related tasks. We also show how a similar architecture achieves state-of-the-art performance for the related task of saliency map prediction. Our source code and trained models available at https://github.com/massens/saliency-360salient-2017.
CitationAssens, M., Giro, X., McGuinness, K., O'Connor, N. Scanpath and saliency prediction on 360 degree images. "Signal processing: image communication", 23 Juny 2018, vol. 69, p. 8-14.
ISSN0923-5965
Publisher versionhttps://www.sciencedirect.com/science/article/pii/S0923596518306209
Files | Description | Size | Format | View |
---|---|---|---|---|
Scanpath2.pdf | 2,641Mb | View/Open |