Zhao, Xin, Lou, Jianxun, Wu, Xinbo, Wu, Yingying, Leveque, Lucie, Liu, Xiaochang, Guo, Pengfei, Qin, Yipeng ORCID: https://orcid.org/0000-0002-1551-9126, Lin, Hanhe, Saupe, Dietmar and Liu, Hantao ORCID: https://orcid.org/0000-0003-4544-3481 2023. CUDAS: Distortion-aware saliency benchmark. IEEE Access 11 , pp. 58025-58036. 10.1109/ACCESS.2023.3283344 |
Preview |
PDF
- Published Version
Available under License Creative Commons Attribution Non-commercial No Derivatives. Download (5MB) | Preview |
Abstract
Visual saliency prediction remains an academic challenge due to the diversity and complexity of natural scenes as well as the scarcity of eye movement data on where people look in images. In many practical applications, digital images are inevitably subject to distortions, such as those caused by acquisition, editing, compression or transmission. A great deal of attention has been paid to predicting the saliency of distortion-free pristine images, but little attention has been given to understanding the impact of visual distortions on saliency prediction. In this paper, we first present the CUDAS database - a new distortion-aware saliency benchmark, where eye-tracking data was collected for 60 pristine images and their corresponding 540 distorted formats. We then conduct a statistical evaluation to reveal the behaviour of state-of-the-art saliency prediction models on distorted images and provide insights on building an effective model for distortion-aware saliency prediction. The new database is made publicly available to the research community.
Item Type: | Article |
---|---|
Date Type: | Publication |
Status: | Published |
Schools: | Computer Science & Informatics |
Publisher: | Institute of Electrical and Electronics Engineers |
ISSN: | 2169-3536 |
Date of First Compliant Deposit: | 14 June 2023 |
Date of Acceptance: | 29 May 2023 |
Last Modified: | 22 Mar 2024 12:27 |
URI: | https://orca.cardiff.ac.uk/id/eprint/160368 |
Actions (repository staff only)
Edit Item |