Lou, Jianxun, Lin, Hanhe, Marshall, David ![]() ![]() ![]() |
![]() |
PDF
- Accepted Post-Print Version
Download (4MB) |
Abstract
Radiologists’ eye-movement during diagnostic image reading reflects their personal training and experience, which means that their diagnostic decisions are related to their perceptual processes. For training, monitoring, and performance evaluation of radiologists, it would be beneficial to be able to automatically predict the spatial distribution of the radiologist’s visual attention on the diagnostic images. The measurement of visual saliency is a well-studied area that allows for prediction of a person’s gaze attention. However, compared with the extensively studied natural image visual saliency (in free viewing tasks), the saliency for diagnostic images is less studied; there could be fundamental differences in eye-movement behaviours between these two domains. Most current saliency prediction models have been optimally developed for natural images, which could lead them to be less adept at predicting the visual attention of radiologists during the diagnosis. In this paper, we propose a method specifically for automatically capturing the visual attention of radiologists during mammogram reading. By adopting high-resolution image representations from both deep and shallow encoders, the proposed method avoids potential detail losses and achieves superior results on multiple evaluation metrics in a large mammogram eye-movement dataset.
Item Type: | Conference or Workshop Item (Paper) |
---|---|
Date Type: | Published Online |
Status: | Published |
Schools: | Schools > Computer Science & Informatics |
Publisher: | IEEE |
ISBN: | 978-1-6654-9621-6 |
ISSN: | 1522-4880 |
Date of First Compliant Deposit: | 25 July 2022 |
Date of Acceptance: | 20 June 2022 |
Last Modified: | 03 Jul 2025 14:28 |
URI: | https://orca.cardiff.ac.uk/id/eprint/150867 |
Actions (repository staff only)
![]() |
Edit Item |