Predicting goal-directed human attention using inverse reinforcement learning

Z Yang, L Huang, Y Chen, Z Wei… - Proceedings of the …, 2020 - openaccess.thecvf.com
Human gaze behavior prediction is important for behavioral vision and for computer vision
applications. Most models mainly focus on predicting free-viewing behavior using saliency …

Towards predicting reading comprehension from gaze behavior

S Ahn, C Kelton, A Balasubramanian… - ACM Symposium on Eye …, 2020 - dl.acm.org
As readers of a language, we all agree to move our eyes in roughly the same way. Yet might
there be hidden within this self-similar behavior subtle clues as to how a reader is …

Benchmarking gaze prediction for categorical visual search

G Zelinsky, Z Yang, L Huang, Y Chen… - Proceedings of the …, 2019 - openaccess.thecvf.com
Movements of human attention during free viewing have received wide interest in the
computer vision community. However, search behavior, where the fixation scanpaths are …

Reading detection in real-time

C Kelton, Z Wei, S Ahn, A Balasubramanian… - Proceedings of the 11th …, 2019 - dl.acm.org
Observable reading behavior, the act of moving the eyes over lines of text, is highly
stereotyped among the users of a language, and this has led to the development of reading …

Deep-BCN: Deep networks meet biased competition to create a brain-inspired model of attention control

H Adeli, G Zelinsky - … of the IEEE conference on computer …, 2018 - openaccess.thecvf.com
The mechanism of attention control is best described by biased-competition theory (BCT),
which suggests that a top-down goal state biases a competition among object …

BiasedWalk: Biased sampling for representation learning on graphs

D Nguyen, FD Malliaros - … Conference on Big Data (Big Data), 2018 - ieeexplore.ieee.org
Network embedding algorithms are able to learn latent feature representations of nodes,
transforming networks into lower dimensional vector representations. Typical key …

Changing perspectives on goal-directed attention control: The past, present, and future of modeling fixations during visual search

GJ Zelinsky, Y Chen, S Ahn, H Adeli - Psychology of learning and …, 2020 - Elsevier
People make eye movements while interacting with objects, and these behaviors are rich
with information about how visual goals are represented in the brain and used to prioritize …

A Priority Map for Vision-and-Language Navigation with Trajectory Plans and Feature-Location Cues

J Armitage, L Impett, R Sennrich - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
In a busy city street, a pedestrian surrounded by distractions can pick out a single sign if it is
relevant to their route. Artificial agents in outdoor Vision-and-Language Navigation (VLN) …

Weighting the factors affecting attention guidance during free viewing and visual search: The unexpected role of object recognition uncertainty

S Chakraborty, D Samaras, GJ Zelinsky - Journal of Vision, 2022 - jov.arvojournals.org
The factors determining how attention is allocated during visual tasks have been studied for
decades, but few studies have attempted to model the weighting of several of these factors …

GazeSearch: Radiology Findings Search Benchmark

TT Pham, TP Nguyen, Y Ikebe, A Awasthi… - arXiv preprint arXiv …, 2024 - arxiv.org
Medical eye-tracking data is an important information source for understanding how
radiologists visually interpret medical images. This information not only improves the …