Towards a richer 2D understanding of hands at scale

T Cheng, D Shan, A Hassen… - Advances in Neural …, 2023 - proceedings.neurips.cc
As humans, we learn a lot about how to interact with the world by observing others
interacting with their hands. To help AI systems obtain a better understanding of hand …

An outlook into the future of egocentric vision

C Plizzari, G Goletto, A Furnari, S Bansal… - International Journal of …, 2024 - Springer
What will the future be? We wonder! In this survey, we explore the gap between current
research in egocentric vision and the ever-anticipated future, where wearable computing …

Deformer: Dynamic fusion transformer for robust hand pose estimation

Q Fu, X Liu, R Xu, JC Niebles… - Proceedings of the …, 2023 - openaccess.thecvf.com
Accurately estimating 3D hand pose is crucial for understanding how humans interact with
the world. Despite remarkable progress, existing methods often struggle to generate …

Localizing active objects from egocentric vision with symbolic world knowledge

TL Wu, Y Zhou, N Peng - arXiv preprint arXiv:2310.15066, 2023 - arxiv.org
The ability to actively ground task instructions from an egocentric view is crucial for AI agents
to accomplish tasks or assist humans virtually. One important step towards this goal is to …

Domain adaptive hand keypoint and pixel localization in the wild

T Ohkawa, YJ Li, Q Fu, R Furuta, KM Kitani… - European Conference on …, 2022 - Springer
We aim to improve the performance of regressing hand keypoints and segmenting pixel-
level hand masks under new imaging conditions (eg., outdoors) when we only have labeled …

Diag-IoU loss for object detection

S Zhang, C Li, Z Jia, L Liu, Z Zhang… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Existing IoU-based loss functions have achieved promising performance for bounding box
regression in object detection. However, they cannot fully reflect the relation between the …

Actionvos: Actions as prompts for video object segmentation

L Ouyang, R Liu, Y Huang, R Furuta, Y Sato - European Conference on …, 2025 - Springer
Delving into the realm of egocentric vision, the advancement of referring video object
segmentation (RVOS) stands as pivotal in understanding human activities. However …

[HTML][HTML] Exploiting multimodal synthetic data for egocentric human-object interaction detection in an industrial scenario

R Leonardi, F Ragusa, A Furnari… - Computer Vision and …, 2024 - Elsevier
In this paper, we tackle the problem of Egocentric Human-Object Interaction (EHOI)
detection in an industrial setting. To overcome the lack of public datasets in this context, we …

Are Synthetic Data Useful for Egocentric Hand-Object Interaction Detection?

R Leonardi, A Furnari, F Ragusa… - European Conference on …, 2025 - Springer
In this study, we investigate the effectiveness of synthetic data in enhancing egocentric hand-
object interaction detection. Via extensive experiments and comparative analyses on three …

Egocentric human-object interaction detection exploiting synthetic data

R Leonardi, F Ragusa, A Furnari… - … Conference on Image …, 2022 - Springer
We consider the problem of detecting Egocentric Human-Object Interactions (EHOIs) in
industrial contexts. Since collecting and labeling large amounts of real images is …