Affordancellm: Grounding affordance from vision language models

S Qian, W Chen, M Bai, X Zhou… - Proceedings of the …, 2024 - openaccess.thecvf.com
Affordance grounding refers to the task of finding the area of an object with which one can
interact. It is a fundamental but challenging task as a successful solution requires the …

Locate: Localize and transfer object parts for weakly supervised affordance grounding

G Li, V Jampani, D Sun… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Humans excel at acquiring knowledge through observation. For example, we can learn to
use new tools by watching demonstrations. This skill is fundamental for intelligent systems to …

Grounding 3d object affordance from 2d interactions in images

Y Yang, W Zhai, H Luo, Y Cao… - Proceedings of the …, 2023 - openaccess.thecvf.com
Grounding 3D object affordance seeks to locate objects'" action possibilities" regions in the
3D space, which serves as a link between perception and operation for embodied agents …

One-shot open affordance learning with foundation models

G Li, D Sun, L Sevilla-Lara… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract We introduce One-shot Open Affordance Learning (OOAL) where a model is trained
with just one example per base object category but is expected to identify novel objects and …

What does CLIP know about peeling a banana?

C Cuttano, G Rosi, G Trivigno… - Proceedings of the …, 2024 - openaccess.thecvf.com
Humans show an innate capability to identify tools to support specific actions. The
association between objects parts and the actions they facilitate is usually named …

Weakly Supervised Multimodal Affordance Grounding for Egocentric Images

L Xu, Y Gao, W Song, A Hao - Proceedings of the AAAI Conference on …, 2024 - ojs.aaai.org
To enhance the interaction between intelligent systems and the environment, locating the
affordance regions of objects is crucial. These regions correspond to specific areas that …

Self-Explainable Affordance Learning with Embodied Caption

Z Zhang, Z Wei, G Sun, P Wang, L Van Gool - arXiv preprint arXiv …, 2024 - arxiv.org
In the field of visual affordance learning, previous methods mainly used abundant images or
videos that delineate human behavior patterns to identify action possibility regions for object …

Learning 2D Invariant Affordance Knowledge for 3D Affordance Grounding

X Gao, P Zhang, D Qu, D Wang, Z Wang, Y Ding… - arXiv preprint arXiv …, 2024 - arxiv.org
3D Object Affordance Grounding aims to predict the functional regions on a 3D object and
has laid the foundation for a wide range of applications in robotics. Recent advances tackle …

INTRA: Interaction Relationship-aware Weakly Supervised Affordance Grounding

JH Jang, H Seo, SY Chun - arXiv preprint arXiv:2409.06210, 2024 - arxiv.org
Affordance denotes the potential interactions inherent in objects. The perception of
affordance can enable intelligent agents to navigate and interact with new environments …

EgoChoir: Capturing 3D Human-Object Interaction Regions from Egocentric Views

Y Yang, W Zhai, C Wang, C Yu, Y Cao… - arXiv preprint arXiv …, 2024 - arxiv.org
Understanding egocentric human-object interaction (HOI) is a fundamental aspect of human-
centric perception, facilitating applications like AR/VR and embodied AI. For the egocentric …