Challenges, evaluation and opportunities for open-world learning

M Kejriwal, E Kildebeck, R Steininger… - Nature Machine …, 2024 - nature.com
Environmental changes can profoundly impact the performance of artificial intelligence
systems operating in the real world, with effects ranging from overt catastrophic failures to …

Aligning cyber space with physical world: A comprehensive survey on embodied ai

Y Liu, W Chen, Y Bai, X Liang, G Li, W Gao… - arXiv preprint arXiv …, 2024 - arxiv.org
Embodied Artificial Intelligence (Embodied AI) is crucial for achieving Artificial General
Intelligence (AGI) and serves as a foundation for various applications that bridge cyberspace …

Foundation models in robotics: Applications, challenges, and the future

R Firoozi, J Tucker, S Tian… - … Journal of Robotics …, 2023 - journals.sagepub.com
We survey applications of pretrained foundation models in robotics. Traditional deep
learning models in robotics are trained on small datasets tailored for specific tasks, which …

Mvimgnet: A large-scale dataset of multi-view images

X Yu, M Xu, Y Zhang, H Liu, C Ye… - Proceedings of the …, 2023 - openaccess.thecvf.com
Being data-driven is one of the most iconic properties of deep learning algorithms. The birth
of ImageNet drives a remarkable trend of" learning from large-scale data" in computer vision …

Physics-informed machine learning: A survey on problems, methods and applications

Z Hao, S Liu, Y Zhang, C Ying, Y Feng, H Su… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent advances of data-driven machine learning have revolutionized fields like computer
vision, reinforcement learning, and many scientific and engineering domains. In many real …

Generating visual scenes from touch

F Yang, J Zhang, A Owens - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
An emerging line of work has sought to generate plausible imagery from touch. Existing
approaches, however, tackle only narrow aspects of the visuo-tactile synthesis problem, and …

Multiply: A multisensory object-centric embodied large language model in 3d world

Y Hong, Z Zheng, P Chen, Y Wang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Human beings possess the capability to multiply a melange of multisensory cues while
actively exploring and interacting with the 3D world. Current multi-modal large language …

The objectfolder benchmark: Multisensory learning with neural and real objects

R Gao, Y Dou, H Li, T Agarwal, J Bohg… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract We introduce the ObjectFolder Benchmark, a benchmark suite of 10 tasks for
multisensory object-centric learning, centered around object recognition, reconstruction, and …

See, hear, and feel: Smart sensory fusion for robotic manipulation

H Li, Y Zhang, J Zhu, S Wang, MA Lee, H Xu… - arXiv preprint arXiv …, 2022 - arxiv.org
Humans use all of their senses to accomplish different tasks in everyday activities. In
contrast, existing work on robotic manipulation mostly relies on one, or occasionally two …

Touching a nerf: Leveraging neural radiance fields for tactile sensory data generation

S Zhong, A Albini, OP Jones… - … on Robot Learning, 2023 - proceedings.mlr.press
Tactile perception is key for robotics applications such as manipulation. However, tactile
data collection is time-consuming, especially when compared to vision. This limits the use of …