Learning to manipulate anywhere: A visual generalizable framework for reinforcement learning
Can we endow visuomotor robots with generalization capabilities to operate in diverse open-
world scenarios? In this paper, we propose\textbf {Maniwhere}, a generalizable framework …
world scenarios? In this paper, we propose\textbf {Maniwhere}, a generalizable framework …
Point-SAM: Promptable 3D Segmentation Model for Point Clouds
The development of 2D foundation models for image segmentation has been significantly
advanced by the Segment Anything Model (SAM). However, achieving similar success in 3D …
advanced by the Segment Anything Model (SAM). However, achieving similar success in 3D …
SAM-E: Leveraging Visual Foundation Model with Sequence Imitation for Embodied Manipulation
Acquiring a multi-task imitation policy in 3D manipulation poses challenges in terms of
scene understanding and action prediction. Current methods employ both 3D representation …
scene understanding and action prediction. Current methods employ both 3D representation …
Policy-shaped prediction: avoiding distractions in model-based reinforcement learning
Model-based reinforcement learning (MBRL) is a promising route to sample-efficient policy
optimization. However, a known vulnerability of reconstruction-based MBRL consists of …
optimization. However, a known vulnerability of reconstruction-based MBRL consists of …
Focus On What Matters: Separated Models For Visual-Based RL Generalization
A primary challenge for visual-based Reinforcement Learning (RL) is to generalize
effectively across unseen environments. Although previous studies have explored different …
effectively across unseen environments. Although previous studies have explored different …