A comprehensive survey on pretrained foundation models: A history from bert to chatgpt
Pretrained Foundation Models (PFMs) are regarded as the foundation for various
downstream tasks with different data modalities. A PFM (eg, BERT, ChatGPT, and GPT-4) is …
downstream tasks with different data modalities. A PFM (eg, BERT, ChatGPT, and GPT-4) is …
On the opportunities and risks of foundation models
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …
Reinforcement learning with action-free pre-training from videos
Recent unsupervised pre-training methods have shown to be effective on language and
vision domains by learning useful representations for multiple downstream tasks. In this …
vision domains by learning useful representations for multiple downstream tasks. In this …
A comprehensive survey of data augmentation in visual reinforcement learning
Visual reinforcement learning (RL), which makes decisions directly from high-dimensional
visual inputs, has demonstrated significant potential in various domains. However …
visual inputs, has demonstrated significant potential in various domains. However …
Getting aligned on representational alignment
Biological and artificial information processing systems form representations that they can
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …
Multi-view masked world models for visual robotic manipulation
Visual robotic manipulation research and applications often use multiple cameras, or views,
to better perceive the world. How else can we utilize the richness of multi-view data? In this …
to better perceive the world. How else can we utilize the richness of multi-view data? In this …
Plastic: Improving input and label plasticity for sample efficient reinforcement learning
Abstract In Reinforcement Learning (RL), enhancing sample efficiency is crucial, particularly
in scenarios when data acquisition is costly and risky. In principle, off-policy RL algorithms …
in scenarios when data acquisition is costly and risky. In principle, off-policy RL algorithms …
Firerisk: A remote sensing dataset for fire risk assessment with benchmarks using supervised and self-supervised learning
S Shen, S Seneviratne, X Wanyan… - … Conference on Digital …, 2023 - ieeexplore.ieee.org
In recent decades, wildfires have caused tremendous property losses, fatalities, and
extensive damage to forest ecosystems. Inspired by the abundance of publicly available …
extensive damage to forest ecosystems. Inspired by the abundance of publicly available …
Masked autoencoding for scalable and generalizable decision making
We are interested in learning scalable agents for reinforcement learning that can learn from
large-scale, diverse sequential data similar to current large vision and language models. To …
large-scale, diverse sequential data similar to current large vision and language models. To …
: Temporal Latent Action-Driven Contrastive Loss for Visual Reinforcement Learning
Despite recent progress in reinforcement learning (RL) from raw pixel data, sample
inefficiency continues to present a substantial obstacle. Prior works have attempted to …
inefficiency continues to present a substantial obstacle. Prior works have attempted to …