A comprehensive survey on pretrained foundation models: A history from bert to chatgpt

C Zhou, Q Li, C Li, J Yu, Y Liu, G Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Pretrained Foundation Models (PFMs) are regarded as the foundation for various
downstream tasks with different data modalities. A PFM (eg, BERT, ChatGPT, and GPT-4) is …

On the opportunities and risks of foundation models

R Bommasani, DA Hudson, E Adeli, R Altman… - arXiv preprint arXiv …, 2021 - arxiv.org
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …

Reinforcement learning with action-free pre-training from videos

Y Seo, K Lee, SL James… - … Conference on Machine …, 2022 - proceedings.mlr.press
Recent unsupervised pre-training methods have shown to be effective on language and
vision domains by learning useful representations for multiple downstream tasks. In this …

A comprehensive survey of data augmentation in visual reinforcement learning

G Ma, Z Wang, Z Yuan, X Wang, B Yuan… - arXiv preprint arXiv …, 2022 - arxiv.org
Visual reinforcement learning (RL), which makes decisions directly from high-dimensional
visual inputs, has demonstrated significant potential in various domains. However …

Getting aligned on representational alignment

I Sucholutsky, L Muttenthaler, A Weller, A Peng… - arXiv preprint arXiv …, 2023 - arxiv.org
Biological and artificial information processing systems form representations that they can
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …

Multi-view masked world models for visual robotic manipulation

Y Seo, J Kim, S James, K Lee… - … on Machine Learning, 2023 - proceedings.mlr.press
Visual robotic manipulation research and applications often use multiple cameras, or views,
to better perceive the world. How else can we utilize the richness of multi-view data? In this …

Plastic: Improving input and label plasticity for sample efficient reinforcement learning

H Lee, H Cho, H Kim, D Gwak, J Kim… - Advances in …, 2024 - proceedings.neurips.cc
Abstract In Reinforcement Learning (RL), enhancing sample efficiency is crucial, particularly
in scenarios when data acquisition is costly and risky. In principle, off-policy RL algorithms …

Firerisk: A remote sensing dataset for fire risk assessment with benchmarks using supervised and self-supervised learning

S Shen, S Seneviratne, X Wanyan… - … Conference on Digital …, 2023 - ieeexplore.ieee.org
In recent decades, wildfires have caused tremendous property losses, fatalities, and
extensive damage to forest ecosystems. Inspired by the abundance of publicly available …

Masked autoencoding for scalable and generalizable decision making

F Liu, H Liu, A Grover, P Abbeel - Advances in Neural …, 2022 - proceedings.neurips.cc
We are interested in learning scalable agents for reinforcement learning that can learn from
large-scale, diverse sequential data similar to current large vision and language models. To …

: Temporal Latent Action-Driven Contrastive Loss for Visual Reinforcement Learning

R Zheng, X Wang, Y Sun, S Ma… - Advances in …, 2024 - proceedings.neurips.cc
Despite recent progress in reinforcement learning (RL) from raw pixel data, sample
inefficiency continues to present a substantial obstacle. Prior works have attempted to …