Making scalable meta learning practical

S Choe, SV Mehta, H Ahn… - Advances in neural …, 2024 - proceedings.neurips.cc
Despite its flexibility to learn diverse inductive biases in machine learning programs, meta
learning (ie,\learning to learn) has long been recognized to suffer from poor scalability due …

Bidirectional learning for offline model-based biological sequence design

C Chen, Y Zhang, X Liu… - … Conference on Machine …, 2023 - proceedings.mlr.press
Offline model-based optimization aims to maximize a black-box objective function with a
static dataset of designs and their scores. In this paper, we focus on biological sequence …

Torchdeq: A library for deep equilibrium models

Z Geng, JZ Kolter - arXiv preprint arXiv:2310.18605, 2023 - arxiv.org
Deep Equilibrium (DEQ) Models, an emerging class of implicit models that maps inputs to
fixed points of neural networks, are of growing interest in the deep learning community …

Nyström method for accurate and scalable implicit differentiation

R Hataya, M Yamada - International Conference on Artificial …, 2023 - proceedings.mlr.press
The essential difficulty of gradient-based bilevel optimization using implicit differentiation is
to estimate the inverse Hessian vector product with respect to neural network parameters …

Torchopt: An efficient library for differentiable optimization

J Ren, X Feng, B Liu, X Pan, Y Fu, L Mai… - Journal of Machine …, 2023 - jmlr.org
Differentiable optimization algorithms often involve expensive computations of various meta-
gradients. To address this, we design and implement TorchOpt, a new PyTorch-based …

Modality-agnostic self-supervised learning with meta-learned masked auto-encoder

H Jang, J Tack, D Choi, J Jeong… - Advances in Neural …, 2024 - proceedings.neurips.cc
Despite its practical importance across a wide range of modalities, recent advances in self-
supervised learning (SSL) have been primarily focused on a few well-curated domains, eg …

Rethinking meta-learning from a learning lens

J Wang, W Qiang, J Li, L Si, C Zheng - arXiv preprint arXiv:2409.08474, 2024 - arxiv.org
Meta-learning has emerged as a powerful approach for leveraging knowledge from previous
tasks to solve new tasks. The mainstream methods focus on training a well-generalized …

Efficient Meta label correction based on Meta Learning and bi-level optimization

S Mallem, A Hasnat, A Nakib - Engineering Applications of Artificial …, 2023 - Elsevier
The design of highly accurate deep learning architectures is related to different parameters
through different optimizations at different levels of the design. While the architectural design …

AutoLoRA: Automatically Tuning Matrix Ranks in Low-Rank Adaptation Based on Meta Learning

R Zhang, R Qiang, SA Somayajula, P Xie - arXiv preprint arXiv …, 2024 - arxiv.org
Large-scale pretraining followed by task-specific finetuning has achieved great success in
various NLP tasks. Since finetuning all parameters of large pretrained models poses …

BiLoRA: A Bi-level Optimization Framework for Overfitting-Resilient Low-Rank Adaptation of Large Pre-trained Models

R Qiang, R Zhang, P Xie - arXiv preprint arXiv:2403.13037, 2024 - arxiv.org
Low-rank adaptation (LoRA) is a popular method for fine-tuning large-scale pre-trained
models in downstream tasks by learning low-rank incremental matrices. Though LoRA and …