Nonlinear nonmodal stability theory

RR Kerswell - Annual Review of Fluid Mechanics, 2018 - annualreviews.org
This review discusses a recently developed optimization technique for analyzing the
nonlinear stability of a flow state. It is based on a nonlinear extension of nonmodal analysis …

Reverse time migration: A prospect of seismic imaging methodology

HW Zhou, H Hu, Z Zou, Y Wo, O Youn - Earth-science reviews, 2018 - Elsevier
Reverse time migration (RTM) is a seismic imaging method to map the subsurface reflectivity
using recorded seismic waveforms. The practice in exploration seismology has long …

Filip: Fine-grained interactive language-image pre-training

L Yao, R Huang, L Hou, G Lu, M Niu, H Xu… - arXiv preprint arXiv …, 2021 - arxiv.org
Unsupervised large-scale vision-language pre-training has shown promising advances on
various downstream tasks. Existing methods often model the cross-modal interaction either …

Efficient large-scale language model training on gpu clusters using megatron-lm

D Narayanan, M Shoeybi, J Casper… - Proceedings of the …, 2021 - dl.acm.org
Large language models have led to state-of-the-art accuracies across several tasks.
However, training these models efficiently is challenging because: a) GPU memory capacity …

Learning transferable visual models from natural language supervision

A Radford, JW Kim, C Hallacy… - International …, 2021 - proceedings.mlr.press
State-of-the-art computer vision systems are trained to predict a fixed set of predetermined
object categories. This restricted form of supervision limits their generality and usability since …

Revisiting stereo depth estimation from a sequence-to-sequence perspective with transformers

Z Li, X Liu, N Drenkow, A Ding… - Proceedings of the …, 2021 - openaccess.thecvf.com
Stereo depth estimation relies on optimal correspondence matching between pixels on
epipolar lines in the left and right images to infer depth. In this work, we revisit the problem …

[HTML][HTML] Combined scaling for zero-shot transfer learning

H Pham, Z Dai, G Ghiasi, K Kawaguchi, H Liu, AW Yu… - Neurocomputing, 2023 - Elsevier
Recent developments in multimodal training methodologies, including CLIP and ALIGN,
obviate the necessity for individual data labeling. These approaches utilize pairs of data and …

Gpipe: Efficient training of giant neural networks using pipeline parallelism

Y Huang, Y Cheng, A Bapna, O Firat… - Advances in neural …, 2019 - proceedings.neurips.cc
Scaling up deep neural network capacity has been known as an effective approach to
improving model quality for several different machine learning tasks. In many cases …

Training deep nets with sublinear memory cost

T Chen, B Xu, C Zhang, C Guestrin - arXiv preprint arXiv:1604.06174, 2016 - arxiv.org
We propose a systematic approach to reduce the memory consumption of deep neural
network training. Specifically, we design an algorithm that costs O (sqrt (n)) memory to train …

Memory-efficient pipeline-parallel dnn training

D Narayanan, A Phanishayee, K Shi… - International …, 2021 - proceedings.mlr.press
Many state-of-the-art ML results have been obtained by scaling up the number of
parameters in existing models. However, parameters and activations for such large models …