[HTML][HTML] A taxonomy of machine learning clustering algorithms, challenges, and future realms
In the field of data mining, clustering has shown to be an important technique. Numerous
clustering methods have been devised and put into practice, and most of them locate high …
clustering methods have been devised and put into practice, and most of them locate high …
I/o access patterns in hpc applications: A 360-degree survey
The high-performance computing I/O stack has been complex due to multiple software
layers, the inter-dependencies among these layers, and the different performance tuning …
layers, the inter-dependencies among these layers, and the different performance tuning …
Compute-efficient deep learning: Algorithmic trends and opportunities
BR Bartoldson, B Kailkhura, D Blalock - Journal of Machine Learning …, 2023 - jmlr.org
Although deep learning has made great progress in recent years, the exploding economic
and environmental costs of training neural networks are becoming unsustainable. To …
and environmental costs of training neural networks are becoming unsustainable. To …
Fluid: Dataset abstraction and elastic acceleration for cloud-native deep learning training jobs
R Gu, K Zhang, Z Xu, Y Che, B Fan… - 2022 IEEE 38th …, 2022 - ieeexplore.ieee.org
Nowdays, it is prevalent to train deep learning (DL) models in cloud-native platforms that
actively leverage containerization and orchestration technologies for high elasticity, low and …
actively leverage containerization and orchestration technologies for high elasticity, low and …
Miso: exploiting multi-instance gpu capability on multi-tenant gpu clusters
GPU technology has been improving at an expedited pace in terms of size and performance,
empowering HPC and AI/ML researchers to advance the scientific discovery process …
empowering HPC and AI/ML researchers to advance the scientific discovery process …
Why globally re-shuffle? Revisiting data shuffling in large scale deep learning
Stochastic gradient descent (SGD) is the most prevalent algorithm for training Deep Neural
Networks (DNN). SGD iterates the input data set in each training epoch processing data …
Networks (DNN). SGD iterates the input data set in each training epoch processing data …
Plumber: Diagnosing and removing performance bottlenecks in machine learning data pipelines
Input pipelines, which ingest and transform input data, are an essential part of training
Machine Learning (ML) models. However, it is challenging to implement efficient input …
Machine Learning (ML) models. However, it is challenging to implement efficient input …
{SHADE}: Enable Fundamental Cacheability for Distributed Deep Learning Training
Deep learning training (DLT) applications exhibit unique I/O workload behaviors that pose
new challenges for storage system design. DLT is I/O intensive since data samples need to …
new challenges for storage system design. DLT is I/O intensive since data samples need to …
Gpu-enabled asynchronous multi-level checkpoint caching and prefetching
Checkpointing is an I/O intensive operation increasingly used by High-Performance
Computing (HPC) applications to revisit previous intermediate datasets at scale. Unlike the …
Computing (HPC) applications to revisit previous intermediate datasets at scale. Unlike the …
High Throughput Training of Deep Surrogates from Large Ensemble Runs
Recent years have seen a surge in deep learning approaches to accelerate numerical
solvers, which provide faithful but computationally intensive simulations of the physical …
solvers, which provide faithful but computationally intensive simulations of the physical …