Graph neural network training and data tiering

SW Min, K Wu, M Hidayetoglu, J Xiong… - Proceedings of the 28th …, 2022 - dl.acm.org
Graph Neural Networks (GNNs) have shown success in learning from graph-structured data,
with applications to fraud detection, recommendation, and knowledge graph reasoning …

Baechi: fast device placement of machine learning graphs

B Jeon, L Cai, P Srivastava, J Jiang, X Ke… - Proceedings of the 11th …, 2020 - dl.acm.org
Machine Learning graphs (or models) can be challenging or impossible to train when either
devices have limited memory, or the models are large. Splitting the model graph across …

CoProver: A recommender system for proof construction

E Yeh, B Hitaj, S Owre, M Quemener… - … Conference on Intelligent …, 2023 - Springer
Abstract Interactive Theorem Provers (ITPs) are an indispensable tool in the arsenal of
formal method experts as a platform for construction and (formal) verification of proofs. The …

Multi-feature vision transformer via self-supervised representation learning for improvement of covid-19 diagnosis

X Qi, DJ Foran, JL Nosher, I Hacihaliloglu - Workshop on Medical Image …, 2022 - Springer
The role of chest X-ray (CXR) imaging, due to being more cost-effective, widely available,
and having a faster acquisition time compared to CT, has evolved during the COVID-19 …

SRL: Scaling Distributed Reinforcement Learning to Over Ten Thousand Cores

Z Mei, W Fu, J Gao, G Wang, H Zhang, Y Wu - arXiv preprint arXiv …, 2023 - arxiv.org
The ever-growing complexity of reinforcement learning (RL) tasks demands a distributed
system to efficiently generate and process a massive amount of data. However, existing …

Communication-Efficient Distributed Deep Learning via Federated Dynamic Averaging

M Theologitis, G Frangias, G Anestis… - arXiv preprint arXiv …, 2024 - arxiv.org
Driven by the ever-growing volume and decentralized nature of data, coupled with the
escalating size of modern models, distributed deep learning (DDL) has been entrenched as …

DistSim: A performance model of large-scale hybrid distributed DNN training

G Lu, R Chen, Y Wang, Y Zhou, R Zhang, Z Hu… - Proceedings of the 20th …, 2023 - dl.acm.org
With the ever-increasing computational demand of DNN training workloads, distributed
training has been widely adopted. A combination of data, model and pipeline parallelism …

TAPS: Topology-Aware Intra-Operator Parallelism Strategy Searching Algorithm for Deep Neural Networks

P Liang, H Zheng, T Su, L Qiao, D Li - arXiv preprint arXiv:2301.04285, 2023 - arxiv.org
TAPS is a Topology-Aware intra-operator Parallelism strategy Searching algorithm that
generates intra-operator parallelism strategies by considering both intra-node and inter …