5G-advanced toward 6G: Past, present, and future

W Chen, X Lin, J Lee, A Toskala, S Sun… - IEEE Journal on …, 2023 - ieeexplore.ieee.org
Since the start of 5G work in 3GPP in early 2016, tremendous progress has been made in
both standardization and commercial deployments. 3GPP is now entering the second phase …

Environmental sustainability and AI in radiology: a double-edged sword

FX Doo, J Vosshenrich, TS Cook, L Moy… - Radiology, 2024 - pubs.rsna.org
According to the World Health Organization, climate change is the single biggest health
threat facing humanity. The global health care system, including medical imaging, must …

Reusing deep learning models: Challenges and directions in software engineering

JC Davis, P Jajal, W Jiang… - 2023 IEEE John …, 2023 - ieeexplore.ieee.org
Deep neural networks (DNNs) achieve state-of-the-art performance in many areas, including
computer vision, system configuration, and question-answering. However, DNNs are …

Compute-efficient deep learning: Algorithmic trends and opportunities

BR Bartoldson, B Kailkhura, D Blalock - Journal of Machine Learning …, 2023 - jmlr.org
Although deep learning has made great progress in recent years, the exploding economic
and environmental costs of training neural networks are becoming unsustainable. To …

Dynamic GPU power capping with online performance tracing for energy efficient GPU computing using DEPO tool

A Krzywaniak, P Czarnul, J Proficz - Future Generation Computer Systems, 2023 - Elsevier
GPU accelerators have become essential to the recent advance in computational power of
high-performance computing (HPC) systems. Current HPC systems' reaching an …

Fusionai: Decentralized training and deploying llms with massive consumer-level gpus

Z Tang, Y Wang, X He, L Zhang, X Pan, Q Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
The rapid growth of memory and computation requirements of large language models
(LLMs) has outpaced the development of hardware, hindering people who lack large-scale …

Cost-effective on-device continual learning over memory hierarchy with Miro

X Ma, S Jeong, M Zhang, D Wang, J Choi… - Proceedings of the 29th …, 2023 - dl.acm.org
Continual learning (CL) trains NN models incrementally from a continuous stream of tasks.
To remember previously learned knowledge, prior studies store old samples over a memory …

{EnvPipe}: Performance-preserving {DNN} training framework for saving energy

S Choi, I Koo, J Ahn, M Jeon, Y Kwon - 2023 USENIX Annual Technical …, 2023 - usenix.org
Energy saving is a crucial mission for data center providers. Among many services, DNN
training and inference are significant contributors to energy consumption. This work focuses …

Towards improved power management in cloud gpus

P Patel, Z Gong, S Rizvi, E Choukse… - IEEE Computer …, 2023 - ieeexplore.ieee.org
As modern server GPUs are increasingly power intensive, better power management
mechanisms can significantly reduce the power consumption, capital costs, and carbon …

DVFO: Learning-Based DVFS for Energy-Efficient Edge-Cloud Collaborative Inference

Z Zhang, Y Zhao, H Li, C Lin… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Due to limited resources on edge and different characteristics of deep neural network (DNN)
models, it is a big challenge to optimize DNN inference performance in terms of energy …