High-accuracy low-precision training
Low-precision computation is often used to lower the time and energy cost of machine
learning, and recently hardware accelerators have been developed to support it. Still, it has …
learning, and recently hardware accelerators have been developed to support it. Still, it has …
A2Q: Accumulator-aware quantization with guaranteed overflow avoidance
I Colbert, A Pappalardo… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present accumulator-aware quantization (A2Q), a novel weight quantization method
designed to train quantized neural networks (QNNs) to avoid overflow when using low …
designed to train quantized neural networks (QNNs) to avoid overflow when using low …
Qpytorch: A low-precision arithmetic simulation framework
Low-precision training reduces computational cost and produces efficient models. Recent
research in developing new low-precision training algorithms often relies on simulation to …
research in developing new low-precision training algorithms often relies on simulation to …
[HTML][HTML] Towards resource-frugal deep convolutional neural networks for hyperspectral image segmentation
Hyperspectral image analysis has been gaining research attention thanks to the current
advances in sensor design which have made acquiring such imagery much more affordable …
advances in sensor design which have made acquiring such imagery much more affordable …
Accelerating Graph Neural Networks on Real Processing-In-Memory Systems
C Giannoula, P Yang, I Fernandez Vega… - arXiv e …, 2024 - ui.adsabs.harvard.edu
Abstract Graph Neural Networks (GNNs) are emerging ML models to analyze graph-
structure data. Graph Neural Network (GNN) execution involves both compute-intensive and …
structure data. Graph Neural Network (GNN) execution involves both compute-intensive and …
MEM-OPT: A scheduling and data re-use system to optimize on-chip memory usage for CNNs on-board FPGAs
In the last years, Convolutional Neural networks (CNNs) found applications in many fields
from computer vision to speech recognition, showing outstanding results in terms of …
from computer vision to speech recognition, showing outstanding results in terms of …
PyGim: An Efficient Graph Neural Network Library for Real Processing-In-Memory Architectures
C Giannoula, P Yang, I Fernandez, J Yang… - Proceedings of the …, 2024 - dl.acm.org
Graph Neural Networks (GNNs) are emerging models to analyze graph-structure data. GNN
execution involves both compute-intensive and memory-intensive kernels. The latter kernels …
execution involves both compute-intensive and memory-intensive kernels. The latter kernels …
How to Do Machine Learning with Small Data?--A Review from an Industrial Perspective
I Kraljevski, YC Ju, D Ivanov, C Tschöpe… - arXiv preprint arXiv …, 2023 - arxiv.org
Artificial intelligence experienced a technological breakthrough in science, industry, and
everyday life in the recent few decades. The advancements can be credited to the ever …
everyday life in the recent few decades. The advancements can be credited to the ever …
Constrained deep neural network architecture search for IoT devices accounting for hardware calibration
Deep neural networks achieve outstanding results for challenging image classification tasks.
However, the design of network topologies is a complex task, and the research community is …
However, the design of network topologies is a complex task, and the research community is …
LC: A flexible, extensible open-source toolkit for model compression
Y Idelbayev, MÁ Carreira-Perpiñán - Proceedings of the 30th ACM …, 2021 - dl.acm.org
The continued increase in memory, runtime and energy consumption of deployed machine
learning models on one side, and the trend to miniaturize intelligent devices and sensors on …
learning models on one side, and the trend to miniaturize intelligent devices and sensors on …