High-accuracy low-precision training

C De Sa, M Leszczynski, J Zhang, A Marzoev… - arXiv preprint arXiv …, 2018 - arxiv.org
Low-precision computation is often used to lower the time and energy cost of machine
learning, and recently hardware accelerators have been developed to support it. Still, it has …

A2Q: Accumulator-aware quantization with guaranteed overflow avoidance

I Colbert, A Pappalardo… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present accumulator-aware quantization (A2Q), a novel weight quantization method
designed to train quantized neural networks (QNNs) to avoid overflow when using low …

Qpytorch: A low-precision arithmetic simulation framework

T Zhang, Z Lin, G Yang, C De Sa - 2019 Fifth Workshop on …, 2019 - ieeexplore.ieee.org
Low-precision training reduces computational cost and produces efficient models. Recent
research in developing new low-precision training algorithms often relies on simulation to …

[HTML][HTML] Towards resource-frugal deep convolutional neural networks for hyperspectral image segmentation

J Nalepa, M Antoniak, M Myller, PR Lorenzo… - Microprocessors and …, 2020 - Elsevier
Hyperspectral image analysis has been gaining research attention thanks to the current
advances in sensor design which have made acquiring such imagery much more affordable …

Accelerating Graph Neural Networks on Real Processing-In-Memory Systems

C Giannoula, P Yang, I Fernandez Vega… - arXiv e …, 2024 - ui.adsabs.harvard.edu
Abstract Graph Neural Networks (GNNs) are emerging ML models to analyze graph-
structure data. Graph Neural Network (GNN) execution involves both compute-intensive and …

MEM-OPT: A scheduling and data re-use system to optimize on-chip memory usage for CNNs on-board FPGAs

G Dinelli, G Meoni, E Rapuano… - IEEE Journal on …, 2020 - ieeexplore.ieee.org
In the last years, Convolutional Neural networks (CNNs) found applications in many fields
from computer vision to speech recognition, showing outstanding results in terms of …

PyGim: An Efficient Graph Neural Network Library for Real Processing-In-Memory Architectures

C Giannoula, P Yang, I Fernandez, J Yang… - Proceedings of the …, 2024 - dl.acm.org
Graph Neural Networks (GNNs) are emerging models to analyze graph-structure data. GNN
execution involves both compute-intensive and memory-intensive kernels. The latter kernels …

How to Do Machine Learning with Small Data?--A Review from an Industrial Perspective

I Kraljevski, YC Ju, D Ivanov, C Tschöpe… - arXiv preprint arXiv …, 2023 - arxiv.org
Artificial intelligence experienced a technological breakthrough in science, industry, and
everyday life in the recent few decades. The advancements can be credited to the ever …

Constrained deep neural network architecture search for IoT devices accounting for hardware calibration

F Scheidegger, L Benini, C Bekas… - Advances in Neural …, 2019 - proceedings.neurips.cc
Deep neural networks achieve outstanding results for challenging image classification tasks.
However, the design of network topologies is a complex task, and the research community is …

LC: A flexible, extensible open-source toolkit for model compression

Y Idelbayev, MÁ Carreira-Perpiñán - Proceedings of the 30th ACM …, 2021 - dl.acm.org
The continued increase in memory, runtime and energy consumption of deployed machine
learning models on one side, and the trend to miniaturize intelligent devices and sensors on …