Weight-oriented approximation for energy-efficient neural network inference accelerators
ZG Tasoulas, G Zervakis… - … on Circuits and …, 2020 - ieeexplore.ieee.org
Current research in the area of Neural Networks (NN) has resulted in performance
advancements for a variety of complex problems. Especially, embedded system applications …
advancements for a variety of complex problems. Especially, embedded system applications …
Adapt: Fast emulation of approximate dnn accelerators in pytorch
Current state-of-the-art employs approximate multipliers to address the highly increased
power demands of deep neural network (DNN) accelerators. However, evaluating the …
power demands of deep neural network (DNN) accelerators. However, evaluating the …
Power-efficient accelerator design for neural networks using computation reuse
Applications of neural networks in various fields of research and technology have expanded
widely in recent years. In particular, applications with inherent tolerance to accuracy loss …
widely in recent years. In particular, applications with inherent tolerance to accuracy loss …
Leveraging the error resilience of neural networks for designing highly energy efficient accelerators
In recent years, inexact computing has been increasingly regarded as one of the most
promising approaches for slashing energy consumption in many applications that can …
promising approaches for slashing energy consumption in many applications that can …
Hardware approximate techniques for deep neural network accelerators: A survey
Deep Neural Networks (DNNs) are very popular because of their high performance in
various cognitive tasks in Machine Learning (ML). Recent advancements in DNNs have …
various cognitive tasks in Machine Learning (ML). Recent advancements in DNNs have …
CANN: Curable approximations for high-performance deep neural network accelerators
Approximate Computing (AC) has emerged as a means for improving the performance, area
and power-/energy-efficiency of a digital design at the cost of output quality degradation …
and power-/energy-efficiency of a digital design at the cost of output quality degradation …
Improving the accuracy and hardware efficiency of neural networks using approximate multipliers
Improving the accuracy of a neural network (NN) usually requires using larger hardware that
consumes more energy. However, the error tolerance of NNs and their applications allow …
consumes more energy. However, the error tolerance of NNs and their applications allow …
Nnest: Early-stage design space exploration tool for neural network inference accelerators
Deep neural network (DNN) has achieved spectacular success in recent years. In response
to DNN's enormous computation demand and memory footprint, numerous inference …
to DNN's enormous computation demand and memory footprint, numerous inference …
Energy-efficient neural network acceleration in the presence of bit-level memory errors
As a result of the increasing demand for deep neural network (DNN)-based services, efforts
to develop hardware accelerators for DNNs are growing rapidly. However, while highly …
to develop hardware accelerators for DNNs are growing rapidly. However, while highly …
Full-stack optimization for accelerating cnns using powers-of-two weights with fpga validation
We present a full-stack optimization framework for accelerating inference of CNNs
(Convolutional Neural Networks) and validate the approach with a field-programmable gate …
(Convolutional Neural Networks) and validate the approach with a field-programmable gate …