DeepBurning-MixQ: An open source mixed-precision neural network accelerator design framework for FPGAs

E Luo, H Huang, C Liu, G Li, B Yang… - 2023 IEEE/ACM …, 2023 - ieeexplore.ieee.org
Mixed-precision neural networks (MPNNs) that enable the use of just enough data width for
a deep learning task promise significant advantages of both inference accuracy and …

Differentiable Neural Architecture, Mixed Precision and Accelerator Co-Search

KT Chitty-Venkata, Y Bian, M Emani… - IEEE …, 2023 - ieeexplore.ieee.org
Quantization, effective Neural Network architecture, and efficient accelerator hardware are
three important design paradigms to maximize accuracy and efficiency. Mixed Precision …

Joint Pruning and Channel-wise Mixed-Precision Quantization for Efficient Deep Neural Networks

BA Motetti, M Risso, A Burrello, E Macii… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
The resource requirements of deep neural networks (DNNs) pose significant challenges to
their deployment on edge devices. Common approaches to address this issue are pruning …

A Context-Awareness and Hardware-Friendly Sparse Matrix Multiplication Kernel for CNN Inference Acceleration

H Wang, Y Ding, Y Liu, W Liu, C Liu… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Sparsification technology is crucial for deploying convolutional neural networks in resource-
constrained environments. However, the efficiency of sparse models is hampered by …

Convergence-aware operator-wise mixed-precision training

W Dai, Z Jia, Y Bai, Q Sun - CCF Transactions on High Performance …, 2024 - Springer
With the support of more precision formats in emerging hardware architectures, mixed-
precision has become a popular approach to accelerate deep learning (DL) training …

Hardware-aware design, search, and optimization of deep neural networks

SS Chitty-Venkata - 2023 - search.proquest.com
Deep Learning has achieved remarkable progress in the last decade due to its powerful
automatic representation capability for a variety of tasks, such as Image Recognition …