AUTO-PRUNE: Automated DNN pruning and mapping for ReRAM-based accelerator

S Yang, W Chen, X Zhang, S He, Y Yin… - Proceedings of the ACM …, 2021 - dl.acm.org
Emergent ReRAM-based accelerators support in-memory computation to accelerate deep
neural network (DNN) inference. Weight matrix pruning of DNNs is a widely used technique …

Sense: Model-hardware codesign for accelerating sparse CNNs on systolic arrays

W Sun, D Liu, Z Zou, W Sun, S Chen… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Sparsity is an intrinsic property of convolutional neural networks (CNNs), worth exploiting for
CNN accelerators. However, the extra processing involved comes with hardware overhead …

APQ: Automated DNN Pruning and Quantization for ReRAM-Based Accelerators

S Yang, S He, H Duan, W Chen… - … on Parallel and …, 2023 - ieeexplore.ieee.org
Emerging ReRAM-based accelerators support in-memory computation to accelerate deep
neural network (DNN) inference. Weight matrix pruning is a widely used technique to reduce …

Kernel Shape Control for Row-Efficient Convolution on Processing-In-Memory Arrays

J Rhe, KE Jeon, JC Lee, S Jeong… - 2023 IEEE/ACM …, 2023 - ieeexplore.ieee.org
Processing-in-memory (PIM) architectures have been highlighted as one of the viable
solutions for faster and more power-efficient convolutional neural networks (CNNs) …

MIME: adapting a single neural network for multi-task inference with memory-efficient dynamic pruning

A Bhattacharjee, Y Venkatesha, A Moitra… - Proceedings of the 59th …, 2022 - dl.acm.org
Recent years have seen a paradigm shift towards multi-task learning. This calls for memory
and energy-efficient solutions for inference in a multi-task scenario. We propose an …

An efficient CNN accelerator for pattern-compressed sparse neural networks on FPGA

Y Zhang, H Wang, Z Pan - Neurocomputing, 2025 - Elsevier
Currently, the sparsity of weights and activations are mainly utilized to improve the energy
efficiency and computational performance of CNN accelerators. However, the irregular …

KERNTROL: Kernel Shape Control Toward Ultimate Memory Utilization for In-Memory Convolutional Weight Mapping

J Rhe, KE Jeon, JC Lee, S Jeong… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Processing-in-memory (PIM) architectures have been highlighted as one of the most viable
options for faster and more power-efficient computation. Paired with a convolutional weight …

HFMRE: Constructing Huffman Tree in Bags to Find Excellent Instances for Distantly Supervised Relation Extraction

M Li, C Shao, G Li, M Zhou - Findings of the Association for …, 2023 - aclanthology.org
Since the introduction of distantly supervised relation extraction methods, numerous
approaches have been developed, the most representative of which is multi-instance …

High area/energy efficiency RRAM CNN accelerator with pattern-pruning-based weight mapping scheme

S Yu, L Zhang, J Wang, J Yue, Z Yuan… - 2021 IEEE 10th Non …, 2021 - ieeexplore.ieee.org
Resistive random access memory (RRAM) is an emerging device for processing-in-memory
(PIM) architecture to accelerate convolutional neural network (CNN). However, due to the …

DSCU: Accelerating CNN inference in FPGAs with dual sizes of compute unit

Z Bao, J Guo, W Zhang, H Dang - Journal of Low Power Electronics and …, 2022 - mdpi.com
FPGA-based accelerators have shown great potential in improving the performance of CNN
inference. However, the existing FPGA-based approaches suffer from a low compute unit …