AUTO-PRUNE: Automated DNN pruning and mapping for ReRAM-based accelerator
Emergent ReRAM-based accelerators support in-memory computation to accelerate deep
neural network (DNN) inference. Weight matrix pruning of DNNs is a widely used technique …
neural network (DNN) inference. Weight matrix pruning of DNNs is a widely used technique …
Sense: Model-hardware codesign for accelerating sparse CNNs on systolic arrays
W Sun, D Liu, Z Zou, W Sun, S Chen… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Sparsity is an intrinsic property of convolutional neural networks (CNNs), worth exploiting for
CNN accelerators. However, the extra processing involved comes with hardware overhead …
CNN accelerators. However, the extra processing involved comes with hardware overhead …
APQ: Automated DNN Pruning and Quantization for ReRAM-Based Accelerators
S Yang, S He, H Duan, W Chen… - … on Parallel and …, 2023 - ieeexplore.ieee.org
Emerging ReRAM-based accelerators support in-memory computation to accelerate deep
neural network (DNN) inference. Weight matrix pruning is a widely used technique to reduce …
neural network (DNN) inference. Weight matrix pruning is a widely used technique to reduce …
Kernel Shape Control for Row-Efficient Convolution on Processing-In-Memory Arrays
Processing-in-memory (PIM) architectures have been highlighted as one of the viable
solutions for faster and more power-efficient convolutional neural networks (CNNs) …
solutions for faster and more power-efficient convolutional neural networks (CNNs) …
MIME: adapting a single neural network for multi-task inference with memory-efficient dynamic pruning
Recent years have seen a paradigm shift towards multi-task learning. This calls for memory
and energy-efficient solutions for inference in a multi-task scenario. We propose an …
and energy-efficient solutions for inference in a multi-task scenario. We propose an …
An efficient CNN accelerator for pattern-compressed sparse neural networks on FPGA
Y Zhang, H Wang, Z Pan - Neurocomputing, 2025 - Elsevier
Currently, the sparsity of weights and activations are mainly utilized to improve the energy
efficiency and computational performance of CNN accelerators. However, the irregular …
efficiency and computational performance of CNN accelerators. However, the irregular …
KERNTROL: Kernel Shape Control Toward Ultimate Memory Utilization for In-Memory Convolutional Weight Mapping
Processing-in-memory (PIM) architectures have been highlighted as one of the most viable
options for faster and more power-efficient computation. Paired with a convolutional weight …
options for faster and more power-efficient computation. Paired with a convolutional weight …
HFMRE: Constructing Huffman Tree in Bags to Find Excellent Instances for Distantly Supervised Relation Extraction
M Li, C Shao, G Li, M Zhou - Findings of the Association for …, 2023 - aclanthology.org
Since the introduction of distantly supervised relation extraction methods, numerous
approaches have been developed, the most representative of which is multi-instance …
approaches have been developed, the most representative of which is multi-instance …
High area/energy efficiency RRAM CNN accelerator with pattern-pruning-based weight mapping scheme
Resistive random access memory (RRAM) is an emerging device for processing-in-memory
(PIM) architecture to accelerate convolutional neural network (CNN). However, due to the …
(PIM) architecture to accelerate convolutional neural network (CNN). However, due to the …
DSCU: Accelerating CNN inference in FPGAs with dual sizes of compute unit
Z Bao, J Guo, W Zhang, H Dang - Journal of Low Power Electronics and …, 2022 - mdpi.com
FPGA-based accelerators have shown great potential in improving the performance of CNN
inference. However, the existing FPGA-based approaches suffer from a low compute unit …
inference. However, the existing FPGA-based approaches suffer from a low compute unit …