MJPNet-S*: Multistyle Joint-Perception Network with Knowledge Distillation for Drone RGB-Thermal Crowd Density Estimation in Smart Cities

W Zhou, X Yang, X Dong, M Fang… - IEEE Internet of Things …, 2024 - ieeexplore.ieee.org
Crowd density estimation has gained significant research interest owing to its potential in
various industries and social applications. Therefore, this article proposes a multistyle joint …

Frequency attention for knowledge distillation

C Pham, VA Nguyen, T Le, D Phung… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Knowledge distillation is an attractive approach for learning compact deep neural
networks, which learns a lightweight student model by distilling knowledge from a complex …

Multi-dataset fusion for multi-task learning on face attribute recognition

H Lu, S Xu, J Wang - Pattern Recognition Letters, 2023 - Elsevier
The goal of face attribute recognition (FAR) is to recognize the attributes of face images,
such as gender, race, etc. Multi-dataset fusion aims to train a network with multiple datasets …

Self-Supervised Quantization-Aware Knowledge Distillation

K Zhao, M Zhao - arXiv preprint arXiv:2403.11106, 2024 - arxiv.org
Quantization-aware training (QAT) and Knowledge Distillation (KD) are combined to achieve
competitive performance in creating low-bit deep learning models. However, existing works …

PURF: Improving teacher representations by imposing smoothness constraints for knowledge distillation

MI Hossain, S Akhter, CS Hong, EN Huh - Applied Soft Computing, 2024 - Elsevier
Abstract Knowledge distillation is one of the most persuasive approaches to model
compression that transfers the representational expertise from large deep-learning teacher …

Difference-Aware Distillation for Semantic Segmentation

J Gou, X Zhou, L Du, Y Zhan… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
In recent years, various distillation methods for semantic segmentation have been proposed.
However, these methods typically train the student model to imitate the intermediate features …

DistillSleepNet: Heterogeneous Multi-Level Knowledge Distillation via Teacher Assistant for Sleep Staging

Z Jia, H Liang, Y Liu, H Wang… - IEEE Transactions on Big …, 2024 - ieeexplore.ieee.org
Accurate sleep staging is crucial for the diagnosis of diseases such as sleep disorders.
Existing sleep staging models with excellent performance are usually large and require a lot …

DSP-KD: dual-stage progressive knowledge distillation for skin disease classification

X Zeng, Z Ji, H Zhang, R Chen, Q Liao, J Wang, T Lyu… - Bioengineering, 2024 - mdpi.com
The increasing global demand for skin disease diagnostics emphasizes the urgent need for
advancements in AI-assisted diagnostic technologies for dermatoscopic images. In current …

Quantized Graph Neural Networks for Image Classification

X Xu, L Ma, T Zeng, Q Huang - Mathematics, 2023 - mdpi.com
Researchers have resorted to model quantization to compress and accelerate graph neural
networks (GNNs). Nevertheless, several challenges remain:(1) quantization functions …

FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation

L Ma, M Sun, Z Shen - arXiv preprint arXiv:2407.07093, 2024 - arxiv.org
This work presents a Fully BInarized Large Language Model (FBI-LLM), demonstrating for
the first time how to train a large-scale binary language model from scratch (not the partial …