MJPNet-S*: Multistyle Joint-Perception Network with Knowledge Distillation for Drone RGB-Thermal Crowd Density Estimation in Smart Cities
W Zhou, X Yang, X Dong, M Fang… - IEEE Internet of Things …, 2024 - ieeexplore.ieee.org
Crowd density estimation has gained significant research interest owing to its potential in
various industries and social applications. Therefore, this article proposes a multistyle joint …
various industries and social applications. Therefore, this article proposes a multistyle joint …
Frequency attention for knowledge distillation
Abstract Knowledge distillation is an attractive approach for learning compact deep neural
networks, which learns a lightweight student model by distilling knowledge from a complex …
networks, which learns a lightweight student model by distilling knowledge from a complex …
Multi-dataset fusion for multi-task learning on face attribute recognition
The goal of face attribute recognition (FAR) is to recognize the attributes of face images,
such as gender, race, etc. Multi-dataset fusion aims to train a network with multiple datasets …
such as gender, race, etc. Multi-dataset fusion aims to train a network with multiple datasets …
Self-Supervised Quantization-Aware Knowledge Distillation
Quantization-aware training (QAT) and Knowledge Distillation (KD) are combined to achieve
competitive performance in creating low-bit deep learning models. However, existing works …
competitive performance in creating low-bit deep learning models. However, existing works …
PURF: Improving teacher representations by imposing smoothness constraints for knowledge distillation
Abstract Knowledge distillation is one of the most persuasive approaches to model
compression that transfers the representational expertise from large deep-learning teacher …
compression that transfers the representational expertise from large deep-learning teacher …
Difference-Aware Distillation for Semantic Segmentation
In recent years, various distillation methods for semantic segmentation have been proposed.
However, these methods typically train the student model to imitate the intermediate features …
However, these methods typically train the student model to imitate the intermediate features …
DistillSleepNet: Heterogeneous Multi-Level Knowledge Distillation via Teacher Assistant for Sleep Staging
Accurate sleep staging is crucial for the diagnosis of diseases such as sleep disorders.
Existing sleep staging models with excellent performance are usually large and require a lot …
Existing sleep staging models with excellent performance are usually large and require a lot …
DSP-KD: dual-stage progressive knowledge distillation for skin disease classification
The increasing global demand for skin disease diagnostics emphasizes the urgent need for
advancements in AI-assisted diagnostic technologies for dermatoscopic images. In current …
advancements in AI-assisted diagnostic technologies for dermatoscopic images. In current …
Quantized Graph Neural Networks for Image Classification
X Xu, L Ma, T Zeng, Q Huang - Mathematics, 2023 - mdpi.com
Researchers have resorted to model quantization to compress and accelerate graph neural
networks (GNNs). Nevertheless, several challenges remain:(1) quantization functions …
networks (GNNs). Nevertheless, several challenges remain:(1) quantization functions …
FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation
This work presents a Fully BInarized Large Language Model (FBI-LLM), demonstrating for
the first time how to train a large-scale binary language model from scratch (not the partial …
the first time how to train a large-scale binary language model from scratch (not the partial …