Efficient acceleration of deep learning inference on resource-constrained edge devices: A review
Successful integration of deep neural networks (DNNs) or deep learning (DL) has resulted
in breakthroughs in many areas. However, deploying these highly accurate models for data …
in breakthroughs in many areas. However, deploying these highly accurate models for data …
[HTML][HTML] Few-shot object detection on aerial imagery via deep metric learning and knowledge inheritance
W Li, J Zhou, X Li, Y Cao, G Jin - … Journal of Applied Earth Observation and …, 2023 - Elsevier
Object detection is crucial in aerial imagery analysis. Previous methods based on
convolutional neural networks (CNNs) require large-scale labeled datasets for training to …
convolutional neural networks (CNNs) require large-scale labeled datasets for training to …
Knowledge distillation with the reused teacher classifier
Abstract Knowledge distillation aims to compress a powerful yet cumbersome teacher model
into a lightweight student model without much sacrifice of performance. For this purpose …
into a lightweight student model without much sacrifice of performance. For this purpose …
Tokens-to-token vit: Training vision transformers from scratch on imagenet
Transformers, which are popular for language modeling, have been explored for solving
vision tasks recently, eg, the Vision Transformer (ViT) for image classification. The ViT model …
vision tasks recently, eg, the Vision Transformer (ViT) for image classification. The ViT model …
Volo: Vision outlooker for visual recognition
Recently, Vision Transformers (ViTs) have been broadly explored in visual recognition. With
low efficiency in encoding fine-level features, the performance of ViTs is still inferior to the …
low efficiency in encoding fine-level features, the performance of ViTs is still inferior to the …
L2g: A simple local-to-global knowledge transfer framework for weakly supervised semantic segmentation
Mining precise class-aware attention maps, aka, class activation maps, is essential for
weakly supervised semantic segmentation. In this paper, we present L2G, a simple online …
weakly supervised semantic segmentation. In this paper, we present L2G, a simple online …
Cross-layer distillation with semantic calibration
Recently proposed knowledge distillation approaches based on feature-map transfer
validate that intermediate layers of a teacher model can serve as effective targets for training …
validate that intermediate layers of a teacher model can serve as effective targets for training …
Channel-wise knowledge distillation for dense prediction
Abstract Knowledge distillation (KD) has been proven a simple and effective tool for training
compact dense prediction models. Lightweight student networks are trained by extra …
compact dense prediction models. Lightweight student networks are trained by extra …
All tokens matter: Token labeling for training better vision transformers
In this paper, we present token labeling---a new training objective for training high-
performance vision transformers (ViTs). Different from the standard training objective of ViTs …
performance vision transformers (ViTs). Different from the standard training objective of ViTs …
General instance distillation for object detection
In recent years, knowledge distillation has been proved to be an effective solution for model
compression. This approach can make lightweight student models acquire the knowledge …
compression. This approach can make lightweight student models acquire the knowledge …