Factorizable graph convolutional networks

Y Yang, Z Feng, M Song… - Advances in Neural …, 2020 - proceedings.neurips.cc
Graphs have been widely adopted to denote structural connections between entities. The
relations are in many cases heterogeneous, but entangled together and denoted merely as …

Task switching network for multi-task learning

G Sun, T Probst, DP Paudel… - Proceedings of the …, 2021 - openaccess.thecvf.com
Abstract We introduce Task Switching Networks (TSNs), a task-conditioned architecture with
a single unified encoder/decoder for efficient multi-task learning. Multiple tasks are …

Frustratingly easy transferability estimation

LK Huang, J Huang, Y Rong… - … on machine learning, 2022 - proceedings.mlr.press
Transferability estimation has been an essential tool in selecting a pre-trained model and
the layers in it for transfer learning, to transfer, so as to maximize the performance on a target …

Interpretability for reliable, efficient, and self-cognitive DNNs: From theories to applications

X Kang, J Guo, B Song, B Cai, H Sun, Z Zhang - Neurocomputing, 2023 - Elsevier
In recent years, remarkable achievements have been made in artificial intelligence tasks
and applications based on deep neural networks (DNNs), especially in the fields of vision …

Model spider: Learning to rank pre-trained models efficiently

YK Zhang, TJ Huang, YX Ding… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Figuring out which Pre-Trained Model (PTM) from a model zoo fits the target task is
essential to take advantage of plentiful model resources. With the availability of numerous …

A Review on Transferability Estimation in Deep Transfer Learning

Y Xue, R Yang, X Chen, W Liu… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Deep transfer learning has become increasingly prevalent in various fields such as industry
and medical science in recent years. To ensure the successful implementation of target …

Progressive network grafting for few-shot knowledge distillation

C Shen, X Wang, Y Yin, J Song, S Luo… - Proceedings of the AAAI …, 2021 - ojs.aaai.org
Abstract Knowledge distillation has demonstrated encouraging performances in deep model
compression. Most existing approaches, however, require massive labeled data to …

Modelgif: Gradient fields for model functional distance

J Song, Z Xu, S Wu, G Chen… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
The last decade has witnessed the success of deep learning and the surge of publicly
released trained models, which necessitates the quantification of the model functional …

Transferability metrics for selecting source model ensembles

A Agostinelli, J Uijlings, T Mensink… - Proceedings of the …, 2022 - openaccess.thecvf.com
We address the problem of ensemble selection in transfer learning: Given a large pool of
source models we want to select an ensemble of models which, after fine-tuning on the …

Ranking neural checkpoints

Y Li, X Jia, R Sang, Y Zhu, B Green… - Proceedings of the …, 2021 - openaccess.thecvf.com
This paper is concerned with ranking many pre-trained deep neural networks (DNNs), called
checkpoints, for the transfer learning to a downstream task. Thanks to the broad use of …