Self-training and multi-task learning for limited data: evaluation study on object detection
Self-training allows a network to learn from the predictions of a more complicated model,
thus often requires well-trained teacher models and mixture of teacher-student data while
multi-task learning jointly optimizes different targets to learn salient interrelationship and
requires multi-task annotations for each training example. These frameworks, despite being
particularly data demanding have potentials for data exploitation if such assumptions can be
relaxed. In this paper, we compare self-training object detection under the deficiency of …
thus often requires well-trained teacher models and mixture of teacher-student data while
multi-task learning jointly optimizes different targets to learn salient interrelationship and
requires multi-task annotations for each training example. These frameworks, despite being
particularly data demanding have potentials for data exploitation if such assumptions can be
relaxed. In this paper, we compare self-training object detection under the deficiency of …
Abstract
Self-training allows a network to learn from the predictions of a more complicated model, thus often requires well-trained teacher models and mixture of teacher-student data while multi-task learning jointly optimizes different targets to learn salient interrelationship and requires multi-task annotations for each training example. These frameworks, despite being particularly data demanding have potentials for data exploitation if such assumptions can be relaxed. In this paper, we compare self-training object detection under the deficiency of teacher training data where students are trained on unseen examples by the teacher, and multi-task learning with partially annotated data, ie single-task annotation per training example. Both scenarios have their own limitation but potentially helpful with limited annotated data. Experimental results show the improvement of performance when using a weak teacher with unseen data for training a multi-task student. Despite the limited setup we believe the experimental results show the potential of multi-task knowledge distillation and self-training, which could be beneficial for future study. Source code and data splits are at https://lhoangan. github. io/multas
openaccess.thecvf.com
以上显示的是最相近的搜索结果。 查看全部搜索结果