Abstract
Multi-task learning (MTL) using convolutional neural networks (CNN) deals with training the network for multiple correlated tasks in concert. For accuracy-critical applications, there are endeavors to boost the model performance by resorting to a deeper network, which also increases the model complexity. However, such burdensome models are difficult to be deployed on mobile or edge devices. To ensure a trade-off between performance and complexity of CNNs in the context of MTL, we introduce the novel paradigm of self-distillation within the network. Different from traditional knowledge distillation (KD), which trains the Student in accordance with a cumbersome Teacher, our self-distilled multi-task CNN model: SD-MTCNN aims at distilling knowledge from deeper CNN layers into the shallow layers. Precisely, we follow a hard-sharing based MTL setup where all the tasks share a generic feature-encoder on top of which separate task-specific decoders are enacted. Under this premise, SD-MTCNN distills the more abstract features from the decoders to the encoded feature space, which guarantees improved multi-task performance from different parts of the network. We validate SD-MTCNN on three benchmark datasets: CityScapes, NYUv2, and Mini-Taskonomy, and results confirm the improved generalization capability of self-distilled multi-task CNNs in comparison to the literature and baselines.
| Original language | English |
|---|---|
| Publication status | Published - 10 Sept 2020 |
| Event | 31st British Machine Vision Conference, BMVC 2020 - Virtual, Online Duration: 7 Sept 2020 → 10 Sept 2020 |
Conference
| Conference | 31st British Machine Vision Conference, BMVC 2020 |
|---|---|
| City | Virtual, Online |
| Period | 7/09/20 → 10/09/20 |
Bibliographical note
Publisher Copyright:© 2020. The copyright of this document resides with its authors. It may be distributed unchanged freely in print or electronic forms.
ASJC Scopus subject areas
- Artificial Intelligence
- Computer Vision and Pattern Recognition