Master Thesis : Online Distillation with Continual Learning for Cyclic Domain Shifts
Houyon, Joachim
Promotor(s) : Van Droogenbroeck, Marc
Date of defense : 26-Jun-2023/27-Jun-2023 • Permalink : http://hdl.handle.net/2268.2/17707
Details
Title : | Master Thesis : Online Distillation with Continual Learning for Cyclic Domain Shifts |
Author : | Houyon, Joachim |
Date of defense : | 26-Jun-2023/27-Jun-2023 |
Advisor(s) : | Van Droogenbroeck, Marc |
Committee's member(s) : | Louppe, Gilles
Cioppa, Anthony |
Language : | English |
Number of pages : | 90 |
Keywords : | [en] Computer Vision [en] Deep Learning [en] Continual Learning |
Discipline(s) : | Engineering, computing & technology > Computer science |
Commentary : | In reference to IEEE copyrighted material which is used with permission in this thesis, the IEEE does not endorse any of the University of Liège’s products or services. Internal or personal use of this material is permitted. If interested in reprinting/republishing IEEE copyrighted material for advertising or promotional purposes or for creating new collective works for resale or redistribution, please go to http://www.ieee.org/publications_standards/publications/rights/rights_link.html to learn how to obtain a License from RightsLink. If applicable, University Microfilms and/or ProQuest Library, or the Archives of Canada may supply single copies of the dissertation. |
Target public : | Researchers Professionals of domain Student |
Institution(s) : | Université de Liège, Liège, Belgique |
Degree: | Master en science des données, à finalité spécialisée |
Faculty: | Master thesis of the Faculté des Sciences appliquées |
Abstract
[en] The technique of online distillation has become increasingly popular in adapting real-
time deep neural networks using a slow and accurate teacher model. However, one of the
most significant challenges encountered with online distillation is catastrophic forgetting,
which happens when the student model is updated with new domain data and loses the
previously learned knowledge.
The main contribution in this thesis is to apply continual learning techniques to mitigate
the problem fo catastrophic forgetting in online distillation. Indeed, continual learning has
shown to be useful in a more general setting, where the model tend to forget knowledge
from previous tasks when learning the current one. The study aims to assess the efficacy
of various state-of-the-art continual learning methods in reducing catastrophic forgetting
when applied to online distillation, particularly in cyclic domain shifts.
The experimental results presented in this study show improved accuracy and robustness
in the context of online distillation when leveraging continual learning methods to reduce
catastrophic forgetting, with potential applications in fields such as video surveillance or
autonomous driving. As such, this work represents a significant contribution to the fields
of online distillation and continual learning, providing new insights and avenues for future
research.
Cite this master thesis
The University of Liège does not guarantee the scientific quality of these students' works or the accuracy of all the information they contain.