Master Thesis : Online Distillation with Continual Learning for Cyclic Domain Shifts
Houyon, Joachim
Promoteur(s) : Van Droogenbroeck, Marc
Date de soutenance : 26-jui-2023/27-jui-2023 • URL permanente : http://hdl.handle.net/2268.2/17707
Détails
Titre : | Master Thesis : Online Distillation with Continual Learning for Cyclic Domain Shifts |
Auteur : | Houyon, Joachim |
Date de soutenance : | 26-jui-2023/27-jui-2023 |
Promoteur(s) : | Van Droogenbroeck, Marc |
Membre(s) du jury : | Louppe, Gilles
Cioppa, Anthony |
Langue : | Anglais |
Nombre de pages : | 90 |
Mots-clés : | [en] Computer Vision [en] Deep Learning [en] Continual Learning |
Discipline(s) : | Ingénierie, informatique & technologie > Sciences informatiques |
Commentaire : | In reference to IEEE copyrighted material which is used with permission in this thesis, the IEEE does not endorse any of the University of Liège’s products or services. Internal or personal use of this material is permitted. If interested in reprinting/republishing IEEE copyrighted material for advertising or promotional purposes or for creating new collective works for resale or redistribution, please go to http://www.ieee.org/publications_standards/publications/rights/rights_link.html to learn how to obtain a License from RightsLink. If applicable, University Microfilms and/or ProQuest Library, or the Archives of Canada may supply single copies of the dissertation. |
Public cible : | Chercheurs Professionnels du domaine Etudiants |
Institution(s) : | Université de Liège, Liège, Belgique |
Diplôme : | Master en science des données, à finalité spécialisée |
Faculté : | Mémoires de la Faculté des Sciences appliquées |
Résumé
[en] The technique of online distillation has become increasingly popular in adapting real-
time deep neural networks using a slow and accurate teacher model. However, one of the
most significant challenges encountered with online distillation is catastrophic forgetting,
which happens when the student model is updated with new domain data and loses the
previously learned knowledge.
The main contribution in this thesis is to apply continual learning techniques to mitigate
the problem fo catastrophic forgetting in online distillation. Indeed, continual learning has
shown to be useful in a more general setting, where the model tend to forget knowledge
from previous tasks when learning the current one. The study aims to assess the efficacy
of various state-of-the-art continual learning methods in reducing catastrophic forgetting
when applied to online distillation, particularly in cyclic domain shifts.
The experimental results presented in this study show improved accuracy and robustness
in the context of online distillation when leveraging continual learning methods to reduce
catastrophic forgetting, with potential applications in fields such as video surveillance or
autonomous driving. As such, this work represents a significant contribution to the fields
of online distillation and continual learning, providing new insights and avenues for future
research.
Citer ce mémoire
L'Université de Liège ne garantit pas la qualité scientifique de ces travaux d'étudiants ni l'exactitude de l'ensemble des informations qu'ils contiennent.