2021 ~ ICCV publication
Rehearsal revealed: The limits and merits of revisiting samples in continual learning

Learning from non-stationary data streams and overcoming catastrophic forgetting still poses a serious challenge for machine learning research. Rather than aiming to improve state-of-the-art, in this work we provide insight into the limits and merits of rehearsal, one of continual learning's most established methods. We hypothesize that models trained sequentially with rehearsal tend to stay in the same low-loss region after a task has finished, but are at risk of overfitting on its sample memory, hence harming generalization. We provide both conceptual and strong empirical evidence on three benchmarks for both behaviors, bringing novel insights into the dynamics of rehearsal and continual learning in general. Finally, we interpret important continual learning works in the light of our findings, allowing for a deeper understanding of their successes.

Eli Verwimp*, Matthias De Lange* and Tinne Tuytelaars (* Equal contribution).
Paper Code

Presented at
  • Seminars with PSI: "Towards adaptive AI with Continual Learning".
  • ContinualAI Reading Group, May 14th, 2021. Youtube (by Eli)
2021 ~ ICCV publication
Continual Prototype Evolution: Learning Online from Non-Stationary Data Streams

Attaining prototypical features to represent class distributions is well established in representation learning. However, learning prototypes online from streams of data proves a challenging endeavor as they rapidly become outdated, caused by an ever-changing parameter space in the learning process. Additionally, continual learning assumes a non-stationary nature of the data stream, typically resulting in catastrophic forgetting of previous knowledge. As a first, we introduce a system addressing both problems, where prototypes evolve continually in a shared latent space, enabling learning and prediction at any point in time. In contrast to the major body of work in continual learning, data streams are processed in an online fashion, without additional task-information, and an efficient memory scheme provides robustness to imbalanced data streams. Besides nearest neighbor based prediction, learning is facilitated by a novel objective function, encouraging cluster density about the class prototype and increased inter-class variance. Furthermore, the latent space quality is elevated by pseudo-prototypes in each batch, constituted by replay of exemplars from memory. We generalize the existing paradigms in continual learning to incorporate data incremental learning from data streams by formalizing a two-agent learner-evaluator framework, and obtain state-of-the-art performance by a significant margin on eight benchmarks, including three highly imbalanced data streams.

Matthias De Lange and Tinne Tuytelaars.
Paper Code

Presented at
  • Seminars with PSI: "Towards adaptive AI with Continual Learning".
  • Blogposts. Leuven.AI Medium
  • Computer Vision Talks, May 28th, 2021. Slides Youtube
  • ContinualAI Reading Group, October 16th, 2020. Slides
2021 ~ TPAMI publication
A continual learning survey: Defying forgetting in classification tasks

Artificial neural networks thrive in solving the classification problem for a particular rigid task, where the network resembles a static entity of knowledge, acquired through generalized learning behaviour from a distinct training phase. However, endeavours to extend this knowledge without targeting the original task usually result in a catastrophic forgetting of this task. Continual learning shifts this paradigm towards a network that can continually accumulate knowledge over different tasks without the need for retraining from scratch, with methods in particular aiming to alleviate forgetting. We focus on task-incremental classification, where tasks arrive in a batch-like fashion, and are delineated by clear boundaries. Our main contributions concern 1) a taxonomy and extensive overview of the state-of-the-art, 2) a novel framework to continually determine stability-plasticity trade-off of the continual learner, 3) a comprehensive experimental comparison of 10 state-of-the-art continual learning methods and 4 baselines. We empirically scrutinize which method performs best, both on balanced Tiny Imagenet and a large-scale unbalanced iNaturalist datasets. We study the influence of model capacity, weight decay and dropout regularization, and the order in which the tasks are presented, and qualitatively compare methods in terms of required memory, computation time and storage.

Matthias De Lange, Rahaf Aljundi, Marc Masana, Sarah Parisot, Xu Jia, Ales Leonardis, Gregory Slabaugh, Tinne Tuytelaars submission at journal IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI).
TPAMI Arxiv Code

2020 ~ CVPR publication
Unsupervised Model Personalization while Preserving Privacy and Scalability: An Open Problem

Personalization of models to local user images is prone to three main problems: scalability towards thousands of users, retaining user-privacy, and labeling local user data. Our Dual User-Adaptation framework (DUA) unveils a novel perspective to tackle all of these practical concerns and enables personalization on both the server and local user edge-device. First, we train a sequence of task-specific models on the server, making the number of models independent of the user-count to greatly enhance scalability. These models are aggregated into a single personalized model solely based on user-specific priors. As a consequence, no sensitive user data is shared and using parameter importance avoids the need for labeled user data. Additionally, the DUA framework invigorates a second round of user-data adaptation, but now locally on the user edge-device. Therefore, DUA manifests two personalization opportunities without sharing any explicit user data. Code and benchmarks are publicly available online.

Matthias De Lange, Sarah Parisot, Xu Jia, Ales Leonardis, Gregory Slabaugh, Tinne Tuytelaars at CVPR conference.
Paper Code

Presented at
  • CVPR 2020. Slides
  • "Image and Speech Processing" seminar in Center of Speech and Image Processing (PSI). Slides
  • June 2020 ContinualAI meetup: "Continual Learning: in the Cloud, at the Edge or Both?" Slides Video

[Jun. 2022] Research Scientist Intern @ Meta Reality Labs Research

Joining Meta Reality Labs as a research scientist for 5 months was an incredible intern experience! I'm forever grateful for joining the team with Karl Ridgeway, Hamid Eghbalzadeh, Michael Iuzzolino, and Franziska Meier to work on Continual Learning for Augmented Reality. There’s nothing like working in an open and collaborative research environment, surrounded by people wanting to build awesome things! During my internship, I focused on Continual Learning for Augmented Reality with Egocentric Vision data, leading to three paper submissions: my personal computer vision project and two co-authored papers in self-supervised learning and human-computer interaction.

[Aug. 2022] Outstanding reviewer @ ECCV 2022

Acknowledged as an outstanding reviewer at the top-tier European Computer Vision Conference, one of the top conferences in computer vision. Reviewing entails understanding of scientific papers loosely related to your domain of expertise and giving constructive feedback to the authors on experiments, scientific methods, and reporting of the results.

[Aug. 2021] Main General chair @ CLVision workshop (3rd edition, CVPR 2022)

The CVPR 2022 Workshop on Continual Learning (CLVision, 3rd Edition) aims to gather researchers and engineers from academia and industry to discuss the latest advances in Continual Learning. In this one-day workshop, we will have regular paper presentations, invited speakers, and technical benchmark challenges to present the current state of the art, as well as the limitations and future directions for Continual Learning, arguably one of the most crucial milestones of AI.

With a focus on interdisciplinarity and inclusiveness, and with our experience from the succesful previous edition (over 600 unique attendees!), we hope to make an even better success of this edition! See you at CVPR 2022! Check out the website for more details!

[Oct. 2021] Featured daily picks at @ICCV 2021

Honoured to be feature in the ICCV daily magazine by RSIPvision! Interested in continual learning? Check out my top paper picks at #ICCV2021! Check it out here!
Don't forget to pass by our posters on "Continual Prototype Evolution" and "Rehearsal Revealed".

[Aug. 2021] Continual Learning Challenge @ ICCV 2021 (Co-organizer)
Do you also feel it's time to bring Continual Learning closer to the industry? :muscle:
In the ICCV2021 CL challenge you can do just that, and win 20,000 USD along the way!For this workshop, Huawei introduces new benchmarks for both lifelong classification and object detection towards real-life autonomous driving! :fire:
All of this is implemented for you in the modular Avalanche framework, so no hassle with dataloaders etc. All you have to focus on is your clever CL solution :sunglasses:

Deadline: 1st of October.

For more information, see Track3: https://sslad2021.github.io/pages/challenge.html
Participation links (CodaLab):
- Track 3A (Continual classification): https://competitions.codalab.org/competitions/33830
- Track 3B (Continual 2D object detection): https://competitions.codalab.org/competitions/33993
[May. 2021] Continual Learning Blog @ Leuven.AI

Many of the contemporary Machine Learning models are trained only once and remain static when deployed. But shouldn't generally intelligent systems evolve with their surrounding world? Find out more in my introduction to the field of Continual Learning.

[Dec. 2020] General chair @ CLVision workshop (2nd edition, CVPR 2021)

I'm excited to announce that our Workshop "Continual Learning in Computer Vision" has been accepted for a second edition at CVPR 2021! The workshop was selected among more than 100 proposals this year with excellent feedback. Last year edition was a great success, we will make sure to make it even better, diverse and inclusive this year! Check out the website for more details!

[Nov. 2020] Winner and runner up @ HackJunction

We won the Just-AI challenge and placed 2nd at Columbia Road challenge with a personalized experience with a conversational AI-based shopping assistant. Hack Junction was an amazing experience with a great team of engineering friends. Hacking and coding our way through the lockdown (:

Check out our winning project Nutreat: Shop Smarter. Watch our video!

Research scientist (intern)
Meta, Reality Labs Research


Research focus: continual learning with neural networks for Augmented Reality

Full-time, paid 5 month contract

3 paper publications (1x main author, 2x co-author)

2022

PhD at Center for Image and Speech Processing (PSI)
Department Electrical Engineering (ESAT)


KU Leuven

PSI Seminar Organizer

Program Committee (Reviewing) @ IEEE TPAMI, IEEE TMI, ICML2020, CVPR2021, IJCAI2021, ICCV2021, CVPR2022, CoLLA 2022 (senior PC)

2018 - Present

Master Engineering Technology Electronics and ICT
SPECIALISATION ICT


Laureate (Highest grade 2018)

Summa Cum Laude

KU Leuven

2017 - 2018

Academical Bachelor Engineering Technology Electronics and ICT
SPECIALISATION ICT


Summa Cum Laude

KU Leuven

2014 - 2017

Highschool Degree
Science-Mathematics


Don Boscocollege Zwijnaarde

2008 - 2014