Lifelong Machine Learning without Lifelong Data Retention

Author(s)
Smith, James Seale
Advisor(s)
Editor(s)
Associated Organization(s)
Organizational Unit
Organizational Unit
School of Interactive Computing
School established in 2007
Supplementary to:
Abstract
Machine learning models suffer from a phenomenon known as catastrophic forgetting when learning novel concepts from continuously shifting training data. Typical solutions for this continual learning problem require extensive replay of previously seen data, which increases memory costs and may violate data privacy. To address these challenges, we first explore replacing this replay data with alternatives: (i) unlabeled data “from the wild” and (ii) synthetic data generated via model inversion. Our work using this alternative replay data boasts strong performance on replay-free continual learning for image classification. Next, we consider an alternative solution to entirely replace replay data: pre-training. Specifically, we leverage strongly pre-trained models and continuously edit them with prompts and low-rank adapters for both (i) image classification and (ii) natural-language visual reasoning. Finally, we extend the idea of continual learning using pre-trained models to the proposed setting of continual customization of text-to-image diffusion models. We hope that our work on enabling models to learn from evolving data distributions and adapt to new tasks will help unlock the full potential of machine learning in addressing emerging real-world challenges.
Sponsor
Date
2023-12-10
Extent
Resource Type
Text
Resource Subtype
Dissertation
Rights Statement
Rights URI