Home

Awesome

Awesome-Forgetting-in-Deep-Learning

Awesome <img src="https://img.shields.io/badge/Contributions-Welcome-278ea5" alt=""/>

A comprehensive list of papers about 'A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning'.

Abstract

Forgetting refers to the loss or deterioration of previously acquired information or knowledge. While the existing surveys on forgetting have primarily focused on continual learning, forgetting is a prevalent phenomenon observed in various other research domains within deep learning. Forgetting manifests in research fields such as generative models due to generator shifts, and federated learning due to heterogeneous data distributions across clients. Addressing forgetting encompasses several challenges, including balancing the retention of old task knowledge with fast learning of new tasks, managing task interference with conflicting goals, and preventing privacy leakage, etc. Moreover, most existing surveys on continual learning implicitly assume that forgetting is always harmful. In contrast, our survey argues that forgetting is a double-edged sword and can be beneficial and desirable in certain cases, such as privacy-preserving scenarios. By exploring forgetting in a broader context, we aim to present a more nuanced understanding of this phenomenon and highlight its potential advantages. Through this comprehensive survey, we aspire to uncover potential solutions by drawing upon ideas and approaches from various fields that have dealt with forgetting. By examining forgetting beyond its conventional boundaries, in future work, we hope to encourage the development of novel strategies for mitigating, harnessing, or even embracing forgetting in real applications.

Citation

If you find our paper or this resource helpful, please consider citing:

@article{Forgetting_Survey_2024,
  title={A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning},
  author={Wang, Zhenyi and Yang, Enneng and Shen, Li and Huang, Heng},
  journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
  year={2024},
  publisher={IEEE}
}

Thanks!


Framework


Harmful Forgetting

Harmful forgetting occurs when we desire the machine learning model to retain previously learned knowledge while adapting to new tasks, domains, or environments. In such cases, it is important to prevent and mitigate knowledge forgetting.

Problem SettingGoalSource of forgetting
Continual Learninglearn non-stationary data distribution without forgetting previous knowledgedata-distribution shift during training
Foundation Modelunsupervised learning on large-scale unlabeled datadata-distribution shift in pre-training, fine-tuning
Domain Adaptationadapt to target domain while maintaining performance on source domaintarget domain sequentially shift over time
Test-time Adaptationmitigate the distribution gap between training and testingadaptation to the test data distribution during testing
Meta-Learninglearn adaptable knowledge to new tasksincrementally meta-learn new classes / task-distribution shift
Generative Modellearn a generator to appriximate real data distributiongenerator shift/data-distribution shift
Reinforcement Learningmaximize accumulate rewardsstate, action, reward and state transition dynamics
Federated Learningdecentralized training without sharing datamodel average; non-i.i.d data; data-distribution shift
<!-- | Self-Supervised Learning | unsupervised pre-training | data-distribution shift | -->

Links: <u> Forgetting in Continual Learning </u> | <u> Forgetting in Foundation Models </u> | <u> Forgetting in Domain Adaptation</u> | <u> Forgetting in Test-Time Adaptation</u> |
<u> Forgetting in Meta-Learning </u>|
<u> Forgetting in Generative Models </u>| <u> Forgetting in Reinforcement Learning</u> | <u> Forgetting in Federated Learning</u>


Forgetting in Continual Learning

<a href="#top">[Back to top]</a>

The goal of continual learning (CL) is to learn on a sequence of tasks without forgetting the knowledge on previous tasks.

Links: <u> Task-aware CL </u>| <u> Task-free CL </u>| <u> Online CL </u>| <u> Semi-supervised CL </u>| <u> Few-shot CL </u>| <u> Unsupervised CL </u>| <u> Theoretical Analysis </u>

Survey and Book

Paper TitleYearConference/Journal
A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning2024TPAMI
Deep Class-Incremental Learning: A Survey2024TPAMI
A Comprehensive Survey of Continual Learning: Theory, Method and Application2024TPAMI
Federated Continual Learning for Edge-AI: A Comprehensive Survey2024Arxiv
Continual Learning with Neuromorphic Computing: Theories, Methods, and Applications2024Arxiv
Recent Advances of Multimodal Continual Learning: A Comprehensive Survey2024Arxiv
Towards General Industrial Intelligence: A Survey on Industrial IoT-Enhanced Continual Large Models2024Arxiv
Towards Lifelong Learning of Large Language Models: A Survey2024Arxiv
Recent Advances of Foundation Language Models-based Continual Learning: A Survey2024Arxiv
Continual Learning of Large Language Models: A Comprehensive Survey2024Arxiv
Continual Learning on Graphs: Challenges, Solutions, and Opportunities2024Arxiv
Continual Learning on Graphs: A Survey2024Arxiv
Continual Learning for Large Language Models: A Survey2024Arxiv
Continual Learning with Pre-Trained Models: A Survey2024Arxiv
A Survey on Few-Shot Class-Incremental Learning2024Neural Networks
Sharpness and Gradient Aware Minimization for Memory-based Continual Learning2023SOICT
A Survey on Incremental Update for Neural Recommender Systems2023Arxiv
Continual Graph Learning: A Survey2023Arxiv
Towards Label-Efficient Incremental Learning: A Survey2023Arxiv
Advancing continual lifelong learning in neural information retrieval: definition, dataset, framework, and empirical evaluation2023Arxiv
How to Reuse and Compose Knowledge for a Lifetime of Tasks: A Survey on Continual Learning and Functional Composition2023Transactions on Machine Learning Research
Online Continual Learning in Image Classification: An Empirical Survey2022Neurocomputing
Class-incremental learning: survey and performance evaluation on image classification2022TPAMI
Towards Continual Reinforcement Learning: A Review and Perspectives2022Journal of Artificial Intelligence Research
An Introduction to Lifelong Supervised Learning2022Arxiv
Continual Learning for Real-World Autonomous Systems: Algorithms, Challenges and Frameworks2022Arxiv
A continual learning survey: Defying forgetting in classification tasks2021TPAMI
Recent Advances of Continual Learning in Computer Vision: An Overview2021Arxiv
Continual Lifelong Learning in Natural Language Processing: A Survey2020COLING
A Comprehensive Study of Class Incremental Learning Algorithms for Visual Tasks2020Neural Networks
Continual Lifelong Learning with Neural Networks: A Review2019Neural Networks
Three scenarios for continual learning2018NeurIPSW
Lifelong Machine Learning2016Book

Task-aware CL

<a href="#top">[Back to top]</a>

Task-aware CL focuses on addressing scenarios where explicit task definitions, such as task IDs or labels, are available during the CL process. Existing methods on task-aware CL have explored five main branches: Memory-based Methods | Architecture-based Methods | Regularization-based Methods | Subspace-based Methods | Bayesian Methods.

Memory-based Methods

<a href="#top">[Back to top]</a>

Memory-based (or Rehearsal-based) method keeps a memory buffer that stores the examples/knowledges from previous tasks and replay those examples during learning new tasks.

Paper TitleYearConference/Journal
Prior-free Balanced Replay: Uncertainty-guided Reservoir Sampling for Long-Tailed Continual Learning2024MM
FTF-ER: Feature-Topology Fusion-Based Experience Replay Method for Continual Graph Learning2024MM
Multi-layer Rehearsal Feature Augmentation for Class-Incremental Learning2024ICML
Gradual Divergence for Seamless Adaptation: A Novel Domain Incremental Learning Method2024ICML
Accelerating String-Key Learned Index Structures via Memoization based Incremental Training2024VLDB
DSLR: Diversity Enhancement and Structure Learning for Rehearsal-based Graph Continual Learning2024WWW
Exemplar-based Continual Learning via Contrastive Learning2024IEEE Transactions on Artificial Intelligence
Saving 100x Storage: Prototype Replay for Reconstructing Training Sample Distribution in Class-Incremental Semantic Segmentation2023NeurIPS
Selective Amnesia: A Continual Learning Approach to Forgetting in Deep Generative Models2023NeurIPS
A Unified Approach to Domain Incremental Learning with Memory: Theory and Algorithm2023NeurIPS
An Efficient Dataset Condensation Plugin and Its Application to Continual Learning2023NeurIPS
Augmented Memory Replay-based Continual Learning Approaches for Network Intrusion Detection2023NeurIPS
Bilevel Coreset Selection in Continual Learning: A New Formulation and Algorithm2023NeurIPS
FeCAM: Exploiting the Heterogeneity of Class Distributions in Exemplar-Free Continual Learning2023NeurIPS
Distributionally Robust Memory Evolution with Generalized Divergence for Continual Learning2023TPAMI
Improving Replay Sample Selection and Storage for Less Forgetting in Continual Learning2023ICCV
Masked Autoencoders are Efficient Class Incremental Learners2023ICCV
Error Sensitivity Modulation based Experience Replay: Mitigating Abrupt Representation Drift in Continual Learning2023ICLR
A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental Learning2023ICLR
DualHSIC: HSIC-Bottleneck and Alignment for Continual Learning2023ICML
DDGR: Continual Learning with Deep Diffusion-based Generative Replay2023ICML
BiRT: Bio-inspired Replay in Vision Transformers for Continual Learning2023ICML
Neuro-Symbolic Continual Learning: Knowledge, Reasoning Shortcuts and Concept Rehearsal2023ICML
Poisoning Generative Replay in Continual Learning to Promote Forgetting2023ICML
Regularizing Second-Order Influences for Continual Learning2023CVPR
Class-Incremental Exemplar Compression for Class-Incremental Learning2023CVPR
A closer look at rehearsal-free continual learning2023CVPRW
Continual Learning by Modeling Intra-Class Variation2023TMLR
Class-Incremental Learning using Diffusion Model for Distillation and Replay2023Arxiv
On the Effectiveness of Lipschitz-Driven Rehearsal in Continual Learning2022NeurIPS
Exploring Example Influence in Continual Learning2022NeurIPS
Navigating Memory Construction by Global Pseudo-Task Simulation for Continual Learning2022NeurIPS
Learning Fast, Learning Slow: A General Continual Learning Method based on Complementary Learning System2022ICLR
Information-theoretic Online Memory Selection for Continual Learning2022ICLR
Memory Replay with Data Compression for Continual Learning2022ICLR
Improving Task-free Continual Learning by Distributionally Robust Memory Evolution2022ICML
GCR: Gradient Coreset based Replay Buffer Selection for Continual Learning2022CVPR
On the Convergence of Continual Learning with Adaptive Methods2022UAI
RMM: Reinforced Memory Management for Class-Incremental Learning2021NeurIPS
Rainbow Memory: Continual Learning with a Memory of Diverse Samples2021CVPR
Prototype Augmentation and Self-Supervision for Incremental Learning2021CVPR
Class-incremental experience replay for continual learning under concept drift2021CVPRW
Always Be Dreaming: A New Approach for Data-Free Class-Incremental Learning2021ICCV
Using Hindsight to Anchor Past Knowledge in Continual Learning2021AAAI
Improved Schemes for Episodic Memory-based Lifelong Learning2020NeurIPS
Dark Experience for General Continual Learning: a Strong, Simple Baseline2020NeurIPS
La-MAML: Look-ahead Meta Learning for Continual Learning2020NeurIPS
GAN Memory with No Forgetting2020NeurIPS
Brain-inspired replay for continual learning with artificial neural networks2020Nature Communications
LAMOL: LAnguage MOdeling for Lifelong Language Learning2020ICLR
Mnemonics Training: Multi-Class Incremental Learning without Forgetting2020CVPR
GDumb: A Simple Approach that Questions Our Progress in Continual Learning2020ECCV
Episodic Memory in Lifelong Language Learning2019NeurIPS
Continual Learning with Tiny Episodic Memories2019ICML
Efficient lifelong learning with A-GEM2019ICLR
Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference2019ICLR
Large Scale Incremental Learning2019CVPR
On Tiny Episodic Memories in Continual Learning2019Arxiv
Memory Replay GANs: learning to generate images from new categories without forgetting2018NeurIPS
Progress & Compress: A scalable framework for continual learning2018ICML
Gradient Episodic Memory for Continual Learning2017NeurIPS
Continual Learning with Deep Generative Replay2017NeurIPS
iCaRL: Incremental Classifier and Representation Learning2017CVPR
Catastrophic forgetting, rehearsal and pseudorehearsal1995Connection Science
Architecture-based Methods

<a href="#top">[Back to top]</a>

The architecture-based approach avoids forgetting by reducing parameter sharing between tasks or adding parameters to new tasks.

Paper TitleYearConference/Journal
Harnessing Neural Unit Dynamics for Effective and Scalable Class-Incremental Learning2024ICML
Revisiting Neural Networks for Continual Learning: An Architectural Perspective2024IJCAI
Recall-Oriented Continual Learning with Generative Adversarial Meta-Model2024AAAI
Divide and not forget: Ensemble of selectively trained experts in Continual Learning2024ICLR
A Probabilistic Framework for Modular Continual Learning2024ICLR
Incorporating neuro-inspired adaptability for continual learning in artificial intelligence2023Nature Machine Intelligence
TriRE: A Multi-Mechanism Learning Paradigm for Continual Knowledge Retention and Promotion2023NeurIPS
ScrollNet: Dynamic Weight Importance for Continual Learning2023ICCV
CLR: Channel-wise Lightweight Reprogramming for Continual Learning2023ICCV
Parameter-Level Soft-Masking for Continual Learning2023ICML
Continual Learning on Dynamic Graphs via Parameter Isolation2023SIGIR
Heterogeneous Continual Learning2023CVPR
Dense Network Expansion for Class Incremental Learning2023CVPR
Achieving a Better Stability-Plasticity Trade-off via Auxiliary Networks in Continual Learning2023CVPR
Forget-free Continual Learning with Winning Subnetworks2022ICML
NISPA: Neuro-Inspired Stability-Plasticity Adaptation for Continual Learning in Sparse Networks2022ICML
Continual Learning with Filter Atom Swapping2022ICLR
SparCL: Sparse Continual Learning on the Edge2022NeurIPS
Learning Bayesian Sparse Networks with Full Experience Replay for Continual Learning2022CVPR
FOSTER: Feature Boosting and Compression for Class-Incremental Learning2022ECCV
BNS: Building Network Structures Dynamically for Continual Learning2021NeurIPS
DER: Dynamically Expandable Representation for Class Incremental Learning2021CVPR
Adaptive Aggregation Networks for Class-Incremental Learning2021CVPR
BatchEnsemble: an Alternative Approach to Efficient Ensemble and Lifelong Learning2020ICLR
Calibrating CNNs for Lifelong Learning2020NeurIPS
Continual Learning of a Mixed Sequence of Similar and Dissimilar Tasks2020NeurIPS
Compacting, Picking and Growing for Unforgetting Continual Learning2019NeurIPS
Superposition of many models into one2019NeurIPS
Reinforced Continual Learning2018NeurIPS
Progress & Compress: A scalable framework for continual learning2018ICML
Overcoming Catastrophic Forgetting with Hard Attention to the Task2018ICML
Lifelong Learning with Dynamically Expandable Networks 2018ICLR
PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning2018CVPR
Expert Gate: Lifelong Learning with a Network of Experts2017CVPR
Progressive Neural Networks2016Arxiv
Regularization-based Methods

<a href="#top">[Back to top]</a>

Regularization-based approaches avoid forgetting by penalizing updates of important parameters or distilling knowledge with previous model as a teacher.

Paper TitleYearConference/Journal
A Statistical Theory of Regularization-Based Continual Learning2024ICML
IMEX-Reg: Implicit-Explicit Regularization in the Function Space for Continual Learning2024TMLR
Contrastive Continual Learning with Importance Sampling and Prototype-Instance Relation Distillation2024AAAI
Elastic Feature Consolidation for Cold Start Exemplar-free Incremental Learning2024ICLR
Fine-Grained Knowledge Selection and Restoration for Non-Exemplar Class Incremental Learning2024AAAI
Prototype-Sample Relation Distillation: Towards Replay-Free Continual Learning2023ICML
Continual Learning via Sequential Function-Space Variational Inference2022ICML
Overcoming Catastrophic Forgetting in Incremental Object Detection via Elastic Response Distillation2022CVPR
Class-Incremental Learning by Knowledge Distillation with Adaptive Feature Consolidation2022CVPR
Class-Incremental Learning via Knowledge Amalgamation2022PKDD
Natural continual learning: success is a journey, not (just) a destination2021NeurIPS
Distilling Causal Effect of Data in Class-Incremental Learning2021CVPR
On Learning the Geodesic Path for Incremental Learning2021CVPR
CPR: Classifier-Projection Regularization for Continual Learning2021ICLR
Few-Shot Class-Incremental Learning via Relation Knowledge Distillation2021AAAI
Continual Learning with Node-Importance based Adaptive Group Sparse Regularization2020NeurIPS
PODNet: Pooled Outputs Distillation for Small-Tasks Incremental Learning2020ECCV
Topology-Preserving Class-Incremental Learning2020ECCV
Uncertainty-based Continual Learning with Adaptive Regularization2019NeurIPS
Learning a Unified Classifier Incrementally via Rebalancing2019CVPR
Learning Without Memorizing2019CVPR
Efficient Lifelong Learning with A-GEM2019ICLR
Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence2018ECCV
Lifelong Learning via Progressive Distillation and Retrospection2018ECCV
Memory Aware Synapses: Learning what (not) to forget2018ECCV
Overcoming catastrophic forgetting in neural networks2017Arxiv
Continual Learning Through Synaptic Intelligence2017ICML
Learning without Forgetting2017TPAMI
Subspace-based Methods

<a href="#top">[Back to top]</a>

Subspace-based methods perform CL in multiple disjoint subspaces to avoid interference between multiple tasks.

Paper TitleYearConference/Journal
Introducing Common Null Space of Gradients for Gradient Projection Methods in Continual Learning2024ACM MM
Improving Data-aware and Parameter-aware Robustness for Continual Learning2024Arxiv
Prompt Gradient Projection for Continual Learning2024ICLR
Hebbian Learning based Orthogonal Projection for Continual Learning of Spiking Neural Networks2024ICLR
Towards Continual Learning Desiderata via HSIC-Bottleneck Orthogonalization and Equiangular Embedding2024AAAI
Orthogonal Subspace Learning for Language Model Continual Learning2023EMNLP
Data Augmented Flatness-aware Gradient Projection for Continual Learning2023ICCV
Rethinking Gradient Projection Continual Learning: Stability / Plasticity Feature Space Decoupling2023CVPR
Building a Subspace of Policies for Scalable Continual Learning2023ICLR
Continual Learning with Scaled Gradient Projection2023AAAI
SketchOGD: Memory-Efficient Continual Learning2023Arxiv
Continual Learning through Networks Splitting and Merging with Dreaming-Meta Weighted Model Fusion2023Arxiv
Beyond Not-Forgetting: Continual Learning with Backward Knowledge Transfer2022NeurIPS
TRGP: Trust Region Gradient Projection for Continual Learning2022ICLR
Continual Learning with Recursive Gradient Optimization2022ICLR
Class Gradient Projection For Continual Learning2022MM
Balancing Stability and Plasticity through Advanced Null Space in Continual Learning2022ECCV
Adaptive Orthogonal Projection for Batch and Online Continual Learning2022AAAI
Natural continual learning: success is a journey, not (just) a destination2021NeurIPS
Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning2021NeurIPS
Gradient Projection Memory for Continual Learning2021ICLR
Training Networks in Null Space of Feature Covariance for Continual Learning2021CVPR
Generalisation Guarantees For Continual Learning With Orthogonal Gradient Descent2021Arxiv
Defeating Catastrophic Forgetting via Enhanced Orthogonal Weights Modification2021Arxiv
Continual Learning in Low-rank Orthogonal Subspaces2020NeurIPS
Orthogonal Gradient Descent for Continual Learning2020AISTATS
Generalisation Guarantees for Continual Learning with Orthogonal Gradient Descent2020Arxiv
Generative Feature Replay with Orthogonal Weight Modification for Continual Learning2020Arxiv
Continual Learning of Context-dependent Processing in Neural Networks2019Nature Machine Intelligence
Bayesian Methods

<a href="#top">[Back to top]</a>

Bayesian methods provide a principled probabilistic framework for addressing Forgetting.

Paper TitleYearConference/Journal
Learning to Continually Learn with the Bayesian Principle2024ICML
A Probabilistic Framework for Modular Continual Learning2023Arxiv
Online Continual Learning on Class Incremental Blurry Task Configuration with Anytime Inference2022ICLR
Continual Learning via Sequential Function-Space Variational Inference2022ICML
Generalized Variational Continual Learning2021ICLR
Variational Auto-Regressive Gaussian Processes for Continual Learning2021ICML
Bayesian Structural Adaptation for Continual Learning2021ICML
Continual Learning using a Bayesian Nonparametric Dictionary of Weight Factors2021AISTATS
Posterior Meta-Replay for Continual Learning2021NeurIPS
Natural continual learning: success is a journey, not (just) a destination2021NeurIPS
Continual Learning with Adaptive Weights (CLAW)2020ICLR
Uncertainty-guided Continual Learning with Bayesian Neural Networks2020ICLR
Functional Regularisation for Continual Learning with Gaussian Processes2020ICLR
Continual Deep Learning by Functional Regularisation of Memorable Past2020NeurIPS
Variational Continual Learning2018ICLR
Online Structured Laplace Approximations for Overcoming Catastrophic Forgetting2018NeurIPS
Overcoming Catastrophic Forgetting by Incremental Moment Matching2017NeurIPS

Task-free CL

<a href="#top">[Back to top]</a>

Task-free CL refers to a specific scenario that the learning system does not have access to any explicit task information.

Paper TitleYearConference/Journal
Task-Free Continual Generation and Representation Learning via Dynamic Expansionable Memory Cluster2024AAAI
Task-Free Dynamic Sparse Vision Transformer for Continual Learning2024AAAI
Doubly Perturbed Task-Free Continual Learning2024AAAI
Loss Decoupling for Task-Agnostic Continual Learning2023NeurIPS
Online Bias Correction for Task-Free Continual Learning2023ICLR
Task-Free Continual Learning via Online Discrepancy Distance Learning2022NeurIPS
Improving Task-free Continual Learning by Distributionally Robust Memory Evolution2022ICML
VariGrow: Variational architecture growing for task-agnostic continual learning based on Bayesian novelty2022ICML
Gradient-based Editing of Memory Examples for Online Task-free Continual Learning2021NeurIPS
Continuous Meta-Learning without Tasks2020NeurIPS
A Neural Dirichlet Process Mixture Model for Task-Free Continual Learning2020ICLR
Online Continual Learning with Maximally Interfered Retrieval2019NeurIPS
Gradient based sample selection for online continual learning2019NeurIPS
Efficient lifelong learning with A-GEM2019ICLR
Task-Free Continual Learning2019CVPR
Continual Learning with Tiny Episodic Memories2019Arxiv

Online CL

<a href="#top">[Back to top]</a>

In online CL, the learner is only allowed to process the data for each task once.

Paper TitleYearConference/Journal
Dealing with Synthetic Data Contamination in Online Continual Learning2024NeurIPS
Random Representations Outperform Online Continually Learned Representations2024NeurIPS
Forgetting, Ignorance or Myopia: Revisiting Key Challenges in Online Continual Learning2024NeurIPS
Mitigating Catastrophic Forgetting in Online Continual Learning by Modeling Previous Task Interrelations via Pareto Optimization2024ICML
ER-FSL: Experience Replay with Feature Subspace Learning for Online Continual Learning2024MM
Dual-Enhanced Coreset Selection with Class-wise Collaboration for Online Blurry Class Incremental Learning2024CVPR
Orchestrate Latent Expertise: Advancing Online Continual Learning with Multi-Level Supervision and Reverse Self-Distillation2024CVPR
Learning Equi-angular Representations for Online Continual Learning2024CVPR
Online Continual Learning For Interactive Instruction Following Agents2024ICLR
Online Continual Learning for Interactive Instruction Following Agents2024ICLR
Summarizing Stream Data for Memory-Constrained Online Continual Learning2024AAAI
Online Class-Incremental Learning For Real-World Food Image Classification2024WACV
Rapid Adaptation in Online Continual Learning: Are We Evaluating It Right?2023ICCV
CBA: Improving Online Continual Learning via Continual Bias Adaptor2023ICCV
Online Continual Learning on Hierarchical Label Expansion2023ICCV
New Insights for the Stability-Plasticity Dilemma in Online Continual Learning2023ICLR
Real-Time Evaluation in Online Continual Learning: A New Hope2023CVPR
PCR: Proxy-based Contrastive Replay for Online Class-Incremental Continual Learning2023CVPR
Dealing with Cross-Task Class Discrimination in Online Continual Learning2023CVPR
Online continual learning through mutual information maximization2022ICML
Online Coreset Selection for Rehearsal-based Continual Learning2022ICLR
New Insights on Reducing Abrupt Representation Change in Online Continual Learning2022ICLR
Online Continual Learning on Class Incremental Blurry Task Configuration with Anytime Inference2022ICLR
Information-theoretic Online Memory Selection for Continual Learning2022ICLR
Continual Normalization: Rethinking Batch Normalization for Online Continual Learning2022ICLR
Navigating Memory Construction by Global Pseudo-Task Simulation for Continual Learning2022NeurIPS
Not Just Selection, but Exploration: Online Class-Incremental Continual Learning via Dual View Consistency2022CVPR
Online Task-free Continual Learning with Dynamic Sparse Distributed Memory2022ECCV
Mitigating Forgetting in Online Continual Learning with Neuron Calibration2021NeurIPS
Online class-incremental continual learning with adversarial shapley value2021AAAI
Online Continual Learning with Natural Distribution Shifts: An Empirical Study with Visual Data2021ICCV
Continual Prototype Evolution: Learning Online from Non-Stationary Data Streams2021ICCV
La-MAML: Look-ahead Meta Learning for Continual Learning2020NeurIPS
Online Learned Continual Compression with Adaptive Quantization Modules2020ICML
Online Continual Learning under Extreme Memory Constraints2020ECCV
Online Continual Learning with Maximally Interfered Retrieval2019NeurIPS
Gradient based sample selection for online continual learning2019NeurIPS
On Tiny Episodic Memories in Continual LearningArxiv2019
Progress & Compress: A scalable framework for continual learning2018ICML

The presence of imbalanced data streams in CL (especially online CL) has drawn significant attention, primarily due to its prevalence in real-world application scenarios.

Paper TitleYearConference/Journal
Joint Input and Output Coordination for Class-Incremental Learning2024IJCAI
Imbalance Mitigation for Continual Learning via Knowledge Decoupling and Dual Enhanced Contrastive Learning2024TNNLS
Overcoming Recency Bias of Normalization Statistics in Continual Learning: Balance and Adaptation2023NeurIPS
Online Bias Correction for Task-Free Continual Learning2023ICLR
Information-theoretic Online Memory Selection for Continual Learning2022ICLR
SS-IL: Separated Softmax for Incremental Learning2021ICCV
Online Continual Learning from Imbalanced Data2020ICML
Maintaining Discrimination and Fairness in Class Incremental Learning2020CVPR
Semantic Drift Compensation for Class-Incremental Learning2020CVPR
Imbalanced Continual Learning with Partitioning Reservoir Sampling2020ECCV
GDumb: A Simple Approach that Questions Our Progress in Continual Learning2020ECCV
Large scale incremental learning2019CVPR
IL2M: Class Incremental Learning With Dual Memory2019ICCV
End-to-end incremental learning2018ECCV

Semi-supervised CL

<a href="#top">[Back to top]</a>

Semi-supervised CL is an extension of traditional CL that allows each task to incorporate unlabeled data as well.

Paper TitleYearConference/Journal
Continual Learning on a Diet: Learning from Sparsely Labeled Streams Under Constrained Computation2024ICLR
Dynamic Sub-graph Distillation for Robust Semi-supervised Continual Learning2024AAAI
Semi-supervised drifted stream learning with short lookback2022SIGKDD
Ordisco: Effective and efficient usage of incremental unlabeled data for semi-supervised continual learning2021CVPR
Memory-Efficient Semi-Supervised Continual Learning: The World is its Own Replay Buffer2021IJCNN
Overcoming Catastrophic Forgetting with Unlabeled Data in the Wild2019ICCV

Few-shot CL

<a href="#top">[Back to top]</a>

Few-shot CL refers to the scenario where a model needs to learn new tasks with only a limited number of labeled examples per task while retaining knowledge from previously encountered tasks.

Paper TitleYearConference/Journal
Wearable Sensor-Based Few-Shot Continual Learning on Hand Gestures for Motor-Impaired Individuals via Latent Embedding Exploitation2024IJCAI
A Bag of Tricks for Few-Shot Class-Incremental Learning2024Arxiv
Analogical Learning-Based Few-Shot Class-Incremental Learning2024IEEE TCSVT
Few-Shot Class-Incremental Learning via Training-Free Prototype Calibration2023NeurIPS
Few-shot Class-incremental Learning: A Survey2023Arxiv
Warping the Space: Weight Space Rotation for Class-Incremental Few-Shot Learning2023ICLR
Neural Collapse Inspired Feature-Classifier Alignment for Few-Shot Class Incremental Learning2023ICLR
Few-Shot Class-Incremental Learning by Sampling Multi-Phase Tasks2022TPAMI
Dynamic Support Network for Few-Shot Class Incremental Learning2022TPAMI
Subspace Regularizers for Few-Shot Class Incremental Learning2022ICLR
MetaFSCIL: A Meta-Learning Approach for Few-Shot Class Incremental Learning2022CVPR
Forward Compatible Few-Shot Class-Incremental Learning2022CVPR
Constrained Few-shot Class-incremental Learning2022CVPR
Few-Shot Class-Incremental Learning via Entropy-Regularized Data-Free Replay2022ECCV
MgSvF: Multi-Grained Slow vs. Fast Framework for Few-Shot Class-Incremental Learning2021TPAMI
Semantic-aware Knowledge Distillation for Few-Shot Class-Incremental Learning2021CVPR
Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning2021CVPR
Few-Shot Incremental Learning with Continually Evolved Classifiers2021CVPR
Synthesized Feature based Few-Shot Class-Incremental Learning on a Mixture of Subspaces2021ICCV
Few-Shot Lifelong Learning2021AAAI
Few-Shot Class-Incremental Learning via Relation Knowledge Distillation2021AAAI
Few-shot Continual Learning: a Brain-inspired Approach2021Arxiv
Few-Shot Class-Incremental Learning2020CVPR
<!-- | [Neural Collapse Terminus: A Unified Solution for Class Incremental Learning and Its Variants](https://arxiv.org/pdf/2308.01746.pdf) | 2023 | Arxiv -->

Unsupervised CL

<a href="#top">[Back to top]</a>

Unsupervised CL (UCL) assumes that only unlabeled data is provided to the CL learner.

Paper TitleYearConference/Journal
Class-Incremental Unsupervised Domain Adaptation via Pseudo-Label Distillation2024TIP
Plasticity-Optimized Complementary Networks for Unsupervised Continual2024WACV
Unsupervised Continual Learning in Streaming Environments2023TNNLS
Representational Continuity for Unsupervised Continual Learning2022ICLR
Probing Representation Forgetting in Supervised and Unsupervised Continual Learning2022CVPR
Unsupervised Continual Learning for Gradually Varying Domains2022CVPRW
Co2L: Contrastive Continual Learning2021ICCV
Unsupervised Progressive Learning and the STAM Architecture2021IJCAI
Continual Unsupervised Representation Learning2019NeurIPS

Theoretical Analysis

<a href="#top">[Back to top]</a>

Theory or analysis of continual learning

Paper TitleYearConference/Journal
Theoretical Insights into Overparameterized Models in Multi-Task and Replay-Based Continual Learning2024Arxiv
An analysis of best-practice strategies for replay and rehearsal in continual learning2024CVPRW
Provable Contrastive Continual Learning2024ICML
A Statistical Theory of Regularization-Based Continual Learning2024ICML
Efficient Continual Finite-Sum Minimization2024ICLR
Provable Contrastive Continual Learning2024ICLR
Understanding Forgetting in Continual Learning with Linear Regression: Overparameterized and Underparameterized Regimes2024ICLR
The Joint Effect of Task Similarity and Overparameterization on Catastrophic Forgetting -- An Analytical Model2024ICLR
A Unified and General Framework for Continual Learning2024ICLR
Continual Learning in the Presence of Spurious Correlations: Analyses and a Simple Baseline2024ICLR
On the Convergence of Continual Learning with Adaptive Methods2023UAI
Does Continual Learning Equally Forget All Parameters? 2023ICML
The Ideal Continual Learner: An Agent That Never Forgets2023ICML
Continual Learning in Linear Classification on Separable Data2023ICML
Theory on Forgetting and Generalization of Continual Learning2023ArXiv
A Theoretical Study on Solving Continual Learning2022NeurIPS
Learning Curves for Continual Learning in Neural Networks: Self-Knowledge Transfer and Forgetting2022ICLR
Continual Learning in the Teacher-Student Setup: Impact of Task Similarity2022ICML
Formalizing the Generalization-Forgetting Trade-off in Continual Learning2021NeurIPS
A PAC-Bayesian Bound for Lifelong Learning2014ICML

Forgetting in Foundation Models

<a href="#top">[Back to top]</a>

Foundation models are large machine learning models trained on a vast quantity of data at scale, such that they can be adapted to a wide range of downstream tasks.

Links: Forgetting in Fine-Tuning Foundation Models | Forgetting in One-Epoch Pre-training | CL in Foundation Model

Forgetting in Fine-Tuning Foundation Models

<a href="#top">[Back to top]</a>

When fine-tuning a foundation model, there is a tendency to forget the pre-trained knowledge, resulting in sub-optimal performance on downstream tasks.

Paper TitleYearConference/Journal
A Practitioner’s Guide to Continual Multimodal Pretraining2024Arxiv
SLCA++: Unleash the Power of Sequential Fine-tuning for Continual Learning with Pre-training2024Arxiv
MoFO: Momentum-Filtered Optimizer for Mitigating Forgetting in LLM Fine-Tuning2024Arxiv
Towards Effective and Efficient Continual Pre-training of Large Language Models2024Arxiv
Revisiting Catastrophic Forgetting in Large Language Model Tuning2024Arxiv
D-CPT Law: Domain-specific Continual Pre-Training Scaling Law for Large Language Models2024Arxiv
Dissecting learning and forgetting in language model finetuning2024ICLR
Understanding Catastrophic Forgetting in Language Models via Implicit Inference2024ICLR
Two-stage LLM Fine-tuning with Less Specialization and More Generalization2024ICLR
What Will My Model Forget? Forecasting Forgotten Examples in Language Model Refinement2024Arxiv
Scaling Laws for Forgetting When Fine-Tuning Large Language Models2024Arxiv
TOFU: A Task of Fictitious Unlearning for LLMs2024Arxiv
Self-regulating Prompts: Foundational Model Adaptation without Forgetting2023ICCV
Speciality vs Generality: An Empirical Study on Catastrophic Forgetting in Fine-tuning Foundation Models2023Arxiv
Continual Pre-Training of Large Language Models: How to (re)warm your model?2023ICMLW
Improving Gender Fairness of Pre-Trained Language Models without Catastrophic Forgetting2023ACL
On The Role of Forgetting in Fine-Tuning Reinforcement Learning Models2023ICLRW
Retentive or Forgetful? Diving into the Knowledge Memorizing Mechanism of Language Models2023Arxiv
Reinforcement Learning with Action-Free Pre-Training from Videos2022ICML
Video PreTraining (VPT): Learning to Act by Watching Unlabeled Online Videos2022NeurIPS
On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting2022NeurIPS
How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial Robustness?2021NeurIPS
Mixout: Effective Regularization to Finetune Large-scale Pretrained Language Models2020ICLR
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting2020EMNLP
Universal Language Model Fine-tuning for Text Classification2018ACL

Forgetting in One-Epoch Pre-training

<a href="#top">[Back to top]</a>

Foundation models often undergo training on a dataset for a single pass. As a result, the earlier examples encountered during pre-training may be overwritten or forgotten by the model more quickly than the later examples.

Paper TitleYearConference/Journal
Exploring Forgetting in Large Language Model Pre-Training2024Arxiv
Measuring Forgetting of Memorized Training Examples2023ICLR
Quantifying Memorization Across Neural Language Models2023ICLR
Analyzing leakage of personally identifiable information in language models2023S&P
How Well Does Self-Supervised Pre-Training Perform with Streaming Data?2022ICLR
The challenges of continuous self-supervised learning2022ECCV
Continual contrastive learning for image classification2022ICME

CL in Foundation Model

<a href="#top">[Back to top]</a>

By leveraging the powerful feature extraction capabilities of foundation models, researchers have been able to explore new avenues for advancing continual learning techniques.

Paper TitleYearConference/Journal
Mitigating Catastrophic Forgetting in Large Language Models with Self-Synthesized Rehearsal2024ACL
Mixture of Experts Meets Prompt-Based Continual Learning2024NeurIPS
SAFE: Slow and Fast Parameter-Efficient Tuning for Continual Learning with Pre-Trained Mode2024NeurIPS
Incremental Learning of Retrievable Skills For Efficient Continual Task Adaptation2024NeurIPS
Vector Quantization Prompting for Continual Learning2024NeurIPS
Dual Low-Rank Adaptation for Continual Learning with Pre-Trained Models2024Arxiv
Is Parameter Collision Hindering Continual Learning in LLMs2024Arxiv
Does RoBERTa Perform Better than BERT in Continual Learning: An Attention Sink Perspective2024COLM
Dual Consolidation for Pre-Trained Model-Based Domain-Incremental Learning2024Arxiv
ICL-TSVD: Bridging Theory and Practice in Continual Learning with Pre-trained Models2024Arxiv
Adaptive Adapter Routing for Long-Tailed Class-Incremental Learning2024Machine Learning Journal
CoIN: A Benchmark of Continual Instruction tuNing for Multimodel Large Language Model2024Arxiv
Continual Instruction Tuning for Large Multimodal Models2024Arxiv
Parameter-Efficient Fine-Tuning for Continual Learning: A Neural Tangent Kernel Perspective2024Arxiv
Class-Incremental Learning with CLIP: Adaptive Representation Adjustment and Parameter Fusion2024ECCV
Model Tailor: Mitigating Catastrophic Forgetting in Multi-modal Large Language Models2024ICML
One Size Fits All for Semantic Shifts: Adaptive Prompt Tuning for Continual Learning2024ICML
HiDe-PET: Continual Learning via Hierarchical Decomposition of Parameter-Efficient Tuning2024Arxiv
Domain Adaptation of Llama3-70B-Instruct through Continual Pre-Training and Model Merging: A Comprehensive Evaluation2024Arxiv
Mitigate Negative Transfer with Similarity Heuristic Lifelong Prompt Tuning2024ACL
Reflecting on the State of Rehearsal-free Continual Learning with Pretrained Models2024CoLLAs
Choice of PEFT Technique in Continual Learning: Prompt Tuning is Not All You Need2024Arxiv
Disperse-Then-Merge: Pushing the Limits of Instruction Tuning via Alignment Tax Reduction2024ACL
Gradient Projection For Parameter-Efficient Continual Learning2024Arxiv
Continual Learning of Large Language Models: A Comprehensive Survey2024Arxiv
Prompt Customization for Continual Learning2024MM
Dynamically Anchored Prompting for Task-Imbalanced Continual Learning2024IJCAI
InfLoRA: Interference-Free Low-Rank Adaptation for Continual Learning2024CVPR
Semantically-Shifted Incremental Adapter-Tuning is A Continual ViTransformer2024CVPR
Evolving Parameterized Prompt Memory for Continual Learning2024AAAI
Expandable Subspace Ensemble for Pre-Trained Model-Based Class-Incremental Learning2024CVPR
Consistent Prompting for Rehearsal-Free Continual Learning2024CVPR
Interactive Continual Learning: Fast and Slow Thinking2024CVPR
HOP to the Next Tasks and Domains for Continual Learning in NLP2024AAAI
OVOR: OnePrompt with Virtual Outlier Regularization for Rehearsal-Free Class-Incremental Learning2024ICLR
Continual Learning for Large Language Models: A Survey2024Arxiv
Continual Learning with Pre-Trained Models: A Survey2024Arxiv
INCPrompt: Task-Aware incremental Prompting for Rehearsal-Free Class-incremental Learning2024ICASSP
P2DT: Mitigating Forgetting in task-incremental Learning with progressive prompt Decision Transformer2024ICASSP
Scalable Language Model with Generalized Continual Learning2024ICLR
Prompt Gradient Projection for Continual Learning2024ICLR
TiC-CLIP: Continual Training of CLIP Models2024ICLR
Hierarchical Prompts for Rehearsal-free Continual Learning2024Arxiv
KOPPA: Improving Prompt-based Continual Learning with Key-Query Orthogonal Projection and Prototype-based One-Versus-All2023Arxiv
RanPAC: Random Projections and Pre-trained Models for Continual Learning2023NeurIPS
Hierarchical Decomposition of Prompt-Based Continual Learning: Rethinking Obscured Sub-optimality2023NeurIPS
A Unified Continual Learning Framework with General Parameter-Efficient Tuning2023ICCV
Generating Instance-level Prompts for Rehearsal-free Continual Learning2023ICCV
Introducing Language Guidance in Prompt-based Continual Learning2023ICCV
Generating Instance-level Prompts for Rehearsal-free Continual Learning2023ICCV
Space-time Prompting for Video Class-incremental Learning2023ICCV
When Prompt-based Incremental Learning Does Not Meet Strong Pretraining2023ICCV
Online Class Incremental Learning on Stochastic Blurry Task Boundary via Mask and Visual Prompt Tuning2023ICCV
SLCA: Slow Learner with Classifier Alignment for Continual Learning on a Pre-trained Model2023ICCV
Progressive Prompts: Continual Learning for Language Models2023ICLR
Continual Pre-training of Language Models2023ICLR
Continual Learning of Language Models2023ICLR
CODA-Prompt: COntinual Decomposed Attention-based Prompting for Rehearsal-Free Continual Learning2023CVPR
PIVOT: Prompting for Video Continual Learning2023CVPR
Do Pre-trained Models Benefit Equally in Continual Learning?2023WACV
Revisiting Class-Incremental Learning with Pre-Trained Models: Generalizability and Adaptivity are All You Need2023Arxiv
First Session Adaptation: A Strong Replay-Free Baseline for Class-Incremental Learning2023Arxiv
Memory Efficient Continual Learning with Transformers2022NeurIPS
S-Prompts Learning with Pre-trained Transformers: An Occam’s Razor for Domain Incremental Learning2022NeurIPS
Pretrained Language Model in Continual Learning: A Comparative Study2022ICLR
Effect of scale on catastrophic forgetting in neural networks2022ICLR
LFPT5: A Unified Framework for Lifelong Few-shot Language Learning Based on Prompt Tuning of T52022ICLR
Learning to Prompt for Continual Learning2022CVPR
Class-Incremental Learning with Strong Pre-trained Models2022CVPR
DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning2022ECCV
ELLE: Efficient Lifelong Pre-training for Emerging Data2022ACL
Fine-tuned Language Models are Continual Learners2022EMNLP
Continual Training of Language Models for Few-Shot Learning2022EMNLP
Continual Learning with Foundation Models: An Empirical Study of Latent Replay2022Conference on Lifelong Learning Agents
Rational LAMOL: A Rationale-Based Lifelong Learning Framework2021ACL
Achieving Forgetting Prevention and Knowledge Transfer in Continual Learning2021NeurIPS
An Empirical Investigation of the Role of Pre-training in Lifelong Learning2021Arxiv
LAnguage MOdeling for Lifelong Language Learning2020ICLR

Forgetting in Domain Adaptation

<a href="#top">[Back to top]</a>

The goal of domain adaptation is to transfer the knowledge from a source domain to a target domain.

Paper TitleYearConference/Journal
Towards Cross-Domain Continual Learning2024ICDE
Continual Source-Free Unsupervised Domain Adaptation2023International Conference on Image Analysis and Processing
CoSDA: Continual Source-Free Domain Adaptation2023Arxiv
Lifelong Domain Adaptation via Consolidated Internal Distribution2022NeurIPS
Online Domain Adaptation for Semantic Segmentation in Ever-Changing Conditions2022ECCV
FRIDA -- Generative Feature Replay for Incremental Domain Adaptation2022CVIU
Unsupervised Continual Learning for Gradually Varying Domains2022CVPRW
Continual Adaptation of Visual Representations via Domain Randomization and Meta-learning2021CVPR
Gradient Regularized Contrastive Learning for Continual Domain Adaptation2021AAAI
Learning to Adapt to Evolving Domains2020NeurIPS
AdaGraph: Unifying Predictive and Continuous Domain Adaptation through Graphs2019CVPR
ACE: Adapting to Changing Environments for Semantic Segmentation2019ICCV
Adapting to Continuously Shifting Domains2018ICLRW

Forgetting in Test-Time Adaptation

<a href="#top">[Back to top]</a>

<!-- <u>[Click back to content outline](#framework)</u> -->

Test time adaptation (TTA) refers to the process of adapting a pre-trained model on-the-fly to unlabeled test data during inference or testing.

Paper TitleYearConference/Journal
PCoTTA: Continual Test-Time Adaptation for Multi-Task Point Cloud Understanding2024NeurIPS
Adaptive Cascading Network for Continual Test-Time Adaptation2024CIKM
Controllable Continual Test-Time Adaptation2024Arxiv
ViDA: Homeostatic Visual Domain Adapter for Continual Test Time Adaptation2024ICLR
Continual Momentum Filtering on Parameter Space for Online Test-time Adaptation2024ICLR
A Comprehensive Survey on Test-Time Adaptation under Distribution Shifts2023Arxiv
MECTA: Memory-Economic Continual Test-Time Model Adaptation2023ICLR
Decorate the Newcomers: Visual Domain Prompt for Continual Test Time Adaptation2023AAAI (Outstanding Student Paper Award)
Robust Mean Teacher for Continual and Gradual Test-Time Adaptation2023CVPR
A Probabilistic Framework for Lifelong Test-Time Adaptation2023CVPR
EcoTTA: Memory-Efficient Continual Test-time Adaptation via Self-distilled Regularization2023CVPR
AUTO: Adaptive Outlier Optimization for Online Test-Time OOD Detection2023Arxiv
Efficient Test-Time Model Adaptation without Forgetting2022ICML
MEMO: Test time robustness via adaptation and augmentation2022NeurIPS
Continual Test-Time Domain Adaptation2022CVPR
Improving test-time adaptation via shift-agnostic weight regularization and nearest source prototypes2022ECCV
Tent: Fully Test-Time Adaptation by Entropy Minimization2021ICLR

Forgetting in Meta-Learning

<a href="#top">[Back to top]</a>

Meta-learning, also known as learning to learn, focuses on developing algorithms and models that can learn from previous learning experiences to improve their ability to learn new tasks or adapt to new domains more efficiently and effectively.

Links: Incremental Few-Shot Learning | Continual Meta-Learning

Incremental Few-Shot Learning

<a href="#top">[Back to top]</a>

Incremental few-shot learning (IFSL) focuses on the challenge of learning new categories with limited labeled data while retaining knowledge about previously learned categories.

Paper TitleYearConference/Journal
Few-Shot Class-Incremental Learning via Training-Free Prototype Calibration2023NeurIPS
Constrained Few-shot Class-incremental Learning2022CVPR
Meta-Learning with Less Forgetting on Large-Scale Non-Stationary Task Distributions2022ECCV
Overcoming Catastrophic Forgetting in Incremental Few-Shot Learning by Finding Flat Minima2021NeurIPS
Incremental Few-shot Learning via Vector Quantization in Deep Embedded Space2021ICLR
XtarNet: Learning to Extract Task-Adaptive Representation for Incremental Few-Shot Learning2020ICML
Incremental Few-Shot Learning with Attention Attractor Networks2019NeurIPS
Dynamic Few-Shot Visual Learning without Forgetting2018CVPR

Continual Meta-Learning

<a href="#top">[Back to top]</a>

The goal of continual meta-learning (CML) is to address the challenge of forgetting in non-stationary task distributions.

Paper TitleYearConference/Journal
Meta Continual Learning Revisited: Implicitly Enhancing Online Hessian Approximation via Variance Reduction2024ICLR
Recasting Continual Learning as Sequence Modeling2023NeurIPS
Adaptive Compositional Continual Meta-Learning2023ICML
Learning to Learn and Remember Super Long Multi-Domain Task Sequence2022CVPR
Meta-Learning with Less Forgetting on Large-Scale Non-Stationary Task Distributions2022ECCV
Variational Continual Bayesian Meta-Learning2021NeurIPS
Meta Learning on a Sequence of Imbalanced Domains with Difficulty Awareness2021ICCV
Addressing Catastrophic Forgetting in Few-Shot Problems2020ICML
Continuous meta-learning without tasks2020NeurIPS
Reconciling meta-learning and continual learning with online mixtures of tasks2019NeurIPS
Fast Context Adaptation via Meta-Learning2019ICML
Online meta-learning2019ICML

Forgetting in Generative Models

<a href="#top">[Back to top]</a>

The goal of a generative model is to learn a generator that can generate samples from a target distribution.

Links: GAN Training is a Continual Learning Problem | Lifelong Learning of Generative Models

GAN Training is a Continual Learning Problem

<a href="#top">[Back to top]</a>

Treating GAN training as a continual learning problem.

Paper TitleYearConference/Journal
Learning to Retain while Acquiring: Combating Distribution-Shift in Adversarial Data-Free Knowledge Distillation2023CVPR
Momentum Adversarial Distillation: Handling Large Distribution Shifts in Data-Free Knowledge Distillation2022NeurIPS
Robust and Resource-Efficient Data-Free Knowledge Distillation by Generative Pseudo Replay2022AAAI
Preventing Catastrophic Forgetting and Distribution Mismatch in Knowledge Distillation via Synthetic Data2022WACV
On Catastrophic Forgetting and Mode Collapse in Generative Adversarial Networks2020IJCNN
Generative adversarial network training is a continual learning problem2018ArXiv

Lifelong Learning of Generative Models

<a href="#top">[Back to top]</a>

The goal is to develop generative models that can continually generate high-quality samples for both new and previously encountered tasks.

Paper TitleYearConference/Journal
KFC: Knowledge Reconstruction and Feedback Consolidation Enable Efficient and Effective Continual Generative Learning2024ICLR
The Curse of Recursion: Training on Generated Data Makes Models Forget2023Arxiv
Forget-Me-Not: Learning to Forget in Text-to-Image Diffusion Models2023Arxiv
Selective Amnesia: A Continual Learning Approach to Forgetting in Deep Generative Models2023Arxiv
Lifelong Generative Modelling Using Dynamic Expansion Graph Model2022AAAI
Continual Variational Autoencoder Learning via Online Cooperative Memorization2022ECCV
Hyper-LifelongGAN: Scalable Lifelong Learning for Image Conditioned Generation2021CVPR
Lifelong Twin Generative Adversarial Networks2021ICIP
Lifelong Mixture of Variational Autoencoders2021TNNLS
Lifelong Generative Modeling2020Neurocomputing
GAN Memory with No Forgetting2020NeurIPS
Lifelong GAN: Continual Learning for Conditional Image Generation2019ICCV

Forgetting in Reinforcement Learning

<a href="#top">[Back to top]</a>

Reinforcement learning is a machine learning technique that allows an agent to learn how to behave in an environment by trial and error, through rewards and punishments.

Paper TitleYearConference/Journal
Data Augmentation for Continual RL via Adversarial Gradient Episodic Memory2024Arxiv
Fine-tuning Reinforcement Learning Models is Secretly a Forgetting Mitigation Problem2024Arxiv
Hierarchical Continual Reinforcement Learning via Large Language Model2024Arxiv
Augmenting Replay in World Models for Continual Reinforcement Learning2024Arxiv
CPPO: Continual Learning for Reinforcement Learning with Human Feedback2024ICLR
Prediction and Control in Continual Reinforcement Learning2023NeurIPS
Replay-enhanced Continual Reinforcement Learning2023TMLR
A Definition of Continual Reinforcement Learning2023Arxiv
Continual Task Allocation in Meta-Policy Network via Sparse Prompting2023ICML
Building a Subspace of Policies for Scalable Continual Learning2023ICLR
Continual Model-based Reinforcement Learning for Data Efficient Wireless Network Optimisation2023ECML
Modular Lifelong Reinforcement Learning via Neural Composition2022ICLR
Disentangling Transfer in Continual Reinforcement Learning2022NeurIPS
Towards continual reinforcement learning: A review and perspectives2022Journal of Artificial Intelligence Research
Reinforced continual learning for graphs2022CIKM
Model-Free Generative Replay for Lifelong Reinforcement Learning: Application to Starcraft-22022Conference on Lifelong Learning Agents
Transient Non-stationarity and Generalisation in Deep Reinforcement Learning2021ICLR
Sharing Less is More: Lifelong Learning in Deep Networks with Selective Layer Transfer2021ICML
Pseudo-rehearsal: Achieving deep reinforcement learning without catastrophic forgetting2021Neurocomputing
Lifelong Policy Gradient Learning of Factored Policies for Faster Training Without Forgetting2020NeurIPS
Policy Consolidation for Continual Reinforcement Learning2019ICML
Exploiting Hierarchy for Learning and Transfer in KL-regularized RL2019Arxiv
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks2017ICML
Progressive neural networks2016Arxiv
Learning a synaptic learning rule1991IJCNN

Forgetting in Federated Learning

<a href="#top">[Back to top]</a>

Federated learning (FL) is a decentralized machine learning approach where the training process takes place on local devices or edge servers instead of a centralized server.

Links: Forgetting Due to Non-IID Data in FL | Federated Continual Learning

Forgetting Due to Non-IID Data in FL

<a href="#top">[Back to top]</a>

This branch pertains to the forgetting problem caused by the inherent non-IID (not identically and independently distributed) data among different clients participating in FL.

Paper TitleYearConference/Journal
Flashback: Understanding and Mitigating Forgetting in Federated Learning2024Arxiv
How to Forget Clients in Federated Online Learning to Rank?2024ECIR
GradMA: A Gradient-Memory-based Accelerated Federated Learning with Alleviated Catastrophic Forgetting2023CVPR
Acceleration of Federated Learning with Alleviated Forgetting in Local Training2022ICLR
Preservation of the Global Knowledge by Not-True Distillation in Federated Learning2022NeurIPS
Learn from Others and Be Yourself in Heterogeneous Federated Learning2022CVPR
Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning2022CVPR
Model-Contrastive Federated Learning2021CVPR
SCAFFOLD: Stochastic Controlled Averaging for Federated Learning2020ICML
Overcoming Forgetting in Federated Learning on Non-IID Data2019NeurIPSW

Federated Continual Learning

<a href="#top">[Back to top]</a>

This branch addresses the issue of continual learning within each individual client in the federated learning process, which results in forgetting at the overall FL level.

Paper TitleYearConference/Journal
Diffusion-Driven Data Replay: A Novel Approach to Combat Forgetting in Federated Class Continual Learning2024ECCV
PIP: Prototypes-Injected Prompt for Federated Class Incremental2024CIKM
Personalized Federated Continual Learning via Multi-granularity Prompt2024KDD
Federated Continual Learning via Prompt-based Dual Knowledge Transfer2024ICML
Text-Enhanced Data-free Approach for Federated Class-Incremental Learning2024CVPR
Federated Continual Learning via Knowledge Fusion: A Survey2024TKDE
Accurate Forgetting for Heterogeneous Federated Continual Learning2024ICLR
Federated Orthogonal Training: Mitigating Global Catastrophic Forgetting in Continual Federated Learning2024ICLR
A Data-Free Approach to Mitigate Catastrophic Forgetting in Federated Class Incremental Learning for Vision Tasks2023NeurIPS
Federated Continual Learning via Knowledge Fusion: A Survey2023Arxiv
A Data-Free Approach to Mitigate Catastrophic Forgetting in Federated Class Incremental Learning for Vision Tasks2023NeurIPS
TARGET: Federated Class-Continual Learning via Exemplar-Free Distillation2023ICCV
FedET: A Communication-Efficient Federated Class-Incremental Learning Framework Based on Enhanced Transformer2023IJCAI
Better Generative Replay for Continual Federated Learning2023ICLR
Don’t Memorize; Mimic The Past: Federated Class Incremental Learning Without Episodic Memory2023ICMLW
Addressing Catastrophic Forgetting in Federated Class-Continual Learning2023Arxiv
Federated Class-Incremental Learning2022CVPR
Continual Federated Learning Based on Knowledge Distillation2022IJCAI
Federated Continual Learning with Weighted Inter-client Transfer2021ICML
A distillation-based approach integrating continual learning and federated learning for pervasive services2021Arxiv

Beneficial Forgetting

<a href="#top">[Back to top]</a> Beneficial forgetting arises when the model contains private information that could lead to privacy breaches or when irrelevant information hinders the learning of new tasks. In these situations, forgetting becomes desirable as it helps protect privacy and facilitate efficient learning by discarding unnecessary information.

Problem SettingGoal
Mitigate Overfittingmitigate memorization of training data through selective forgetting
Debias and Forget Irrelevant Informationforget biased information to achieve better performance or remove irrelevant information to learn new tasks
Machine Unlearningforget some specified training data to protect user privacy

Links: <u>Combat Overfitting Through Forgetting</u> | <u>Learning New Knowledge Through Forgetting Previous Knowledge</u> | <u>Machine Unlearning</u>

Forgetting Irrelevant Information to Achieve Better Performance

<a href="#top">[Back to top]</a>

Combat Overfitting Through Forgetting

<a href="#top">[Back to top]</a>

Overfitting in neural networks occurs when the model excessively memorizes the training data, leading to poor generalization. To address overfitting, it is necessary to selectively forget irrelevant or noisy information.

Paper TitleYearConference/Journal
"Forgetting" in Machine Learning and Beyond: A Survey2024Arxiv
The Effectiveness of Random Forgetting for Robust Generalization2024ICLR
Sample-Efficient Reinforcement Learning by Breaking the Replay Ratio Barrier2023ICLR
The Primacy Bias in Deep Reinforcement Learning2022ICML
The Impact of Reinitialization on Generalization in Convolutional Neural Networks2021Arxiv
Learning with Selective Forgetting2021IJCAI
SIGUA: Forgetting May Make Learning with Noisy Labels More Robust2020ICML
Invariant Representations through Adversarial Forgetting2020AAAI
Forget a Bit to Learn Better: Soft Forgetting for CTC-based Automatic Speech Recognition2019Interspeech

Learning New Knowledge Through Forgetting Previous Knowledge

<a href="#top">[Back to top]</a>

"Learning to forget" suggests that not all previously acquired prior knowledge is helpful for learning new tasks.

Paper TitleYearConference/Journal
"Forgetting" in Machine Learning and Beyond: A Survey2024Arxiv
Improving Language Plasticity via Pretraining with Active Forgetting2023NeurIPS
ReFactor GNNs: Revisiting Factorisation-based Models from a Message-Passing Perspective2022NeurIPS
Fortuitous Forgetting in Connectionist Networks2022ICLR
Skin Deep Unlearning: Artefact and Instrument Debiasing in the Context of Melanoma Classification2022ICML
Near-Optimal Task Selection for Meta-Learning with Mutual Information and Online Variational Bayesian Unlearning2022AISTATS
AFEC: Active Forgetting of Negative Transfer in Continual Learning2021NeurIPS
Knowledge Evolution in Neural Networks2021CVPR
Active Forgetting: Adaptation of Memory by Prefrontal Control2021Annual Review of Psychology
Learning to Forget for Meta-Learning2020CVPR
The Forgotten Part of Memory2019Nature
Learning Not to Learn: Training Deep Neural Networks with Biased Data2019CVPR
Inhibiting your native language: the role of retrieval-induced forgetting during second-language acquisition2007Psychological Science

Machine Unlearning

<a href="#top">[Back to top]</a>

Machine unlearning, a recent area of research, addresses the need to forget previously learned training data in order to protect user data privacy.

Paper TitleYearConference/Journal
Unlearning during Learning: An Efficient Federated Machine Unlearning Method2024IJCAI
Label-Agnostic Forgetting: A Supervision-Free Unlearning in Deep Models2024ICLR
Machine Unlearning: A Survey2023ACM Computing Surveys
Deep Unlearning via Randomized Conditionally Independent Hessians2022CVPR
Eternal Sunshine of the Spotless Net: Selective Forgetting in Deep Networks2022CVPR
PUMA: Performance Unchanged Model Augmentation for Training Data Removal2022AAAI
ARCANE: An Efficient Architecture for Exact Machine Unlearning2022IJCAI
Learn to Forget: Machine Unlearning via Neuron Masking2022IEEE TDSC
Backdoor Defense with Machine Unlearning2022IEEE INFOCOM
Markov Chain Monte Carlo-Based Machine Unlearning: Unlearning What Needs to be Forgotten2022ASIA CCS
Machine Unlearning2021SSP
Remember What You Want to Forget: Algorithms for Machine Unlearning2021NeurIPS
Machine Unlearning via Algorithmic Stability2021COLT
Variational Bayesian Unlearning2020NeurIPS
Rapid retraining of machine learning models2020ICML
Certified Data Removal from Machine Learning Models2020ICML
Making AI Forget You: Data Deletion in Machine Learning2019NeurIPS
Lifelong Anomaly Detection Through Unlearning2019CCS
The EU Proposal for a General Data Protection Regulation and the Roots of the ‘Right to Be Forgotten’2013Computer Law & Security Review

Star History

Star History Chart


Contact

We welcome all researchers to contribute to this repository 'forgetting in deep learning'.

Email: wangzhenyineu@gmail.com | ennengyang@stumail.neu.edu.cn