Learning Continually from Low-shot Data Stream
@article{Le2019LearningCF, title={Learning Continually from Low-shot Data Stream}, author={Canyu Le and Xihan Wei and Biao Wang and L. Zhang}, journal={ArXiv}, year={2019}, volume={abs/1908.10223} }
While deep learning has achieved remarkable results on various applications, it is usually data hungry and struggles to learn over non-stationary data stream. To solve these two limits, the deep learning model should not only be able to learn from a few of data, but also incrementally learn new concepts from data stream over time without forgetting the previous knowledge. Limited literature simultaneously address both problems. In this work, we propose a novel approach, MetaCL, which enables…
Figures and Tables from this paper
One Citation
Continual Few-Shot Learning for Text Classification
- Computer ScienceEMNLP
- 2021
This work proposes a continual few-shot learning (CFL) task, in which a system is challenged with a difficult phenomenon and asked to learn to correct mistakes with only a few training examples.
33 References
Matching Networks for One Shot Learning
- Computer ScienceNIPS
- 2016
This work employs ideas from metric learning based on deep neural features and from recent advances that augment neural networks with external memories to learn a network that maps a small labelled support set and an unlabelled example to its label, obviating the need for fine-tuning to adapt to new class types.
Meta-Learning with Memory-Augmented Neural Networks
- Computer ScienceICML
- 2016
The ability of a memory-augmented neural network to rapidly assimilate new data, and leverage this data to make accurate predictions after only a few samples is demonstrated.
Learning to Compare: Relation Network for Few-Shot Learning
- Computer Science2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition
- 2018
A conceptually simple, flexible, and general framework for few-shot learning, where a classifier must learn to recognise new classes given only few examples from each, which is easily extended to zero- shot learning.
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
- Computer ScienceICML
- 2017
We propose an algorithm for meta-learning that is model-agnostic, in the sense that it is compatible with any model trained with gradient descent and applicable to a variety of different learning…
Dynamic Few-Shot Visual Learning Without Forgetting
- Computer Science2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition
- 2018
This work proposes to extend an object recognition system with an attention based few-shot classification weight generator, and to redesign the classifier of a ConvNet model as the cosine similarity function between feature representations and classification weight vectors.
Learning without Forgetting
- Computer ScienceIEEE Transactions on Pattern Analysis and Machine Intelligence
- 2018
This work proposes the Learning without Forgetting method, which uses only new task data to train the network while preserving the original capabilities, and performs favorably compared to commonly used feature extraction and fine-tuning adaption techniques.
Lifelong Learning with Dynamically Expandable Networks
- Computer ScienceICLR
- 2018
The obtained network fine-tuned on all tasks obtained significantly better performance over the batch models, which shows that it can be used to estimate the optimal network structure even when all tasks are available in the first place.
Overcoming Catastrophic Forgetting by Incremental Moment Matching
- Computer ScienceNIPS
- 2017
IMM incrementally matches the moment of the posterior distribution of the neural network which is trained on the first and the second task, respectively to make the search space of posterior parameter smooth.
Overcoming catastrophic forgetting in neural networks
- Computer ScienceProceedings of the National Academy of Sciences
- 2017
It is shown that it is possible to overcome the limitation of connectionist models and train networks that can maintain expertise on tasks that they have not experienced for a long time and selectively slowing down learning on the weights important for previous tasks.
Efficient Lifelong Learning with A-GEM
- Computer ScienceICLR
- 2019
An improved version of GEM is proposed, dubbed Averaged GEM (A-GEM), which enjoys the same or even better performance as GEM, while being almost as computationally and memory efficient as EWC and other regularization-based methods.