Corpus ID: 8909022

Matching Networks for One Shot Learning

@inproceedings{Vinyals2016MatchingNF,
  title={Matching Networks for One Shot Learning},
  author={Oriol Vinyals and Charles Blundell and Timothy P. Lillicrap and Koray Kavukcuoglu and Daan Wierstra},
  booktitle={NIPS},
  year={2016}
}
Learning from a few examples remains a key challenge in machine learning. [...] Key Method We then define one-shot learning problems on vision (using Omniglot, ImageNet) and language tasks. Our algorithm improves one-shot accuracy on ImageNet from 87.6% to 93.2% and from 88.0% to 93.8% on Omniglot compared to competing approaches. We also demonstrate the usefulness of the same model on language modeling by introducing a one-shot task on the Penn Treebank.Expand
A Neural Framework for Low-Shot Learning
There has been a growing interest in developing machine learning models that are capable of low-shot learning, the machine learning problem of learning from little data. Progress on low-shot learningExpand
Memory Matching Networks for One-Shot Image Recognition
TLDR
Memory Matching Networks (MM-Net) is presented - a novel deep architecture that explores the training procedure, following the philosophy that training and test conditions must match, and improves one-shot accuracy on Omniglot and miniImageNet. Expand
Adaptive Learning Knowledge Networks for Few-Shot Learning
TLDR
This paper proposes a new framework called Adaptive Learning Knowledge Networks (ALKN) for few-shot learning, which learns the knowledge of different classes from the features of labeled samples and store the learned knowledge into memory which will be dynamically updated during the learning process. Expand
A Dual Attention Network with Semantic Embedding for Few-Shot Learning
TLDR
This work proposes a novel meta-learning method for few-shot classification based on two simple attention mechanisms: one is a spatial attention to localize relevant object regions and the other is a task attention to select similar training data for label prediction. Expand
Local Contrast Learning
TLDR
This work proposed a novel deep learning approach named Local Contrast Learning (LCL) based on the key insight about a human cognitive behavior that human recognizes the objects in a specific context by contrasting the Objects in the context or in her/his memory. Expand
Learning to Support: Exploiting Structure Information in Support Sets for One-Shot Learning
TLDR
A novel meta-learner is proposed which shows state-of-the-art performance on common benchmarks for one/few shot classification and can be viewed as an approximation to an ensemble, which saves the factor of in training and test times and in the storage of the final model. Expand
Multi-attention Network for One Shot Learning
TLDR
A neural network architecture is designed which takes the semantic embedding of the class tag to generate attention maps and uses those attention maps to create the image features for one-shot learning. Expand
Learning from Few Samples: A Survey
TLDR
This paper provides a taxonomy for the techniques and categorize them as data-augmentation, embedding, optimization and semantics based learning for few-shot, one-shot and zero-shot settings, and describes the seminal work done in each category and discusses their approach towards solving the predicament of learning from few samples. Expand
Learning to Compare: Relation Network for Few-Shot Learning
TLDR
A conceptually simple, flexible, and general framework for few-shot learning, where a classifier must learn to recognise new classes given only few examples from each, which is easily extended to zero- shot learning. Expand
Robust Compare Network for Few-Shot Learning
TLDR
This work argues that it is desirable to learn a robust encoder that can draw inferences about other cases from one example, and improves the accuracy of few-shot learning by mining the internal mechanism of deep networks, which can leverage label information more effectively. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 34 REFERENCES
Siamese Neural Networks for One-Shot Image Recognition
TLDR
A method for learning siamese neural networks which employ a unique structure to naturally rank similarity between inputs and is able to achieve strong results which exceed those of other deep learning models with near state-of-the-art performance on one-shot classification tasks. Expand
Sequence to Sequence Learning with Neural Networks
TLDR
This paper presents a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure, and finds that reversing the order of the words in all source sentences improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier. Expand
One shot learning of simple visual concepts
TLDR
A generative model of how characters are composed from strokes is introduced, where knowledge from previous characters helps to infer the latent strokes in novel characters, using a massive new dataset of handwritten characters. Expand
Deep Metric Learning Using Triplet Network
TLDR
This paper proposes the triplet network model, which aims to learn useful representations by distance comparisons, and demonstrates using various datasets that this model learns a better representation than that of its immediate competitor, the Siamese network. Expand
DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition
TLDR
DeCAF, an open-source implementation of deep convolutional activation features, along with all associated network parameters, are released to enable vision researchers to be able to conduct experimentation with deep representations across a range of visual concept learning paradigms. Expand
Zero-Shot Learning by Convex Combination of Semantic Embeddings
TLDR
A simple method for constructing an image embedding system from any existing image classifier and a semantic word embedding model, which contains the $\n$ class labels in its vocabulary is proposed, which outperforms state of the art methods on the ImageNet zero-shot learning task. Expand
Rethinking the Inception Architecture for Computer Vision
TLDR
This work is exploring ways to scale up networks in ways that aim at utilizing the added computation as efficiently as possible by suitably factorized convolutions and aggressive regularization. Expand
Meta-Learning with Memory-Augmented Neural Networks
TLDR
The ability of a memory-augmented neural network to rapidly assimilate new data, and leverage this data to make accurate predictions after only a few samples is demonstrated. Expand
Convolutional Deep Belief Networks on CIFAR-10
We describe how to train a two-layer convolutional Deep Belief Network (DBN) on the 1.6 million tiny images dataset. When training a convolutional DBN, one must decide what to do with the edge pixelsExpand
ImageNet classification with deep convolutional neural networks
TLDR
A large, deep convolutional neural network was trained to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes and employed a recently developed regularization method called "dropout" that proved to be very effective. Expand
...
1
2
3
4
...