Graceful Forgetting II. Data as a Process
@article{Cheveigne2022GracefulFI, title={Graceful Forgetting II. Data as a Process}, author={Alain de Cheveign'e}, journal={ArXiv}, year={2022}, volume={abs/2211.15441} }
This is the second part of a two-part essay on memory and its inseparable nemesis, forgetting . It looks at memory from a computational perspective in terms of function and constraints, in the rational spirit of Marr (1982) or Anderson (1989). The core question is: How to fit an infinite past into finite storage? The requirements and benefits of such a “scalable” data store are analyzed and the consequences explored, the main of which is that conserving data should be seen as a process…
Figures from this paper
References
SHOWING 1-10 OF 59 REFERENCES
Memory Networks
- Computer ScienceICLR
- 2015
This work describes a new class of learning models called memory networks, which reason with inference components combined with a long-term memory component; they learn how to use these jointly.
Hybrid computing using a neural network with dynamic external memory
- Computer ScienceNature
- 2016
A machine learning model called a differentiable neural computer (DNC), which consists of a neural network that can read from and write to an external memory matrix, analogous to the random-access memory in a conventional computer.
Episodic Memory in Lifelong Language Learning
- Computer ScienceNeurIPS
- 2019
This work proposes an episodic memory model that performs sparse experience replay and local adaptation to mitigate catastrophic forgetting in a lifelong language learning setup where a model needs to learn from a stream of text examples without any dataset identifier.
End-To-End Memory Networks
- Computer ScienceNIPS
- 2015
A neural network with a recurrent attention model over a possibly large external memory that is trained end-to-end, and hence requires significantly less supervision during training, making it more generally applicable in realistic settings.
Long Short-Term Memory
- Computer ScienceNeural Computation
- 1997
A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Predictability , Complexity , and Learning
- Computer Science
- 2002
It is argued that the divergent part of Ipred(T), the mutual information between the past and the future of a time series, provides the unique measure for the complexity of dynamics underlying aTime series.
CONDITIONS FOR VERSATILE LEARNING , HELMHOLTZ ’ S UNCONSCIOUS INFERENCE , AND THE TASK OF PERCEPTION
- Psychology
- 2002
It is a mistake to consider perception and learning separately because what one learns is strongly constrained by what one perceives, and what one perceives depends on what one bas experienced. I…
Episodic Memory Reader: Learning What to Remember for Question Answering from Streaming Data
- Computer ScienceACL
- 2019
A novel end-to-end deep network model for reading comprehension called Episodic Memory Reader (EMR) that sequentially reads the input contexts into an external memory, while replacing memories that are less important for answering unseen questions is proposed.
Of bits and wows: A Bayesian theory of surprise with applications to attention
- Computer ScienceNeural Networks
- 2010
Memory as Perception of the Past: Compressed Time inMind and Brain
- Biology, PsychologyTrends in Cognitive Sciences
- 2018