• Published 2018

On Evaluating the Generalization of LSTM Models in Formal Languages

@inproceedings{Paulson2018OnET,
  title={On Evaluating the Generalization of LSTM Models in Formal Languages},
  author={John A. Paulson},
  year={2018}
}
Recurrent Neural Networks (RNNs) are theoretically Turing-complete and established themselves as a dominant model for language processing. Yet, there still remains an uncertainty regarding their language learning capabilities. In this paper, we empirically evaluate the inductive learning capabilities of Long Short-Term Memory networks, a popular extension of simple RNNs, to learn simple formal languages, in particular ab, abc, and abcd. We investigate the influence of various aspects of… CONTINUE READING

Figures and Tables from this paper.

References

Publications referenced by this paper.
SHOWING 1-10 OF 25 REFERENCES

A Recurrent Neural Network that Learns to Count

VIEW 5 EXCERPTS
HIGHLY INFLUENTIAL

Automatic differentiation in PyTorch

VIEW 1 EXCERPT

LSTM: A Search Space Odyssey

VIEW 2 EXCERPTS

On learning context-free and context-sensitive languages