Corpus ID: 9816245

Generating News Headlines with Recurrent Neural Networks

@article{Lopyrev2015GeneratingNH,
  title={Generating News Headlines with Recurrent Neural Networks},
  author={Konstantin Lopyrev},
  journal={ArXiv},
  year={2015},
  volume={abs/1512.01712}
}
We describe an application of an encoder-decoder recurrent neural network with LSTM units and attention to generating headlines from the text of news articles. [...] Key Method Furthermore, we study how the neural network decides which input words to pay attention to, and specifically we identify the function of the different neurons in a simplified attention mechanism. Interestingly, our simplified attention mechanism performs better that the more complex attention mechanism on a held out set of articles.Expand
A Survey on Recurrent Neural Network Architectures for Sequential Learning
TLDR
This paper presents an approach to text captioning using recurrent neural networks which comprise of an encoder–decoder model and results show that the model performs better when the input is fed with the summary as compared to the original article itself. Expand
Summary++: Summarizing Chinese News Articles with Attention
TLDR
This paper presents Summary++, the model that competed in NLPCC2018’s Summary task, which uses a neural encoder decoder attention model with pointer generator network to focus on words attented to rather than words predicted. Expand
Auto Generation of Arabic News Headlines
We describe two RNN models to generate Arabic news headlines from given news articles. The first model is a standard sequence-to-sequence architecture, and the second model is a sequence-to-sequenceExpand
Abstractive headline generation for spoken content by attentive recurrent neural networks with ASR error modeling
TLDR
This paper proposes an AsR error modeling approach to learn the underlying structure of ASR error patterns and incorporate this model in an Attentive Recurrent Neural Network (ARNN) architecture, so that the model for abstractive headline generation for spoken content can be learned from abundant text data and the ASR data for some recognizers. Expand
Headline Generation: Learning from Decomposed Document Titles
TLDR
A novel method for generating titles for unstructured text documents is proposed and the results of a randomized double-blind trial in which subjects were unaware of which titles were human or machine-generated are presented. Expand
Headline Generation: Learning from Decomposable Document Titles
TLDR
A novel method for generating titles for unstructured text documents is proposed and the results of a randomized double-blind trial in which subjects were unaware of which titles were human or machine-generated are presented. Expand
Distraction-Based Neural Networks for Document Summarization
TLDR
This paper proposes neural models to train computers not just to pay attention to specific regions and content of input documents with attention models, but also distract them to traverse between different content of a document so as to better grasp the overall meaning for summarization. Expand
Distraction-based neural networks for modeling documents
TLDR
This paper proposes neural models to train computers not just to pay attention to specific regions and content of input documents with attention models, but also distract them to traverse between different content of a document so as to better grasp the overall meaning for summarization. Expand
Automatic Lecture Video Content Summarizationwith Attention-based Recurrent Neural Network
TLDR
An automatic summarization method for a lecture video transcript that uses attention based Recurrent Neural Network (RNN) to capture the content of the lecture to show a significant improvement in the term of ROUGE score compared to the baseline models. Expand
Recurrent neural network for abstractive summarization of documents
TLDR
The proposed model principally comprises of two sections - an encoder and a decoder - each one of which is a Recurrent Neural Network (RNN). Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 20 REFERENCES
Recurrent Neural Network Regularization
TLDR
This paper shows how to correctly apply dropout to LSTMs, and shows that it substantially reduces overfitting on a variety of tasks. Expand
A Neural Attention Model for Abstractive Sentence Summarization
TLDR
This work proposes a fully data-driven approach to abstractive sentence summarization by utilizing a local attention-based model that generates each word of the summary conditioned on the input sentence. Expand
Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks
TLDR
This work proposes a curriculum learning strategy to gently change the training process from a fully guided scheme using the true previous token, towards a less guided scheme which mostly uses the generated token instead. Expand
Effective Approaches to Attention-based Neural Machine Translation
TLDR
A global approach which always attends to all source words and a local one that only looks at a subset of source words at a time are examined, demonstrating the effectiveness of both approaches on the WMT translation tasks between English and German in both directions. Expand
Speech recognition with deep recurrent neural networks
TLDR
This paper investigates deep recurrent neural networks, which combine the multiple levels of representation that have proved so effective in deep networks with the flexible use of long range context that empowers RNNs. Expand
Sequence to Sequence Learning with Neural Networks
TLDR
This paper presents a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure, and finds that reversing the order of the words in all source sentences improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier. Expand
Teaching Machines to Read and Comprehend
TLDR
A new methodology is defined that resolves this bottleneck and provides large scale supervised reading comprehension data that allows a class of attention based deep neural networks that learn to read real documents and answer complex questions with minimal prior knowledge of language structure to be developed. Expand
Listen, Attend and Spell
TLDR
A neural network that learns to transcribe speech utterances to characters without making any independence assumptions between the characters, which is the key improvement of LAS over previous end-to-end CTC models. Expand
Deep visual-semantic alignments for generating image descriptions
TLDR
A model that generates natural language descriptions of images and their regions based on a novel combination of Convolutional Neural Networks over image regions, bidirectional Recurrent Neural networks over sentences, and a structured objective that aligns the two modalities through a multimodal embedding is presented. Expand
Deep Visual-Semantic Alignments for Generating Image Descriptions
  • A. Karpathy, Li Fei-Fei
  • Computer Science, Medicine
  • IEEE Transactions on Pattern Analysis and Machine Intelligence
  • 2017
TLDR
A model that generates natural language descriptions of images and their regions based on a novel combination of Convolutional Neural Networks over image regions, bidirectional Recurrent Neural networks over sentences, and a structured objective that aligns the two modalities through a multimodal embedding is presented. Expand
...
1
2
...