Using Pre-Trained Transformer for Better Lay Summarization
@inproceedings{Kim2020UsingPT, title={Using Pre-Trained Transformer for Better Lay Summarization}, author={Seungwon Kim}, booktitle={SDP}, year={2020} }
In this paper, we tack lay summarization tasks, which aim to automatically produce lay summaries for scientific papers, to participate in the first CL-LaySumm 2020 in SDP workshop at EMNLP 2020. We present our approach of using Pre-training with Extracted Gap-sentences for Abstractive Summarization (PEGASUS; Zhang et al., 2019b) to produce the lay summary and combining those with the extractive summarization model using Bidirectional Encoder Representations from Transformers (BERT; Devlin et al…
One Citation
Overview and Insights from the Shared Tasks at Scholarly Document Processing 2020: CL-SciSumm, LaySumm and LongSumm
- Computer ScienceSDP
- 2020
The quality and quantity of the submissions show that there is ample interest in scholarly document summarization, and the state of the art in this domain is at a midway point between being an impossible task and one that is fully resolved.
References
SHOWING 1-10 OF 35 REFERENCES
PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization
- Computer ScienceICML
- 2020
This work proposes pre-training large Transformer-based encoder-decoder models on massive text corpora with a new self-supervised objective, PEGASUS, and demonstrates it achieves state-of-the-art performance on all 12 downstream datasets measured by ROUGE scores.
Text Summarization with Pretrained Encoders
- Computer ScienceEMNLP
- 2019
This paper introduces a novel document-level encoder based on BERT which is able to express the semantics of a document and obtain representations for its sentences and proposes a new fine-tuning schedule which adopts different optimizers for the encoder and the decoder as a means of alleviating the mismatch between the two.
TLDR: Extreme Summarization of Scientific Documents
- Computer ScienceFINDINGS
- 2020
This work introduces SCITLDR, a new multi-target dataset of 5.4K TLDRs over 3.2K papers, and proposes CATTS, a simple yet effective learning strategy for generatingTLDRs that exploits titles as an auxiliary training signal.
Learning by Semantic Similarity Makes Abstractive Summarization Better
- Computer ScienceArXiv
- 2020
This paper proposes Semantic Similarity strategy that can consider semantic meanings of generated summaries while training, and achieves a new state-of-the-art performance, ROUGE-L score of 41.5 on CNN/DM dataset.
Get To The Point: Summarization with Pointer-Generator Networks
- Computer ScienceACL
- 2017
A novel architecture that augments the standard sequence-to-sequence attentional model in two orthogonal ways, using a hybrid pointer-generator network that can copy words from the source text via pointing, which aids accurate reproduction of information, while retaining the ability to produce novel words through the generator.
A Deep Reinforced Model for Abstractive Summarization
- Computer ScienceICLR
- 2018
A neural network model with a novel intra-attention that attends over the input and continuously generated output separately, and a new training method that combines standard supervised word prediction and reinforcement learning (RL) that produces higher quality summaries.
Neural Summarization by Extracting Sentences and Words
- Computer ScienceACL
- 2016
This work develops a general framework for single-document summarization composed of a hierarchical document encoder and an attention-based extractor that allows for different classes of summarization models which can extract sentences or words.
SummaRuNNer: A Recurrent Neural Network Based Sequence Model for Extractive Summarization of Documents
- Computer ScienceAAAI
- 2017
We present SummaRuNNer, a Recurrent Neural Network (RNN) based sequence model for extractive summarization of documents and show that it achieves performance better than or comparable to…
Headline Generation: Learning from Decomposable Document Titles
- Computer Science
- 2019
A novel method for generating titles for unstructured text documents is proposed and the results of a randomized double-blind trial in which subjects were unaware of which titles were human or machine-generated are presented.
Headline Generation: Learning from Decomposed Document Titles
- Computer ScienceArXiv
- 2019
A novel method for generating titles for unstructured text documents is proposed and the results of a randomized double-blind trial in which subjects were unaware of which titles were human or machine-generated are presented.