• Corpus ID: 237532277

RetrievalSum: A Retrieval Enhanced Framework for Abstractive Summarization

  title={RetrievalSum: A Retrieval Enhanced Framework for Abstractive Summarization},
  author={Chen An and Ming Zhong and Zhichao Geng and Jianqiang Yang and Xipeng Qiu},
Existing summarization systems mostly generate summaries purely relying on the content of the source document. However, even for humans, we usually need some references or exemplars to help us fully understand the source document and write summaries in a particular format. But how to find the high-quality exemplars and incorporate them into summarization systems is still challenging and worth exploring. In this paper, we propose RETRIEVALSUM, a novel retrieval enhanced abstractive summarization… 

Unsupervised Summarization with Customized Granularities

This paper proposes the first unsupervised multi-granularity summarization framework, GranuSum, which takes events as the basic semantic units of the source documents and proposes to rank these events by their salience, and develops a model to summarize input documents with given events as anchors and hints.

Training Data is More Valuable than You Think: A Simple and Effective Method by Retrieving from Training Data

Experimental results show that this simple method can achieve significantly better performance on a variety of NLU and NLG tasks, including summarization, machine translation, language modeling, and question answering tasks.

Improving the Factual Accuracy of Abstractive Clinical Text Summarization using Multi-Objective Optimization

This study proposes a framework for improving the factual accuracy of abstractive summarization of clinical text using knowledge-guided multi-objective optimization and experiment with three transformer encoder-decoder architectures to demonstrate that optimizing different loss functions leads to improved performance in terms of entity-level factual accuracy.

Entity-driven Fact-aware Abstractive Summarization of Biomedical Literature

It is demonstrated that injecting knowledge into the training/inference phase of these models enables the models to achieve significantly better performance than the standard source document-tosummary setting in terms of entity-level factual accuracy, N-gram novelty, and semantic equivalence while performing comparably on ROUGE metrics.

COLO: A Contrastive Learning based Re-ranking Framework for One-Stage Summarization

Traditional training paradigms for extractive and abstractive summarization systems always only use token-level or sentence-level training objectives. However, the output summary is always evaluated



Extractive Summarization as Text Matching

This paper forms the extractive summarization task as a semantic text matching problem, in which a source document and candidate summaries will be matched in a semantic space to create a semantic matching framework.

Text Summarization with Pretrained Encoders

This paper introduces a novel document-level encoder based on BERT which is able to express the semantics of a document and obtain representations for its sentences and proposes a new fine-tuning schedule which adopts different optimizers for the encoder and the decoder as a means of alleviating the mismatch between the two.

Abstractive Summarization of Reddit Posts with Multi-level Memory Networks

This work collects Reddit TIFU dataset, consisting of 120K posts from the online discussion forum Reddit, and proposes a novel abstractive summarization model named multi-level memory networks (MMN), equipped with multi- level memory to store the information of text from different levels of abstraction.

Knowledge Graph-Augmented Abstractive Summarization with Semantic-Driven Cloze Reward

ASGARD is presented, a novel framework for Abstractive Summarization with Graph-Augmentation and semantic-driven RewarD, and proposes the use of dual encoders—a sequential document encoder and a graph-structured encoder—to maintain the global context and local characteristics of entities, complementing each other.

PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization

This work proposes pre-training large Transformer-based encoder-decoder models on massive text corpora with a new self-supervised objective, PEGASUS, and demonstrates it achieves state-of-the-art performance on all 12 downstream datasets measured by ROUGE scores.

Evaluating the Factual Consistency of Abstractive Text Summarization

A weakly-supervised, model-based approach for verifying factual consistency and identifying conflicts between source documents and a generated summary substantially outperforms previous models, including those trained with strong supervision using standard datasets for natural language inference and fact checking.

How to Write Summaries with Patterns? Learning towards Abstractive Summarization through Prototype Editing

Extensive experiments conducted on a large-scale real-world text summarization dataset show that PESG achieves the state-of-the-art performance in terms of both automatic metrics and human evaluations.

GSum: A General Framework for Guided Neural Abstractive Summarization

A general and extensible guided summarization framework that can effectively take different kinds of external guidance as input is proposed and demonstrated, and how different types of guidance generate qualitatively different summaries is demonstrated, lending a degree of controllability to the learned models.

A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents

This work proposes the first model for abstractive summarization of single, longer-form documents (e.g., research papers), consisting of a new hierarchical encoder that models the discourse structure of a document, and an attentive discourse-aware decoder to generate the summary.

Retrieve, Rerank and Rewrite: Soft Template Based Neural Summarization

This paper uses a popular IR platform to use existing summaries as soft templates to guide the seq2seq model, and extends the framework to jointly conduct template Reranking and template-aware summary generation (Rewriting).