Extractive Summarization using Continuous Vector Space Models

@inproceedings{Kragebck2014ExtractiveSU,
  title={Extractive Summarization using Continuous Vector Space Models},
  author={Mikael Krageb{\"a}ck and Olof Mogren and Nina Tahmasebi and Devdatt P. Dubhashi},
  booktitle={CVSC@EACL},
  year={2014}
}
Automatic summarization can help users extract the most important pieces of information from the vast amount of text digitized into electronic form everyday. Central to automatic summarization is the notion of similarity between sentences in text. In this paper we propose the use of continuous vector representations for semantically aware representations of sentences as a basis for measuring similarity. We evaluate different compositions for sentence representation on a standard dataset using… Expand
Extractive Text Summarization Using Word Vector Embedding
TLDR
This paper has proposed an approach to extract a good set of features followed by neural network for supervised extractive summarization and experimental results show the effectiveness of the proposed method against various online extractive text summarizers. Expand
Centroid-based Text Summarization through Compositionality of Word Embeddings
TLDR
This paper proposes a centroidbased method for text summarization that exploits the compositional capabilities of word embeddings and achieves good performance even in comparison to more complex deep learning models. Expand
Central Embeddings for Extractive Summarization Based on Similarity
TLDR
This work aims to find evidence that semantic information is kept in word embeddings and this representation is subject to be grouped based on their similarity, so that main ideas can be identified in sets of documents. Expand
Extractive Summarization Using Multi-Task Learning with Document Classification
TLDR
A general framework for summarization that extracts sentences from a document using externally related information that is addressed in the framework of multi-task learning using curriculum learning for sentence extraction and document classification is proposed. Expand
Text document summarization using word embedding
TLDR
This paper proposes an automatic summarizer using the distributional semantic model to capture semantics for producing high-quality summaries and concludes that usage of semantic as a feature for text summarization provides improved results and helps to further reduce redundancies from the input source. Expand
A Survey on Neural Network-Based Summarization Methods
TLDR
This literature review surveys the recent work on neural-based models in automatic text summarization and examines in detail ten state-of-the-art neural- based summarizers: five abstractive models and five extractive models. Expand
Neural Extractive Summarization with Side Information
TLDR
It is shown that extractive summarization with side information consistently outperforms its counterpart that does not use any side information, in terms of both informativeness and fluency. Expand
A semantic space approach for automatic summarization of documents
TLDR
It is shown that simply cumulatively adding semantic space vectors to represent sentence level meaning already yields comparable performance to the state-of-the-art in document summarization. Expand
An unsupervised method for extractive multi-document summarization based on centroid approach and sentence embeddings
TLDR
An unsupervised method for generic extractive multi-document summarization based on the sentence embedding representations and the centroid approach that outperforms several state-of-the-art methods and achieves promising results compared to the best performing methods including supervised deep learning based methods. Expand
The combination of similarity measures for extractive summarization
TLDR
This paper proposes the combination of similarity measures for sentence comparison and outperforms the recent works in English with the significant improvement and achieves the competitive result in Vietnamese. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 30 REFERENCES
Extractive summarisation via sentence removal: condensing relevant sentences into a short summary
TLDR
The main contribution of this paper is the definition of an algorithm for sentence removal, developed to maximise the score between the summary and the original document. Expand
ROUGE: A Package for Automatic Evaluation of Summaries
TLDR
Four different RouGE measures are introduced: ROUGE-N, ROUge-L, R OUGE-W, and ROUAGE-S included in the Rouge summarization evaluation package and their evaluations. Expand
Efficient Estimation of Word Representations in Vector Space
TLDR
Two novel model architectures for computing continuous vector representations of words from very large data sets are proposed and it is shown that these vectors provide state-of-the-art performance on the authors' test set for measuring syntactic and semantic word similarities. Expand
Word Representations: A Simple and General Method for Semi-Supervised Learning
TLDR
This work evaluates Brown clusters, Collobert and Weston (2008) embeddings, and HLBL (Mnih & Hinton, 2009) embeds of words on both NER and chunking, and finds that each of the three word representations improves the accuracy of these baselines. Expand
Centroid-based summarization of multiple documents
TLDR
A multi-document summarizer, MEAD, is presented, which generates summaries using cluster centroids produced by a topic detection and tracking system and an evaluation scheme based on sentence utility and subsumption is applied. Expand
A Class of Submodular Functions for Document Summarization
TLDR
A class of submodular functions meant for document summarization tasks which combine two terms, one which encourages the summary to be representative of the corpus, and the other which positively rewards diversity, which means that an efficient scalable greedy optimization scheme has a constant factor guarantee of optimality. Expand
Distributed Representations of Words and Phrases and their Compositionality
TLDR
This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling. Expand
Dynamic Pooling and Unfolding Recursive Autoencoders for Paraphrase Detection
TLDR
This work introduces a method for paraphrase detection based on recursive autoencoders (RAE) and unsupervised RAEs based on a novel unfolding objective and learns feature vectors for phrases in syntactic trees to measure word- and phrase-wise similarity between two sentences. Expand
Opinosis: A Graph Based Approach to Abstractive Summarization of Highly Redundant Opinions
TLDR
A novel graph-based summarization framework (Opinosis) that generates concise abstractive summaries of highly redundant opinions that have better agreement with human summaries compared to the baseline extractive method. Expand
Learning Continuous Phrase Representations and Syntactic Parsing with Recursive Neural Networks
Natural language parsing has typically been done with small sets of discrete categories such as NP and VP, but this representation does not capture the full syntactic nor semantic richness ofExpand
...
1
2
3
...