Discontinuous Incremental Shift-reduce Parsing

@inproceedings{Maier2015DiscontinuousIS,
  title={Discontinuous Incremental Shift-reduce Parsing},
  author={Wolfgang Maier},
  booktitle={Annual Meeting of the Association for Computational Linguistics},
  year={2015}
}
  • Wolfgang Maier
  • Published in
    Annual Meeting of the…
    1 July 2015
  • Computer Science
We present an extension to incremental shift-reduce parsing that handles discontinuous constituents, using a linear classifier and beam search. We achieve very high parsing speeds (up to 640 sent./sec.) and accurate results (up to 79.52 F1 on TiGer). 

Figures and Tables from this paper

Discontinuous Constituent Parsing as Sequence Labeling

This paper reduces discontinuous parsing to sequence labeling and proposes to encode tree discontinuities as nearly ordered permutations of the input sequence as well as studies whether such discontinuous representations are learnable.

A Derivational Model of Discontinuous Parsing

The notion of latent-variable probabilistic context-free derivation of syntactic structures is enhanced to allow heads and unrestricted discontinuities and lends itself to intrinsic evaluation in terms of perplexity, as shown in experiments.

Incremental Discontinuous Phrase Structure Parsing with the GAP Transition

A novel transition system for discontinuous lexicalized constituent parsing called SR-GAP is introduced, an extension of the shift-reduce algorithm with an additional gap transition that outperforms the previous best transition-based discontinuous parser by a large margin.

Transition-Based Left-Corner Parsing for Identifying PTB-Style Nonlocal Dependencies

This paper proposes a left-corner parser which can identify nonlocal dependencies and uses a structured perceptron which enables the parser to utilize global features captured by non local dependencies.

Reducing Discontinuous to Continuous Parsing with Pointer Network Reordering

A Pointer Network capable of accurately generating the continuous token arrangement for a given input sentence and define a bijective function to recover the original order is developed.

Discontinuous Combinatory Constituency Parsing

A pair of continuous combinator-based constituency parsers (one binary and one multi-branching) are extended into a discontinuous pair and achieved state-of-the-art discontinuous accuracy with a significant speed advantage.

A Unifying Theory of Transition-based and Sequence Labeling Parsing

A mapping from transition-based parsing algorithms that read sentences from left to right to sequence labeling encodings of syntactic trees is defined, and applying it to dependency parsing, sequence labeling versions of four algorithms are implemented.

Edinburgh Research Explorer Discontinuous Constituency Parsing with a Stack-Free Transition System and a Dynamic Oracle

A novel transition system for discontinuous constituency parsing uses a set of parsing items, with constant-time random access, which makes it possible to construct any discontinuous constituencies tree in exactly 4 n − 2 transitions for a sentence of length n, whereas existing systems need a quadratic number of transitions to derive some structures.

Generic refinement of expressive grammar formalisms with an application to discontinuous constituent parsing

We formulate a generalization of Petrov et al. (2006)’s split/merge algorithm for interpreted regular tree grammars (Koller and Kuhlmann, 2011), which capture a large class of grammar formalisms. We

Advances in Using Grammars with Latent Annotations for Discontinuous Parsing

It is found that the grammars presented are more accurate than previous approaches based on discontinuous grammar formalisms and early instances of the discriminative models but inferior to recent discriminatives parsers.

Discontinuity Revisited: An Improved Conversion to Context-Free Representations

A labeled dependency evaluation shows that the new conversion method leads to better results by preserving local relationships and introducing fewer inconsistencies into the training data.

Language-Independent Parsing with Empty Elements

We present a simple, language-independent method for integrating recovery of empty elements into syntactic parsing. This method outperforms the best published method we are aware of on English and a

An Improved Oracle for Dependency Parsing with Online Reordering

An improved training strategy for dependency parsers that use online reordering to handle non-projective trees improves both efficiency and accuracy by reducing the number of swap operations performed on non- projective trees by up to 80%.

Parsing as Reduction

This work reduces phrase-representation parsing to dependency parsing, and shows that any off-the-shelf, trainable dependency parser can be used to produce constituents, and can perform discontinuous parsing in a very natural manner.

Antecedent Recovery: Experiments with a Trace Tagger

This paper develops both a two- step approach which combines a trace tagger with a state-of-the-art lexicalized parser and a one-step approach which finds nonlocal dependencies while parsing.

Shift-Reduce CCG Parsing

It is shown that the shift-reduce parser gives competitive accuracies compared to C&C, and given the high ambiguity levels in an automatically-extracted grammar and the amount of information in the CCG lexical categories which form the shift actions, this is a surprising result.

Efficient parsing with Linear Context-Free Rewriting Systems

This work shows that parsing long sentences with such an optimally binarized grammar remains infeasible, and introduces a technique which removes this length restriction, while maintaining a respectable accuracy.

Non-Projective Dependency Parsing in Expected Linear Time

A novel transition system for dependency parsing, which constructs arcs only between adjacent words but can parse arbitrary non-projective trees by swapping the order of words in the input, shows state-of-the-art accuracy.

Discontinuous Parsing with an Efficient and Accurate DOP Model

We present a discontinuous variant of tree-substitution grammar (tsg) based on Linear Context-Free Rewriting Systems. We use this formalism to instantiate a Data-Oriented Parsing model applied to

A Classifier-Based Parser with Linear Run-Time Complexity

It is shown that, with an appropriate feature set used in classification, a very simple one-path greedy parser can perform at the same level of accuracy as more complex parsers.