One Semantic Parser to Parse Them All: Sequence to Sequence Multi-Task Learning on Semantic Parsing Datasets

  title={One Semantic Parser to Parse Them All: Sequence to Sequence Multi-Task Learning on Semantic Parsing Datasets},
  author={Marco Damonte and Emilio Monti},
Semantic parsers map natural language utterances to meaning representations. The lack of a single standard for meaning representations led to the creation of a plethora of semantic parsing datasets. To unify different datasets and train a single model for them, we investigate the use of Multi-Task Learning (MTL) architectures. We experiment with five datasets (Geoquery, NLMaps, TOP, Overnight, AMR). We find that an MTL architecture that shares the entire network across datasets yields… 

Figures and Tables from this paper

ATP: AMRize Then Parse! Enhancing AMR Parsing with PseudoAMRs

This work proposes a principled method to involve auxiliary tasks to boost AMR parsing and shows that this method achieves new state-of-the-art performance on different benchmarks especially in topology-related scores.

State-of-the-art generalisation research in NLP: a taxonomy and review

A taxonomy for characterising and understanding generalisation research in NLP is presented, a taxonomy is used to present a comprehensive map of published generalisation studies, and recommendations for which areas might deserve attention in the future are made.

GraphQ IR: Unifying the Semantic Parsing of Graph Query Languages with One Intermediate Representation

GraphQ IR has a natural-language-like expression that bridges the semantic gap and formally defined syntax that maintains the graph structure and can more precisely convert user queries into GraphQ IR, which can be later losslessly compiled into various downstream graph query languages.

GraphQ IR: Unifying Semantic Parsing of Graph Query Language with Intermediate Representation

A natural-language-like representation that bridges the semantic gap and its formally defined syntax that maintains the graph structure, neural semantic parser can more ef-fectively convert user queries into the authors' GraphQ IR, which can be later automatically compiled into different downstream graph query languages.

An Analysis of Data Processing for Big Data Analytics

This article presents HC processing techniques, Big Data Analytics (BDA), large dataset instruments, and some classic ML and DM methodologies, and the use of deep learning to Data Analytics is investigated.



Measuring Compositional Generalization: A Comprehensive Method on Realistic Data

A novel method to systematically construct compositional generalization benchmarks by maximizing compound divergence while guaranteeing a small atom divergence between train and test sets is introduced, and it is demonstrated how this method can be used to create new compositionality benchmarks on top of the existing SCAN dataset.

BERT and PALs: Projected Attention Layers for Efficient Adaptation in Multi-Task Learning

Using new adaptation modules, PALs or `projected attention layers', this work matches the performance of separately fine-tuned models on the GLUE benchmark with roughly 7 times fewer parameters, and obtains state-of-the-art results on the Recognizing Textual Entailment dataset.

Multitask Parsing Across Semantic Representations

It is shown that multitask learning significantly improves UCCA parsing in both in-domain and out-of-domain settings, and uniform transition-based system and learning architecture for all parsing tasks are experiment on.

Neural Machine Translation by Jointly Learning to Align and Translate

It is conjecture that the use of a fixed-length vector is a bottleneck in improving the performance of this basic encoder-decoder architecture, and it is proposed to extend this by allowing a model to automatically (soft-)search for parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.

Learning to Parse Database Queries Using Inductive Logic Programming

Experimental results with a complete database-query application for U.S. geography show that CHILL is able to learn parsers that outperform a preexisting, hand-crafted counterpart, and provide direct evidence of the utility of an empirical approach at the level of a complete natural language application.

Abstract Meaning Representation for Sembanking

A sembank of simple, whole-sentence semantic structures will spur new work in statistical natural language understanding and generation, like the Penn Treebank encouraged work on statistical parsing.

An Overview of Multi-task Learning

Many areas, including computer vision, bioinformatics, health informatics, speech, natural language processing, web applications and ubiquitous computing, use MTL to improve the performance of the applications involved and some representative works are reviewed.

A Hierarchical Multi-task Approach for Learning Embeddings from Semantic Tasks

A hierarchical model trained in a multi-task learning setup on a set of carefully selected semantic tasks achieves state-of-the-art results on a number of tasks, namely Named Entity Recognition, Entity Mention Detection and Relation Extraction without hand-engineered features or external NLP tools like syntactic parsers.

Semantic Parsing for Task Oriented Dialog using Hierarchical Representations

This work proposes a hierarchical annotation scheme for semantic parsing that allows the representation of compositional queries, and can be efficiently and accurately parsed by standard constituency parsing models.

GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding

A benchmark of nine diverse NLU tasks, an auxiliary dataset for probing models for understanding of specific linguistic phenomena, and an online platform for evaluating and comparing models, which favors models that can represent linguistic knowledge in a way that facilitates sample-efficient learning and effective knowledge-transfer across tasks.