ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
- Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut
- Computer ScienceInternational Conference on Learning…
- 26 September 2019
This work presents two parameter-reduction techniques to lower memory consumption and increase the training speed of BERT, and uses a self-supervised loss that focuses on modeling inter-sentence coherence.
SummScreen: A Dataset for Abstractive Screenplay Summarization
- Mingda Chen, Z. Chu, Sam Wiseman, Kevin Gimpel
- Computer ScienceAnnual Meeting of the Association for…
- 14 April 2021
Human evaluation and qualitative analysis reveal that the non-oracle models are competitive with their oracle counterparts in terms of generating faithful plot events and can benefit from better content selectors.
A Multi-Task Approach for Disentangling Syntax and Semantics in Sentence Representations
- Mingda Chen, Qingming Tang, Sam Wiseman, Kevin Gimpel
- Computer ScienceNorth American Chapter of the Association for…
- 2 April 2019
We propose a generative model for a sentence that uses two latent variables, with one intended to represent the syntax of the sentence and the other to represent its semantics. We show we can achieve…
Controllable Paraphrase Generation with a Syntactic Exemplar
- Mingda Chen, Qingming Tang, Sam Wiseman, Kevin Gimpel
- Computer ScienceAnnual Meeting of the Association for…
- 3 June 2019
This work proposes a novel task, where the syntax of a generated sentence is controlled rather by a sentential exemplar, and develops a variational model with a neural module specifically designed for capturing syntactic knowledge and several multitask training objectives to promote disentangled representation learning.
Evaluation Benchmarks and Learning Criteria for Discourse-Aware Sentence Representations
- Mingda Chen, Z. Chu, Kevin Gimpel
- Computer ScienceConference on Empirical Methods in Natural…
- 1 August 2019
This work proposes DiscoEval, a test suite of tasks to evaluate whether sentence representations include broader context information, and proposes a variety of training objectives that makes use of natural annotations from Wikipedia to build sentence encoders capable of modeling discourse.
Variational Sequential Labelers for Semi-Supervised Learning
- Mingda Chen, Qingming Tang, Karen Livescu, Kevin Gimpel
- Computer ScienceConference on Empirical Methods in Natural…
- 23 June 2019
A family of multitask variational methods for semi-supervised sequence labeling that combines a latent-variable generative model and a discriminative labeler, and explores several latent variable configurations, including ones with hierarchical structure.
WikiTableT: A Large-Scale Data-to-Text Dataset for Generating Wikipedia Article Sections
- Mingda Chen, Sam Wiseman, Kevin Gimpel
- Computer ScienceFindings
- 29 December 2020
Qualitative analysis shows that the best approaches can generate fluent and high quality texts but they struggle with coherence and factuality, showing the potential for the WIKITABLET dataset to inspire future work on long-form generation.
EntEval: A Holistic Evaluation Benchmark for Entity Representations
- Mingda Chen, Z. Chu, Yang Chen, K. Stratos, Kevin Gimpel
- Computer ScienceConference on Empirical Methods in Natural…
- 1 August 2019
This work proposes EntEval: a test suite of diverse tasks that require nontrivial understanding of entities including entity typing, entity similarity, entity relation prediction, and entity disambiguation, and develops training techniques for learning better entity representations by using natural hyperlink annotations in Wikipedia.
How to Ask Better Questions? A Large-Scale Multi-Domain Dataset for Rewriting Ill-Formed Questions
- Z. Chu, Mingda Chen, Xiance Si
- Computer ScienceAAAI Conference on Artificial Intelligence
- 21 November 2019
A large-scale dataset for the task of rewriting an ill-formed natural language question to a well-formed one andTrain sequence-to-sequence neural models on the constructed dataset and obtain an improvement of 13.2% in BLEU-4 over baseline methods built from other data resources.
Controllable Paraphrasing and Translation with a Syntactic Exemplar
- Mingda Chen, Sam Wiseman, Kevin Gimpel
- Computer ScienceArXiv
- 12 October 2020
The authors' single proposed model can perform four tasks: controlled paraphrase generation in both languages and controlled machine translation in both language directions, and analysis shows that their models learn to disentangle semantics and syntax in their latent representations.
...
...