• Publications
  • Influence
Style Transfer in Text: Exploration and Evaluation
TLDR
Two models are explored to learn style transfer with non-parallel data to learn separate content representations and style representations using adversarial networks, and a novel evaluation metrics which measure two aspects of style transfer: transfer strength and content preservation. Expand
Named Entity Recognition for Chinese Social Media with Jointly Trained Embeddings
TLDR
A new corpus of Weibo messages annotated for both name and nominal mentions is presented and a joint training objective for the embeddings that makes use of both (NER) labeled and unlabeled raw text is proposed. Expand
Cross-Sentence N-ary Relation Extraction with Graph LSTMs
TLDR
A general relation extraction framework based on graph long short-term memory networks (graph LSTMs) that can be easily extended to cross-sentence n-ary relation extraction is explored, demonstrating its effectiveness with both conventional supervised learning and distant supervision. Expand
Stack-Pointer Networks for Dependency Parsing
TLDR
A novel architecture for dependency parsing: stack-pointer networks (StackPtr), which first reads and encodes the whole sentence, then builds the dependency tree top-down in a depth-first fashion, yielding an efficient decoding algorithm with O(n^2) time complexity. Expand
Plan-And-Write: Towards Better Automatic Storytelling
TLDR
Experiments show that with explicit storyline planning, the generated stories are more diverse, coherent, and on topic than those generated without creating a full plan, according to both automatic and human evaluations. Expand
Better Automatic Evaluation of Open-Domain Dialogue Systems with Contextualized Embeddings
TLDR
Using contextualized word embeddings to compute more accurate relatedness scores and thus better evaluation metrics is explored, and experiments show that the evaluation metrics outperform RUBER, which is trained on staticembeddings. Expand
On Difficulties of Cross-Lingual Transfer with Order Differences: A Case Study on Dependency Parsing
TLDR
Investigating crosslingual transfer and posit that an orderagnostic model will perform better when transferring to distant foreign languages shows that RNN-based architectures transfer well to languages that are close to English, while self-attentive models have better overall cross-lingualtransferability and perform especially well on distant languages. Expand
Connecting the Dots: A Knowledgeable Path Generator for Commonsense Question Answering
TLDR
This paper augments a general commonsense QA framework with a knowledgeable path generator by extrapolating over existing paths in a KG with a state-of-the-art language model, which learns to connect a pair of entities in text with a dynamic, and potentially novel, multi-hop relational path. Expand
Improving Named Entity Recognition for Chinese Social Media with Word Segmentation Representation Learning
TLDR
It is shown that new state-of-the-art word segmentation systems use neural models to learn representations for predicting word boundaries, and these same representations, jointly trained with an NER system, yield significant improvements in NER for Chinese social media. Expand
The Woman Worked as a Babysitter: On Biases in Language Generation
TLDR
The notion of the regard towards a demographic is introduced, the varying levels of regard towards different demographics are used as a defining metric for bias in NLG, and the extent to which sentiment scores are a relevant proxy metric for regard is analyzed. Expand
...
1
2
3
4
5
...