• Corpus ID: 240353977

Automatic Knowledge Augmentation for Generative Commonsense Reasoning

@article{Seo2021AutomaticKA,
  title={Automatic Knowledge Augmentation for Generative Commonsense Reasoning},
  author={Jaehyung Seo and Chanjun Park and Sugyeong Eo and Hyeonseok Moon and Heuiseok Lim},
  journal={ArXiv},
  year={2021},
  volume={abs/2111.00192}
}
Generative commonsense reasoning is the capability of a language model to generate a sentence with a given concept-set that is based on commonsense knowledge. However, generative language models still struggle to provide outputs, and the training set does not contain patterns that are sufficient for generative commonsense reasoning. In this paper, we propose a data-centric method that uses automatic knowledge augmentation to extend commonsense knowledge using a machine knowledge generator. This… 

Tables from this paper

Active-Learning-as-a-Service: An Efficient MLOps System for Data-Centric AI
TLDR
An efficient MLOps system for AL, named ALaaS (Active-Learning-as-a-Service), which abstracts an AL process to several components and provides rich APIs for advanced users to extend the system to new scenarios.

References

SHOWING 1-10 OF 17 REFERENCES
G-DAug: Generative Data Augmentation for Commonsense Reasoning
TLDR
This work proposes a novel generative data augmentation technique, G-DAUGˆC, that aims to achieve more accurate and robust learning in a low-resource setting and produces a diverse set of fluent training examples, demonstrating that its selection and training approaches are important for performance.
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks
TLDR
A general-purpose fine-tuning recipe for retrieval-augmented generation (RAG) -- models which combine pre-trained parametric and non-parametric memory for language generation, and finds that RAG models generate more specific, diverse and factual language than a state-of-the-art parametric-only seq2seq baseline.
Language Models are Unsupervised Multitask Learners
TLDR
It is demonstrated that language models begin to learn these tasks without any explicit supervision when trained on a new dataset of millions of webpages called WebText, suggesting a promising path towards building language processing systems which learn to perform tasks from their naturally occurring demonstrations.
SPICE: Semantic Propositional Image Caption Evaluation
There is considerable interest in the task of automatically generating image captions. However, evaluation is challenging. Existing automatic evaluation metrics are primarily sensitive to n-gram
Generalization without Systematicity: On the Compositional Skills of Sequence-to-Sequence Recurrent Networks
TLDR
This paper introduces the SCAN domain, consisting of a set of simple compositional navigation commands paired with the corresponding action sequences, and tests the zero-shot generalization capabilities of a variety of recurrent neural networks trained on SCAN with sequence-to-sequence methods.
Bleu: a Method for Automatic Evaluation of Machine Translation
TLDR
This work proposes a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run.
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
TLDR
BART is presented, a denoising autoencoder for pretraining sequence-to-sequence models, which matches the performance of RoBERTa on GLUE and SQuAD, and achieves new state-of-the-art results on a range of abstractive dialogue, question answering, and summarization tasks.
METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments
TLDR
METEOR is described, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machineproduced translation and human-produced reference translations and can be easily extended to include more advanced matching strategies.
A machine learning approach for improved BM25 retrieval
TLDR
A machine learning approach to BM25-style retrieval is developed that learns, using LambdaRank, from the input attributes of BM25, and significantly improves retrieval effectiveness over BM25 and BM25F.
CommonGen: A Constrained Text Generation Challenge for Generative Commonsense Reasoning
TLDR
A constrained text generation task, CommonGen associated with a benchmark dataset, to explicitly test machines for the ability of generative commonsense reasoning, and demonstrates that the learned generative Commonsense reasoning capability can be transferred to improve downstream tasks such as CommonsenseQA by generating additional context.
...
...