On the Opportunities and Risks of Foundation Models
- Rishi Bommasani, Drew A. Hudson, Percy Liang
- Computer ScienceArXiv
- 16 August 2021
This report provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities, to their applications, and what they are even capable of due to their emergent properties.
Diffusion-LM Improves Controllable Text Generation
- Xiang Lisa Li, John Thickstun, Ishaan Gulrajani, Percy Liang, Tatsunori Hashimoto
- Computer ScienceArXiv
- 27 May 2022
A new non-autoregressive language model based on continuous diffusions that iteratively denoises a sequence of Gaussian vectors into word vectors, yielding a sequences of intermediate latent variables that enables a simple gradient-based algorithm to perform complex, controllable generation tasks.
Emergent Abilities of Large Language Models
Scaling up language models has been shown to predictably improve performance and sample efficiency on a wide range of downstream tasks. This paper instead discusses an unpredictable phenomenon that we…
Holistic Evaluation of Language Models
- Percy Liang, Rishi Bommasani, Yuta Koreeda
- Computer ScienceArXiv
- 16 November 2022
Holistic Evaluation of Language Models (HELM) is presented to improve the transparency of language models and intends for HELM to be a living benchmark for the community, continuously updated with new scenarios, metrics, and models.
Jury Learning: Integrating Dissenting Voices into Machine Learning Models
- Mitchell L. Gordon, Michelle S. Lam, Michael S. Bernstein
- Computer ScienceInternational Conference on Human Factors in…
- 7 February 2022
A deep learning architecture that models every annotator in a dataset, samples from annotators’ models to populate the jury, then runs inference to classify enables juries that dynamically adapt their composition, explore counterfactuals, and visualize dissent.
Contrastive Decoding: Open-ended Text Generation as Optimization
- Xiang Lisa Li, Ari Holtzman, M. Lewis
- Computer ScienceArXiv
- 27 October 2022
Contrastive decoding (CD), a more reliable search objective that returns the difference between likelihood under a large LM and a small LM, is proposed, which requires zero training, and produces higher quality text than decoding from the larger LM alone.
Is Importance Weighting Incompatible with Interpolating Classifiers?
- Ke Alexander Wang, Niladri S. Chatterji, Saminul Haque, Tatsunori Hashimoto
- Computer ScienceInternational Conference on Learning…
- 24 December 2021
Surprisingly, the theory shows that using weights that are obtained by exponentiating the classical unbiased importance weights can improve performance, and the loss function can outperform reweighted cross-entropy by as much as 9% in test accuracy.
Is a Caption Worth a Thousand Images? A Controlled Study for Representation Learning
- Shibani Santurkar, Yann Dubois, Rohan Taori, Percy Liang, Tatsunori Hashimoto
- Computer ScienceArXiv
- 15 July 2022
This work studies this question through a carefully controlled comparison of two approaches in terms of their ability to learn representations that generalize to downstream classification tasks, finding that when the pre-training dataset meets certain criteria—it is suf ficiently large and contains descriptive captions with low variability—image-only methods do not match CLIP’s transfer performance, even when they are trained with more image data.
Easily Accessible Text-to-Image Generation Amplifies Demographic Stereotypes at Large Scale
- Federico Bianchi, Pratyusha Kalluri, Aylin Caliskan
- Computer ScienceArXiv
- 7 November 2022
This work investigates image-generation models and finds that they amplify dangerous and complex stereotypes, which are difficult to predict and not easily mitigated by users or model owners.
Language modeling via stochastic processes
- Rose E. Wang, Esin Durmus, Noah D. Goodman, Tatsunori Hashimoto
- Computer ScienceInternational Conference on Learning…
- 21 March 2022
Compared to domain-specific methods and fine-tuning GPT2 across a variety of text domains, TC improves performance on text infilling and discourse coherence and preserves the text structure on long text generation settings.
...
...