Automatic differentiation in PyTorch
@inproceedings{Paszke2017AutomaticDI, title={Automatic differentiation in PyTorch}, author={Adam Paszke and Sam Gross and Soumith Chintala and Gregory Chanan and Edward Yang and Zach DeVito and Zeming Lin and Alban Desmaison and Luca Antiga and Adam Lerer}, year={2017} }
In this article, we describe an automatic differentiation module of PyTorch — a library designed to enable rapid research on machine learning models. It builds upon a few projects, most notably Lua Torch, Chainer, and HIPS Autograd [4], and provides a high performance environment with easy access to automatic differentiation of models executed on different devices (CPU and GPU). To make prototyping easier, PyTorch does not follow the symbolic approach used in many other deep learning frameworks…
10,068 Citations
Torchmeta: A Meta-Learning library for PyTorch
- Computer ScienceArXiv
- 2019
The constant introduction of standardized benchmarks in the literature has helped accelerating the recent advances in meta-learning research. They offer a way to get a fair comparison between…
Auto-Pytorch: Multi-Fidelity MetaLearning for Efficient and Robust AutoDL
- Computer ScienceIEEE Transactions on Pattern Analysis and Machine Intelligence
- 2021
This paper introduces Auto-PyTorch, which combines multi-fidelity optimization with portfolio construction for warmstarting and ensembling of deep neural networks (DNNs) and common baselines for tabular data to enable fully automated deep learning (AutoDL).
A Brief Introduction to Automatic Differentiation for Machine Learning
- Computer ScienceArXiv
- 2021
This report describes automatic differentiation, its motivations, and different implementation approaches, and briefly describes dataflow programming as it relates to AD.
An Introduction to Automatic Differentiation for Machine Learning
- Computer Science
- 2021
This report describes automatic differentiation, its motivations, and different implementation approaches, and briefly describes dataflow programming as it relates to AD.
PyTorch: An Imperative Style, High-Performance Deep Learning Library
- Computer ScienceNeurIPS
- 2019
This paper details the principles that drove the implementation of PyTorch and how they are reflected in its architecture, and explains how the careful and pragmatic implementation of the key components of its runtime enables them to work together to achieve compelling performance.
fastai: A Layered API for Deep Learning
- Computer ScienceInf.
- 2020
This paper has used this library to successfully create a complete deep learning course, which was able to write more quickly than using previous approaches, and the code was more clear.
Torchbearer: A Model Fitting Library for PyTorch
- Computer ScienceArXiv
- 2018
We introduce torchbearer, a model fitting library for pytorch aimed at researchers working on deep learning or differentiable programming. The torchbearer library provides a high level metric and…
Pixyz: a library for developing deep generative models
- Computer ScienceArXiv
- 2021
A new DGM library called Pixyz is proposed that is faster than existing probabilistic modeling languages in learning simple DGMs and can be used to implement complex DGMs in a simple and concise manner, which is difficult to do with existing libraries.
BackPACK: Packing more into backprop
- Computer ScienceICLR
- 2020
BackPACK is introduced, an efficient framework built on top of PyTorch that extends the backpropagation algorithm to extract additional information from first-and second-order derivatives to address the problem of automatic differentiation frameworks not supporting other quantities such as the variance of the mini-batch gradients.
Differentiation and Weighted Model Integration
- Computer Science
- 2019
This note shows how to perform gradient based optimization on discrete-continuous probabilistic models, expressed as weighted model integration problems, by means of auto-differentiation, which provides a new powerful tool for inference, learning and optimization on (deep) discrete-Continuous models.
References
SHOWING 1-7 OF 7 REFERENCES
Modeling, Inference and Optimization With Composable Differentiable Procedures
- Computer Science
- 2016
It is shown that early stopping and ensembling, popular tricks for avoiding overfitting, can be interpreted as variational Bayesian inference, and how to compute gradients of cross-validation loss with respect to hyperparameters of learning algorithms, with both time and memory efficiency.
DyNet: The Dynamic Neural Network Toolkit
- Computer ScienceArXiv
- 2017
DyNet is a toolkit for implementing neural network models based on dynamic declaration of network structure that has an optimized C++ backend and lightweight graph representation and is designed to allow users to implement their models in a way that is idiomatic in their preferred programming language.
TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems
- Computer ScienceArXiv
- 2016
The TensorFlow interface and an implementation of that interface that is built at Google are described, which has been used for conducting research and for deploying machine learning systems into production across more than a dozen areas of computer science and other fields.
Evaluating derivatives - principles and techniques of algorithmic differentiation, Second Edition
- BiologyFrontiers in applied mathematics
- 2000
This second edition has been updated and expanded to cover recent developments in applications and theory, including an elegant NP completeness argument by Uwe Naumann and a brief introduction to scarcity, a generalization of sparsity.
Compiling fast partial derivatives of functions given by algorithms
- Computer Science
- 1980
The system Jake described produces gradients significantly faster than numerical differencing for n > 8 and can handle algorithms Af with arbitrary flow of control.
TensorFlow: Large-scale machine learning on heterogeneous systems
- 2015