• Corpus ID: 235436267

Causal Navigation by Continuous-time Neural Networks

@inproceedings{Vorbach2021CausalNB,
  title={Causal Navigation by Continuous-time Neural Networks},
  author={Charles Vorbach and Ramin M. Hasani and Alexander Amini and Mathias Lechner and Daniela Rus},
  booktitle={NeurIPS},
  year={2021}
}
Imitation learning enables high-fidelity, vision-based learning of policies within rich, photorealistic environments. However, such techniques often rely on traditional discrete-time neural models and face difficulties in generalizing to domain shifts by failing to account for the causal relationships between the agent and the environment. In this paper, we propose a theoretical and experimental framework for learning causal representations using continuous-time neural networks, specifically… 

Figures and Tables from this paper

Latent Imagination Facilitates Zero-Shot Transfer in Autonomous Racing

TLDR
This paper investigates how model-based agents capable of learning in imagination substantially outperform model-free agents with respect to performance, sample efficiency, successful task completion, and generalization in real-world autonomous vehicle control tasks, where advanced model- free deep RL algorithms fail.

Closed-form Continuous-Depth Models

TLDR
This paper presents a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster while exhibiting equally strong modeling abilities compared to their ODE-based counterparts.

Closed-form Continuous-time Neural Models

TLDR
It is shown it is possible to closely approximate the interaction between neurons and synapses – the building blocks of natural and artificial neural networks – constructed by liquid time-constant networks (LTCs) efficiently in closed-form.

Sparse Flows: Pruning Continuous-depth Models

TLDR
This work designs a framework to decipher the internal dynamics of these continuous depth models by pruning their network architectures, and empirical results suggest that pruning improves generalization for neural ODEs in generative modeling.

BarrierNet: A Safety-Guaranteed Layer for Neural Networks

TLDR
These novel safety layers, termed a BarrierNet, can be used in conjunction with any neural network-based controller, and can be trained by gradient descent, which allows the safety constraints of a neural controller be adaptable to changing environments.

Entangled Residual Mappings

TLDR
While entangled mappings can preserve the iterative refinement of features across various deep models, they influence the representation learning process in convolutional networks differently than attention-based models and recurrent neural networks.

GoTube: Scalable Statistical Verification of Continuous-Depth Models

TLDR
A new statistical verification algorithm that formally quantifies the behavioral robustness of any time-continuous process formulated as a continuous-depth model, and is stable and sets the state-of-the-art in terms of its ability to scale to time horizons well beyond what has been previously possible.

GoTube: Scalable Stochastic Verification of Continuous-Depth Models

TLDR
This work introduces a new stochastic verification algorithm that formally quantifies the behavioral robustness of any timecontinuous process formulated as a continuous-depth model and calls it GoTube, which is stable and sets the state of the art in terms of its ability to scale to time horizons well beyond what has been previously possible.

References

SHOWING 1-10 OF 104 REFERENCES

Conditional Affordance Learning for Driving in Urban Environments

TLDR
This work proposes a direct perception approach which maps video input to intermediate representations suitable for autonomous navigation in complex urban environments given high-level directional inputs, and is the first to handle traffic lights and speed signs by using image-level labels only.

One-Shot Hierarchical Imitation Learning of Compound Visuomotor Tasks

TLDR
This work proposes a method that learns both how to learn primitive behaviors from video demonstrations and how to dynamically compose these behaviors to perform multi-stage tasks by "watching" a human demonstrator.

Model-based versus Model-free Deep Reinforcement Learning for Autonomous Racing Cars

TLDR
It is shown that model-based agents capable of learning in imagination, substantially outperform model-free agents with respect to performance, sample efficiency, successful task completion, and generalization, and that the generalization ability of model- based agents strongly depends on the observationmodel choice.

Deep Imitative Models for Flexible Inference, Planning, and Control

TLDR
This paper proposes Imitative Models, probabilistic predictive models of desirable behavior able to plan interpretable expert-like trajectories to achieve specified goals, and derives families of flexible goal objectives that can be used to successfully direct behavior.

Gershgorin Loss Stabilizes the Recurrent Neural Network Compartment of an End-to-end Robot Learning Scheme

TLDR
A new regularization loss component is introduced together with a learning algorithm that improves the stability of the learned autonomous system, by forcing the eigenvalues of the internal state updates of an LDS to be negative reals.

Learning by Cheating

TLDR
This work shows that this challenging learning problem can be simplified by decomposing it into two stages and uses the presented approach to train a vision-based autonomous driving system that substantially outperforms the state of the art on the CARLA benchmark and the recent NoCrash benchmark.

Learning to Control PDEs with Differentiable Physics

TLDR
It is shown that by using a differentiable PDE solver in conjunction with a novel predictor-corrector scheme, this work can train neural networks to understand and control complex nonlinear physical systems over long time frames.

Lipschitz Recurrent Neural Networks

TLDR
This work proposes a recurrent unit that describes the hidden state's evolution with two parts: a well-understood linear component plus a Lipschitz nonlinearity, which is more robust with respect to input and parameter perturbations as compared to other continuous-time RNNs.

Closed-form Continuous-Depth Models

TLDR
This paper presents a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster while exhibiting equally strong modeling abilities compared to their ODE-based counterparts.

Generative Adversarial Imitation Learning

TLDR
A new general framework for directly extracting a policy from data, as if it were obtained by reinforcement learning following inverse reinforcement learning, is proposed and a certain instantiation of this framework draws an analogy between imitation learning and generative adversarial networks.
...