Viability Theorem for Deterministic Mean Field Type Control Systems

@article{Averboukh2016ViabilityTF,
  title={Viability Theorem for Deterministic Mean Field Type Control Systems},
  author={Yurii V. Averboukh},
  journal={Set-Valued and Variational Analysis},
  year={2016},
  volume={26},
  pages={993-1008}
}
  • Y. Averboukh
  • Published 31 December 2016
  • Mathematics
  • Set-Valued and Variational Analysis
A mean field type control system is a dynamical system in the Wasserstein space describing an evolution of a large population of agents with mean-field interaction under a control of a unique decision maker. We develop the viability theorem for the mean field type control system. To this end we introduce a set of tangent elements to the given set of probabilities. Each tangent element is a distribution on the tangent bundle of the phase space. The viability theorem for mean field type control… 

A stability property in mean field type differential games

  • Y. Averboukh
  • Mathematics
    Journal of Mathematical Analysis and Applications
  • 2021

Extremal Shift Rule and Viability Property for Mean Field-Type Control Systems

We investigate when a mean field-type control system can fulfill a given constraint. Namely, given a closed set of probability measures on the torus, starting from any initial probability measure

Impulsive control of nonlocal transport equations

Optimal control of nonlocal continuity equations: numerical solution

Nonlocal continuity equations on the spaces of probability measures arise as macroscopic mathematical models of multi-agent dynamical systems describing the time evolution of large ensembles (beams,

Nonlocal balance equations with parameters in the space of signed measures

A parametric family of nonlocal balance equations in the space of signed measures is studied. Under assumptions that cover a number of known conceptual models we establish the existence of the

Generalized dynamic programming principle and sparse mean-field control problems

Attainability property for a probabilistic target in wasserstein spaces

In this paper we establish an attainability result for the minimum time function of a control problem in the space of probability measures endowed with Wasserstein distance. The dynamics is provided

O C ] 1 A ug 2 01 9 Impulsive control of nonlocal transport equations

The paper extends an impulsive control-theoretical framework towards dynamical systems in the space of measures. We consider a transport equation describing the time-evolution of a conservative

Compatibility of state constraints and dynamics for multiagent control systems

The aim of the paper is to provide a necessary and sufficient condition for a given constraint to be compatible with the controlled continuity equation, and to obtain a new comparison theorem for evolutionary Hamilton–Jacobi equations in the Wasserstein space.

References

SHOWING 1-10 OF 30 REFERENCES

Dynamic Programming for Mean-Field Type Control

A Hamilton–Jacobi–Bellman fixed-point algorithm is compared to a steepest descent method issued from calculus of variations and an extended Bellman’s principle is derived by a different argument.

A General Stochastic Maximum Principle for SDEs of Mean-field Type

We study the optimal control for stochastic differential equations (SDEs) of mean-field type, in which the coefficients depend on the state of the solution process as well as of its expected value.

A Maximum Principle for SDEs of Mean-Field Type

We study the optimal control of a stochastic differential equation (SDE) of mean-field type, where the coefficients are allowed to depend on some functional of the law as well as the state of the

Viability with Probabilistic Knowledge of Initial Condition, Application to Optimal Control

In this paper we provide an extension of the Viability and Invariance Theorems in the Wasserstein metric space of probability measures with finite quadratic moments in ℝ d for controlled systems of

Forward-Backward Stochastic Differential Equations and Controlled McKean Vlasov Dynamics

The purpose of this paper is to provide a detailed probabilistic analysis of the optimal control of nonlinear stochastic dynamical systems of McKean–Vlasov type. Motivated by the recent interest in

Existence of optimal controls for systems governed by mean-field stochastic differential equations

In this paper, we study the existence of an optimal control for systems, governed by stochastic dierential equations of mean-eld type. For non linear systems, we prove the existence of an optimal

Control of McKean–Vlasov dynamics versus mean field games

We discuss and compare two investigation methods for the asymptotic regime of stochastic differential games with a finite number of players as the number of players tends to the infinity. These two

Bellman equation and viscosity solutions for mean-field stochastic control problem

We consider the stochastic optimal control problem of McKean-Vlasov stochastic differential equation where the coefficients may depend upon the joint law of the state and control. By using feedback

Limit Theory for Controlled McKean-Vlasov Dynamics

  • D. Lacker
  • Mathematics
    SIAM J. Control. Optim.
  • 2017
Along the way it is shown that a large class of McKean-Vlasov control problems admit optimal Markovian controls, and any distribution on the set of optimal control-state pairs for the McKeans Vlasov problem can be realized as a limit in this manner.

Dynamic Programming for Optimal Control of Stochastic McKean-Vlasov Dynamics

A dynamic programming principle for the value function in the Wasserstein space of probability measures, which is proved from a flow property of the conditional law of the controlled state process, and the viscosity property is proved together with a uniqueness result for thevalue function.