1 Introduction
Mean field games (MFGs for short) have been introduced in the pioneering works of Lasry–Lions and Huang–Malhamé–Caines (see [Reference Lasry and Lions24, Reference Huang, Malhamé and Caines21]). The main motivation of both groups was to model strategic decision making in systems involving a large number of rational agents, arising from (stochastic) differential games. Ever since, this theory witnessed a great success, both from the theoretical viewpoint and the point of view of applications. We refer to [Reference Carmona and Delarue11, Reference Carmona and Delarue12, Reference Cardaliaguet and Porretta15] for a thorough, relatively up-to-date description of the evolution of this field, from the probabilistic and analytic aspects.
Already early on, Lions in his lecture series at Collège de France ([Reference Lions23]) has introduced the so-called master equation, associated to MFGs. This is a nonlocal and nonlinear PDE of hyperbolic type set on
$\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)$
, where
$\mathbb {R}^d$
models the state space of a typical agent, while
${\mathscr P}_2(\mathbb {R}^d)$
(the set of Borel probability measures with finite second moment, supported on
$\mathbb {R}^d$
) encodes the distribution of the agents. One of the main motivations for the solvability of the master equation is that it provides a deep link between games with finite, but large number of agents and the corresponding MFG: classical solutions to the master equation serve as great tools to obtain quantitative rates of convergence of closed loop Nash equilibria of games with finite number of agents, when the number of agents tends to infinity.
The master equation that we consider in this paper writes as follows. As data, we are given a Hamiltonian
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
and a final cost
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
. We emphasise that throughout the text we assume that H and G are smooth enough (we detail the specific assumptions later), and in particular they are defined and finite at any probability measure with finite second moment. Therefore, they will be assumed to be non-local and regularising in the measure variable. Furthermore, we are given a time horizon
$T>0$
and the intensities of the Brownian idiosyncratic and common noises
$\beta ,\beta _0\in \mathbb {R}$
, respectively. Then, the master equation, written for the unknown function
$V:(0,T)\times \mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
reads as
$$ \begin{align} \left\{ \begin{array}{rl} -\partial_t V(t,x,\mu) + H(x,\mu,\partial_x V) + \mathcal NV(t,x,\mu)\\ -\frac{\beta^2}{2} \Delta_{\mathrm{ind}}V - \frac{\beta_0^2}{2} \Delta_{\mathrm{com}} V(t,x,\mu) &= 0,\\[3pt] & \mathrm{{in}}\ (0,T)\times\mathbb{R}^d\times{\mathscr P}_2(\mathbb{R}^d),\\[3pt] V(T, x, \mu) &= G(x, \mu),\\[3pt] & \mathrm{{in}}\ \mathbb{R}^d\times{\mathscr P}_2(\mathbb{R}^d), \end{array} \right. \end{align} $$
where
$$\begin{align*}\mathcal NV(t,x,\mu) &= \int_{\mathbb{R}^d} \partial_\mu V(t,x,\mu,\tilde x) \cdot \partial_pH(\tilde x, \mu, \partial_x V(t, \tilde x, \mu)) {\mathrm{d}}\mu(\tilde x)\\ \Delta_{\mathrm{ind}}V &= \operatorname{\mathrm{tr}}(\partial_{x x} V(t,x,\mu)) + \int_{\mathbb{R}^d} \operatorname{\mathrm{tr}}(\partial_{\tilde x \mu} V(t,x,\mu,\tilde x)) {\mathrm{d}}\mu(\tilde x) \end{align*}$$
and
$$\begin{align*}\Delta_{\mathrm{com}}V &= \operatorname{\mathrm{tr}}(\partial_{x x} V(t,x,\mu)) + \int_{\mathbb{R}^d} \operatorname{\mathrm{tr}}(\partial_{\tilde x \mu} V(t,x,\mu,\tilde x)) {\mathrm{d}}\mu(\tilde x)\\ &\quad + 2 \int_{\mathbb{R}^d} \operatorname{\mathrm{tr}}(\partial_{x \mu} V(t,x,\mu,\tilde x)) {\mathrm{d}}\mu(\tilde x) \\ &\quad + \int_{\mathbb{R}^d\times\mathbb{R}^d} \operatorname{\mathrm{tr}}(\partial_{\mu \mu} V(t,x,\mu,\tilde x, \bar x)) {\mathrm{d}}\mu(\tilde x) {\mathrm{d}}\mu(\bar x). \end{align*}$$
Here
$\partial _\mu V$
stands for the so-called Wasserstein gradient whose definition is given later in the text.
The search for well-posedness theories for (1) has initiated a great program in the theory. In general, this poses great challenges because of the non-local and infinite-dimensional character of the PDE. In particular, this PDE does not possess a comparison principle which means that the consideration of viscosity solutions, for instance, would not be feasible in this setting. Therefore, notions of suitable weak solutions could lead to debates, especially if these lack uniqueness principles. However, there is no ambiguity regarding classical solutions. Our focus in this paper will also be on classical solutions, and so, unless otherwise specified, the term well-posedness should be understood in the sense of classical solutions. Similarly to the theory of finite-dimensional conservations laws, when aiming for global classical solutions, it is quite clear that these should be expected only under suitable monotonicity conditions on the data H and G. Such monotonicity conditions are also strongly related to the uniqueness of MFG Nash equilibria.
Literature review on the well-posedness of master equations. To date, there have been different notions of monotonicity conditions proposed on the data H and G, which could serve as sufficient conditions for the global well-posedness theory of (1). The diversity and richness of these conditions are deeply related to the geometry under the lens of which we look at
${\mathscr P}_2(\mathbb {R}^d)$
. For instance,
${\mathscr P}_2(\mathbb {R}^d)$
can be seen as a flat convex space, but it is natural to look at it also as a non-negatively curved infinite-dimensional manifold, when equipped with suitable metrics. Historically, the so-called Lasry–Lions (LL) monotonicity condition was the first one, introduced already in the seminal work [Reference Lasry and Lions24]. Geometrically, this is linked to the flat geometry, imposed on
${\mathscr P}_2(\mathbb {R}^d).$
When it comes to nonlocal Hamiltonians, this notion has been defined and exploited so far only for so-called separable Hamiltonians, that is, the ones which have the structure
for some
$H_0$
and F. An alternative monotonicity condition is the so-called displacement monotonicity condition, which does not require the separable structural assumption on H. This stems from the notion of displacement convexity, used widely in the context of optimal transport theory. Thus, this is linked to the curved geometry on
${\mathscr P}_2(\mathbb {R}^d)$
. We now give a brief overview of the well-posedness theories for (1) in these settings and we also mention some alternative, more recently proposed notions of monotonicity conditions.
In [Reference Carmona and Delarue12, Theorem 5.46] the authors have shown that the master equation (1) is globally well-posed if the data are LL monotone and possess additional regularity assumptions. Several other works provide similar conclusions. We refer to [Reference Cardaliaguet, Delarue, Lasry and Lions14, Theorem 2.4.5] for the case when the physical space is the flat torus instead of
$\mathbb {R}^d$
and to [Reference Chassagneux, Crisan and Delarue9, Theorems 56 and 58] to the case without common noise (i.e.,
$\beta _0=0$
). We refer also to [Reference Jakobsen and Rutkowski22] for new results and clarifications regarding the results from [Reference Cardaliaguet, Delarue, Lasry and Lions14]. However, [Reference Carmona and Delarue12, Theorem 5.46] is the closest result for our purposes.
It is also important to mention that all these global well-posedness results in the context of Lasry–Lions monotonicity impose both the separable structure on the Hamiltonian and the presence of a non-degenerate idiosyncratic noise.
In the context of displacement monotonicity global in time well-posedness have been obtained chronologically as follows. [Reference Gangbo and Mészáros16] provided this in the context of deterministic and potential (in particular
$\beta =\beta _0=0$
and H separable) games (for similar results, see also [Reference Bensoussan, Graber and Yam6]). [Reference Gangbo, Mészáros, Mou and Zhang19] provided the first global in time well-posedness result in the case of non-separable displacement monotone Hamiltonians and non-degenerate idiosyncratic noise (i.e.,
$\beta \neq 0$
). Finally, [Reference Bansil, Mészáros and Mou8] provided the result in the case of degenerate idiosyncratic noise (i.e.,
$\beta =0$
) and compared to [Reference Gangbo, Mészáros, Mou and Zhang19], under lower level regularity assumptions on the data, and the weaker version of the displacement monotonicity condition on H.
Recently, in [Reference Mou and Zhang26] and [Reference Mou and Zhang27] the authors have proposed a notion of anti-monotonicity condition on final data of master equations, which together with other sufficient structural conditions on the Hamiltonian resulted in the global in time well-posedness of the master equation. We would like to emphasise that for this to hold, the anti-monotonicity condition on the final data has to be carefully chosen in line with the structural conditions on the Hamiltonian. As we show below, this framework can entirely be embedded into our main results under the umbrella of our newly proposed canonical transformation.
Several other recent developments have seen the light in the context of the well-posedness of MFG master equations. For a non-exhaustive list we refer to [Reference Ambrose and Mészáros3, Reference Bertucci5, Reference Cardaliaguet, Cirant and Porretta10, Reference Cecchin and Delarue13, Reference Graber and Mészáros18, Reference Graber and Mészáros17].
Our contributions. In this paper our main objective is to explore some geometric features of Hamiltonian systems which could lead to the global well-posedness of the master equation (1). The heart of our analysis consist of so-called canonical transformations which in particular reveal new perspectives on existing and new monotonicity conditions on the Hamiltonians and final data associated to (1), and in turn lead to new well-posedness theories. The values of the noise intensities,
$\beta , \beta _0$
will not be significant in our consideration, and our main results hold true also for degenerate problems, that is, when
$\beta =0$
or
$\beta _0 = 0$
.
In classical Hamiltonian mechanics, canonical transformations are coordinate transformations on the phase space, which preserve the structure of Hamilton’s equations. In symplectic geometry, canonical transforms are known as symplectomorphisms (where the phase space is a cotangent bundle and the symplectic form is the canonical 2-form). Since in our setting we are only concerned with Euclidean space we do not use the symplectic terminology. However, it would be interesting to study how symplectomorphisms could potentially generate new well-posedness theories for Hamilton–Jacobi equations and the master equation in more general settings (i.e., when the underlying space is not Euclidean). We refer the reader to [Reference Arnol’d4] for an introduction to applications of symplectic geometry in classical mechanics. We refer also to our companion short note [Reference Bansil and Mészáros7], where we explain the regularisation effect of such transformations in the case of deterministic finite-dimensional HJB equations.
As the master equation has in particular a natural character arising from infinite-dimensional Hamiltonian dynamics, we will show below that such transformations play a deep role in revealing hidden features of it.
Let us describe the driving idea behind our results. For Hamiltonians
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
and final data
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
we consider a family of prototypical linear canonical transformations as follows. Let
$\alpha \in \mathbb {R}$
and define
$H_\alpha : \mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
and
$G_\alpha : \mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
as
In particular, this means that the corresponding canonical transformation has the form of
This is a ‘finite-dimensional’ transformation, as there is no change in the measure variable
$\mu $
. Having defined these transformations, the heart of our analysis is based on the following observation: fix any
$\alpha \in \mathbb {R}$
, then the master equation with data
$(H,G)$
is well-posed if and only if it is well-posed with data
$(H_\alpha , G_\alpha )$
(see Theorem 3.2; in particular the solutions to the corresponding master equations differ only by an explicit function of
$(t,x)$
, parametrised by
$\alpha $
).
The message of this result is that if one produces a well-posedness theory for the master equation, this will lead to a whole one parameter family of well-posedness theories, with the transformed data. A deeper consequence of this theorem is the opposite implication. Suppose that one is given the data
$(H,G)$
. If one is able to find a suitable range of the parameter
$\alpha $
such that
$(H_\alpha , G_\alpha )$
satisfies some well-known monotonicity conditions, then the problem with the original data must be well-posed. This second one will be the direction that we investigate in this paper.
Fix
$\alpha \in \mathbb {R}$
. It is easy to see that G is LL monotone, if and only if
$G_\alpha $
is LL monotone and the situation is the same for separable H. However, as we will show below, this phenomenon is much different in the displacement monotone regime. Therefore the previously described result has powerful applications in the context of displacement monotonicity but not for LL monotonicity.
In the main theorem of this paper, Theorem 3.6, we propose easily verifiable sufficient conditions on H to ensure that
$H_\alpha $
is displacement monotone. As a consequence, we discover new regimes of global well-posedness of the master equation. In an informal way, this result can be summarised as follows (we refer to Theorem 3.6 for the precise statement).
Theorem 1.1. Suppose that
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
is twice continuously differentiable with uniformly bounded second-order derivatives. Suppose moreover that H is strongly convex in the p-variable.
Suppose that the symmetric part of
$\partial _{xp}H$
is bounded below by an explicit quantity depending on the other second derivatives of H. Then,
$H_{\alpha }$
is displacement monotone for a suitable range of
$\alpha \in \mathbb {R}$
, depending on the size of the second derivatives of H in a precise way.
Furthermore, if
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
is twice continuously differentiable and displacement
$\alpha $
-monotone for such specific
$\alpha $
, then the master equation is globally well-posed.
This theorem has an immediate implication, coming from a sort of ‘regularisation phenomenon’ of
$\partial _{xp}H$
. This can informally be formulated as follows.
Corollary 1.2. Suppose that
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
and
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
are twice continuously differentiable with uniformly bounded second-order derivatives. Suppose moreover that H is strongly convex in the p-variable.
We have that there exists
$C>0$
depending on second derivatives of H and G (but independent of T) so that if
$\alpha \geq C$
, then the master equation is globally well-posed with data
$(\tilde H, G)$
, where
$\tilde H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
is given by
Hence even if we did not know that the original master equation was solvable, the modified master equation is solvable for
$\alpha $
large enough. One can compare the Hamiltonian
$\tilde H$
with the one in [Reference Mou and Zhang27, Example 7.2].
Remark 1.3. Corollary 1.2 has a deep message: if the Hamiltonian is such that
$\partial _{xp}H$
is sufficiently large compared to other second-order derivatives of H and the second-order derivatives of G, then we have a global well-posedness theory for the master equation. Therefore
$\partial _{xp}H$
, and in particular adding suitable multiples of the function
$(x,p,\mu )\mapsto p\cdot x$
to H can produce a ‘regularisation effect’ for the master equation, independently of
$T>0$
. By carefully examining Lemma 3.4, we see that what is going on is that the
$p\cdot x$
term is transformed into a multiple of
$\frac {\left |x\right |{}^2}2$
, which provides displacement monotonicity for the problem and hence regularises the master equation. It is easy to see that adding a suitable multiple of the term
$\frac {\left |x\right |{}^2}2$
to H produces displacement monotonicity. Clearly, these regularisation effects are independent of the noise intensities.
Remark 1.4. We emphasise that the regularisation provided by the function
$(x,p,\mu )\mapsto \alpha p\cdot x$
in the statement of Corollary 1.2 produces indeed a genuinely new class of data, not covered in the literature before, for which the master equation is globally well-posed. In particular, if we take an arbitrary pair of data
$(H,G)$
, not satisfying any monotonicity condition (either displacement or LL, if H is separable), it is immediate to check that
$\tilde H$
will satisfy neither displacement monotonicity nor LL monotonicity. Therefore, the monotonicity of the pair
$(\tilde H, G)$
is indeed hidden.
Further implications of our main results. Having our main results in hand, we have revisited some previous well-posedness results from the literature.
When G is displacement semi-monotone, then the well-posedness of (1) can be guaranteed if
$H_\alpha $
is displacement monotone for sufficiently large
$\alpha $
. It turns out that our characterisation for this given in Proposition 3.4 coincides with the respective assumptions on H discovered recently in [Reference Mou and Zhang26].
In the recent paper [Reference Mou and Zhang27], the authors proposed a notion of anti-monotonicity for final data G. They have described some sufficient conditions on H and G which result in a global well-posedness theory of (1), if
$\beta \neq 0$
, and G is suitably anti-monotone. There was an emphasis on the fact that G needed to be ‘sufficiently’ anti-monotone.
It turns out that these well-posedness results from [Reference Mou and Zhang27], under the additional assumptions that H is strictly convex in the p-variable fall directly into the framework of the canonical transformations and they are an easy consequence of our main results, in particular Corollary 1.2. More precisely, first in Proposition 3.8 we show that if G is
$\lambda $
-anti-monotone, this implies that it is displacement semi-monotone with a constant which depends only on
$\lambda $
(in particular, the displacement semi-monotonicity constant is independent of the second derivative bounds of G). Having strong convexity of H in the p-variable, which has also bounded second derivatives allows us to use our Corollary 1.2. The Hamiltonian considered in [Reference Mou and Zhang27] has the form of
for some constant matrix
$A_0\in \mathbb {R}^{d\times d}$
. This is slightly different than
$\tilde H$
from our Corollary 1.2, but the term
$\langle A_0p, x\rangle $
has exactly the same role as
$\alpha p\cdot x$
in our consideration. Therefore, for completeness, as our last contributions, in Proposition 3.12 and Remark 3.13 we show that the assumptions from the main theorem in [Reference Mou and Zhang27] essentially imply our assumptions. Furthermore, in the case of Hamiltonians which are strongly convex in the p-variable, our results need less and weaker assumption, and they hold true without the presence of a non-degenerate idiosyncratic noise. In particular, we demonstrate that the emphasis on the sufficient anti-monotonicity of G in [Reference Mou and Zhang27] is misleading, and this is not needed. Specifically, in [Reference Mou and Zhang27] it is remarked: ‘…we will need to require our data to be sufficiently anti-monotone in appropriate sense’. However we will see that anti-monotonicty is not needed (as anti-monotonicity implies semi-monotonicity) and that [Reference Mou and Zhang27] has other, more essential assumptions on H which are what really give the well-posedness result.
We would like to emphasise that in this paper we provide a general mechanism leading to a global well-posedness theory of master equations, beyond [Reference Mou and Zhang27], and the main results from this reference are a consequence of this general theory.
Some concluding remarks.
-
• For simplicity and transparency of our main ideas, in this manuscript we have decided to focus only on linear canonical transformations of the form
$\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\ni (x,\mu ,p)\mapsto (x,\mu , x-\alpha p).$
Without much philosophical effort but with significant technical effort, one could consider canonical transformations of the form where
$$ \begin{align*}\mathbb{R}^d\times{\mathscr P}_2(\mathbb{R}^d)\times\mathbb{R}^d\ni (x,\mu,p)\mapsto (x,\mu, x-\nabla \varphi(x)), \end{align*} $$
$\varphi :\mathbb {R}^d\to \mathbb {R}$
is any given smooth potential function, with bounded second derivatives. In the case of noise, this transformation would lead to the modified Hamiltonians and final data as and
$$ \begin{align*}H_\varphi(x,\mu,p): = H(x,\mu, x-\nabla \varphi(x)) + \frac{\beta^2+\beta_0^2}{2}\Delta\varphi(x)\end{align*} $$
It is easy to see that Theorem 3.2 holds true if in its statement
$$ \begin{align*}G_\varphi(x,\mu):= G(x,\mu) + \varphi(x). \end{align*} $$
$(H_\alpha ,G_\alpha )$
is replaced with
$(H_\varphi ,G_\varphi ).$
However, in order to obtain new global well-posedness theory (in the case of potentially degenerate noise), we would need to have a ‘convexifying regularisation’ on
$G_\varphi $
, which means that
$\varphi $
would need to be taken to be convex with sufficiently large Hessian eigenvalues. From this point of view,
$\varphi (x)=\frac {\alpha }{2}|x|^2$
would be a natural choice, and this is why we have decided to reduce our study to this particular family of potentials.
We remark that in general Hamiltonians are only defined up to an additive constant. In classical mechanics, this is saying that we may pick any value to correspond to the ‘zero energy’. In the presence of noise the attentive reader will notice that our
$H_\alpha $
is not the same as the
$H_\varphi $
defined above, when
$\varphi (x)$
is taken to be
$\frac {\alpha }{2}|x|^2$
. However, this is not an issue as the difference between the two is a constant. In particular, the two Hamiltonians are equivalent. Thus, we could have defined our
$H_\alpha $
as
$H_\alpha (x,\mu ,p):= H(x,\mu ,p-\alpha x) + {\frac {(\beta ^2+\beta _0^2)d}{2}}\alpha $
which would then be the exactly the same as
$H_\varphi $
defined above, however this would introduce unnecessary notational clutter. -
• In this paper we have considered only ‘finite-dimensional’ canonical transformations (where the measure component stayed fixed). These have proved to have a deep effect on new global well-posedness theories for the master equation. It is a very interesting, but seemingly challenging task to analyse truly infinite-dimensional canonical transformations in the context of MFG master equations. In particular it seems that the infinite-dimensional canonical transformations do not preserve the structure of MFG, they only preserve the structure of optimal control problems. In this we see a significant difference between games and variational problems.
Remark 1.5. If the Hamiltonian H has an associated Lagrangian with bounded second derivatives we must have that H is strongly convex in p. Similarly, the master equation only corresponds to a game, when H is convex in p. To the best of the authors’ knowledge there is no motivation for the master equation outside of this case.
We remark that if one is interested in the case of non-convex H in p then one can adapt our results by using the Hamiltonian system directly. We refer to the Lagrangian purely for pedagogical reasons and it is not needed for any technical reason. In particular our canonical transformation and main theorem, Theorem 3.2, holds regardless of the convexity of H in p.
2 Preliminaries
In order to keep this discussion self-contained, let us recall some definitions and notations.
Let
$p\ge 1$
. Based on [Reference Ambrosio, Gigli and Savaré1], we recall that the p-Wasserstein between
$\mu ,\nu \in {\mathscr P}_p(\mathbb {R}^d)$
(probability measures with finite p-order moment supported on
$\mathbb {R}^d$
) is defined as
$$ \begin{align*}W_p^p(\mu,\nu):=\inf\left\{\int_{\mathbb{R}^d\times\mathbb{R}^d}|x-y|^p{\mathrm{d}}\gamma(x,y):\ \gamma\in\Pi(\mu,\nu)\right\}, \end{align*} $$
where
$\Pi (\mu ,\nu ):=\left \{\gamma \in {\mathscr P}_p(\mathbb {R}^d\times \mathbb {R}^d):\ (p^x)_\sharp \gamma = \mu ,\ (p^y)_\sharp \gamma = \nu \right \}$
stands for the set of admissible transport plans in the transportation of
$\mu $
onto
$\nu $
, and
$p^x,p^y:\mathbb {R}^d\times \mathbb {R}^d\to \mathbb {R}^d$
denote the canonical projection operators, that is,
$p^x(a,b) = a$
and
$p^y(a,b) = b$
. We refer to the metric space
$({\mathscr P}_p(\mathbb {R}^d),W_p)$
as the Wasserstein space.
We refer to [Reference Ambrosio, Gigli and Savaré1, Reference Gangbo and Tudorascu20] and to [Reference Carmona and Delarue11, Chapter 5] for the notion of Wasserstein differentiability and fully
$C^k$
functions defined on the Wasserstein space, respectively. Based on [Reference Ahuja2, Reference Carmona and Delarue11, Reference Gangbo, Mészáros, Mou and Zhang19, Reference Mészáros and Mou25] we recall the notion of displacement monotonicity.
Definition 2.1. Let
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
be a fully
$C^1$
function.
-
1. We say that G is displacement monotone if
for any
$$ \begin{align*}\int_{\mathbb{R}^d\times\mathbb{R}^d} [\partial_xG(x,\mu)-\partial_x G(y,\nu)]\cdot (x-y){\mathrm{d}} \gamma(x,y)\ge 0, \end{align*} $$
$\gamma \in \Pi (\mu ,\nu )$
and for any
$\mu ,\nu \in {\mathscr P}_2(\mathbb {R}^d)$
. If G is more regular, say fully
$C^2$
, this definition is equivalent to for all
$$ \begin{align*} &\int_{\mathbb{R}^d}\langle\partial_{xx}G(x,\mu)\xi(x),\xi(x)\rangle{\mathrm{d}}\mu(x)\\ & + \int_{\mathbb{R}^d\times\mathbb{R}^d}\langle\partial_{x\mu}G(x,\mu,\tilde x)\xi(x),\xi(\tilde x)\rangle{\mathrm{d}}\mu(x){\mathrm{d}}\mu(\tilde x)\ge 0, \end{align*} $$
$\mu \in {\mathscr P}_2(\mathbb {R}^d)$
and for all
$\xi \in C_c(\mathbb {R}^d;\mathbb {R}^d)$
.
-
2. Based on [Reference Gangbo, Mészáros, Mou and Zhang19, Definition 2.7], we say that G is displacement semi-monotone or displacement
$\alpha $
-monotone, if there exists
$\alpha \in \mathbb {R}$
such that
$(x,\mu )\mapsto G(x,\mu )+\frac {\alpha }{2}|x|^2$
is displacement monotone.
For the corresponding Hamiltonians, we can define the displacement monotonicity condition as follows.
Definition 2.2. Let
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
be such that
$H(\cdot ,\mu ,\cdot )\in C^1(\mathbb {R}^d\times \mathbb {R}^d)$
for all
$\mu \in {\mathscr P}_2(\mathbb {R}^d)$
. We say that H is displacement monotone, if
$$ \begin{align} &-\int_{\mathbb{R}^d\times\mathbb{R}^d} [\partial_xH(x,\mu,p^1(x))-\partial_x H(y,\nu,p^2(y))]\cdot (x-y){\mathrm{d}} \gamma(x,y)\\ \nonumber&+ \int_{\mathbb{R}^d\times\mathbb{R}^d} [\partial_p H(x,\mu,p^1(x))-\partial_p H(y,\nu,p^2(x))]\cdot (p^1(x)-p^2(y)){\mathrm{d}} \gamma(x,y), \end{align} $$
for all
$\mu ,\nu \in {\mathscr P}_2(\mathbb {R}^d)$
,
$\gamma \in \Pi (\mu ,\nu )$
and for all
$p^1,p^2 \in C_b(\mathbb {R}^d;\mathbb {R}^d)$
.
Remark 2.3.
-
1. Suppose that
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
is fully
$C^2$
, strictly convex in the p-variable and satisfies (5)for all
$$ \begin{align} &\int_{\mathbb{R}^d\times\mathbb{R}^d}\left[\partial_{x\mu}H(x,\mu,\tilde x,p(x))v(\tilde x)+\partial_{xx}H(x,\mu,p(x))v(x)\right]\cdot v(x){\mathrm{d}}\mu(x){\mathrm{d}}\mu(\tilde x)\\ \nonumber&+\frac{1}{4}\int_{\mathbb{R}^d}\left\{\Big|[\partial_{pp}H(x,\mu,p(x))]^{-\frac12}\int_{\mathbb{R}^d}\partial_{p\mu}H(x,\mu,\tilde x,p(x))v(\tilde x){\mathrm{d}}\mu(\tilde x)\Big|^2\right\}{\mathrm{d}} \mu(x)\\ \nonumber&\le 0, \end{align} $$
$\mu \in {\mathscr P}_2(\mathbb {R}^d)$
, for all
$p\in C(\mathbb {R}^d;\mathbb {R}^d)$
and for all
$v\in L^2_{\mu }(\mathbb {R}^d;\mathbb {R}^d)$
. Then H satisfies the displacement monotonicity condition from Definition 2.2. For the proof of this fact we refer to [Reference Mészáros and Mou25, Lemma 2.7].
Definition 2.4 [Reference Mou and Zhang27, Definition 3.8], [Reference Mou and Zhang26, Definition 3.4].
Let
$\lambda = (\lambda _0,\lambda _1, \lambda _2,\lambda _3)\in \mathbb {R}^4$
be such that
$\lambda _0>0$
,
$\lambda _1\in \mathbb {R}, \lambda _2>0$
and
$\lambda _3\ge 0$
. Let
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
be fully
$C^2$
. It is said that G is
$\lambda $
-anti-monotone, if
$$ \begin{align*} &\lambda_0 \int_{\mathbb{R}^d}\langle\partial_{xx}G(x,\mu)\xi(x),\xi(x)\rangle{\mathrm{d}}\mu(x)\\ & + \lambda_1\int_{\mathbb{R}^d\times\mathbb{R}^d}\langle\partial_{x\mu}G(x,\mu,\tilde x)\xi(x),\xi(\tilde x)\rangle{\mathrm{d}}\mu(x){\mathrm{d}}\mu(\tilde x)\\ &+ \int_{\mathbb{R}^d}\left|\partial_{xx}G(x,\mu)\xi(x)\right|{}^2{\mathrm{d}}\mu(x) + \lambda_2 \int_{\mathbb{R}^d}\Big|\int_{\mathbb{R}^d}\partial_{x\mu}G(x,\mu,\tilde x)\xi(\tilde x){\mathrm{d}}\mu(\tilde x)\Big|^2 {\mathrm{d}} \mu(x) \\ & \leq \lambda_3\int_{\mathbb{R}^d} \left|\xi(x)\right|{}^2 {\mathrm{d}} \mu(x) \end{align*} $$
for all
$\mu \in {\mathscr P}_2(\mathbb {R}^d)$
and for all
$\xi \in L^2_{\mu }(\mathbb {R}^d;\mathbb {R}^d)$
.
3 New Well-Posedness theories for MFG and master equations
We impose a set of assumptions which are going to be imposed for our main results. These are relatively standard assumptions, which appear naturally in the literature on the well-posedness theories for master equations.
Assumption 1. Suppose that
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
is fully
$C^2$
, bounded below and is such that
-
•
$\partial _{xx}G$
is uniformly continuous and it is uniformly bounded by
$L^G$
on
$\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)$
; -
•
$\partial _{x\mu }G$
is uniformly continuous and it is uniformly bounded by
$L^G$
on
$\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d$
,
for some
$L^G>0$
.
Assumption 2. Suppose that
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
is fully
$C^2$
and satisfies the following:
-
•
$\partial _{pp}H$
is uniformly continuous and
$\partial _{pp}H(x,\mu ,p) \geq c_0^{-1}I$
, for some
$c_0>0$
and for all
$(x,\mu ,p)\in \mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d$
; -
•
$\partial _{xp}H$
,
$\partial _{pp}H, \partial _{xx}H$
are continuous and are uniformly bounded by
$L^H$
on
$\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d$
; -
•
$\partial _{p\mu }H, \partial _{x\mu }H$
are uniformly continuous and are uniformly bounded by
$L^H$
on
$\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\times \mathbb {R}^d$
; -
•
$\partial _p H(x,\mu ,p)\cdot p - H(x,\mu ,p)\ge -L^H$
for all
$(x,\mu ,p)\in \mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d,$
for some
$L^H>0$
.
Remark 3.1.
-
1. When continuity of functions is assumed in the measure variable, this is with respect to the
$W_2$
metric. -
2. Assumptions 1 and 2 from above are the standing assumptions imposed in [Reference Bansil, Mészáros and Mou8].
Let us now restate our crucial observation from the introduction in form of a theorem.
Theorem 3.2. Fix any
$\alpha \in \mathbb {R}$
. The master equation with data
$(H,G)$
is well-posed if and only if it is well-posed with data
$(H_\alpha , G_\alpha )$
.
Proof. Via direct computation we can verify that V is a solution of the master equation with data
$(H,G)$
if and only if
$\tilde V(t,x,\mu ) := V(t,x,\mu ) + \frac {\alpha }{2}\left |x\right |{}^2 - \frac {(\beta _0^2+\beta ^2)\alpha d}{2}(t-T)$
is a solution of the master equation with data
$({H_\alpha , G_\alpha })$
.
Remark 3.3. Because of the connection between the solvability of the master equation with data
$(H,G)$
and
$({H_\alpha , G_\alpha })$
described in Theorem 3.2, the same connection holds true for the solutions to the corresponding finite-dimensional mean field games systems as well.
Recall the definition (3). Now we give some sufficient conditions on Hamiltonians H which would result in the displacement monotonicity of the transformed Hamiltonians
$H_\alpha $
.
Lemma 3.4. Let H be fully
$C^2$
. Then
$H_\alpha $
is displacement monotone if and only if
$$ \begin{align} &\int_{\mathbb{R}^d}\left[\Big(\partial_{xx}H(x,\mu,p(x)) - 2\alpha\partial_{xp}H(x,\mu,p(x))\Big)v(x)\right]\cdot v(x){\mathrm{d}}\mu(x)\\ \nonumber &+ \int_{\mathbb{R}^d\times\mathbb{R}^d} \left[ \Big(\partial_{x\mu}H(x,\mu,\tilde x,p(x)) - 2\alpha\partial_{p\mu}H(x,\mu,\tilde x,p(x)) \Big)v(\tilde x)\right]\cdot v(x) {\mathrm{d}}\mu(x){\mathrm{d}}\mu(\tilde x) \\ \nonumber&+\frac{1}{4}\int_{\mathbb{R}^d}\bigg\{\bigg|[\partial_{pp}H(x,\mu,p(x))]^{-\frac12}\bigg[\int_{\mathbb{R}^d}\partial_{p\mu}H(x,\mu,\tilde x,p(x))v(\tilde x){\mathrm{d}}\mu(\tilde x)\\ \nonumber&+2\alpha\partial_{pp}H(x,\mu,p(x))v(x)\bigg]\bigg|^2\bigg\}{\mathrm{d}} \mu(x)\\ \nonumber&\le 0, \end{align} $$
for all
$\mu \in {\mathscr P}_2(\mathbb {R}^d)$
, for all
$p\in C(\mathbb {R}^d;\mathbb {R}^d)$
and for all
$v\in L^2_{\mu }(\mathbb {R}^d;\mathbb {R}^d)$
.
Proof. We readily compute
$$ \begin{align*} \partial_{xx} \tilde H(x,\mu, p) &= \partial_{xx} H(x,\mu,p-\alpha x) - 2\alpha \operatorname{\mathrm{Re}} (\partial_{xp} H(x,\mu,p-\alpha x))\\ &\quad + \alpha^2 \partial_{pp} H(x,\mu,p-\alpha x), \\ \partial_{x\mu} \tilde H(x,\mu,\cdot, p) &= \partial_{x\mu} H(x,\mu,\cdot,p-\alpha x) - \alpha \partial_{p\mu} H(x,\mu,\cdot,p-\alpha x),\\ \partial_{p\mu} \tilde H(x,\mu,\cdot, p) &= \partial_{p\mu} H(x,\mu,\cdot, p-\alpha x),\\ \partial_{pp} \tilde H(x,\mu, p) &= \partial_{pp} H(x,\mu, p-\alpha x). \end{align*} $$
The result now immediately follows by writing the inequality (5) for
$\tilde H$
in terms of H, after noting that we may replace
$\operatorname {\mathrm {Re}} (\partial _{xp} H)$
with
$\partial _{xp} H$
since the quadratic form induced by a skew-symmetric operator is null.
Remark 3.5. The inequality in (6) can be equivalently rewritten as
$$ \begin{align} &\int_{\mathbb{R}^d\times\mathbb{R}^d}\left[\partial_{x\mu}H(x,\mu,\tilde x,p(x))v(\tilde x) -\alpha \partial_{p\mu}H(x,\mu,\tilde x,p(x))v(\tilde x)\right]\cdot v(x){\mathrm{d}}\mu(x){\mathrm{d}}\mu(\tilde x)\\ \nonumber&+\int_{\mathbb{R}^d}\big[\partial_{xx}H(x,\mu,p(x))v(x)-2\alpha \partial_{xp}H(x,\mu,p(x))v(x)\\ \nonumber& + \alpha^2\partial_{pp}H(x,\mu,p(x))v(x) \big]\cdot v(x){\mathrm{d}}\mu(x)\\ \nonumber&+\frac{1}{4}\int_{\mathbb{R}^d}\left\{\Big|[\partial_{pp}H(x,\mu,p(x))]^{-\frac12}\int_{\mathbb{R}^d}\partial_{p\mu}H(x,\mu,\tilde x,p(x))v(\tilde x){\mathrm{d}}\mu(\tilde x)\Big|^2\right\}{\mathrm{d}} \mu(x)\\ \nonumber &\le 0, \end{align} $$
for all
$\mu \in {\mathscr P}_2(\mathbb {R}^d)$
, for all
$p\in C(\mathbb {R}^d;\mathbb {R}^d)$
and for all
$v\in L^2_{\mu }(\mathbb {R}^d;\mathbb {R}^d)$
. This is the exact same condition as [Reference Mou and Zhang26, (5.10)].
We introduce the following notations.
where for
$A\in \mathbb {R}^{d\times d}$
, we adopt the notation
$\operatorname {\mathrm {Re}} (A) := (A+A^\top )/2$
and for
$A\in \mathbb {R}^{d\times d}$
symmetric
$\lambda _{\min }(A)$
stands for its smallest eigenvalue. Furthermore, to denote the suprema of the standard
$2$
-matrix norms, we use the notation
$$ \begin{align*} &\left|\partial_{x\mu} H\right|:= \sup_{(x,\mu,p,\tilde x)\in\mathbb{R}^d\times{\mathscr P}_2(\mathbb{R}^d)\times\mathbb{R}^d\times\mathbb{R}^d}\left|\partial_{x\mu} H(x,\mu,p,\tilde x)\right|;\\ &\left|\partial_{p\mu} H\right|:= \sup_{(x,\mu,p,\tilde x)\in\mathbb{R}^d\times{\mathscr P}_2(\mathbb{R}^d)\times\mathbb{R}^d\times\mathbb{R}^d}\left|\partial_{p\mu} H(x,\mu,p,\tilde x)\right|;\\ &\left|\partial_{xx} H\right|:=\sup_{(x,\mu,p)\in\mathbb{R}^d\times{\mathscr P}_2(\mathbb{R}^d)\times\mathbb{R}^d}\left|\partial_{xx} H(x,\mu,p)\right|, \end{align*} $$
and so on for similar quantities. Now, we can formulate the second main result of our paper.
Theorem 3.6. Suppose that
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
satisfies
for some
$c_0>0$
and for all
$(x,\mu ,p)\in \mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d$
. Suppose that
$\underline {\kappa }(\partial _{xp}H), \left |\partial _{pp} H\right |, \left |\partial _{xx} H\right |, \left |\partial _{p\mu } H\right |$
and
$\left |\partial _{x\mu } H\right |$
are finite. Define
Suppose that
$\underline {\kappa }(\partial _{xp} H) \geq \frac 12\left |\partial _{p\mu }H\right | + \sqrt {\left |\partial _{pp} H\right | L_{our}^{H}}$
. Then
$H_\alpha $
is displacement monotone for any
where
$$ \begin{align*}\alpha^H_\pm:= \frac{\underline{\kappa}(\partial_{xp} H) - \frac12\left|\partial_{p\mu}H\right|\pm \sqrt{\left(\underline\kappa(\partial_{xp} H) - \frac12\left|\partial_{p\mu}H\right|\right)^2-\left|\partial_{pp} H\right| L_{our}^{H}}}{\left|\partial_{pp} H\right|}. \end{align*} $$
In particular we have the result for
$\alpha := \frac {\underline {\kappa }(\partial _{xp} H) - \frac 12\left |\partial _{p\mu }H\right |}{\left |\partial _{pp} H\right |}$
.
Proof. For
$\alpha \in \left [\alpha ^H_-,\alpha ^H_+\right ]$
,
$\mu \in {\mathscr P}_2(\mathbb {R}^d)$
,
$p\in C(\mathbb {R}^d;\mathbb {R}^d)$
and for
$v\in L^2_{\mu }(\mathbb {R}^d;\mathbb {R}^d)$
normalized, that is,
$\int _{\mathbb {R}^d}|v(x)|^2 d\mu =1$
, we compute
$$\begin{align*}&\int_{\mathbb{R}^d\times\mathbb{R}^d}\left[\partial_{x\mu}H(x,\mu,\tilde x,p(x))v(\tilde x) -\alpha \partial_{p\mu}H(x,\mu,\tilde x,p(x))v(\tilde x)\right]\cdot v(x){\mathrm{d}}\mu(x){\mathrm{d}}\mu(\tilde x) \\ & +\int_{\mathbb{R}^d}\big[\partial_{xx}H(x,\mu,p(x))v(x) -2\alpha \partial_{xp}H(x,\mu,p(x))v(x)\\ & + \alpha^2\partial_{pp}H(x,\mu,p(x))v(x) \big]\cdot v(x){\mathrm{d}}\mu(x)\\ \nonumber&+\frac{1}{4}\int_{\mathbb{R}^d}\left\{\Big|[\partial_{pp}H(x,\mu,p(x))]^{-\frac12}\int_{\mathbb{R}^d}\partial_{p\mu}H(x,\mu,\tilde x,p(x))v(\tilde x){\mathrm{d}}\mu(\tilde x)\Big|^2\right\}{\mathrm{d}} \mu(x) \\ &\leq \int_{\mathbb{R}^d\times\mathbb{R}^d}[\left|\partial_{x\mu}H\right| +\alpha \left|\partial_{p\mu}H\right| +\left|\partial_{xx}H\right| -2\alpha \underline{\kappa}(\partial_{xp}H) + \alpha^2\left|\partial_{pp}H\right| ]{\mathrm{d}}\mu(x){\mathrm{d}}\mu(\tilde x)\\ \nonumber&+\frac{c_0}{4}\int_{\mathbb{R}^d}\left\{\Big|\int_{\mathbb{R}^d}\left|\partial_{p\mu}H\right|{\mathrm{d}}\mu(\tilde x)\Big|^2\right\}{\mathrm{d}} \mu(x) \\ &= \left|\partial_{xx}H\right| -2\alpha \underline{\kappa}(\partial_{xp}H) + \alpha^2\left|\partial_{pp}H\right| + \left|\partial_{x\mu}H\right| +\alpha \left|\partial_{p\mu}H\right|+\frac{c_0\left|\partial_{p\mu}H\right|{}^2}{4} \\ &= \left|\partial_{pp}H\right|\alpha^2 -2\left(\underline{\kappa}(\partial_{xp}H) - \frac12\left|\partial_{p\mu}H\right|\right)\alpha + \left|\partial_{xx}H\right| + \left|\partial_{x\mu}H\right|+\frac{c_0\left|\partial_{p\mu}H\right|{}^2}{4} \\ &= \left|\partial_{pp}H\right|\alpha^2 -2 \left(\underline{\kappa}(\partial_{xp}H) - \frac12\left|\partial_{p\mu}H\right|\right)\alpha + L_{our}^{H} \\ & \le 0, \end{align*}$$
where in the last inequality we used the sign of the quadratic expression.
As an immediate consequence of Theorem 3.6, we have the well-posedness result in Corollary 1.2.
Proof of Corollary 1.2.
We see that all second-order derivatives of
$\tilde H$
and H match, except the ones involving
$\partial _{xp}$
, for which we have
By the uniform bounds on the corresponding second-order derivatives of H, we see that for
$\alpha $
sufficiently large,
$\tilde H$
fulfils the assumptions of Theorem 3.6. Increasing
$\alpha $
further if necessary, we can ensure that G is displacement
$\alpha $
-monotone. Having G displacement
$\alpha $
-monotone and
$H_\alpha $
displacement monotone would result via Theorem 3.2 in the desired global well-posedness result for the master equation.
3.1 Our results and previous results on the master equation involving displacement semi-monotone data
We notice that the inequality (6) is precisely the inequality (5.10) from [Reference Mou and Zhang26]. This means in particular that [Reference Mou and Zhang26, Theorem 5.6] is a direct consequence of Theorem 3.2 and Remark 3.4 above.
We note that Theorem 3.2 shows that we have a global well-posedness theory for the master equation as long as G is displacement semi-monotone and the corresponding
$\tilde H$
is displacement monotone. In particular, it is enough for these to satisfy the ‘first-order’ monotonicity conditions, in the sense of Definition 2.1(1) and (4). Therefore, Theorem 3.2 together with the well-posedness results from [Reference Bansil, Mészáros and Mou8] provide a more general result than the one in [Reference Mou and Zhang26, Theorem 5.6].
3.2 Our results and previous results on the master equation involving anti-monotone data
Our first objective in this subsection is to show that any function
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
which is
$\lambda $
-anti-monotone in the sense of Definition 2.4 is actually displacement
$\alpha $
-monotone in the sense of Definition 2.1(2), where
$\alpha $
can be computed explicitly in terms of
$\lambda =(\lambda _0,\lambda _1, \lambda _2,\lambda _3)$
. We start with some preparatory results.
Remark 3.7.
G is
$\lambda $
-anti-monotone in the sense of Definition 2.4 with
$\lambda =(\lambda _0,\lambda _1, \lambda _2,\lambda _3)$
if and only if
$$ \begin{align*} &\int_{\mathbb{R}^d} \bigg\{\left|\partial_{xx}G(x,\mu)\xi(x) + \frac {\lambda_0} 2 \xi(x)\right|^2\\&+ \lambda_2\left| \int_{\mathbb{R}^d} \partial_{x\mu}G(x,\mu,\tilde x)\xi(\tilde x){\mathrm{d}}\mu(\tilde x) + \frac {\lambda_1} {2\lambda_2} \xi(x)\right|^2\bigg\} {\mathrm{d}}\mu(x) \\ \nonumber&\leq \left(\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2\right)\int_{\mathbb{R}^d} \left|\xi(x)\right|{}^2 {\mathrm{d}} \mu(x). \end{align*} $$
Proof. This is immediate by an algebraic manipulation after computing the squares.
Proposition 3.8. If G is
$\lambda $
-anti monotone in the sense of Definition 2.4 with
$\lambda =(\lambda _0,\lambda _1, \lambda _2,\lambda _3)$
, then
$$\begin{align*}&\left|\int_{\mathbb{R}^d\times\mathbb{R}^d}\langle\partial_{x\mu}G(x,\mu,\tilde x)\xi(x),\xi(\tilde x)\rangle{\mathrm{d}}\mu(x){\mathrm{d}}\mu(\tilde x)\right|\\ &\qquad \qquad\leq \left(\frac {\left|\lambda_1\right|} {2\lambda_2} + \sqrt{\frac{\lambda_3}{\lambda_2} + \frac{{\lambda_0} ^2}{4\lambda_2} + {\left(\frac {\lambda_1} {2\lambda_2}\right)^2}} \right)\int_{\mathbb{R}^d} \left|\xi(x)\right|{}^2 {\mathrm{d}} \mu(x) \end{align*}$$
and
$$\begin{align*}&\left|\int_{\mathbb{R}^d} \langle \partial_{xx}G(x,\mu)\xi(x), \rangle{ \xi(x)}{\mathrm{d}}\mu(x)\right|\\ &\qquad \qquad \leq \left(\frac{\left|\lambda_0\right|}2 + \sqrt{\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2} \right) \int_{\mathbb{R}^d} {\mathrm{d}}\mu(x) \left|\xi(x)\right|{}^2. \end{align*}$$
In particular G is displacement
$\alpha _\lambda $
-monotone, with
$$ \begin{align*}\alpha_\lambda\ge\max\left\{\frac {\left|\lambda_1\right|} {2\lambda_2} + \sqrt{\frac{\lambda_3}{\lambda_2} + \frac{{\lambda_0} ^2}{4\lambda_2} + {\left(\frac {\lambda_1} {2\lambda_2}\right)^2}}; \frac{\left|\lambda_0\right|}2 + \sqrt{\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2}\right\}. \end{align*} $$
Proof. Let us recall that in the definition of
$\lambda $
-anti-monotonicity we have
$\lambda _0>0$
,
$\lambda _2>0$
,
$\lambda _3\ge 0$
and there is no sign restriction on
$\lambda _1$
.
First, let us suppose that
$\lambda _1 \neq 0$
.
Note that for any
$v, w\in \mathbb {R}^d$
and any
$C> 0$
we have
With the choice of
$v:=\frac {\lambda _1} {2\lambda _2} \xi (x)$
and
$w:=\int _{\mathbb {R}^d} \partial _{x\mu }G(x,\mu ,\tilde x)\xi (\tilde x){\mathrm {d}}\mu (\tilde x)$
, we obtain
$$\begin{align*}&\int_{\mathbb{R}^d} \left|\left\langle\int_{\mathbb{R}^d} \partial_{x\mu}G(x,\mu,\tilde x)\xi(\tilde x){\mathrm{d}}\mu(\tilde x),\frac {\lambda_1} {2\lambda_2} \xi(x)\right\rangle\right|{\mathrm{d}}\mu(x) \\&\leq \int_{\mathbb{R}^d} \bigg\{ \left(\frac {C}2 + 1\right) \left|\frac {\lambda_1} {2\lambda_2} \xi(x)\right|^2\\& + \frac{1}{2C}\left|\int_{\mathbb{R}^d} \partial_{x\mu}G(x,\mu,\tilde x)\xi(\tilde x){\mathrm{d}}\mu(\tilde x) + \frac {\lambda_1} {2\lambda_2} \xi(x)\right|^2\bigg\}{\mathrm{d}}\mu(x) \\&= \int_{\mathbb{R}^d} \bigg\{ \left(\frac {C}2 + 1\right) \left|\frac {\lambda_1} {2\lambda_2} \xi(x)\right|^2\\& + \frac{1}{2C\lambda_2} \left(\lambda_2 \left|\int_{\mathbb{R}^d} \partial_{x\mu}G(x,\mu,\tilde x)\xi(\tilde x){\mathrm{d}}\mu(\tilde x) + \frac {\lambda_1} {2\lambda_2} \xi(x)\right|^2 \right)\bigg\}{\mathrm{d}}\mu(x) \\&\leq \int_{\mathbb{R}^d} \bigg\{ \left(\frac {C}2 + 1\right) \left|\frac {\lambda_1} {2\lambda_2} \xi(x)\right|^2\\&+ \frac{1}{2C \lambda_2} \left(\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2\right) \left|\xi(x)\right|{}^2 \bigg\}{\mathrm{d}}\mu(x) \end{align*}$$
where the last inequality follows from Proposition 3.7. Hence,
$$\begin{align*}&\int_{\mathbb{R}^d} \left|\left\langle\int_{\mathbb{R}^d} \partial_{x\mu}G(x,\mu,\tilde x)\xi(\tilde x){\mathrm{d}}\mu(\tilde x),\xi(x)\right\rangle\right| {\mathrm{d}}\mu(x) \\ &\leq \left( \left(\frac {C}2 + 1\right) \frac {\left|\lambda_1\right|} {2\lambda_2} + \frac{1}{C {\left|\lambda_1\right|}} \left(\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2\right) \right)\int_{\mathbb{R}^d} \left|\xi(x)\right|{}^2 {\mathrm{d}} \mu(x) \\ &= \left( \frac {\left|\lambda_1\right|} {2\lambda_2} + \frac {C{\left|\lambda_1\right|}} {4\lambda_2} + \frac{1}{C {\left|\lambda_1\right|}} \left(\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2\right) \right)\int_{\mathbb{R}^d} \left|\xi(x)\right|{}^2 {\mathrm{d}} \mu(x). \end{align*}$$
We now take
$C = \frac {1}{\left |\lambda _1\right |}\sqrt {\left ({\lambda _3 + \left (\frac {\lambda _0} 2\right )^2 + \lambda _2\left (\frac {\lambda _1} {2\lambda _2}\right )^2}\right ){(4\lambda _2)}}$
to obtain
$$\begin{align*}&\int_{\mathbb{R}^d} \left|\left\langle\int_{\mathbb{R}^d} \partial_{x\mu}G(x,\mu,\tilde x)\xi(\tilde x){\mathrm{d}}\mu(\tilde x),\xi(x)\right\rangle\right| {\mathrm{d}}\mu(x) \\ &\leq \left(\frac {\left|\lambda_1\right|} {2\lambda_2} +2 \sqrt{\frac{\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2}{4\lambda_2}} \right)\int_{\mathbb{R}^d} \left|\xi(x)\right|{}^2 {\mathrm{d}} \mu(x) \\ &= \left(\frac {\left|\lambda_1\right|} {2\lambda_2} + \sqrt{\frac{\lambda_3}{\lambda_2} + \frac{{\lambda_0} ^2}{4\lambda_2} + {\left(\frac {\lambda_1} {2\lambda_2}\right)^2}} \right)\int_{\mathbb{R}^d} \left|\xi(x)\right|{}^2 {\mathrm{d}} \mu(x) \end{align*}$$
Now, as the left-hand side of this estimate is continuous at
$\lambda _1=0$
, we can send
$\lambda _1\to 0$
, and conclude the claim for general
$\lambda _1\in \mathbb {R}.$
In the same manner with the choice of
$v:=\frac {\lambda _0}2 \xi (x)$
and
$w:=\partial _{xx}G(x,\mu )\xi (x)$
, for
$C>0$
arbitrary we get
$$\begin{align*}&\int_{\mathbb{R}^d} \left|\langle \partial_{xx}G(x,\mu)\xi(x), \rangle{ \xi(x)}\right|{\mathrm{d}}\mu(x) \\ &\leq \frac 2{\left|\lambda_0\right|} \int_{\mathbb{R}^d} \left(\frac{C+2}2 \left|\frac{\lambda_0}2 \xi(x)\right|^2 + \frac{1}{2C} \left|\partial_{xx}G(x,\mu)\xi(x) + \frac {\lambda_0} 2 \xi(x)\right|^2 \right){\mathrm{d}}\mu(x) \\ &\leq \frac 2{\left|\lambda_0\right|} \left(\frac{C+2}2 \left(\frac{\lambda_0^2}4\right) + \frac{1}{2C} \left(\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2\right) \right) \int_{\mathbb{R}^d} {\mathrm{d}}\mu(x) \left|\xi(x)\right|{}^2 \\ &= \left(\frac{\left|\lambda_0\right|}2 + \frac{C \left|\lambda_0\right|}4 + \frac{1}{\left|\lambda_0\right|C} \left(\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2\right) \right) \int_{\mathbb{R}^d} {\mathrm{d}}\mu(x) \left|\xi(x)\right|{}^2 \end{align*}$$
By taking
$C = \frac 2{\left |\lambda _0\right |}\sqrt {\left ({\lambda _3 + \left (\frac {\lambda _0} 2\right )^2 + \lambda _2\left (\frac {\lambda _1} {2\lambda _2}\right )^2}\right )}$
we obtain the result.
Remark 3.9. In Proposition 3.8 we see that the estimates, and hence the conclusion regarding the displacement
$\alpha $
-monotonicity, hold true even for
$\lambda _0\le 0$
. Therefore, we might drop the requirement
$\lambda _0>0$
, and our claims from below will remain true.
Corollary 3.10. Let
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
be
$\lambda $
-anti-monotone which satisfies Assumption 1. Suppose that
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
satisfies Assumption 2 and it is such that
$H_{\alpha _\lambda }$
is displacement monotone, where the constant
$\alpha _\lambda $
is given in Proposition 3.8. Then, the master equation (1) with data
$(H,G)$
is globally well-posed.
We would like to conclude our paper by showing that, if H is strictly convex in the p-variable, then the main theorem on the global well-posedness of the master equation from [Reference Mou and Zhang27, Theorem 7.1] is a particular case of our main results from Corollary 3.10. For completeness, we informally state this here.
Theorem 3.11 [Reference Mou and Zhang27, Theorem 7.1].
Suppose that
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)$
is smooth enough with uniformly bounded second-, third- and fourth-order derivatives. Suppose that the Hamiltonian
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
has the specific factorisation
for a constant matrix
$A_0\in \mathbb {R}^{d\times d}$
and
$H_0:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
smooth enough. Suppose furthermore that G is
$\lambda $
-anti-monotone and that a special set of specific assumption take place jointly for
$\lambda =(\lambda _0,\lambda _1, \lambda _2,\lambda _3)$
, the matrix
$A_0$
and
$H_0$
. Then the master equation (1) is globally well-posed for any
$T>0$
, in the classical sense.
Proposition 3.12. Suppose that
$G:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\to \mathbb {R}$
is
$\lambda $
-anti monotone and satisfies Assumption 1. Suppose that
$H:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
is given by
with
$H_0:\mathbb {R}^d\times {\mathscr P}_2(\mathbb {R}^d)\times \mathbb {R}^d\to \mathbb {R}$
satisfying Assumption 2 and
$A_0\in \mathbb {R}^{d\times d}$
is a given constant matrix. Let
$K_H := c_0\left |\partial _{pp} H\right | = c_0\left |\partial _{pp} H_0\right |$
be the condition number of
$\partial _{pp} H$
. Suppose that
$$ \begin{align} \underline{\kappa}(A_0) \geq\max\left\{ \left(\frac72+\frac{\sqrt{K_H}}{2}\right) L_2^{H_0} + \sqrt{\left|\partial_{pp}H\right| \left|\partial_{xx}H_0\right| }; \left({\frac32} + f(\lambda)\right)L_2^{H_0}\right\}, \end{align} $$
where
$\lambda = (\lambda _0,\lambda _1,\lambda _2,\lambda _2)$
, we have set
$$ \begin{align*} f(\lambda) &:= \frac {5{\left|\lambda_1\right|}} {4\lambda_2} + 1 + {\frac{\lambda_3}{2\lambda_2}} + \frac{\lambda_0}{4\lambda_2} + \frac{5\lambda_0}{4}+ \frac{\lambda_3}2 + \frac{{\left|\lambda_1\right|}}{4}\\ & = 1 + \frac12\left(\frac{5\lambda_0}{2}+ \frac{{\left|\lambda_1\right|}}{2} + \lambda_3\right) + \frac{1}{2\lambda_2}\left(\frac{\lambda_0}{2} + \frac {5{\left|\lambda_1\right|}} {2} + \lambda_3\right), \end{align*} $$
and
$L_2^{H_0}>0$
is a constant associated to
$H_0$
, satisfying
Then the master equation is globally well-posed.
Proof. Let us note that by the definition of
$L_2^{H_0}$
and by the definition of
$L^{H_0}_{our}$
, we have that
As
$\underline {\kappa }(\partial _{xp}H) \geq \underline {\kappa }(A_0) - \left |\partial _{xp} H_0\right |$
, we see that the assumption
$\underline {\kappa }(A_0) \geq (\frac 72+\frac {\sqrt {K_H}}{2}) L_2^{H_0} + \sqrt {\left |\partial _{pp}H\right | \left |\partial _{xx}H_0\right | }$
and (9) imply
$$ \begin{align*} \underline{\kappa}({\partial_{xp}H})& \geq \underline{\kappa}(A_0) - \left|\partial_{xp} H_0\right| \ge 3 L_2^{H_0} + \frac12\left|\partial_{p\mu}H\right| - \left|\partial_{xp} H_0\right| + \frac{\sqrt{K_H}}{2} L_2^{H_0}\\ & + \sqrt{\left|\partial_{pp}H\right| \left|\partial_{xx}H_0\right| } \\ &\ge 2 L_2^{H_0} + \frac12\left|\partial_{p\mu}H\right| + \frac{\sqrt{K_H}}{2} L_2^{H_0} + \sqrt{\left|\partial_{pp}H\right| \left|\partial_{xx}H_0\right| }\\ & = \frac12\left|\partial_{p\mu}H\right| + \sqrt{\frac{c_0}{4}|\partial_{pp}H_0|\left(L_2^{H_0}\right)^2}+ \sqrt{4\left( L_2^{H_0}\right)^2} + \sqrt{\left|\partial_{pp}H\right| \left|\partial_{xx}H_0\right| }\\ &\ge \frac12\left|\partial_{p\mu}H\right| + \sqrt{\frac{c_0}{4}|\partial_{pp}H_0|\left(L_2^{H_0}\right)^2}+ \sqrt{4|\partial_{pp}H_0| L_2^{H_0}} + \sqrt{\left|\partial_{pp}H\right| \left|\partial_{xx}H_0\right| }\\ &\geq \frac12\left|\partial_{p\mu}H_0\right| + \sqrt{\left|\partial_{pp} H_0\right| L_{our}^{H_0}}\\ &= \frac12\left|\partial_{p\mu}H\right| + \sqrt{\left|\partial_{pp} H\right| L_{our}^{H}} \end{align*} $$
and so we can apply Theorem 3.6. We get that H is displacement
$\alpha $
-monotone with
$$ \begin{align*} \alpha &= \frac{\underline{\kappa}(\partial_{xp} {H}) - \frac12 \left|\partial_{p\mu}H\right|}{\left|\partial_{pp} H\right|}\\ &\ge { \frac{\underline{\kappa}(A_0) - |\partial_{xp}H_0| - \frac12 \left|\partial_{p\mu}H_0\right|}{\left|\partial_{pp} H_0\right|}}\\ &\ge { \frac{ \left(\frac32 + f(\lambda)\right)L_2^{H_0}- |\partial_{xp}H_0| - \frac12 \left|\partial_{p\mu}H_0\right|}{\left|\partial_{pp} H_0\right|}}\\ & \geq f(\lambda). \end{align*} $$
From Proposition 3.8 we see that G is semi-monotone with constant
$$\begin{align*}\eta &:= \frac {{\left|\lambda_1\right|}} {2\lambda_2} + \sqrt{\frac{\lambda_3}{\lambda_2} + \frac{{\lambda_0} ^2}{4\lambda_2} + {\left(\frac {\lambda_1} {2\lambda_2}\right)^2}} + \frac{\lambda_0}2 + \sqrt{\lambda_3 + \left(\frac {\lambda_0} 2\right)^2 + \lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2} \\ &\leq \frac {{\left|\lambda_1\right|}} {2\lambda_2} + \sqrt{\frac{\lambda_3}{\lambda_2}} + \sqrt{\frac{{\lambda_0} ^2}{4\lambda_2}} + \sqrt{{\left(\frac {\lambda_1} {2\lambda_2}\right)^2}} + \frac{\lambda_0}2 + \sqrt{\lambda_3} + \sqrt{\left(\frac {\lambda_0} 2\right)^2} + \sqrt{\lambda_2\left(\frac {\lambda_1} {2\lambda_2}\right)^2} \\ &\leq \frac {{\left|\lambda_1\right|}} {\lambda_2} + \sqrt{\frac{\lambda_3}{\lambda_2}} + \sqrt{\frac{{\lambda_0} ^2}{4\lambda_2}} + {\lambda_0} + \sqrt{\lambda_3} + \sqrt{\frac {\lambda_1^2} {4\lambda_2}} \\ &\leq \frac {{\left|\lambda_1\right|}} {\lambda_2} + \frac12 + {\frac{\lambda_3}{2\lambda_2}} + \frac{\lambda_0}{4\lambda_2} + \frac{\lambda_0}{4}+ {\lambda_0} + \frac12 + \frac{\lambda_3}2 + {\frac {{\left|\lambda_1\right|}} {4\lambda_2}} + \frac{{\left|\lambda_1\right|}}{4}\\ &= \frac {5{\left|\lambda_1\right|}} {4\lambda_2} + 1 + {\frac{\lambda_3}{2\lambda_2}} + \frac{\lambda_0}{4\lambda_2} + \frac{5\lambda_0}{4}+ \frac{\lambda_3}2 + \frac{{\left|\lambda_1\right|}}{4} \\ &= f(\lambda) \end{align*}$$
and so the result follows.
Remark 3.13. We compare Proposition 3.12 with [Reference Mou and Zhang27, Theorem 7.1]. This theorem has many assumptions. We show that up to constants (depending only on
$K_H$
) only a few of these many assumptions imply our assumptions. First, we recall that the definition of the
$3\times 3$
matrices
$A_1,A_2$
from formula [Reference Mou and Zhang27, (4.3)]. These are not constructed from
$A_0$
above, and they involve constants coming in particular from
$\lambda =(\lambda _0,\lambda _1,\lambda _2,\lambda _3).$
Furthermore, for
$A\in \mathbb {R}^{d\times d}$
,
$\bar \kappa (A)$
stands for the largest eigenvalue of
$\operatorname {\mathrm {Re}}(A).$
To continue we need the assumption
In [Reference Mou and Zhang27, Theorem 7.1] (specifically the second item of (7.1)) it is assumed that
although they probably meant to assume (10)Footnote 1 .
We can formulate the following statement.
Claim.
The assumptions of [Reference Mou and Zhang27, Theorem 7.1], up to a multiplicative constant depending on
$K_H$
, imply (8).
Proof of claim.
By definition, we have that
$\bar {\kappa }(A_1^{-1}A_2) \geq v^\top A_1^{-1}A_2 v$
for any unit vector
$v\in \mathbb {R}^3$
. Taking
$v = \frac {1}{\sqrt 3}(1,1,1)^\top $
and using the explicit form of
$A_1, A_2$
given in [Reference Mou and Zhang27, (4.3)] together with the fact that all the entries of these matrices are non-negative, by direct computation we obtain
$$\begin{align*}\bar{\kappa}(A_1^{-1}A_2) &\geq \frac{1}{3} \left( \frac14 \left( \lambda_0 + \lambda_0 + \left|\lambda_0 - \frac12 \lambda_1\right| + \lambda_3 \right) + \frac1{{2}\lambda_2} \left( \lambda_0 + \left|\lambda_1\right| + (\frac12 \left|\lambda_1\right| + \lambda_2 + \lambda_3) \right) \right) \\ &\geq \frac{1}{3} \left( \frac14 \left( \lambda_0 + \lambda_0 - \left|\lambda_0\right| + \frac12 \left|\lambda_1\right| + \lambda_3 \right) + \frac1{{2}\lambda_2} \left( \lambda_0 + \left|\lambda_1\right| + (\frac12 \left|\lambda_1\right| + \lambda_2 + \lambda_3) \right) \right) \\ &= \frac{1}{3} \left( \frac14 \left( \lambda_0 + \frac12 \left|\lambda_1\right| + \lambda_3 \right) + \frac1{{2}\lambda_2} \left( \lambda_0 + \left|\lambda_1\right| + (\frac12 \left|\lambda_1\right| + \lambda_2 + \lambda_3) \right) \right) \\ &\geq \frac{1}{{15}} f(\lambda), \end{align*}$$
so (10) implies that
Furthermore we see from the second inequality in [Reference Mou and Zhang27, (7.2)] that
By the assumption (i) of [Reference Mou and Zhang27, Theorem 7.1] we have that
$\bar {\gamma }$
satisfies [Reference Mou and Zhang27, (4.2)] in which the first inequality implies that
$\lambda _0> \frac {\bar {\gamma }^2}{4 \underline {\gamma }} - \frac {8\lambda _3}{4\underline {\gamma }}$
. Hence we obtain
$(4 \underline {\gamma } \lambda _0 + 8 \lambda _3) \geq \bar {\gamma }^2$
. It is clear that
$2f(\lambda ) \geq \lambda _0$
and
$2f(\lambda ) \geq \lambda _3$
, therefore we get
$16f(\lambda )(1 + \underline {\gamma }) \geq \bar {\gamma }^2$
. Since
$\underline {\gamma } < \bar {\gamma }$
by assumption (i) of [Reference Mou and Zhang27, Theorem 7.1] and
$1 < \bar {\gamma }$
by the same assumption we get
$2\bar {\gamma } \geq 1 + \underline {\gamma }$
and so we obtain
$32f(\lambda ) \geq \bar {\gamma }$
. Hence we get
$$\begin{align*}\underline{\kappa}(A_0)^2 \geq \frac{L_2^{H_0}}{15} f(\lambda) \underline{\kappa}(A_0) \geq \frac{L_2^{H_0}}{{15\cdot 32}} \bar{\gamma} \underline{\kappa}(A_0) \geq \frac{\left|\partial_{pp} H\right|}{{15\cdot 32}} \left|\partial_{xx} H\right| \end{align*}$$
and so we obtain
$\underline {\kappa }(A_0) \geq \frac {1}{{4\sqrt {30}}} \sqrt {\left |\partial _{pp} H\right | \left |\partial _{xx} H\right |}$
.
Moreover, (10) implies that
$\underline {\kappa }(A_0) \geq L_2^{H_0}$
and so we get
To summarise, the assumptions of [Reference Mou and Zhang27, Theorem 7.1] imply (11) and (12) which in turn imply that
$$\begin{align*}\underline{\kappa}(A_0) \geq \frac{1}{{8\sqrt{30}} + \sqrt{K_H}}\max\left\{ \left(\frac72+\frac{\sqrt{K_H}}{2}\right) L_2^{H_0} + \sqrt{\left|\partial_{pp}H\right| \left|\partial_{xx}H_0\right| }; \left({\frac32} + f(\lambda)\right)L_2^{H_0}\right\}. \end{align*}$$
This, aside from the constant of
$\frac {1}{{8\sqrt {30}} + \sqrt {K_H}}$
in front, is the exact assumption (8) of our Proposition 3.12.
Acknowledgements
The authors are grateful to Wilfrid Gangbo for valuable remarks and constructive comments.
Data availability statement
No data was generated for the purposes of this research.
Competing interests
The authors have no competing interest to declare.
Funding statement
MB’s work was supported by the National Science Foundation Graduate Research Fellowship under Grant No. DGE-1650604 and by the Air Force Office of Scientific Research under Award No. FA9550-18-1-0502. ARM has been partially supported by the EPSRC New Investigator Award ‘Mean Field Games and Master equations’ under award no. EP/X020320/1 and by the King Abdullah University of Science and Technology Research Funding (KRF) under award no. ORA-2021-CRG10-4674.2. Both authors acknowledge the partial support of the Heilbronn Institute for Mathematical Research and the UKRI/EPSRC Additional Funding Programme for Mathematical Sciences through the focused research grant ‘The master equation in Mean Field Games’.