Professional Documents
Culture Documents
field games
Introduction
Mean field games have become a powerful mathematical tool to model the dynamics of agents in economics, finance and the social sciences. Different settings have been considered in the literature such as discrete and continuous
in time or finite and continuous state space. Originally finite state mean field
games, see [Lio11, GMS10, Gue11a, Gue11b, GMS13, FG13], were studied as
an attempt to understand the more general continuous state problems introduced by Lasry & Lions in [LL06a, LL06b, LL07] as well as by Huang et al. in
[HMC06, HCM07]. For additional information see also the recent surveys such
as [LLG10, Car11, Ach13, GS13, BFY13].
In this paper we apply finite state mean field games to two classical problems
in socio-economic sciences: consumer choice behaviour and paradigm shifts in
a scientific community, see [BD14]. These mean field game models give rise
to systems of hyperbolic partial differential equations, for which we develop a
new numerical method. The mathematical modelling is based on the following
situation. Let us consider a system of N + 1 identical players or agents, which
can switch between d N different states. Each player is in a state i I =
{1, . . . , d} and can choose a switching strategy to any other state j I. The
only information available to each player, in addition to its own state, is the
number nj of players he/she sees in the different states j I. The fraction of
players in each state i I is denoted by i = nNi and
P we define the probability
vector = (1 , . . . , d ), P(I) = { Rd :
i i = 1, i 0}, which
encodes the statistical information about the ensemble of players. Each player
faces an optimisation problem over all possible switching strategies. Here the
key question is the existence of a Nash equilibrium and to determine the limit
as the number of players tends to infinity. Gomes et al. showed in [GMS13]
1
We start with a more detailed presentation of finite state mean field games and
the formal derivation of (1).
2.1
N + 1 player problem
We assume for the moment that all players (except the reference player) have
d
chosen an identical Markovian strategy (n, i, t) (R+
0 ) . The reference player
d
is allowed to choose a possibly different strategy (n, i, t) (R+
0 ) . Given
these strategies, the joint process (it , nt ) is a time inhomogeneous Markov chain
(neither nt nor it are Markovian when considered separately). The generator of
this process can be written as
i
i
i
Ain = A
0 n + A1 n + A2 n ,
where the definition of each term will be given in what follows. Let ek be the
k th vector of the canonical basis of Rd and ejk = ej ek . Then
X
i
j jn in ,
A
0 n =
jI
A1 in =
j,kI
n,i
,jk
in+ejk in ,
and
i
A
2 n =
X n
X jk nj nk
ij j
in+ejk + in+ekj 2in +
jn + in+eij 2in .
2
2
N
N
jI
j,kI
The terms A
0 and A1 correspond to the transitions due to the switching strategies and . Select one of the players distinct from the reference player. Denote
by kt its position at time t, and call mt the vector mt = nt + eit kt , the process
that records the number of other players in any state from the point of view of
this player. Suppose further that there are no interactions (A
2 = 0). Then for
j 6= k, we have
P kt+ = jkmt = m, kt = k = j (m, k, t) + o() .
(2)
The transitions between different states due to interactions give rise to the term
A
2 . Its particular structure comes from the assumption that any two players
with distinct states j and k can meet with rate Njk (with kj = jk 0). As a
result of this interaction, either both end in state j or state k (with probability
1
2 respectively).
We assume that all players have the same running cost determined by a function
d
c : I P(I) (R+
0 ) R as well as an identical terminal cost (), which is
3
(3)
c(i, , )
.
kk
Let us fix a reference player and set the Markovian strategy for the remaining
N players. The objective of the reference player is to minimise its total cost,
whose minimum over all Markovian strategies is given by
#
"Z
n
T
ns
T
,
iT
i,
.
(4)
un (t) = inf E(it ,nt )=(i,n)
c is , , (s) ds +
N
N
t
We define i n (t) = 1n (t) in (t), . . . , dn (t) in (t) . The generalised Legendre transform of c is given by
h(z, , i) =
min c(i, , ) + i z.
d
(R+
0 )
(5)
Note that h only depends on the differences between coordinates of the variable
z, that is, if i z = i z then h(z, , i) = h(
z , , i).
The function ui,
is
the
solution
to
the
ODE
n
ui,
n
n
i,
= h ui,
,
, i + A1 ui,
n
n + A2 un .
t
N
(6)
d
(R+
0 )
If h is differentiable, for j 6= i,
j (i z, , i) =
h (i z, , i)
.
z j
(7)
(8)
(n, i, t) = uin , , i .
N
X n,i
uin
n
= h uin , , i +
jk uin+ejk uin
(9)
t
N
j,kI
X n
X jk nj nk
ij j
i
i
i
i
j
i
2u
u
+
u
2u
+
u
+
u
+
n ,
n+ejk
n+ekj
n
n
n+eij
N2
N2
jI
j,kI
where we define
2.2
n,i
jk
(t) = nk
j (n + eik , k, t) .
(10)
j,kI
i
A
2 un =
h
1+
i k
N
i
j (k U, , k)
(j k ) +
(j k )2
2N
Ui + O
1
N2
X
X jk
ij (U j U i )
j k 2j j U i + 2k k U i 22j k U i +
N
jI
j,kI
X
1
.
+
ij j i U i j U i + O
N2
jI
2
We observe that for fixed j and k the operators k 2 j j k , and
2
jk j k j k are degenerate elliptic operators. The first one is degenerate because k , j 0; the second because jk 0. Hence their sum is also
a degenerate operator. Therefore the combination of the second order terms in
the expansion of A
1 and A2 can be written as
X
l,mI
blm 2l m =
X k j + 2jk j k
2
j k
j,kI
2
jI
1 X
blm (U, ) 2l m U i ,
N
l,mI
(11)
This implies that the limit of (11) is (1), which does not depend on the interaction between players (jk ). Note that
X
X X
X X
gj (U, ) =
i j (U, , i) =
i
j (U, , i) = 0,
(13)
jI
since
jI
jI iI
iI
jI
(14)
2.3
Next we consider a special class of mean field games, in which system (1) can
be written as the gradient of a Hamilton-Jacobi equation. Suppose that
i) + f (i, ),
h(u, , i) = h(u,
i I,
(15)
(16)
kI
(, t) = H ( , ) ,
t
(17)
(, T ) = 0 ().
k,jI
jI
In this section we present several applications of finite state mean field games to
socio-economic sciences. In order to keep the presentation simple we consider
only two-state problems. We start by stating the explicit equations, where
agents can choose between two options. The classical examples discussed here
are the consumer choice behaviour and a paradigm shift model in the scientific
community. Note that the number of choices can be increased, but we focus on
two-state applications for reasons of clarity and readability.
3.1
Two-state problems
Consider a two-state mean field game, where the fraction of players in either
state, 1 or 2, is given by i , i = 1, 2 with 1 + 2 = 1, and i 0. Since the limit
equation (1) does not depend on the interactions (although the N + 1 player
model does), we set = 0. Note that 6= 0 would result in different numerical
methods (and potentially different solutions) for (1). We suppose further that
the running cost c = c(i, , ) in (4) depends quadratically on the switching rate
, i.e.
2
1X 2
j .
2
(18)
j6=i
Then
h(z, , 1) = f (1, )
2
2
1
1
(z 1 z 2 )+ and h(z, , 2) = f (2, )
(z 2 z 1 )+ .
2
2
(19)
1
f (2, ) + 21 + ( 12 )
2
0
z 2 z 1
z 1 z 2
0
2 (z, , 1) = (z 1 z 2 )+ ,
1 (z, , 2) = (z 2 z 1 )+ .
Since
+
1 (U, , 1) = U 1 U 2 ;
+
1 (U, , 2) = U 2 U 1 ;
2 (U, , 1) =
U1 U2
2 (U, , 2) = U 2 U
+
1 +
;
,
(20)
H(p, ) = F ()
1 ((p1 p2 )+ ) + 2 ((p2 p1 )+ )
.
2
(21)
2
X
du1n
n
n,1
1
1
u
+
h
u
,
,1 ,
1
n
n+e
n
jl
jl
dt
N
j,l=1
(22)
2
n
du2n
n,2
2
2
+
h
u
,
=
u
,2 ,
2
n
n+e
n
jl
jl
dt
N
j,l=1
1
du1n
12
= (N n1 ) 1 2 un+e12 , n+e
un+e12 u1n
N ,2
dt
n
n
, 1 u1n+e21 u1n + h 1 un , N
,1 ,
+ n1 2 1 un , N
du2n
, 2 u2n+e12 u2n
= (N n1 ) 1 2 un , N
dt
2
n
21
+ n1 2 1 un+e21 , n+e
un+e21 u2n + h 2 un , N
,2 .
N ,1
(23)
Since 1 + 2 = 1 we use = (, 1 ), for [0, 1]. We split the domain [0, 1]
k
into N equidistant subintervals and define k = N
, 0 k N, k N. The
variable k corresponds to the fraction of players in state 1. Then the fraction
of players in state 2 is given by 1 k = NNk . Consequently (23) takes the form
du1
dt
du2
dt
= N (1 k ) u2k+1 u1k+1
+N k u1k u2k
+
+
u1k+1 u1k
u1k1 u1k + f (1, k )
= N (1 k ) u2k u1k
+N k u1k1 u2k1
+
+
u2k+1 u2k
1
2
u1k u2k
u2k1 u2k + f (2, 1 k )
1
2
+ 2
u2k u1k
(24)
+ 2
Note that in system (24) the terms k and (1 k ) vanish for k = 0 and k = N ,
thus no particular care has to be taken concerning the ghost points at N +1 and
1 . This is the discrete analogue to not imposing boundary conditions on (1).
A similar situation occurs in state constrained problems for Hamilton-Jacobi
equations.
3.2
Paradigm shift
f (1, ) = [a1 1 r + (1 a1 ) (1 1 )r ] r ,
(25)
1
r
f (2, ) = [a2 (1 2 )r + (1 a2 ) 2 r ] .
These functions are called productivity functions with constant elasticity of substitution and are commonly used in economics to combine two or more productive inputs (in our case scientific activities in the different fields) to an output
quantity. The constant r R, r 6= 0, denotes the elasticity of substitution, and
it measures how easy one can substitute one input for the other. The constants
ai [0, 1] measure the dependence of paradigm i with respect to the other. If ai
is close to one, the field is more autonomous and little influenced by the activity
in the other field.
3.3
Consumer choice
i 1 + s , > 0, 6= 1,
i
1
f (i, ) =
ln( ) + s ,
= 1,
i
i
(26)
Numerical simulations
We illustrate the behaviour of the discrete system (24) with several examples.
Let N = 100, i.e. the interval [0, 1] is discretized into 100 equidistant intervals.
Each grid point corresponds to the the percentage of players being in state 1.
System (24) is solved using an explicit in time discretization with time steps of
size t = 104 . In all examples in this section the terminal time T is set to
T = 10 if not stated otherwise.
4.1
Numerical examples
Example I (Shock formation): In this first example we would like to illustrate the formation of shocks, a phenomena well known for Hamilton-Jacobi
equations. We choose a terminal cost of the form
u1 (, 10) = 1
1
1
and u2 (, 10) = 2 ,
2
2
1.5
u1( ) at t=0
1.5
t=10
t=9
u1( ) at T=10
1
t=8
u ( ) at t=0
1
t=7
0.5
t=6
u ( ) at T=10
1
t=0
0
0.5
0.5
0
1
0.5
0.1
0.2
0.3
0.4
0.5
1
0.6
0.7
0.8
0.9
1.5
0.1
0.2
0.3
0.4
0.5
1
0.6
0.7
0.8
0.9
10
1
2,
a2 =
9
10 ,
r =
(27)
3
4.
In figure 2
2
1
u ( ) at t=0
u1( ) at T=10
1
u ( ) at t=0
1
u (1) at T=10
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
u ( ) at t=0
u ( ) at t=0
u1( ) at T=10
4.5
u1( ) at T=10
4.5
u ( ) at t=0
u ( ) at t=0
u2( ) at T=10
3.5
3.5
2.5
2.5
1.5
1.5
0.5
0.5
u2( ) at T=10
0.1
0.2
0.3
0.4
0.5
1
0.6
0.7
0.8
0.9
0.1
0.2
0.3
0.4
0.5
1
0.6
0.7
0.8
0.9
4.2
1
1
1
1
f (i, ) =
1
2
+
.
F (1 , 2 ) and 0 (1 , 2 ) =
i
2
2
2
2
In order to simplify the numerical implementation, we perform a dimensionality
reduction. Define (1 , 2 ) = (, 1 ), with [0, 1]. Let be the solution to
(17) and define
(, t) = (, 1 , t).
We observe that
, ),
= H(
t
(28)
where
, ) = 1 ( )+ 2 1 (1 ) ( )+ 2 + F (, 1 ).
H(
2
2
(29)
1.5
0.5
0.5
1.5
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
(30)
( + ) ()
() ( )
and + () =
,
in (30) is given by
and H
H(,
, ) =
), if ;
min H(q,
q
), if .
max H(q,
q
13
Finally we consider two-state problems and investigate in detail the shock structure. For this purpose we perform a reduction of the dimension (since the key
issues are related to the difference of the values, more than to its proper value)
and obtain a hyperbolic scalar equation. Then we introduce a related conservation law, which yields a Rankine-Hugoniot condition for possible shocks. This
new formulation allows us to study finite state mean field games from another
numerical perspective and gives new insights into the shock structure and the
qualitative behaviour of solutions to (1).
5.1
(31)
Then h and g, given by (5) and (14), can be written as h(U 1 , U 2 , 1 , 2 , i) =
h(w(, t), 0, , 1 , i) and g1 (U 1 , U 2 , 1 , 2 ) = g1 (w(, t), 0, , 1 ). For twostate problems equation (13) gives g2 = g1 . Hence we obtain
wt = g1 (w(, t), 0, , 1 ) 1 U 1 U 2 2 U 1 U 2
h(w(, t), 0, , 1 , 1) + h(w(, t), 0, , 1 , 2).
Define r and q by
r(w(, t), ) = g1 (w(, t), 0, , 1 ),
q(w(, t), ) = h(w(, t), 0, , 1 , 1) h(w(, t), 0, , 1 , 2),
w(,t)
and denote w
= 1 2 U 1 U 2 |(,1) . Then equation (31)
by
5.2
(32)
Consider the following conservation law associated to (32) on the interval [0, 1],
Pt (, t) + (r(w(, t), ) P (, t)) = 0,
(33)
supplemented with the boundary condition P (0, t) = P (1, t) = 0 for all times
t [0, T ].
If P is a sufficiently smooth solution to (33) and P (, 0) R 0, then the maximum principle
implies that P (, t) 0. Furthermore, if P (, 0)d = 1, we
R
have that P (, t)d = 1. Assuming that P (, 0) is a probability distribution,
we can regard P (, t) as a probability distribution on the set P(I), I = {1, 2},
14
as we have a natural identification for two-state problems: P(I) [0, 1]. Hence
equation (33) describes an evolution of a probability distribution on P(I). Uncertainty in the initial distribution of the mean field can be encoded in the
initial condition P (, 0) and propagated through (33).
Since equation (33) may not have globally smooth solutions, we use the RankineHugoniot condition to characterise certain possibly discontinuous solutions. Let
s : [0, T ] [0, 1] be a C 1 curve and suppose that P is a C 1 function on both
0 < < s(t) and s(t) < < 1, for t [0, T ]. Assume further that (33) holds in
this set. Let B : [0, 1] [0, T ] R. We denote by [B] the jump of B across the
curve s(t), that is, [B] = B(s+ (t), t) B(s (t), t). Equation (33) leads to the
Rankine-Hugoniot condition of the form
[P ]s = [r P ].
(34)
C
.
(t t0 )
dt
Z
C1 S 3/2 P (, t)d + C2 ,
where we use the fact that w is bounded, see [GMS13], in the last step. Then
Z
Z T
S(, t) P (, t) d C +
kS(, s)k3/2
ds + kS(, T )k .
t
This estimate does not give global bounds, but a nonlinear version of Gronwalls
inequality yields the existence of t0 .
15
5.3
Finally we discuss the particular formulation of equations (32) and (33) as well
as the numerical realisation of example I presented in section 4. Equation (32)
can be written as
(1 2)|w| w
1
w(, t)+ |w| w[f (1, , 1 ) f (2, , 1 )] .
2
2
(35)
using that h and g are given by (19) and (20) respectively. Similarly, (33) takes
the form
(1 2) |w| w
P (, t) = 0.
(36)
Pt (, t) +
2
wt (, t) =
70
60
50
40
30
20
0
2
10
4
0
1
6
0.8
0.6
8
0.4
0.2
10
Time
16
Conclusions
In this paper we present effective numerical methods for two-state mean field
games and discuss a number of illustrative examples in socio-economic sciences.
We compare our method with numerical results obtained from classical and well
established schemes for Hamilton-Jacobi equations. As presented in the examples, our method captures shocks effectively. We analyse the shock structure
using an associated conservation law and prove a local Lipschitz estimate for
the solutions of (1).
Several challenging and interesting mathematical questions will be addressed in
a future paper, like the development of numerical schemes based on the dual
variable method and the Lions transversal variable method, see [Lio11].
References
[Abg03]
Remi Abgrall.
Numerical discretization of boundary conditions for first order Hamilton-Jacobi equations.
SIAM
J.
Numer.
Anal.,
41(6):22332261
(electronic),
2003.
http://dx.doi.org/10.1137/S0036142998345980.
[Ach13]
[BD14]
[FG13]
R. Ferreira and D. Gomes. On the convergence of finite state meanfield games through convergence. Preprint - To appear in Journal
of Mathematical Analysis and Applications, 2013.
[GMS10] D. Gomes, J. Mohr, and R. R. Souza. Discrete time, finite state space
mean field games. Journal de Mathematiques Pures et Appliquees,
93(2):308328, 2010.
[GMS13] D. Gomes, J. Mohr, and R. R. Souza. Continuous time finite state
mean-field games. Appl. Math. and Opt., 68(1):99143, 2013.
[GS13]
[Gue11a] O. Gueant. An existence and uniqueness result for mean field games
with congestion effect on graphs. Preprint, 2011.
17
[Gue11b] O. Gueant. From infinity to one: The reduction of some mean field
games to a global control problem. Preprint, 2011.
[HCM07] M. Huang, P. E. Caines, and R. P. Malhame.
Largepopulation cost-coupled LQG problems with nonuniform agents:
individual-mass behavior and decentralized -Nash equilibria.
IEEE Trans. Automat. Control, 52(9):15601571, 2007.
http://dx.doi.org/10.1109/TAC.2007.904450.
[HMC06] M. Huang, R. P. Malhame, and P. E. Caines.
Large population stochastic dynamic games:
closed-loop McKeanVlasov
systems
and
the
Nash
certainty
equivalence
principle.
Commun. Inf. Syst.,
6(3):221251,
2006.
http://projecteuclid.org/getRecord?id=euclid.cis/1183728987.
[KNP01] Alexander Kurganov, Sebastian Noelle, and Guergana Petrova.
Semidiscrete central-upwind schemes for hyperbolic conservation laws
and HamiltonJacobi equations. SIAM Journal on Scientific Computing, 23(3):707740, 2001.
[Kuh70] Thomas S. Kuhn. The structure of scientific revolutions. University
of Chicago Press, Chicago, 1970.
[Lio11]
[LL06a]
J.-M. Lasry and P.-L. Lions. Jeux `a champ moyen. I. Le cas stationnaire. C. R. Math. Acad. Sci. Paris, 343(9):619625, 2006.
[LL06b]
J.-M. Lasry and P.-L. Lions. Jeux `a champ moyen. II. Horizon fini
et contr
ole optimal. C. R. Math. Acad. Sci. Paris, 343(10):679684,
2006.
[LL07]
J.-M. Lasry and P.-L. Lions. Mean field games. Jpn. J. Math.,
2(1):229260, 2007.
[LLG10] J.-M. Lasry, P.-L. Lions, and O. Gueant. Mean field games and applications. Paris-Princeton lectures on Mathematical Finance, 2010.
[Waa08] Knut Waagan.
Convergence rate of monotone numerical
schemes for Hamilton-Jacobi equations with weak boundary
conditions.
SIAM J. Numer. Anal., 46(5):23712392, 2008.
http://dx.doi.org/10.1137/050632932.
Acknowledgements
DG was partially supported by CAMGSD-LARSys (FCT-Portugal), PTDC/MATCAL/0749/2012 (FCT-Portugal), and KAUST SRI, Uncertainty Quantification
Center in Computational Science and Engineering. RMV was partially supported by CNPq - Brazil through a PhD scholarship - Program Science without
Borders. MTW acknowledges support from the Austrian Academy of Sciences
OAW
via the New Frontiers Project NST-001.
18