1. Introduction
The first traffic assignment model was proposed for about 70 years ago in the work of M. Beckmann [
1], see also [
2]. Nowadays Beckmann’s type models are rather well studied [
3,
4,
5,
6,
7]. The entropy-based origin–destination matrix models are also well developed nowadays [
7,
8,
9]. Moreover, as was mentioned in [
10], both of these two types of models can be considered as macrosystem equilibrium for logit (best-response) dynamics in corresponding congestion games [
11].
In this paper, we popularise the results of [
10] for English-reading people (The paper [
10] was written in Russian and has not been translated yet) and refine the results on the convergence rate. Moreover, we propose superposition of the considered dynamics to describe equilibrium in two-stage traffic assignment model [
12,
13].
One of the main results of the paper is Theorem 1, where it is proved that the natural logit-choice and best-response Markovian population dynamics in traffic assignment model (congested population game) converge to equilibrium. By using Cheeger’s inequality we first time show that mixing time (the time required to reach equilibrium) of these dynamics
is proportional to
, where
N is a total number of agents. Note, that in related works analogues of this theorem were proved without estimating of
[
9,
11,
13]. We confirm Theorem 1 by numerical experiments.
Another important result is a saddle-point reformulation of two-stages traffic assignment model. We explain how to apply results of Theorem 1 to this model.
2. Traffic Assignment: Problem Statement
Following [
14] we describe the problem statement.
Let the urban road network be represented by a directed graph
, where vertices
V correspond to intersections or centroids [
4] and edges
E correspond to roads, respectively. Suppose we are given the travel demands: namely, let
(veh/h) be a trip rate for an origin–destination pair
w from the set
. Here,
is the set of all possible origins of trips, and
is the set of destination nodes. For OD pair
denote by
the set of all simple paths from
i to
j. Respectively,
is the set of all possible routes for all OD pairs. Agents travelling from node
i to node
j are distributed among paths from
, i.e., for any
there is a flow
(i.e.,
) along the path
p, and
. Flows from vertices from the set
O to vertices from the set
D create the traffic in the entire network
G, which can be represented by an element of
Note that set X can have extremely large number of paths (routes): e.g., for Manhattan network up to a logarithmic factor. Note, that to describe a state of the network we do not need to know an entire vector x, but only flows on arcs: , where . Let us introduce a matrix such that for , , so in vector notation we have . To describe an equilibrium we use both path- (route-) and link-based notations or .
Beckmann model. An important idea behind the Beckmann model is that the cost (e.g., travel time) of passing a link e is the same for all agents and depends only on the flow along it. We denote this cost for a given flow by . Another essential point is a behavioral assumption (the first Wardrop’s principle): each agent knows the state of the whole network and chooses a path p minimizing the total cost
We consider
to be continuous, non-decreasing, and non-negative. In this case
, is an equilibrium state, i.e., it satisfies conditions
if, and only if,
is a minimum of the potential function:
and
[
2].
According to [
5,
13], we can construct a dual problem for the potential function in the following way:
where
is the Legendre—Fenchel conjugate function of
,
. At the end we obtain the dual problem, which solution is
:
We can reconstruct primal variable
f from the current dual variable
t:
This condition reflects the fact that every driver choose the shortest route [
5]. Another condition
can be equivalently rewrite as
. This condition with the condition
form the optimization problem (
1).
If
Beckmann’s model will turn into Nesterov–de Palma model [
13,
15].
Population games dynamic for (stochastic) Beckmann model. Let us consider each driver to be an agent in population game, where
,
is a set of types of agents. All agent (drivers) of type
can choose one of the strategy
with cost function
. Assume that every driver/agent independently of anything (in particular of any other drivers) is considering the opportunity to reconsider his choice of route/strategy
p in time interval
with probability
, where
is the same for all drivers/agents. It means that with each driver we relate its own Poisson process with parameter
. If in moment of time t (when the flow distribution vector is
) the the driver of type
decides to reconsider his route, than he choose the route
with probability
where
are i.i.d. and satisfy Gumbel max convergence theorem [
16] when
with the parameter
(e.g.,
has (sub)exponential tails at
∞). It means that
asymptotically (when
) has Gumbel distribution
, where
is Euler constant. Note that
,
. In (
2) it means that every driver try to choose the best route. However, the only available information are noise corrupted values
. So the driver try to choose the best route focused on the worst forecasts for each route.
One of the main results of Discrete Choice Theory is as follows [
17]
where
was previously defined in (
2).
Note that the described above dynamic degenerates into the best-response dynamic when
[
11].
Theorem 1. Let . For all there exists such a constant that for all and :where Proof. The first important observation is that the described Markov process is reversible. That is, it satisfies Kolmogorov’s detailed balance condition (see also [
18]) with stationary (invariant) measure
where
[
11]. The result of type (
4) for
holds true due to Hoeffding’s inequality in a Hilbert space [
19]. We can apply this inequality for multinomial part
. The rest part may only strength the concentration phenomenon, especially when
is small. The Sanov’s theorem [
20] says that
from (
5) asymptotically (
) describe the proportions in maximum probability state, that is
To estimate the mixing time
of the considered Markov process we will put it in accordance with this continuous-time process discrete-time process with step
, which corresponds to the expectation time between two nearest events in continuous-time dynamic. Additionally, we consider this discrete Markov chain as a random walk on a proper graph
with starting point corresponds to the vertex
s and transition probability matrix
. According to a Cheeger’s inequality mixing time
for such a random walk, which approximate stationary measure
with accuracy
(in this case
), is
where Cheeger’s constant is determined as
where
[
21]. Since
G and
P correspond to reversible Markov chain with stationary measure
that exponentially concentrate around
one can prove that isoperimetric problem of finding optimal set of vertexes
S has the following solution, which we described below roughly up to a numerical constant:
S is a set of such states
that
. Since the the ratio of sphere volume of radius
to the volume of the ball of the same radius is
, we can obtain that
. So up to a
(we put it into
) mixing time is indeed
. □
Note that the describe above approach assumes that we first
and then
. If we firstly take
than due to Kurtz’s theorem [
22]
satisfies (for all
,
)
where
,
. Note that Sanov’s type function
from (
5) will be Boltzmann–Lyapunov type function for this system of ordinary differential equations (SODE), that is decrees along the trajectory of SODE. This result is a particular case of the general phenomenon: Sanov’s type function for invariant measure obtained from Markovian dynamics is Boltzmann–Lyapunov type function for deterministic Kurtz’s kinetic dynamics [
18,
23,
24].
3. Origin–Destination Matrix Estimation
Origin–destination matrix estimation model can be considered as a particular case of the traffic assignment model. The following interpretation goes back to [
10,
13]. Indeed, let us consider fictive origin
o and fictive destination
d. So
,
. Let us draw fictive edges from
o to real origins of trips
O. The cost of the trip at edge
is
—an average price that each agent pays to live at this origin region
. Analogously, let us draw edges from the vertexes of the real destination set
D to
d. The cost of the trip at edge
is
, minus average salary that each agent obtains in destination region
. So the set of all possible routes (trips) from
o to
d can be described by pairs
. Each route consist of three edges
with cost
, edge
with cost
(is available as an input of the model) and edge
with cost
. So equilibrium origin-destination matrix
(up to a scaling factor) can be find from entropy-linear programming problem
In real life,
and
are typically unknown. However, at the same time, the following agglomeration characters are available
The key observation is that (
6) can be considered as Lagrange multipliers principle for constraint entropy-linear programming problem
where
and
are Lagrange multipliers for (
7) and (
8) correspondingly. The last model is called Wilson’s entropy origin–destination matrix model [
8,
9].
The result of Theorem 1 can be applied to this model due to the mentioned above reduction.
4. Two-Stages Traffic Assignment Model
From the
Section 2 we may know that Beckmann’s model requires origin-destination matrix as an input
. So Beckmann’s model allows to calculate
. At the same time, from
Section 3, we may know that Wilson’s entropy origin–destination model requires cost matrix
as an input, where
. So Wilson’s model allows to calculate
. The solution of the system
is called two-stage traffic assignment model [
12]. Following [
7,
13] we can reduce this problem to the following one (see (
1) and (
9))
The problem (
10) can be rewritten as a convex-concave (if
) saddle-point problem (SPP)
This SPP can be efficiently solved numerically [
7].
Note that, if we consider best-response dynamics from
Section 2, with the parameter
and logit dynamic with the parameter
for the origin–destination matrix estimation and assume that
than such a dynamic will converge to the stationary (invariant) measure that is concentrated around the solution of the SPP problem (
11). This result can be derived from the more general result related with hierarchical congested population games [
7].
5. Numerical Experiments
The main result of the paper is Theorem 1. The main new result of this theorem is a statement that mixing time of the considered Markovian logit-choice and best-response dynamics is approximately , where N is a number of agents.
We consider Braess’s paradox example [
25], see
Figure 1. This picture is taken from Wikipedia. Here, Origin is START and Destination is END. We have one OD-pair and put
, the number of agents. The <<paradox>> arises when
. In this case when there is no road from A to B we have two routes (START, A, END) and (START, B, END) with 2000 agents at each route. So the equilibrium time costs at each route will be 65. When the road AB is present (this road has time costs 0) all agents will use the route (START, A, B, END) and this equilibrium has time costs 80. That is paradoxically larger than it was without road AB.
Numerical experiments confirm Theorem 1. Note that in [
9] it was described a real-life experiment oraganized with MIPT students in Experimental Economics Lab. The students were agents and play in repeated Braess’s paradox game. The result of experiments from [
9] is also well agreed with the described above numerical experiments.