Tag Archives: game theory

Is competition for losers in bikesharing?

The rise and fall of the bikesharing industry in China offers a cautionary tale about the risks of an unregulated market with a low entry barrier. It is well known that, while low entry barriers can promote competition and innovation, they may also lead to higher market volatility and potential challenges in achieving profitability due to intensified rivalry . There are also limited economies of scale to be had, making it exceedingly difficult to establish a monopoly. As Peter Thiel noted, “competition is for losers”‘ in such markets and good entrepreneurs should simply stay away from them.   However, writing off the bikesharing industry as unprofitable cannot be the only story here. After all, bikesharing has a genuinely positive societal impact and should have its place in many of our cities that are haunted by the disease of auto-dependency. The question is what, if anything, can be done to foster a healthy bikesharing market that is attractive to both users and private investors.  We set up to answer this question here.  You may download a preprint here, or read the abstract below.


Abstract: We model inter-operator competition in a dockless bikesharing (DLB) market as a non-cooperative game. To play the game, a DLB operator sets a strategic target (e.g., maximizing profit or maximizing ridership) and makes tactical decisions (e.g., pricing and fleet sizing). As each operator’s payoff and decision set are influenced by its own decisions as well as those of its competitors, the outcome of the game is a generalized Nash equilibrium (GNE). To analyze how competition may shape the choice of strategic targets, we further augment the game framework with a ranking scheme to properly evaluate the preference for different targets. Using a model calibrated with empirical data, we show that, if an operator is committed to maximizing its market share with a budget constraint, all other operators must respond in kind. Otherwise, they would be driven out of the market. When all operators compete for market dominance, Moreover, even if all operators agree to focus on making money rather than ruinously seeking dominance, profitability still plunges quickly with the number of players. Taken together, the results explain why the unregulated DLB market is often oversupplied and prone to collapse under competition. We also show this market failure may be prevented by a fleet cap policy, which sets an upper limit on each operator’s fleet size.

Wardrop equilibrium can be boundedly rational

As one of the most fundamental concepts in transportation science, Wardrop equilibrium (WE) was the cornerstone of countless large mathematical models that were built in the past six decades to plan, design, and operate transportation systems around the world. However, like Nash Equilibrium, its more famous cousin, WE has always had a somewhat flimsy behavioral foundation. The efforts to beef up this foundation have largely centered on reckoning with the imperfections in human decision-making processes, such as the lack of accurate information, limited computing power, and sub-optimal choices. This retreat from behavioral perfectionism was typically accompanied by a conceptual expansion of equilibrium. In place of WE, for example, transportation researchers had defined such generalized equilibrium concepts as stochastic user equilibrium (SUE) and boundedly rational user equilibrium (BRUE). Invaluable as these alternatives are to enriching our understanding of equilibrium and advancing modeling and computational tools, they advocate for the abandonment of WE, predicated on its incompatibility with more realistic behaviors. Our study aims to demonstrate that giving up perfect rationality need not force a departure from WE, since WE may be reached with global stability in a routing game played by boundedly rational travelers. To this end, we construct a day-to-day (DTD) dynamical model that mimics how travelers gradually adjust their valuations of routes, hence the choice probabilities, based on past experiences.

Our model, called cumulative logit (CULO), resembles the classical DTD models but makes a crucial change: whereas the classical models assume routes are valued based on the cost averaged over historical data, ours values the routes based on the cost accumulated. To describe route choice behaviors, the CULO model only uses two parameters, one accounting for the rate at which the future route cost is discounted in the valuation relative to the past ones (the passivity measure) and the other describing the sensitivity of route choice probabilities to valuation differences (the dispersion parameter).  We prove that the CULO model always converges to WE, regardless of the initial point, as long as the passivity measure either shrinks to zero as time proceeds at a sufficiently slow pace or is held at a sufficiently small constant value. Importantly, at the aggregate (i.e., link flow) level, WE is independent of the behavioral parameters. Numerical experiments confirm that a population of travelers behaving differently reaches the same aggregate WE as a homogeneous population, even though in the heterogeneous population, travelers’ route choices may differ considerably at WE.

By equipping WE with a route choice theory compatible with bounded rationality, we uphold its role as a benchmark in transportation systems analysis. Compared to the incumbents, our theory requires no modifications of WE as a result of behavioral accommodation. This simplicity helps avoid the complications that come with a “moving benchmark”, be it caused by a multitude of equilibria or the dependence of equilibrium on certain behavioral traits. Moreover, by offering a plausible explanation for travelers’ preferences among equal-cost routes at WE, the theory resolves the theoretical challenge posed by Harsanyi‘s instability problem. Note that we lay no claim on the behavioral truth about route choices. Real-world routing games take place in such complicated and ever-evolving environments that they may never reach a true stationary state, much less the prediction of a mathematical model riddled with a myriad of assumptions. Indeed, a relatively stable traffic pattern in a transportation network may be explained as a point in a BRUE set, an SUE tied to properly calibrated behavioral parameters, or simply a crude WE according to the CULO model. More empirical research is still needed to compare and vet these competing theories for target applications. However, one should no longer write off WE just because it has no reasonable behavioral foundation.

A preprint can be downloaded at ArXiv or SSRN.

Allocation problem for the platform of platforms

Another joint work with Ruijie Li, built on our previous research of ridesharing, including A-PASS and Pricing carpool.

In this study, we consider a general problem called the Allocation Problem for the Platform of Platforms – dubbed AP3.  Such a problem might arise  in a two-sided service market, where a third-party integrator tries to allocate customers to workers separately controlled by a set of online platforms in a manner that satisfies all stakeholders.   The integrator, as a leader, influences the outcome of the game by pricing the service, whereas the platforms (followers) are given the freedom to accept or reject customers to maximize their own profit, given the prices set by the integrator (see the plot below for an illustration).  A set of nonlinear constraints are imposed on the leader’s problem to eliminate artificial scarcity, orignated from the integrator’s monopoly power.  We formulate AP3 as a Stackelberg bipartite matching problem, which is known to be NP-hard in general.  Our main result concerns the proof that AP3 can be reduced to a polynomially solvable problem by taking advantage of, somewhat paradoxically, the hard requirement of ruling out artificial scarcity.

A preprint can be downloaded here.


Abstract: We study the Allocation Problem for the Platform of Platforms (abbreviated as AP3) in a two-sided service market, where a third-party integrator tries to allocate customers to workers separately controlled by a set of online platforms in a manner that satisfies all stakeholders. AP3 is a natural Stackelberg game. The integrator, as a leader, influences the outcome of the game by pricing the service, whereas the platforms (followers) are given the freedom to accept or reject customers to maximize their own profit, given the prices set by the integrator. A set of nonlinear constraints are imposed on the leader’s problem to eliminate artificial scarcity, derived from the integrator’s monopoly power. We formulate AP3 as a Stackelberg bipartite matching problem, which is known to be NP-hard in general. Our main result concerns the proof that AP3 can be reduced to a polynomially solvable problem by taking advantage of, somewhat paradoxically, the “hard” requirement of ruling out artificial scarcity. Numerical experiments are conducted using the ride-hail service market as a case study. We find artificial scarcity negatively affects the number of customers served, although the magnitude of the effect varies with market conditions. In most cases, the integrator takes the lion’s share of the profit, but the need to eliminate artificial scarcity sometimes forces them to concede the benefits of collaboration to the platforms. The tighter the supply relative to the demand, the more the platforms benefit from removing artificial scarcity. In an over-supplied market, however, the integrator has a consistent and overwhelming advantage bestowed by its monopoly position.

Differentiable bilevel programming

A Stackelberg congestion game (SCG) is a bilevel program in which a leader aims to maximize their own gain by anticipating and manipulating the equilibrium state at which followers settle by playing a congestion game. Large-scale SCGs are well known for their intractability and complexity.   In this study,  we attempt to marry the latest developments in machine learning with traditional methodologies — notably bilevel optimization and game theory — to forge an integrative approach based on differentiable programming.  Among other advantages,  the approach enables us to treat the equilibration of a congestion game as a deep neural network (see below), so that a suite of computational tools, notably automatic differentiation, can be easily applied.  You may download a preprint here.


Abstract: A Stackelberg congestion game (SCG) is a bilevel program in which a leader aims to maximize their own gain by anticipating and manipulating the equilibrium state at which followers settle by playing a congestion game. Large-scale SCGs are well known for their intractability and complexity. This study approaches SCGs through differentiable programming, which marries the latest developments in machine learning with conventional methodologies. The core idea centers on representing the lower-level equilibrium problem using an evolution path formed by the imitative logit dynamics. It enables the use of automatic differentiation over the evolution path towards equilibrium, leading to a double-loop gradient descent algorithm. We further show the fixation on the lower-level equilibrium may be a self-imposed computational obstacle. Instead, the leader may only look ahead along the followers’ evolution path for a few steps, while updating their decisions in sync with the followers through a co-evolution process. The revelation gives rise to a single-loop algorithm that is more efficient in terms of both memory consumption and computation time. Through numerical experiments that cover a wide range of benchmark problems, we find the single-loop algorithm consistently strikes a good balance between solution quality and efficiency, outperforming not only the standard double-loop implementation but also other methods from the literature. Importantly, our results highlight both the wastefulness of “full anticipation” and the peril of “zero anticipation”. If a quick-and-dirty heuristic is needed for solving a really large SCG, the proposed single-loop algorithm with a one-step look-ahead makes an ideal candidate.