About the Series
The the workshop series is organized by Ronen Gradwohl (Kellogg, MEDS), Jason Hartline (Computer Science), and Marciano Siniscalchi (Economics). Funding for the series is provided by the Shaw Family Supporting Organization CS+X Fund.
Synopsis
The first edition of this workshop will be on the theme of Learning as a Solution Concept. Traditional approaches in game theory separately describe outcomes of static equilibrium concepts and conditions under which models of learning converge to the static equilibrium concepts. Recent work, however, has shown the value of understanding the outcomes of strategic interaction directly from assumptions on learning processes and regardless of whether they reach a static equilibrium. One example is “no-regret learning,” which arises from many natural algorithms. The speakers are Drew Fudenberg, Alex Peysakhovich, Vasilis Syrgkanis, and Matt Weinberg.
Logistics
- Date: Friday, November 9, 2018.
- Location: Kellogg Global Hub 5101 (map) and Seeley Mudd 3514 (map), Northwestern U, Evanston, IL 60208.
- Transit: Noyes St. Purple Line (map).
- Parking: Validation for North Campus Parking Garage (map) available at workshop.
- Registration: None required. Please bring your own name badge from past conference.
Schedule
Kellogg Global Hub 5101
- 8:30-9:00: Continental Breakfast
- 9:00-9:05: Opening Remarks
- 9:05-9:45: Matt Weinberg (Princeton):
Selling to a no-regret buyer - 9:45-9:50: Matt Weinberg Q/A
- 9:50-10:30: Vasilis Syrgkanis (Microsoft Research):
No-regret Learning with Recency Bias: Fast and Last-Iterate Convergence - 10:30-10:35: Vasilis Syrgkanis Q/A
- 10:35-11:05: Coffee Break
Seeley Mudd 3514
- 11:05-11:45: Drew Fudenberg (MIT):
Player Compatible Equilibrium - 11:45-11:50: Drew Fudenberg Q/A
- 11:50-12:30:
Game Theory and Deep Reinforcement Learning
: - 12:30-12:35: Q/A
- 12:35-1:30: Lunch
Seeley Mudd 3014
- 2:00-4:00: Student meetings with speakers
Titles and Abstracts
Speaker: Matt Weinberg
Title: Selling to a no-regret buyer
Abstract: We consider the problem of a single seller repeatedly selling a single item to a single buyer (specifically, the buyer has a value drawn fresh from known distribution D in every round). Prior work assumes that the buyer is fully rational and will perfectly reason about how their bids today affect the seller’s decisions tomorrow. In this work we initiate a different direction: the buyer simply runs a no-regret learning algorithm over possible bids. We provide a fairly complete characterization of optimal auctions for the seller in this domain. Specifically:
- If the buyer bids according to EXP3 (or any “mean-based” learning algorithm), then the seller can extract expected revenue arbitrarily close to the expected welfare. This auction is independent of the buyer’s valuation D, but somewhat unnatural as it is sometimes in the buyer’s interest to overbid.
- There exists a learning algorithm A such that if the buyer bids according to A then the optimal strategy for the seller is simply to post the Myerson reserve for D every round.
- If the buyer bids according to EXP3 (or any “mean-based” learning algorithm), but the seller is restricted to “natural” auction formats where overbidding is dominated (e.g. Generalized First-Price or Generalized Second-Price), then the optimal strategy for the seller is a pay-your-bid format with decreasing reserves over time. Moreover, the seller’s optimal achievable revenue is characterized by a linear program, and can be unboundedly better than the best truthful auction yet simultaneously unboundedly worse than the expected welfare.
Title: No-regret Learning with Recency Bias: Fast and Last-Iterate Convergence
Abstract: We investigate the behavior of large families of no-regret dynamics in game theoretic settings in terms of their speed of convergence to small regret solutions and in terms of last-iterate behavior. We show that natural classes of regularized learning algorithms with a form of recency bias achieve faster convergence rates to approximate efficiency and to coarse correlated equilibria in multiplayer normal form games. Our results extend those of [Rakhlin and Shridharan 2013] and [Daskalakis et al. 2014], who only analyzed two-player zero-sum games for specific algorithms. Finally, inspired by empirical findings we investigate the long-term stability properties of such dynamics and show that in a family of zero-sum games they avoid limit cycles and converge pointwise to an equilibrium. Avoiding limit cycles is desirable from both a behavioral and an algorithmic perspective and we discuss implications to the field of adversarial machine learning.
Speaker: Drew Fudenberg
Title: Player Compatible Equilibrium
Abstract: This talk will summarize recent work with Kevin He on the learning foundations of equilibrium refinements. I will focus on our paper “player compatible equilibrium,” which develops a trembles-based equilibrium refinement where the restrictions on the trembles correspond to the relative frequencies of various “experiments” under either Bayesian optimization or the Bayes upper confidence bounds heuristic.
Speaker: Alex Peysakhovich
Title: Game Theory and Deep Reinforcement Learning
Abstract: Modern advances in deep learning have led to agents that have superhuman performance in games like Go, chess, and poker. In this talk I will introduce the basics of deep reinforcement learning (deep RL) and discuss how to push these advances beyond two player zero sum games. I will cover recent work that combines ideas from game theory with deep RL to construct agents that can cooperate, coordinate, and communicate with others.