Slides CFM Imperial
Slides CFM Imperial
CFM-Imperial Conference
Roadmap
Introduction
Algo MMs
Role of Experimentation
Price Discovery
Conclusion
Algorithms in Securities Markets
Introduction
Algo MMs
Role of Experimentation
Price Discovery
Conclusion
The Market Making Game (“RFQ”)
▶ If a trade occurs, the dealers posting the best quote earn (in
aggregate) (amin − ṽ ) and zero otherwise.
The Client’s Demand
▶ Two Cases:
1 1
Π̄(aX , aY ) = ( D(ai , vH )(ai − vH ) + D(aX , vL )(aX − vL )),
2 2
⇔
Cov (D(aX , ṽ ), ṽ )
Π̄(aX , aY ) = D̄(ai ) (aX − E(ṽ )) − ,
D̄(aX )
| {z }
AdverseSelection Cost
∆×∆D
where Cov (D(aX , ṽ ), ṽ ) = 2 > 0.
▶ No Adverse Selection Case: Cov (D(amin , v ), v )) = 0
because the likelihood of a buy does not vary with v .
Benchmark (Glosten-Milgrom (1985))
σ 0.5 1 3 5 7
(1) Quoted Spread 2.00 2.00 1.24 0.68 0.47
(2) Adverse Sel. Cost 2.00 2.00 1.24 0.68 0.47
(3) Realized Spread (1)-(2) 0 0 0 0 0
∆v 0 2 4 6 8
Quoted Spread 0 0.16 0.68 1.65 3.02
Adverse Sel. Cost 0 0.16 0.68 1.65 3.02
Realized spread 0 0 0 0 0
Roadmap
Introduction
Algo MMs
Role of Experimentation
Price Discovery
Conclusion
Why Reinforcement Learning
▶ t = 1: ṽ = vL = 0, ϵ1 = 0.90
Explore
a=3
Trade occurs (vL + L̃ ≥ 3).
QY 1 (3) = α × [3 − vL ] + (1 − α) × QY 0 (3) = 1.5.
QY 1 (3.1) = QY 0 (3.1) = 0.01
Q-Learning Algorithm - Example
▶ t = 2: ṽ = vL = 0, ϵ2 = 0.82
Explore
a = 3.1
Trade occurs (vL + L̃ ≥ 3.1).
QY 2 (3.1) = α × [3.1 − vL ] + (1 − α) × QY 1 (3.1) = 1.55.
Q-Learning Algorithm - Example
1.5
QY 2 =
1.55
▶ t = 3: ṽ = vH = 4, ϵ3 = 0.74
Explore
a=3
Trade occurs.
QY 3 (3) = α × [3 − 4] + (1 − α) × QY 2 (3) = 0.25
Q-Learning Algorithm - Example
0.25
QY 3 =
1.55
▶ t = 4: ṽ = vH = 4, ϵ4 = 0.67
Exploit
Greedy-price: a = 3.1
Trade does not occur,
QY 4 (3.1) = α × 0 + (1 − α) × QY 3 (3.1) = 0.775.
▶ Etc T times
▶ Will AM Y eventually learn that the true expected payoff of
a = 3 is higher...? No...Not necessarily: T is finite and
experimentation is less likely over time.
Roadmap
Introduction
Algo MMs
Role of Experimentation
Price Discovery
Conclusion
Implementation
Nash
Equilibrium
Observation 1: AMs’ prices are not competitive
Observation 2: AM’s prices account for adverse selection
Observation 3: Adverse selection reduces AMs’ rents
Observation 5: AMs’ spreads and rents increase with
Volatility
Observation 6: AMs’ spreads and rents decrease with the
number of AMs
Evidence
0.4
0.2
0
0 0.9 4.9
Realized Profit at aY = 4.9 (aX = 5)
Probability
0.4
0.2
⇒0
0 0.9 4.9
Realized Profit at aY = 4.9 (aX = 5)
Introduction
Algo MMs
Role of Experimentation
Price Discovery
Conclusion
More Experimentation, More competitive outcomes
Figure: We vary β and compute average per period profits over various
time windows.
Roadmap
Introduction
Algo MMs
Role of Experimentation
Price Discovery
Conclusion
Learning from Order Flow
1. They raise their offer for the second client after observing a
buy from the first (because a buy is more likely if v = vH ).
2. They lower their offer for the second client if the first does not
trade (because the first client is more likely not to trade if
v = vL ).
Example
▶ Parameters: Adverse selection, ∆ = 4, Lt ∼ N (0, 5). a2T (
a2NT ): 2nd client price after a trade (No trade) with the first
client
σ
0.5 1 3 5 7
Introduction
Algo MMs
Role of Experimentation
Price Discovery
Conclusion
Conclusion
▶ The behavior of algorithmic market makers using
Q-learning algorithm is significantly different from that
predicted by the standard equilibrium analysis of the
market making game.
1. Non competitive prices (not collusion, just imperfect estimates
of actual expected profits).
2. An increase in adverse selection costs make algo market
makers more competitive.
▶ The variance of profits is important even though AMs
are not penalized for risk taking in our experiments.
1. Parameters that increases this variance makes learning true
expected payoffs and optimal actions more difficult for AMs.
▶ AMs’s behavior is more consistent with the predictions
of the Nash equilibrium if they experiment more but this
is costly.
▶ Next step: Check whether our findings are robust with more
complex algorithms.
Thank You!