Contact
- Eitan Farchi
DE, AOT, Software Testing Analysis and Reviews (STAR), IBM Research - Haifa
Reinforcement Learning
These ideas are at the heart of reinforcement learning.
Markov chain video (English):
MC/TD strategies in reinforcement learning:
Reinforcement learning - Recurring chicken game
- The chicken game
- Nash equilibrium f_1(x, y) >= f_1(z, y), f_2(x, y) >= f_2(x, z)
- (2, 7), (7, 2) - Nash equilibrium in the chicken game
- The minmax/maxmin is 2. The first player can ensure he gets something regardless of what the other player does.
- If the row player plays a dare, the worst that can happen is 0. If she plays chicken, the worst that can happen is 2, so by playing chicken she can ensure a score of 2. The same is true for the column player.
- We'll repeat the game an infinite number of times. The payment of the row play is ri at stage. We'll define the payment to be the liminf of the average.
Sum_{1 to k}(ri)/k as k goes to infinite. Same thing with the column player - The folk theorem in game theory - https://en.wikipedia.org/wiki/Folk_theorem_(game_theory)
- Objective - obtain an equilibrium of (1/2)(0,0) + (1/2)(6, 6). In fact, any convex combination of the possible game payments can be obtained.
- Strategy players play (0, 0) in odd stages and (6, 6) in even stages. If the column player changes from the prescribed behavior, the row player plays dare always, ensuring the player gets no more than 2. Same with the column player.
Payments when nobody deviates - same for both players. For row player the payment is
0 6 0 6 0 6 ....
0 3 2 3 12/5 3 ....
Liminf of the series is 3
https://en.wikipedia.org/wiki/Limit_superior_and_limit_inferior
Assume a deviation of the column play. At the deviation, the column player gets a maximal addition of 2. Then moving forward after the deviation, the column player gets a maximum of 2.
Payments of column player are bounded by -
0 6 0 6 2 2 2 2 2 2 ....
Assume the length of the prefix is 2k before the deviation. The payment for the column player is then bounded by
Lim [k6 + (n - 2k)2]/n = lim 2 = 2
Which is less than 3.
Therefore, the above strategy is a Nash equilibrium.
Can we marry repeated games with reinforcement learning? See some additional interesting reading.
In reinforcement learning, we deal with infinite sequences of actions. We thus need to learn how to think about their probability.
A specific scheme of Monte Carlo control is monotonic if Q(a, pi) is well estimated by the exploration stage.
In the context of reinforcement learning, we show that a specific scheme of Monte Carlo control is monotonic if Q(a, pi) is well estimated by the exploration stage.