1. What is Reinforcement Learning
概述:
举个栗子:
再举一个:
2. Markov Decision Process
- Mathematical formulation of the RL problem
- Markov property: Current state completely characterises the state of the world
处理流程:
The optimal policy π*
3. Q-learning
Definitions: Value function and Q-value function:
Bellman equation:
优化策略:
Solving for the optimal policy: Q-learning
举个栗子:Playing Atari Games
Q-network Architecture
Training the Q-network: Experience Replay
Deep Q-Learning with Experience Replay
4. Policy Gradients
Intuition:
Variance reduction:
Variance reduction: Baseline
How to choose the baseline?
A better baseline: Want to push up the probability of an action from a state, if this action was better than the expected value of what we should get from that state
Actor-Critic Algorithm
5. REINFORCE 的运用
5.1 Recurrent Attention Model (RAM)
效果示意图:
5.2 AlphaGo
6. Summary
- Policy gradients: very general but suffer from high variance so requires a lot of samples.
Challenge: sample-efficiency - Q-learning: does not always work but when it works, usually more sample-efficient. Challenge: exploration
- Guarantees:
Policy Gradients: Converges to a local minima of J(θ), often good enough!
Q-learning: Zero guarantees since you are approximating Bellman equation with a complicated function approximator