Paper Sharing one - Hindsight Experience Replay

Name

Hindsight Experience Replay

Attachment

  1. Original: arxiv
  2. Code: Github
  3. Video: 油管
  4. Two Minute Papers : 油管
  5. 愛可可-愛生活老師Two Minute Papers的微博視頻分享: 微博

Problem or Challenge:

Dealing with sparse rewards, Reward function engineering, Sample efficiency.

Assumptions or hypotheses:

  1. Multiple different goals (a single-goal also works).
  2. Sparse and binary rewards.
  3. Off-policy RL algorithms.

Inspiration or Source:

Human can learn almost as much from achieving an undesired outcome as from the desired one.

Methods or Solutions:

  1. HER can combined with any off-policy RL algorithm.
  2. Training universal policies[2]: Input = the current state + a goal state.
  3. Key idea: Whether an episode is good for the training depends on the final goal. As for off-policy RL algorithms we can choose additional goals.

Story time

    1. Assume you bought a robot and now its in a new house of which it doesnt know the whole environment al all.
    2. You ask it to go to room 1 in 30 seconds.
    3. After a while walking around, It ends in room 2, for traditional RL algorithm, your robot cant learn much experience from this episode as your rewards are all 0.
    4. But what if you change the goal to room 2 and add the full episode with reward 1 to your replay buffer?
    5. You may know that this is an good episode for your robot to learn the policy to go to room 2. This is called hindsight experience replay.
    6. Algorithm:

Experiments or Results:

  1. MuJoCo[3]
  2. Three different tasks: pushing, sliding, and pick-and-place.
  3. For HER they store each transition in the replay buffer twice:

1. Once with the goal used for the generation of the episode.

2. Once with the goal corresponding to the final state from the episode (they call this strategy final).

4.

Learning curves for multi-goal setup

5. DQN without HER was not able to solve any of them.

6. HER learns faster if training episodes contain multiple goals than single goal.

7. HER + Reward shaping.

Surprisingly neither DDPG, nor DDPG+HER was able to successfully solve any of the tasks with any of these reward functions (Two reasons).

8. K goals and three strategies: future(best), episode, random.

9. Strategy future on physical robot, Initially success rate 2/5, after retraining, up to 5/5.

Limitation or Weakness:

  1. Reward shaping does not work well (it requires a lot of domain knowledge).
  2. The goal substitution changes the distribution of experience in an unprincipled way. This bias can in theory lead to instabilities[4].
  3. Cant combine on-policy algorithms.

Summary

  1. It is the first time so complicated behaviors were learned using only sparse, binary rewards.
  2. It can combine any off-policy RL algorithms.

Reference

[1] M. Andrychowicz, F. Wolski, A. Ray, J. Schneider, R. Fong, P. Welinder, B. Mcgrew, J. Tobin, P. Abbeel, and W. Zaremba, 「Hindsight Experience Replay,」 2017.

[2] Schaul T, Dan H, Gregor K, et al. Universal Value Function Approximators[C]//

International Conference on Machine Learning. 2015.

[3] E. Todorov, T. Erez, and Y. Tassa, "MuJoCo: A physics engine for model-based control." pp. 5026-5033.

[4] blog.openai.com/ingredi

[5] 標題圖片來源:blog.openai.com/content


推薦閱讀:

TAG:強化學習ReinforcementLearning | 深度學習DeepLearning | 人工智慧 |