Myths of Reinforcement Learning

Each myth/misstatement is discussed on its own page

  1. Large state spaces are hard for RL
  2. RL is slow
  3. RL does not have (m)any success stories since TDgammon
  4. RL does not work well with function approximation
  5. Value function approximation does not work (and so we should do something else - the favorite alternative seems to be policy search)
  6. Non-Markovianness invalidates standard RL methods
  7. POMDPs are hard for RL to deal with
  8. RL is about learning optimal policies


The following old myths are also unfortunately still around and still damaging for the field