I'm reading the Sutton & Barto's book "Reinforcement Learning: An Introduction" (2nd Edition), as the classes I took were a long time ago, and I'm struggling to understand this part (p. 12):
To do this, it had to have a model of the game that allowed it to foresee how its environment would change in response to moves that it might never make. Many problems are like this, but in others even a short-term model of the effects of actions is lacking. Reinforcement learning can be applied in either case.
I don't understand how can games not have a model. If there is randomness in the environment's reaction to the agent's event, we still have a model don't we? If we have a game that does not have a model with states and values, how can RL be incorporated?
So my question would be:
Do you know examples where "even a short-term model of the effects of actions is lacking", i.e. where an MDP cannot be drawn or estimated?
I hope this is not too broad of a question. I've mainly done RL on Markov decision processes, and I'm struggling to see how a game would not have the possibility to be modelled by one (even if very big!).