Input Window Size for Deep Recurrent Reinforcement Learning

Input Window Size for Deep Recurrent Reinforcement Learning

Deep Recurrent Reinforcement Learning makes use of a Recurrent Neural Network (RNN), such as Long Short-Term Memory (LSTM) or Gated Recurrent Unit (GRU) based networks, for learning a value function that maps environment states to action values. Recurrent Neural Networks are useful for modeling time-series data since the network maintains a memory, learning to retain useful information from inputs of prior model inferences. Every time the model is called, the memory is updated in correspondence with the current inputs.

Read the rest of the article at Mindboard’s Medium channel.

About the author

Eric Muccino administrator