Tag: experience-replay

14 How does LSTM in deep reinforcement learning differ from experience replay? 2018-08-27T01:58:20.250

6 What is experience replay in laymen's terms? 2018-05-30T19:09:05.100

4 Which kind of prioritized experience replay should I use? 2019-05-05T10:05:47.557

4 Why do DQNs tend to forget? 2020-07-27T11:51:00.447

3 How large should the replay buffer be? 2019-04-04T14:40:34.553

3 Experience Replay Not Always Giving Better Results 2019-04-29T15:30:20.570

3 Why do authors track $\gamma_t$ in Prioritized Experience Replay Paper? 2019-05-31T02:47:46.293

3 What is the difference between random and sequential sampling from the reply memory? 2019-09-19T13:10:13.547

3 Can experience replay be used for training after completing every single epoch? 2020-03-06T06:58:21.853

3 How does the optimization process in hindsight experience replay exactly work? 2020-03-12T10:19:55.543

3 Is this a good approach to solving Atari's "Montezuma's Revenge"? 2020-03-13T10:58:40.600

3 On-policy preventing us from using the replay buffer with the PG? 2020-05-12T15:17:21.890

3 How to handle the final state in experience replay? 2020-06-24T02:59:18.853

2 What information should be cached in experience replay for actor-critic? 2019-01-31T23:35:08.680

2 Why experience reply memory in DQN instead of a RNN memory? 2019-04-22T12:00:35.690

2 New transition priorities in Prioritized Experience Replay? 2019-06-01T02:26:03.443

2 Intutitive explanation of why Experience Replay is used in a Deep Q Network? 2020-03-01T19:12:16.580

2 Could we update the policy network with previous trajectories using supervised learning? 2020-04-12T10:08:58.267

2 Prioritised Remembering in Experience Replay (Q-Learning) 2020-07-17T07:09:59.120

2 Why is sampling non-uniformly from the replay memory an issue? (Prioritized experience replay) 2020-08-27T11:05:48.977

1 Should we multiply the target of actor by the importance sampling ratio when prioritized replay is applied to DDPG? 2019-03-12T01:11:11.133

1 Reinforcement Learning with limited number of episodes 2019-03-26T09:57:58.457

1 Do we need to use the experience replay buffer with the A3C algorithm? 2019-05-02T10:25:44.270

1 Can I apply experience on naive actor critic directly? Should it work? 2020-02-15T18:47:34.243

1 Implementing Actor-Critic with Experience Replay for Continuous Action Spaces 2020-02-15T18:52:53.767

1 In a DQN, can Prioritized Experience Replay actually perform worse than a regular Experience Replay? 2020-06-04T22:14:01.853

1 What would happen if we sampled only one tuple from the experience replay? 2020-06-22T19:39:43.160

1 What do the state features of KukaGymEnv represent? 2020-08-18T06:36:10.197

0 My Double DQN with Experience Replay produces a no-action decision most of the time. Why? 2020-06-11T22:02:31.657

0 What is the advantage of using experience replay (as opposed to feeding it sequential data)? 2020-07-27T10:46:56.330

0 DDPG with Hindsight Experience Replay not converging properly 2020-09-01T06:54:14.883