Make An Experience Replay Buffer In One Line Of Code
This article is for those of us who have gotten stuck implementing an experience replay buffer as a beginner. That includes me. I see a lot of DRL agent implementations online using experience replay buffers made out of a bunch of numpy arrays. This data strategy isn’t new, and there’s a reason it runs nice and fast (https://en.wikipedia.org/wiki/AoS_and_SoA). It is actually a pretty good solution if your batch sizes are big. So you should definitely just copy and paste it…