Closed Kaixhin closed 6 years ago
Using a preallocated tensor (the size of which is known) in the replay memory instead of doing CUDA casts and hence allocating new memory should provide a speed boost (maybe)?
Closed as the batch size changes due to invalid samples from the prioritised replay memory.
Using a preallocated tensor (the size of which is known) in the replay memory instead of doing CUDA casts and hence allocating new memory should provide a speed boost (maybe)?