quiver-team / torch-quiver

PyTorch Library for Low-Latency, High-Throughput Graph Learning on GPUs.
https://torch-quiver.readthedocs.io/en/latest/
Apache License 2.0
293 stars 36 forks source link

Use memory pool for memory allocation when sampling && feature collecting #115

Closed eedalong closed 2 years ago

eedalong commented 2 years ago

We create a tensor through pytorch pytorch already use CudaCachingAllocator for memory allocation on device. I also implemented a simple test to check tensor's memory allocator and it shows that a cuda tensor indeed uses cudaCachingAllocator for memory allocation.

lgarithm commented 2 years ago

What about the performance of cudaCachingAllocator?

I didnt measure it, but I think it's a pretty good memory pool. but I'll do more research on this CudaCachingAllocator.

eedalong commented 2 years ago

We did some experiments on cudaCachingAllocator and it shows great performance. So we will close this issue.