NVIDIA / open-gpu-kernel-modules

NVIDIA Linux open GPU kernel module source
Other
15.09k stars 1.26k forks source link

In another update, activate RTX IO for deep learning #17

Open johnnynunez opened 2 years ago

johnnynunez commented 2 years ago

For people who work with deep learning, it would be amazing that you would enable RTX IO for Turing and Ampere. My GPU is 3090 FE and the clusters with 7-8 that I works too. https://www.nvidia.com/es-es/data-center/magnum-io/ https://www.nvidia.com/en-us/geforce/news/rtx-io-gpu-accelerated-storage-technology/

PAR2020 commented 2 years ago

Thanks for the suggestion. We are exploring whether or not this is possible, and will get back to you.

johnnynunez commented 2 years ago

https://github.com/NVIDIA/MagnumIO/issues/5 It should be supporting by RTX Turing and Ampere as my RTX 3090 FE, because RTX IO is the same function for gaming...

PAR2020 commented 2 years ago

Still digging into this, @johnnynunez. As you pointed out, RTX IO couples with Microsoft Direct Storage for gaming, but MagnumIO works with RTX IO GPUs. Please be patient a little while longer as we sort through all of our issues/PRs and get to this info for you. Thanks.

johnnynunez commented 2 years ago

@PAR2020 Yes, don't worry, I'm not in a hurry, as long as it's in the pipeline for future versions. I think it's something that can benefit the community a lot.

weareai commented 2 years ago

@johnnynunez I'd like to learn more about your use case with RTXIO and deep learning. Are you predominantly on Windows or Linux ? Looking forward to hearing from you.

johnnynunez commented 2 years ago

@weareai Linux always. But now that linux gaming is also booming, it should be able to be used for both cases.

weareai commented 2 years ago

@johnnynunez got it. It seems to be that you are interested in accelerating data transfer from storage to the GPU, right ? Have you considered GPU Direct Storage for CUDA on Linux ? Let me know as I am trying to make sure we capture your request if its not already supported for future releases :-)

johnnynunez commented 2 years ago

@johnnynunez got it. It seems to be that you are interested in accelerating data transfer from storage to the GPU, right ? Have you considered GPU Direct Storage for CUDA on Linux ? Let me know as I am trying to make sure we capture your request if its not already supported for future releases :-)

@weareai I understand that magnum IO is not compatible with RTX, only quadro, tesla, A100, etc, in resume, professional GPU's as indicated on the page. I have also seen the fact of using DALI, that is compatible with directstorage, and it doesn't differ from what gpu you have, as long as it's turing or ampere., although I think the idea or my idea is that the developer has to touch little code and that it comes integrated directly with Pytorch, for example.

weareai commented 2 years ago

@johnnynunez sorry missed this response. DALI is supported on many GPUs. In some cases, DALI relies on GDS. GeForce GPUs like the 3090FE don't support GDS in the CUDA user mode driver or runtime. In such cases, DALI wont be able to support these features (DALI use of GDS). Please let me know how we can help to enable your use case and perhaps find alternate solutions to your application.