Closed pytorch-bot[bot] closed 4 weeks ago
Resolving the issue because the test is not flaky anymore after 2550 reruns without any failures and the issue hasn't been updated in 14 days. Please reopen the issue to re-disable the test if you think this is a false positive
Platforms: linux, slow
This test was disabled because it is failing in CI. See recent examples and the most recent trunk workflow logs.
Over the past 3 hours, it has been determined flaky in 6 workflow(s) with 6 failures and 6 successes.
Debugging instructions (after clicking on the recent samples link): DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets:
test_compile_backward_fmod_cuda_float32
Sample error message
``` Traceback (most recent call last): File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 2918, in wrapper method(*args, **kwargs) File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 2918, in wrapper method(*args, **kwargs) File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 2917, in wrapper with policy(): File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 2348, in __exit__ raise RuntimeError(msg) RuntimeError: CUDA driver API confirmed a leak in __main__.TestNestedTensorOpInfoCUDA.test_compile_backward_fmod_cuda_float32! Caching allocator allocated memory was 0 and is now reported as 4608 on device 0. CUDA driver allocated memory was 388366336 and is now 390463488. To execute this test, run the following from the base repo dir: PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_nestedtensor.py -k TestNestedTensorOpInfoCUDA.test_compile_backward_fmod_cuda_float32 This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 ```Test file path:
test_nestedtensor.py
cc @clee2000 @cpuhrsch @jbschlosser @bhosmer @drisspg @soulitzer @davidberard98 @YuqingJ