NVIDIA / nvidia-docker

Build and run Docker containers leveraging NVIDIA GPUs
Apache License 2.0
17.25k stars 2.03k forks source link

docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed #1648

Closed zyr-NULL closed 2 years ago

zyr-NULL commented 2 years ago

Issue or feature description

when i use docker to create container, i get this error

docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.

Steps to reproduce the issue

here is some Information

-- WARNING, the following logs are for debugging purposes only --

I0621 07:07:51.735875 4789 nvc.c:376] initializing library context (version=1.10.0, build=395fd41701117121f1fd04ada01e1d7e006a37ae) I0621 07:07:51.735941 4789 nvc.c:350] using root / I0621 07:07:51.735947 4789 nvc.c:351] using ldcache /etc/ld.so.cache I0621 07:07:51.735963 4789 nvc.c:352] using unprivileged user 1000:1000 I0621 07:07:51.735984 4789 nvc.c:393] attempting to load dxcore to see if we are running under Windows Subsystem for Linux (WSL) I0621 07:07:51.736064 4789 nvc.c:395] dxcore initialization failed, continuing assuming a non-WSL environment W0621 07:07:51.739205 4791 nvc.c:273] failed to set inheritable capabilities W0621 07:07:51.739329 4791 nvc.c:274] skipping kernel modules load due to failure I0621 07:07:51.739807 4793 rpc.c:71] starting driver rpc service W0621 07:08:16.774958 4789 rpc.c:121] terminating driver rpc service (forced) I0621 07:08:20.481845 4789 rpc.c:135] driver rpc service terminated with signal 15 nvidia-container-cli: initialization error: driver rpc error: timed out I0621 07:08:20.481972 4789 nvc.c:434] shutting down library context

*  Kernel version
`Linux gpu-server 4.15.0-187-generic NVIDIA/nvidia-docker#198-Ubuntu SMP Tue Jun 14 03:23:51 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux`

* Driver information

==============NVSMI LOG==============

Timestamp : Tue Jun 21 07:13:57 2022 Driver Version : 515.48.07 CUDA Version : 11.7

Attached GPUs : 4 GPU 00000000:01:00.0 Product Name : NVIDIA A100-SXM4-40GB Product Brand : NVIDIA Product Architecture : Ampere Display Mode : Disabled Display Active : Disabled Persistence Mode : Disabled MIG Mode Current : Disabled Pending : Disabled Accounting Mode : Disabled Accounting Mode Buffer Size : 4000 Driver Model Current : N/A Pending : N/A Serial Number : 1561221014674 GPU UUID : GPU-b67da01e-feba-d839-62c5-2773d4e963f0 Minor Number : 0 VBIOS Version : 92.00.19.00.13 MultiGPU Board : No Board ID : 0x100 GPU Part Number : 692-2G506-0202-002 Module ID : 3 Inforom Version Image Version : G506.0202.00.02 OEM Object : 2.0 ECC Object : 6.16 Power Management Object : N/A GPU Operation Mode Current : N/A Pending : N/A GSP Firmware Version : 515.48.07 GPU Virtualization Mode Virtualization Mode : None Host VGPU Mode : N/A IBMNPU Relaxed Ordering Mode : N/A PCI Bus : 0x01 Device : 0x00 Domain : 0x0000 Device Id : 0x20B010DE Bus Id : 00000000:01:00.0 Sub System Id : 0x144E10DE GPU Link Info PCIe Generation Max : 4 Current : 4 Link Width Max : 16x Current : 16x Bridge Chip Type : N/A Firmware : N/A Replays Since Reset : 0 Replay Number Rollovers : 0 Tx Throughput : 0 KB/s Rx Throughput : 0 KB/s Fan Speed : N/A Performance State : P0 Clocks Throttle Reasons Idle : Active Applications Clocks Setting : Not Active SW Power Cap : Not Active HW Slowdown : Not Active HW Thermal Slowdown : Not Active HW Power Brake Slowdown : Not Active Sync Boost : Not Active SW Thermal Slowdown : Not Active Display Clock Setting : Not Active FB Memory Usage Total : 40960 MiB Reserved : 571 MiB Used : 0 MiB Free : 40388 MiB BAR1 Memory Usage Total : 65536 MiB Used : 1 MiB Free : 65535 MiB Compute Mode : Default Utilization Gpu : 0 % Memory : 0 % Encoder : 0 % Decoder : 0 % Encoder Stats Active Sessions : 0 Average FPS : 0 Average Latency : 0 FBC Stats Active Sessions : 0 Average FPS : 0 Average Latency : 0 Ecc Mode Current : Enabled Pending : Enabled ECC Errors Volatile SRAM Correctable : 0 SRAM Uncorrectable : 0 DRAM Correctable : 0 DRAM Uncorrectable : 0 Aggregate SRAM Correctable : 0 SRAM Uncorrectable : 0 DRAM Correctable : 0 DRAM Uncorrectable : 0 Retired Pages Single Bit ECC : N/A Double Bit ECC : N/A Pending Page Blacklist : N/A Remapped Rows Correctable Error : 0 Uncorrectable Error : 0 Pending : No Remapping Failure Occurred : No Bank Remap Availability Histogram Max : 640 bank(s) High : 0 bank(s) Partial : 0 bank(s) Low : 0 bank(s) None : 0 bank(s) Temperature GPU Current Temp : 32 C GPU Shutdown Temp : 92 C GPU Slowdown Temp : 89 C GPU Max Operating Temp : 85 C GPU Target Temperature : N/A Memory Current Temp : 34 C Memory Max Operating Temp : 95 C Power Readings Power Management : Supported Power Draw : 54.92 W Power Limit : 400.00 W Default Power Limit : 400.00 W Enforced Power Limit : 400.00 W Min Power Limit : 100.00 W Max Power Limit : 400.00 W Clocks Graphics : 1080 MHz SM : 1080 MHz Memory : 1215 MHz Video : 975 MHz Applications Clocks Graphics : 1095 MHz Memory : 1215 MHz Default Applications Clocks Graphics : 1095 MHz Memory : 1215 MHz Max Clocks Graphics : 1410 MHz SM : 1410 MHz Memory : 1215 MHz Video : 1290 MHz Max Customer Boost Clocks Graphics : 1410 MHz Clock Policy Auto Boost : N/A Auto Boost Default : N/A Voltage Graphics : 731.250 mV Processes : None

GPU 00000000:41:00.0 Product Name : NVIDIA A100-SXM4-40GB Product Brand : NVIDIA Product Architecture : Ampere Display Mode : Disabled Display Active : Disabled Persistence Mode : Disabled MIG Mode Current : Disabled Pending : Disabled Accounting Mode : Disabled Accounting Mode Buffer Size : 4000 Driver Model Current : N/A Pending : N/A Serial Number : 1561221014888 GPU UUID : GPU-6ca82e47-c63a-1bea-38ad-d3af9e1dc26b Minor Number : 1 VBIOS Version : 92.00.19.00.13 MultiGPU Board : No Board ID : 0x4100 GPU Part Number : 692-2G506-0202-002 Module ID : 1 Inforom Version Image Version : G506.0202.00.02 OEM Object : 2.0 ECC Object : 6.16 Power Management Object : N/A GPU Operation Mode Current : N/A Pending : N/A GSP Firmware Version : 515.48.07 GPU Virtualization Mode Virtualization Mode : None Host VGPU Mode : N/A IBMNPU Relaxed Ordering Mode : N/A PCI Bus : 0x41 Device : 0x00 Domain : 0x0000 Device Id : 0x20B010DE Bus Id : 00000000:41:00.0 Sub System Id : 0x144E10DE GPU Link Info PCIe Generation Max : 4 Current : 4 Link Width Max : 16x Current : 16x Bridge Chip Type : N/A Firmware : N/A Replays Since Reset : 0 Replay Number Rollovers : 0 Tx Throughput : 0 KB/s Rx Throughput : 0 KB/s Fan Speed : N/A Performance State : P0 Clocks Throttle Reasons Idle : Active Applications Clocks Setting : Not Active SW Power Cap : Not Active HW Slowdown : Not Active HW Thermal Slowdown : Not Active HW Power Brake Slowdown : Not Active Sync Boost : Not Active SW Thermal Slowdown : Not Active Display Clock Setting : Not Active FB Memory Usage Total : 40960 MiB Reserved : 571 MiB Used : 0 MiB Free : 40388 MiB BAR1 Memory Usage Total : 65536 MiB Used : 1 MiB Free : 65535 MiB Compute Mode : Default Utilization Gpu : 0 % Memory : 0 % Encoder : 0 % Decoder : 0 % Encoder Stats Active Sessions : 0 Average FPS : 0 Average Latency : 0 FBC Stats Active Sessions : 0 Average FPS : 0 Average Latency : 0 Ecc Mode Current : Enabled Pending : Enabled ECC Errors Volatile SRAM Correctable : 0 SRAM Uncorrectable : 0 DRAM Correctable : 0 DRAM Uncorrectable : 0 Aggregate SRAM Correctable : 0 SRAM Uncorrectable : 0 DRAM Correctable : 0 DRAM Uncorrectable : 0 Retired Pages Single Bit ECC : N/A Double Bit ECC : N/A Pending Page Blacklist : N/A Remapped Rows Correctable Error : 0 Uncorrectable Error : 0 Pending : No Remapping Failure Occurred : No Bank Remap Availability Histogram Max : 640 bank(s) High : 0 bank(s) Partial : 0 bank(s) Low : 0 bank(s) None : 0 bank(s) Temperature GPU Current Temp : 30 C GPU Shutdown Temp : 92 C GPU Slowdown Temp : 89 C GPU Max Operating Temp : 85 C GPU Target Temperature : N/A Memory Current Temp : 40 C Memory Max Operating Temp : 95 C Power Readings Power Management : Supported Power Draw : 57.45 W Power Limit : 400.00 W Default Power Limit : 400.00 W Enforced Power Limit : 400.00 W Min Power Limit : 100.00 W Max Power Limit : 400.00 W Clocks Graphics : 915 MHz SM : 915 MHz Memory : 1215 MHz Video : 780 MHz Applications Clocks Graphics : 1095 MHz Memory : 1215 MHz Default Applications Clocks Graphics : 1095 MHz Memory : 1215 MHz Max Clocks Graphics : 1410 MHz SM : 1410 MHz Memory : 1215 MHz Video : 1290 MHz Max Customer Boost Clocks Graphics : 1410 MHz Clock Policy Auto Boost : N/A Auto Boost Default : N/A Voltage Graphics : 700.000 mV Processes : None

GPU 00000000:81:00.0 Product Name : NVIDIA A100-SXM4-40GB Product Brand : NVIDIA Product Architecture : Ampere Display Mode : Disabled Display Active : Disabled Persistence Mode : Disabled MIG Mode Current : Disabled Pending : Disabled Accounting Mode : Disabled Accounting Mode Buffer Size : 4000 Driver Model Current : N/A Pending : N/A Serial Number : 1561221015040 GPU UUID : GPU-7e4b55d2-75fc-8ab5-e212-09e69e84704b Minor Number : 2 VBIOS Version : 92.00.19.00.13 MultiGPU Board : No Board ID : 0x8100 GPU Part Number : 692-2G506-0202-002 Module ID : 2 Inforom Version Image Version : G506.0202.00.02 OEM Object : 2.0 ECC Object : 6.16 Power Management Object : N/A GPU Operation Mode Current : N/A Pending : N/A GSP Firmware Version : 515.48.07 GPU Virtualization Mode Virtualization Mode : None Host VGPU Mode : N/A IBMNPU Relaxed Ordering Mode : N/A PCI Bus : 0x81 Device : 0x00 Domain : 0x0000 Device Id : 0x20B010DE Bus Id : 00000000:81:00.0 Sub System Id : 0x144E10DE GPU Link Info PCIe Generation Max : 4 Current : 4 Link Width Max : 16x Current : 16x Bridge Chip Type : N/A Firmware : N/A Replays Since Reset : 0 Replay Number Rollovers : 0 Tx Throughput : 0 KB/s Rx Throughput : 0 KB/s Fan Speed : N/A Performance State : P0 Clocks Throttle Reasons Idle : Active Applications Clocks Setting : Not Active SW Power Cap : Not Active HW Slowdown : Not Active HW Thermal Slowdown : Not Active HW Power Brake Slowdown : Not Active Sync Boost : Not Active SW Thermal Slowdown : Not Active Display Clock Setting : Not Active FB Memory Usage Total : 40960 MiB Reserved : 571 MiB Used : 0 MiB Free : 40388 MiB BAR1 Memory Usage Total : 65536 MiB Used : 1 MiB Free : 65535 MiB Compute Mode : Default Utilization Gpu : 0 % Memory : 0 % Encoder : 0 % Decoder : 0 % Encoder Stats Active Sessions : 0 Average FPS : 0 Average Latency : 0 FBC Stats Active Sessions : 0 Average FPS : 0 Average Latency : 0 Ecc Mode Current : Enabled Pending : Enabled ECC Errors Volatile SRAM Correctable : 0 SRAM Uncorrectable : 0 DRAM Correctable : 0 DRAM Uncorrectable : 0 Aggregate SRAM Correctable : 0 SRAM Uncorrectable : 0 DRAM Correctable : 0 DRAM Uncorrectable : 0 Retired Pages Single Bit ECC : N/A Double Bit ECC : N/A Pending Page Blacklist : N/A Remapped Rows Correctable Error : 0 Uncorrectable Error : 0 Pending : No Remapping Failure Occurred : No Bank Remap Availability Histogram Max : 640 bank(s) High : 0 bank(s) Partial : 0 bank(s) Low : 0 bank(s) None : 0 bank(s) Temperature GPU Current Temp : 32 C GPU Shutdown Temp : 92 C GPU Slowdown Temp : 89 C GPU Max Operating Temp : 85 C GPU Target Temperature : N/A Memory Current Temp : 33 C Memory Max Operating Temp : 95 C Power Readings Power Management : Supported Power Draw : 54.65 W Power Limit : 400.00 W Default Power Limit : 400.00 W Enforced Power Limit : 400.00 W Min Power Limit : 100.00 W Max Power Limit : 400.00 W Clocks Graphics : 1080 MHz SM : 1080 MHz Memory : 1215 MHz Video : 975 MHz Applications Clocks Graphics : 1095 MHz Memory : 1215 MHz Default Applications Clocks Graphics : 1095 MHz Memory : 1215 MHz Max Clocks Graphics : 1410 MHz SM : 1410 MHz Memory : 1215 MHz Video : 1290 MHz Max Customer Boost Clocks Graphics : 1410 MHz Clock Policy Auto Boost : N/A Auto Boost Default : N/A Voltage Graphics : 712.500 mV Processes : None

GPU 00000000:C1:00.0 Product Name : NVIDIA A100-SXM4-40GB Product Brand : NVIDIA Product Architecture : Ampere Display Mode : Disabled Display Active : Disabled Persistence Mode : Disabled MIG Mode Current : Disabled Pending : Disabled Accounting Mode : Disabled Accounting Mode Buffer Size : 4000 Driver Model Current : N/A Pending : N/A Serial Number : 1561221014695 GPU UUID : GPU-66ba085a-5496-d204-d4da-aa9f112d3fd8 Minor Number : 3 VBIOS Version : 92.00.19.00.13 MultiGPU Board : No Board ID : 0xc100 GPU Part Number : 692-2G506-0202-002 Module ID : 0 Inforom Version Image Version : G506.0202.00.02 OEM Object : 2.0 ECC Object : 6.16 Power Management Object : N/A GPU Operation Mode Current : N/A Pending : N/A GSP Firmware Version : 515.48.07 GPU Virtualization Mode Virtualization Mode : None Host VGPU Mode : N/A IBMNPU Relaxed Ordering Mode : N/A PCI Bus : 0xC1 Device : 0x00 Domain : 0x0000 Device Id : 0x20B010DE Bus Id : 00000000:C1:00.0 Sub System Id : 0x144E10DE GPU Link Info PCIe Generation Max : 4 Current : 4 Link Width Max : 16x Current : 16x Bridge Chip Type : N/A Firmware : N/A Replays Since Reset : 0 Replay Number Rollovers : 0 Tx Throughput : 0 KB/s Rx Throughput : 0 KB/s Fan Speed : N/A Performance State : P0 Clocks Throttle Reasons Idle : Active Applications Clocks Setting : Not Active SW Power Cap : Not Active HW Slowdown : Not Active HW Thermal Slowdown : Not Active HW Power Brake Slowdown : Not Active Sync Boost : Not Active SW Thermal Slowdown : Not Active Display Clock Setting : Not Active FB Memory Usage Total : 40960 MiB Reserved : 571 MiB Used : 0 MiB Free : 40388 MiB BAR1 Memory Usage Total : 65536 MiB Used : 1 MiB Free : 65535 MiB Compute Mode : Default Utilization Gpu : 0 % Memory : 0 % Encoder : 0 % Decoder : 0 % Encoder Stats Active Sessions : 0 Average FPS : 0 Average Latency : 0 FBC Stats Active Sessions : 0 Average FPS : 0 Average Latency : 0 Ecc Mode Current : Enabled Pending : Enabled ECC Errors Volatile SRAM Correctable : 0 SRAM Uncorrectable : 0 DRAM Correctable : 0 DRAM Uncorrectable : 0 Aggregate SRAM Correctable : 0 SRAM Uncorrectable : 0 DRAM Correctable : 0 DRAM Uncorrectable : 0 Retired Pages Single Bit ECC : N/A Double Bit ECC : N/A Pending Page Blacklist : N/A Remapped Rows Correctable Error : 0 Uncorrectable Error : 0 Pending : No Remapping Failure Occurred : No Bank Remap Availability Histogram Max : 640 bank(s) High : 0 bank(s) Partial : 0 bank(s) Low : 0 bank(s) None : 0 bank(s) Temperature GPU Current Temp : 30 C GPU Shutdown Temp : 92 C GPU Slowdown Temp : 89 C GPU Max Operating Temp : 85 C GPU Target Temperature : N/A Memory Current Temp : 38 C Memory Max Operating Temp : 95 C Power Readings Power Management : Supported Power Draw : 59.01 W Power Limit : 400.00 W Default Power Limit : 400.00 W Enforced Power Limit : 400.00 W Min Power Limit : 100.00 W Max Power Limit : 400.00 W Clocks Graphics : 1080 MHz SM : 1080 MHz Memory : 1215 MHz Video : 975 MHz Applications Clocks Graphics : 1095 MHz Memory : 1215 MHz Default Applications Clocks Graphics : 1095 MHz Memory : 1215 MHz Max Clocks Graphics : 1410 MHz SM : 1410 MHz Memory : 1215 MHz Video : 1290 MHz Max Customer Boost Clocks Graphics : 1410 MHz Clock Policy Auto Boost : N/A Auto Boost Default : N/A Voltage Graphics : 737.500 mV Processes : None

* docker version

Client: Docker Engine - Community Version: 20.10.17 API version: 1.41 Go version: go1.17.11 Git commit: 100c701 Built: Mon Jun 6 23:02:56 2022 OS/Arch: linux/amd64 Context: default Experimental: true

Server: Docker Engine - Community Engine: Version: 20.10.11 API version: 1.41 (minimum version 1.12) Go version: go1.16.9 Git commit: 847da18 Built: Thu Nov 18 00:35:16 2021 OS/Arch: linux/amd64 Experimental: false containerd: Version: 1.6.6 GitCommit: 10c12954828e7c7c9b6e0ea9b0c02b01407d3ae1 nvidia: Version: 1.1.2 GitCommit: v1.1.2-0-ga916309 docker-init: Version: 0.19.0 GitCommit: de40ad0

* NVIDIA container library version

cli-version: 1.10.0 lib-version: 1.10.0 build date: 2022-06-13T10:39+00:00 build revision: 395fd41701117121f1fd04ada01e1d7e006a37ae build compiler: x86_64-linux-gnu-gcc-7 7.5.0 build platform: x86_64 build flags: -D_GNU_SOURCE -D_FORTIFY_SOURCE=2 -DNDEBUG -std=gnu11 -O2 -g -fdata-sections -ffunction-sections -fplan9-extensions -fstack-protector -fno-strict-aliasing -fvisibility=hidden -Wall -Wextra -Wcast-align -Wpointer-arith -Wmissing-prototypes -Wnonnull -Wwrite-strings -Wlogical-op -Wformat=2 -Wmissing-format-attribute -Winit-self -Wshadow -Wstrict-prototypes -Wunreachable-code -Wconversion -Wsign-conversion -Wno-unknown-warning-option -Wno-format-extra-args -Wno-gnu-alignof-expression -Wl,-zrelro -Wl,-znow -Wl,-zdefs -Wl,--gc-sections

*  NVIDIA packages version

Desired=Unknown/Install/Remove/Purge/Hold | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad) ||/ Name Version Architecture Description +++-====================================-=======================-=======================-============================================================================= ii libnvidia-container-tools 1.10.0-1 amd64 NVIDIA container runtime library (command-line tools) ii libnvidia-container1:amd64 1.10.0-1 amd64 NVIDIA container runtime library un nvidia-container-runtime (no description available) un nvidia-container-runtime-hook (no description available) ii nvidia-container-toolkit 1.10.0-1 amd64 NVIDIA container runtime hook un nvidia-docker (no description available) ii nvidia-docker2 2.7.0-1 all nvidia-docker CLI wrapper

elezar commented 2 years ago

@zyr-NULL to confirm whether this is a regression or not, would you be able to repeat the test after downgrading to nvidia-container-toolkit=1.9.0-1?

zyr-NULL commented 2 years ago

@elezar i make nvidia-container-toolkit=1.9.0-1,but output remain the same.

elezar commented 2 years ago

Sorry, I should have made it clearer. Could you also downgrade libnvidia-container-tools and libnvidia-container1 and repeat.

zyr-NULL commented 2 years ago

@elezar this the same error

gpu-server@gpu-server:~$ dpkg -l '*nvidia*'
Desired=Unknown/Install/Remove/Purge/Hold
| Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
|/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
||/ Name                                 Version                 Architecture            Description
+++-====================================-=======================-=======================-=============================================================================
ii  libnvidia-container-tools            1.9.0-1                 amd64                   NVIDIA container runtime library (command-line tools)
ii  libnvidia-container1:amd64           1.9.0-1                 amd64                   NVIDIA container runtime library
un  nvidia-container-runtime             <none>                  <none>                  (no description available)
un  nvidia-container-runtime-hook        <none>                  <none>                  (no description available)
ii  nvidia-container-toolkit             1.9.0-1                 amd64                   NVIDIA container runtime hook
un  nvidia-docker                        <none>                  <none>                  (no description available)
ii  nvidia-docker2                       2.7.0-1                 all                     nvidia-docker CLI wrapper
gpu-server@gpu-server:~$ systemctl restart docker 
==== AUTHENTICATING FOR org.freedesktop.systemd1.manage-units ===
Authentication is required to restart 'docker.service'.
Authenticating as: gpu-server
Password: 
==== AUTHENTICATION COMPLETE ===
gpu-server@gpu-server:~$ sudo docker run --rm --gpus all nvidia/cuda:11.0.3-base-ubuntu20.04 nvidia-smi
docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: nvidia-container-cli: initialization error: driver rpc error: timed out: unknown.
klueska commented 2 years ago

What happens when you run nvidia-smi directly on the host?

Also, this warning in the log seems suspicious:

W0621 07:07:51.739205 4791 nvc.c:273] failed to set inheritable capabilities

Is there something unconventional about how you are running docker on your host? For example, rootless, in a snap sandbox, etc.

zyr-NULL commented 2 years ago

@klueska when i run this nvidia-smi on the host, it will two minutes for expected results

gpu-server@gpu-server:~$ nvidia-smi
Tue Jun 21 13:47:51 2022       
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 515.48.07    Driver Version: 515.48.07    CUDA Version: 11.7     |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|                               |                      |               MIG M. |
|===============================+======================+======================|
|   0  NVIDIA A100-SXM...  Off  | 00000000:01:00.0 Off |                    0 |
| N/A   32C    P0    52W / 400W |      0MiB / 40960MiB |      0%      Default |
|                               |                      |             Disabled |
+-------------------------------+----------------------+----------------------+
|   1  NVIDIA A100-SXM...  Off  | 00000000:41:00.0 Off |                    0 |
| N/A   30C    P0    53W / 400W |      0MiB / 40960MiB |      0%      Default |
|                               |                      |             Disabled |
+-------------------------------+----------------------+----------------------+
|   2  NVIDIA A100-SXM...  Off  | 00000000:81:00.0 Off |                    0 |
| N/A   31C    P0    52W / 400W |      0MiB / 40960MiB |      0%      Default |
|                               |                      |             Disabled |
+-------------------------------+----------------------+----------------------+
|   3  NVIDIA A100-SXM...  Off  | 00000000:C1:00.0 Off |                    0 |
| N/A   30C    P0    52W / 400W |      0MiB / 40960MiB |      0%      Default |
|                               |                      |             Disabled |
+-------------------------------+----------------------+----------------------+

+-----------------------------------------------------------------------------+
| Processes:                                                                  |
|  GPU   GI   CI        PID   Type   Process name                  GPU Memory |
|        ID   ID                                                   Usage      |
|=============================================================================|
|  No running processes found                                                 |
+-----------------------------------------------------------------------------+

i run docker in root model and i used sudo gpasswd -a gpu-server docker, then run docker ,in addition i did not installed docker with snap.

klueska commented 2 years ago

If it’s taking that long to get a result from nvidia-smi on the host then I could understand why the RPC might time out in the nvidia-container-cli when trying to get a result from the driver RPC call.

Im assuming this is happening because (1) you are not running the nvidia-persistenced daemon and (2) your GPUs are not in persistence mode.

Both (1) and (2) achieve the same thing, but (1) is the preferred method to keep the GPU driver alive even when no clients are attached.

Try enabling one of these methods and report back.

zyr-NULL commented 2 years ago

@klueska Thanks, my problem has been solved through enable persistence model.

elezar commented 2 years ago

@zyr-NULL given your comment above I am closing this issue. Please reopen if there is still a problem.

alkavan commented 2 years ago

@klueska Yeah dude thanks =) That's exactly it, running nvidia-persistenced as root without any additional parameters solved it. You can confirm with nvidia-smi that the GPU units load faster.

LukeLIN-web commented 2 years ago

I met this problem when I use shell to run , after I add #!/bin/bash at the top of shell script. This error disappear.

montmejat commented 1 year ago

If it’s taking that long to get a result from nvidia-smi on the host then I could understand why the RPC might time out in the nvidia-container-cli when trying to get a result from the driver RPC call.

Im assuming this is happening because (1) you are not running the nvidia-persistenced daemon and (2) your GPUs are not in persistence mode.

Both (1) and (2) achieve the same thing, but (1) is the preferred method to keep the GPU driver alive even when no clients are attached.

Try enabling one of these methods and report back.

I tried using nvidia-smi -i 0 -pm ENABLED but not luck. I'm having the same error as @zyr-NULL. I then tried using:

# nvidia-persistenced --user aurelien
nvidia-persistenced failed to initialize. Check syslog for more details.

Here are the logs:

# cat /var/log/syslog
Feb  6 14:15:46 pop-os nvidia-persistenced: device 0000:01:00.0 - persistence mode disabled.
Feb  6 14:15:46 pop-os nvidia-persistenced: device 0000:01:00.0 - NUMA memory offlined.
Feb  6 14:16:06 pop-os nvidia-persistenced: Failed to change ownership of /var/run/nvidia-persistenced: Operation not permitted
Feb  6 14:16:06 pop-os nvidia-persistenced: Shutdown (16491)
Feb  6 14:16:38 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb  6 14:16:38 pop-os nvidia-persistenced: Shutdown (16516)
Feb  6 14:16:46 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb  6 14:16:48 pop-os rtkit-daemon[1589]: message repeated 5 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb  6 14:17:01 pop-os nvidia-persistenced: device 0000:01:00.0 - persistence mode enabled.
Feb  6 14:17:01 pop-os nvidia-persistenced: device 0000:01:00.0 - NUMA memory onlined.
Feb  6 14:17:01 pop-os CRON[16652]: (root) CMD (   cd / && run-parts --report /etc/cron.hourly)
Feb  6 14:17:03 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb  6 14:17:03 pop-os nvidia-persistenced: Shutdown (16656)
Feb  6 14:18:03 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb  6 14:18:04 pop-os rtkit-daemon[1589]: message repeated 3 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb  6 14:18:33 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb  6 14:18:33 pop-os nvidia-persistenced: Shutdown (16799)
Feb  6 14:18:38 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb  6 14:18:53 pop-os rtkit-daemon[1589]: message repeated 7 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb  6 14:19:17 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb  6 14:19:17 pop-os nvidia-persistenced: Shutdown (16994)
Feb  6 14:20:04 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb  6 14:20:07 pop-os rtkit-daemon[1589]: message repeated 5 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb  6 14:20:56 pop-os systemd[1]: Starting Cleanup of Temporary Directories...
Feb  6 14:20:56 pop-os systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully.
Feb  6 14:20:56 pop-os systemd[1]: Finished Cleanup of Temporary Directories.
Feb  6 14:21:03 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb  6 14:21:07 pop-os rtkit-daemon[1589]: message repeated 3 times: [ Supervising 8 threads of 5 processes of 1 users.]
Feb  6 14:21:15 pop-os systemd[1]: Reloading.
Feb  6 14:23:27 pop-os chronyd[1007]: NTS-KE session with 52.10.183.132:4460 (oregon.time.system76.com) timed out
Feb  6 14:23:33 pop-os chronyd[1007]: NTS-KE session with 18.228.202.30:4460 (brazil.time.system76.com) timed out
Feb  6 14:23:33 pop-os chronyd[1007]: NTS-KE session with 15.237.97.214:4460 (paris.time.system76.com) timed out
Feb  6 14:23:34 pop-os chronyd[1007]: NTS-KE session with 3.134.129.152:4460 (ohio.time.system76.com) timed out
Feb  6 14:23:36 pop-os chronyd[1007]: NTS-KE session with 3.220.42.39:4460 (virginia.time.system76.com) timed out
Feb  6 14:24:21 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb  6 14:24:21 pop-os rtkit-daemon[1589]: Supervising 8 threads of 5 processes of 1 users.
Feb  6 14:25:04 pop-os nvidia-persistenced: Failed to open PID file: Permission denied
Feb  6 14:25:04 pop-os nvidia-persistenced: Shutdown (17372)

Any ideas @klueska? :smiling_face_with_tear:

Here's the error by the way:

$ sudo docker run --rm --gpus all nvidia/cuda:11.6.2-base-ubuntu20.04 nvidia-smi
docker: Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: container error: cgroup subsystem devices not found: unknown.
Metrora commented 1 year ago

I have the same problem with @aurelien-m when I run

$ sudo docker run --rm --runtime=nvidia --gpus all nvidia/cuda:11.6.2-base-ubuntu20.04 nvidia-smi

fighterhit commented 1 year ago

I met this problem too. My GPU is A30 and GPU driver is 525.85.12.

 State:      Terminated
      Reason:   StartError
      Message:  failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: detection error: nvml error: unknown error: unknown
[Tue Feb 21 21:10:41 2023] NVRM: Xid (PCI:0000:22:00): 119, pid=3660230, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
elezar commented 1 year ago

@fighterhit this may be related to changes in the GSP firmware paths and should be addressed in the v1.13.0-rc.1 release. Would you be able to give it a try?

fighterhit commented 1 year ago

@fighterhit this may be related to changes in the GSP firmware paths and should be addressed in the v1.13.0-rc.1 release. Would you be able to give it a try?

Hi @elezar , yes I can, but do I need to restart the node or the containers on the node? Because it may need to be deployed in our production environment.

Or does this problem exist in previous versions, I can also accept lower versions :).

elezar commented 1 year ago

Updating the nvidia-container-toolkit should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers.

If this is the firmware issue that I think it is, then a lower version would not work.

What is the output of

ls -al /firmware/nvidia/525.85*/*

If there is a single gsp.bin file there, then a v1.13 release is not required and any NVIDIA Container Toolkit version after v1.8.1 should work. If there are multiple gsp_*.bin files there then v1.13.0-rc.1 would be required to address this.

fighterhit commented 1 year ago

Updating the nvidia-container-toolkit should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers.

If this is the firmware issue that I think it is, then a lower version would not work.

What is the output of

ls -al /firmware/nvidia/525.85*/*

If there is a single gsp.bin file there, then a v1.13 release is not required and any NVIDIA Container Toolkit version after v1.8.1 should work. If there are multiple gsp_*.bin files there then v1.13.0-rc.1 would be required to address this.

@elezar The output is:

ls: cannot access '/firmware/nvidia/525.85*/*': No such file or directory

It seems that there is no such /firmware directory on our all GPU nodes. FYI, our driver is installed by NVIDIA-Linux-x86_64-525.85.12.run.

fighterhit commented 1 year ago

Updating the nvidia-container-toolkit should not require a restart. The components of the NVIDIA container stack are only invoked when containers are created and would not affect running containers. If this is the firmware issue that I think it is, then a lower version would not work. What is the output of

ls -al /firmware/nvidia/525.85*/*

If there is a single gsp.bin file there, then a v1.13 release is not required and any NVIDIA Container Toolkit version after v1.8.1 should work. If there are multiple gsp_*.bin files there then v1.13.0-rc.1 would be required to address this.

@elezar The output is:

ls: cannot access '/firmware/nvidia/525.85*/*': No such file or directory

It seems that there is no such /firmware directory on our all GPU nodes. FYI, our driver is installed by NVIDIA-Linux-x86_64-525.85.12.run.

Hi @elezar , do you have any more suggestions? Thanks!

klueska commented 1 year ago

The path is /lib/firmware/…

fighterhit commented 1 year ago

The path is /lib/firmware/…

Thanks @klueska 😅, there are indeed two gsp_*.bin files in this directory.

-r--r--r-- 1 root root 25M Feb 18 11:34 /lib/firmware/nvidia/525.85.12/gsp_ad10x.bin
-r--r--r-- 1 root root 36M Feb 18 11:34 /lib/firmware/nvidia/525.85.12/gsp_tu10x.bin
fighterhit commented 1 year ago

Hi @klueska , can the solution provided by the nvidia-container-toolkit v1.13.0-rc.1 or the method of using the persistent mode you mentioned before solve this problem? Not sure if I also have to install this latest version since I have turned on persistent mode as you suggested.

klueska commented 1 year ago

The new RC adds support for detecting multiple GSP firmware files, which is required for container support to work correctly on the 525 driver.

The persistenced issue is still relevant but this is a new one related to the latest NVIDIA driver.

fighterhit commented 1 year ago

Thanks @klueska , how can I install this version v1.13.0-rc.1? I searched by using apt-cache show nvidia-container-toolkit but there seems to be no package for this version.

The persistenced issue is still relevant but this is a new one related to the latest NVIDIA driver.

Does the persistenced issue only appear on the latest driver (525.85.12)? Can I solve it by downgrading to a certain driver version?

klueska commented 1 year ago

Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.

That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.

Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.

Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit

fighterhit commented 1 year ago

Thank you for your help @klueska. What confuses me is that in the past, our GPU clusters (1080Ti, 2080Ti, 3090, A30, A100) have not turned on the persistent mode and there is no such problem, only A30 nodes (with driver 525.85.12) have this problem, none of the other types of GPU nodes(1080Ti, 2080Ti, 3090 with driver 525.78.01) have this problem.

fighterhit commented 1 year ago

Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading.

That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers.

Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it.

Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit

Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1 , my system distribution is debian11.

root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
      && curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
      && curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
         sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
         sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
# Unsupported distribution!
# Check https://nvidia.github.io/libnvidia-container
elezar commented 1 year ago

Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading. That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers. Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it. Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit

Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1 , my system distribution is debian11.

root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
      && curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
      && curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
         sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
         sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
# Unsupported distribution!
# Check https://nvidia.github.io/libnvidia-container

@fighterhit what distribution are you using? Please ensure that the distribution variable is set to ubuntu18.04 and try again.

klueska commented 1 year ago

He said he’s on debian11.

fighterhit commented 1 year ago

Persistenced is always needed, but the firmware issue could be „resolved“ by downgrading. That said, I’d recommend updating the nvidia-container-toolkit to ensure compatibility with all future drivers. Since the latest toolkit is still an RC it is not yet in our stable apt repo. You will need to configure our experimental repo to get access to it. Instructions here: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#setting-up-nvidia-container-toolkit

Hi @klueska @elezar , when I configure to use the experimental repo I get the following error and can't install the v1.13.0-rc.1 , my system distribution is debian11.

root@xxx:/etc/apt/sources.list.d# distribution=$(. /etc/os-release;echo $ID$VERSION_ID) \
      && curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg \
      && curl -s -L https://nvidia.github.io/libnvidia-container/experimental/$distribution/libnvidia-container.list | \
         sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
         sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
# Unsupported distribution!
# Check https://nvidia.github.io/libnvidia-container

@fighterhit what distribution are you using? Please ensure that the distribution variable is set to ubuntu18.04 and try again.

@elezar My distribution is Debian GNU/Linux 11 (bullseye).

fighterhit commented 1 year ago

Hi @klueska @elezar , I test the v1.13.0-rc.1 but it still report some errors:

  Warning  Failed     12s (x3 over 64s)  kubelet            Error: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: ldcache error: open failed: /sbin/ldconfig.real: no such file or directory: unknown
  Warning  BackOff  11s (x3 over 44s)  kubelet  Back-off restarting failed container

Maybe /sbin/ldconfig.real should be /sbin/ldconfig?

klueska commented 1 year ago

For debian, yes, it should be @/sbin/ldconfig (without the .real), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15

This file gets installed under/etc/nvidia-container-runtime/config.toml

The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?

fighterhit commented 1 year ago

For debian, yes, it should be @/sbin/ldconfig (without the .real), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15

This file gets installed under/etc/nvidia-container-runtime/config.toml

The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?

Yes @klueska , I failed to installed the latest version using the experimental repo(https://github.com/NVIDIA/nvidia-docker/issues/1648#issuecomment-1445585843), so I followed @elezar advice and set the distro to ubuntu18.04. Can I manually modify /etc/nvidia-container-runtime/config.toml to make it work? But I'm not sure if this will have any other effects.

elezar commented 1 year ago

For debian, yes, it should be @/sbin/ldconfig (without the .real), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15 This file gets installed under/etc/nvidia-container-runtime/config.toml The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?

Yes @klueska , I failed to installed the latest version using the experimental repo(#1648 (comment)), so I followed @elezar advice and set the distro to ubuntu18.04. Can I manually modify /etc/nvidia-container-runtime/config.toml to make it work? But I'm not sure if this will have any other effects.

Sorry about that. I was making an assumption about the distribution you are using. You can install the ubuntu18.04 package and then update the /etc/nvidia-container-runtime/config.toml value after the fact. This should have no other effects.

fighterhit commented 1 year ago

For debian, yes, it should be @/sbin/ldconfig (without the .real), as seen here: https://github.com/NVIDIA/nvidia-container-toolkit/blob/main/config/config.toml.debian#L15 This file gets installed under/etc/nvidia-container-runtime/config.toml The correct config file should have been selected automatically based on your distribution, were you not able to install the debian11 one directly?

Yes @klueska , I failed to installed the latest version using the experimental repo(#1648 (comment)), so I followed @elezar advice and set the distro to ubuntu18.04. Can I manually modify /etc/nvidia-container-runtime/config.toml to make it work? But I'm not sure if this will have any other effects.

Sorry about that. I was making an assumption about the distribution you are using. You can install the ubuntu18.04 package and then update the /etc/nvidia-container-runtime/config.toml value after the fact. This should have no other effects.

Thanks @elezar ! I have tried it and it works fine. It may be better if debian11 could be supported to use the experimental repo.

elezar commented 1 year ago

@fighterhit I have just double-checked our repo configuration and the issue is that for debian10 (to which debian11 redirects) the repo only had a libnvidia-container-experimental.list and no libnvidia-container.list file which is referred to by the instructions.

I have created a link / redirect for this now and the official instructions should work as expected. (It may take about 30 minutes for the changes to reflect in the repo though).

fighterhit commented 1 year ago

@fighterhit I have just double-checked our repo configuration and the issue is that for debian10 (to which debian11 redirects) the repo only had a libnvidia-container-experimental.list and no libnvidia-container.list file which is referred to by the instructions.

I have created a link / redirect for this now and the official instructions should work as expected. (It may take about 30 minutes for the changes to reflect in the repo though).

Thanks for your confirmation, it works now. @elezar

vikramelango commented 1 year ago

I am trying to install NVIDIA container toolkit on Amazon Linux 2, created a new ec2 instance and followed the instructions in this page and running into below issue. I get the same error when i tried this on a Ubuntu instance as well.

nvidia-container-cli: initialization error: load library failed: libnvidia-ml.so.1: cannot open shared object file: no such file or directory: unknown.
ERRO[0000] error waiting for container: context canceled 

@elezar @klueska please advise how to fix this issue, appreciate your inputs. Thanks

fighterhit commented 1 year ago

Hi @elezar @klueska , unfortunately I have used the latest toolkit but this problem reappeared, I think this may be related to the driver, and I asked the driver community for help but got no more reply (https://github.com/NVIDIA/open-gpu-kernel-modules/issues/446), https://forums.developer.nvidia.com/t/timeout-waiting-for-rpc-from-gsp/244789 could you communicate with the driver team about this issue? Because some users in the community also encountered the same problem.Thanks!

[Mon Mar  6 21:09:45 2023] NVRM: GPU at PCI:0000:61:00: GPU-81f5d81e-7906-c145-3def-82e281d7b260
[Mon Mar  6 21:09:45 2023] NVRM: GPU Board Serial Number: 1322621149674
[Mon Mar  6 21:09:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a00b0 0x0).
[Mon Mar  6 21:09:45 2023] CPU: 15 PID: 977720 Comm: nvidia-smi Tainted: P           OE     5.10.0-20-amd64 NVIDIA/nvidia-docker#1 Debian 5.10.158-2
[Mon Mar  6 21:09:45 2023] Hardware name: Inspur NF5468A5/YZMB-02382-101, BIOS 4.02.12 01/28/2022
[Mon Mar  6 21:09:45 2023] Call Trace:
[Mon Mar  6 21:09:45 2023]  dump_stack+0x6b/0x83
[Mon Mar  6 21:09:45 2023]  _nv011231rm+0x39d/0x470 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv011168rm+0x62/0x2e0 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv040022rm+0xdb/0x140 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv041148rm+0x2ce/0x3a0 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv015451rm+0x788/0x800 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv039541rm+0xac/0xe0 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv041150rm+0xac/0x140 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv041149rm+0x37a/0x4d0 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv039443rm+0xc9/0x150 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv039444rm+0x42/0x70 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? rm_cleanup_file_private+0x128/0x180 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv000554rm+0x49/0x60 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? _nv000694rm+0x7fb/0xc80 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar  6 21:09:45 2023]  ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar  6 21:09:45 2023]  ? do_syscall_64+0x33/0x80
[Mon Mar  6 21:09:45 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar  6 21:10:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0050 0x0).
[Mon Mar  6 21:11:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977720, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0040 0x0).
[Mon Mar  6 21:12:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
[Mon Mar  6 21:12:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
[Mon Mar  6 21:13:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
[Mon Mar  6 21:14:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
[Mon Mar  6 21:15:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
[Mon Mar  6 21:15:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
[Mon Mar  6 21:16:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
[Mon Mar  6 21:17:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=905052, name=python, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801702 0x4).
[Mon Mar  6 21:18:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
[Mon Mar  6 21:18:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
[Mon Mar  6 21:19:30 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801348 0x410).
[Mon Mar  6 21:20:15 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:21:00 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x0 0x6c).
[Mon Mar  6 21:21:45 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x80 0x38).
[Mon Mar  6 21:22:08 2023] INFO: task python:976158 blocked for more than 120 seconds.
[Mon Mar  6 21:22:08 2023]       Tainted: P           OE     5.10.0-20-amd64 NVIDIA/nvidia-docker#1 Debian 5.10.158-2
[Mon Mar  6 21:22:08 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[Mon Mar  6 21:22:08 2023] task:python          state:D stack:    0 pid:976158 ppid:835062 flags:0x00000000
[Mon Mar  6 21:22:08 2023] Call Trace:
[Mon Mar  6 21:22:08 2023]  __schedule+0x282/0x880
[Mon Mar  6 21:22:08 2023]  schedule+0x46/0xb0
[Mon Mar  6 21:22:08 2023]  rwsem_down_write_slowpath+0x246/0x4d0
[Mon Mar  6 21:22:08 2023]  os_acquire_rwlock_write+0x31/0x40 [nvidia]
[Mon Mar  6 21:22:08 2023]  _nv038505rm+0xc/0x30 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039453rm+0x18d/0x1d0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv041182rm+0x45/0xd0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv041127rm+0x142/0x2b0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039415rm+0x15a/0x2e0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039416rm+0x5b/0x90 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039416rm+0x31/0x90 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv012688rm+0x1d/0x30 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039431rm+0xb0/0xb0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv012710rm+0x54/0x70 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv011426rm+0xc4/0x120 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv000659rm+0x63/0x70 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv000582rm+0x2c/0x40 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv000694rm+0x86c/0xc80 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar  6 21:22:08 2023]  ? do_syscall_64+0x33/0x80
[Mon Mar  6 21:22:08 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar  6 21:22:08 2023] INFO: task nvidia-smi:993195 blocked for more than 120 seconds.
[Mon Mar  6 21:22:08 2023]       Tainted: P           OE     5.10.0-20-amd64 NVIDIA/nvidia-docker#1 Debian 5.10.158-2
[Mon Mar  6 21:22:08 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[Mon Mar  6 21:22:08 2023] task:nvidia-smi      state:D stack:    0 pid:993195 ppid: 58577 flags:0x00000004
[Mon Mar  6 21:22:08 2023] Call Trace:
[Mon Mar  6 21:22:08 2023]  __schedule+0x282/0x880
[Mon Mar  6 21:22:08 2023]  schedule+0x46/0xb0
[Mon Mar  6 21:22:08 2023]  rwsem_down_write_slowpath+0x246/0x4d0
[Mon Mar  6 21:22:08 2023]  os_acquire_rwlock_write+0x31/0x40 [nvidia]
[Mon Mar  6 21:22:08 2023]  _nv038505rm+0xc/0x30 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039453rm+0x18d/0x1d0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv041182rm+0x45/0xd0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv041127rm+0x142/0x2b0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039415rm+0x15a/0x2e0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039416rm+0x5b/0x90 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv039416rm+0x31/0x90 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv000560rm+0x59/0x70 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv000560rm+0x33/0x70 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? _nv000694rm+0x4ae/0xc80 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar  6 21:22:08 2023]  ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar  6 21:22:08 2023]  ? do_syscall_64+0x33/0x80
[Mon Mar  6 21:22:08 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar  6 21:22:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
[Mon Mar  6 21:23:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:24:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:24:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:25:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:26:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080852e 0x208).
[Mon Mar  6 21:27:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20808513 0x598).
[Mon Mar  6 21:27:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a068 0x4).
[Mon Mar  6 21:28:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a618 0x181c).
[Mon Mar  6 21:29:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a612 0xd98).
[Mon Mar  6 21:30:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20809009 0x8).
[Mon Mar  6 21:30:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar  6 21:31:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar  6 21:32:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x2080 0x4).
[Mon Mar  6 21:33:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar  6 21:33:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
[Mon Mar  6 21:34:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:35:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x208f 0x0).
[Mon Mar  6 21:36:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x208f1105 0x8).
[Mon Mar  6 21:36:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
[Mon Mar  6 21:37:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=905052, name=python, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20801702 0x4).
[Mon Mar  6 21:38:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:39:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:39:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a00b0 0x0).
[Mon Mar  6 21:40:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:41:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0050 0x0).
[Mon Mar  6 21:42:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:42:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080852e 0x208).
[Mon Mar  6 21:43:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20808513 0x598).
[Mon Mar  6 21:44:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a068 0x4).
[Mon Mar  6 21:45:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a618 0x181c).
[Mon Mar  6 21:45:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977772, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 10 (FREE) (0xa55a0040 0x0).
[Mon Mar  6 21:46:18 2023] INFO: task python:976158 blocked for more than 120 seconds.
[Mon Mar  6 21:46:18 2023]       Tainted: P           OE     5.10.0-20-amd64 NVIDIA/nvidia-docker#1 Debian 5.10.158-2
[Mon Mar  6 21:46:18 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[Mon Mar  6 21:46:18 2023] task:python          state:D stack:    0 pid:976158 ppid:835062 flags:0x00000000
[Mon Mar  6 21:46:18 2023] Call Trace:
[Mon Mar  6 21:46:18 2023]  __schedule+0x282/0x880
[Mon Mar  6 21:46:18 2023]  schedule+0x46/0xb0
[Mon Mar  6 21:46:18 2023]  rwsem_down_write_slowpath+0x246/0x4d0
[Mon Mar  6 21:46:18 2023]  os_acquire_rwlock_write+0x31/0x40 [nvidia]
[Mon Mar  6 21:46:18 2023]  _nv038505rm+0xc/0x30 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv039453rm+0x18d/0x1d0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv041182rm+0x45/0xd0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv041127rm+0x142/0x2b0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv039415rm+0x15a/0x2e0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv039416rm+0x5b/0x90 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv039416rm+0x31/0x90 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv012688rm+0x1d/0x30 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? task_numa_fault+0x2a3/0xb70
[Mon Mar  6 21:46:18 2023]  ? _nv039431rm+0xb0/0xb0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv012710rm+0x54/0x70 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv011426rm+0xc4/0x120 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv000659rm+0x63/0x70 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv000582rm+0x2c/0x40 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv000694rm+0x86c/0xc80 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar  6 21:46:18 2023]  ? do_syscall_64+0x33/0x80
[Mon Mar  6 21:46:18 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar  6 21:46:18 2023] INFO: task nvidia-smi:980026 blocked for more than 120 seconds.
[Mon Mar  6 21:46:18 2023]       Tainted: P           OE     5.10.0-20-amd64 NVIDIA/nvidia-docker#1 Debian 5.10.158-2
[Mon Mar  6 21:46:18 2023] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[Mon Mar  6 21:46:18 2023] task:nvidia-smi      state:D stack:    0 pid:980026 ppid:980025 flags:0x00000000
[Mon Mar  6 21:46:18 2023] Call Trace:
[Mon Mar  6 21:46:18 2023]  __schedule+0x282/0x880
[Mon Mar  6 21:46:18 2023]  ? psi_task_change+0x88/0xd0
[Mon Mar  6 21:46:18 2023]  schedule+0x46/0xb0
[Mon Mar  6 21:46:18 2023]  rwsem_down_read_slowpath+0x18e/0x500
[Mon Mar  6 21:46:18 2023]  os_acquire_rwlock_read+0x31/0x40 [nvidia]
[Mon Mar  6 21:46:18 2023]  _nv038503rm+0xc/0x30 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv039453rm+0x64/0x1d0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv041182rm+0x45/0xd0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv041133rm+0xfd/0x2b0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv012728rm+0x59a/0x690 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv039431rm+0x53/0xb0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv011404rm+0x52/0xa0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? _nv000694rm+0x5ae/0xc80 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? rm_ioctl+0x54/0xb0 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? nvidia_ioctl+0x6cd/0x830 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? nvidia_frontend_unlocked_ioctl+0x37/0x50 [nvidia]
[Mon Mar  6 21:46:18 2023]  ? __x64_sys_ioctl+0x8b/0xc0
[Mon Mar  6 21:46:18 2023]  ? do_syscall_64+0x33/0x80
[Mon Mar  6 21:46:18 2023]  ? entry_SYSCALL_64_after_hwframe+0x61/0xc6
[Mon Mar  6 21:46:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080a612 0xd98).
[Mon Mar  6 21:47:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20809009 0x8).
[Mon Mar  6 21:48:01 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar  6 21:48:46 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar  6 21:49:31 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800a4c 0x4).
[Mon Mar  6 21:50:16 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800110 0x84).
[Mon Mar  6 21:51:02 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
[Mon Mar  6 21:51:47 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x20800157 0x0).
[Mon Mar  6 21:52:32 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 103 (GSP_RM_ALLOC) (0x208f 0x0).
[Mon Mar  6 21:53:17 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=977823, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x208f1105 0x8).
[Mon Mar  6 21:54:02 2023] NVRM: Xid (PCI:0000:61:00): 119, pid=980026, name=nvidia-smi, Timeout waiting for RPC from GSP! Expected function 76 (GSP_RM_CONTROL) (0x2080014b 0x5).
lemketron commented 1 year ago

I was getting similar errors trying to run nvidia-smi after updating my Ubuntu 22.04 system.

I noticed that sudo systemctl status nvidia-persistenced was not enabled, so I tried to enable it with sudo systemctl enable nvidia-persistenced but that failed as well.

I then ran across the following forum post which led me to think that perhaps my drivers had somehow been disabled/corrupted by a recent system and/or kernel upgrade so I decided to reinstall (and upgrade) the nvidia driver using the PPA, and nvidia-smi is working again. Hope this helps someone else...

https://forums.developer.nvidia.com/t/nvidia-smi-has-failed-because-it-couldnt-communicate-with-the-nvidia-driver-make-sure-that-the-latest-nvidia-driver-is-installed-and-running/197141/2

Anvilondre commented 1 year ago

I was having the same issue on Ubuntu Server 22.04 and docker-compose. Reinstalling docker with apt (instead of snap) solved my problem.

Haseeb707 commented 1 year ago

Uninstalling docker desktop and installing using apt worked for me As mentioned in NVIDIA/nvidia-container-toolkit#229

bkocis commented 1 year ago

for some reason, reinstalling docker help me as well by executing (ubuntu 22.04): sudo apt-get install docker-ce docker-ce-cli containerd.io docker-buildx-plugin docker-compose-plugin

JosephKuchar commented 1 year ago

@bkocis Thanks, that worked! I uninstalled docker with the official docker instructions, then reinstalled from the same link and it's now all working again.

han9rykim commented 10 months ago

In my case, if there is another docker installed through snap, this kind of issue happened. After removing the docker by sudo snap remove --purge docker figured out this problem.