jackyjsy / CVPR21Chal-SLR

This repo contains the official code of our work SAM-SLR which won the CVPR 2021 Challenge on Large Scale Signer Independent Isolated Sign Language Recognition.
Creative Commons Zero v1.0 Universal
205 stars 50 forks source link

Trouble running the Docker image #31

Open ch3njust1n opened 1 year ago

ch3njust1n commented 1 year ago

I loaded the Docker image with

cat cvpr2021cha_code.tar | sudo docker load

When I checked docker ps, I didn't see anything running and Docker logs showed

sudo docker logs 422402c07084b60ace48a6881017f9c2370b67fcd2a877a575fed557e2468db4
Welcome to SMILELAB SLR Code, please enter
source ./setup_env.sh
/bin/sh: 1: nvidia-smi: not found

I'm not sure which nvidia-smi to install, but these are what my system suggested when I tried running nvidia-smi: sudo apt install nvidia-340 # version 340.108-0ubuntu5.20.04.2, or sudo apt install nvidia-utils-390 # version 390.157-0ubuntu0.20.04.1 sudo apt install nvidia-utils-450-server # version 450.248.02-0ubuntu0.20.04.1 sudo apt install nvidia-utils-470 # version 470.199.02-0ubuntu0.20.04.1 sudo apt install nvidia-utils-470-server # version 470.199.02-0ubuntu0.20.04.1 sudo apt install nvidia-utils-525 # version 525.125.06-0ubuntu0.20.04.3 sudo apt install nvidia-utils-525-server # version 525.125.06-0ubuntu0.20.04.2 sudo apt install nvidia-utils-535 # version 535.86.05-0ubuntu0.20.04.2 sudo apt install nvidia-utils-535-server # version 535.54.03-0ubuntu0.20.04.1 sudo apt install nvidia-utils-435 # version 435.21-0ubuntu7 sudo apt install nvidia-utils-440 # version 440.82+really.440.64-0ubuntu6 sudo apt install nvidia-utils-418-server # version 418.226.00-0ubuntu0.20.04.2

I installed nvidia-340 from the list above and got

NVIDIA-SMI has failed because it couldn't communicate with the NVIDIA driver. Make sure that the latest NVIDIA driver is installed and running.

Questions:

  1. Could someone point me in the right direction for this nvidia-smi issue?
  2. Once I get the container running, how do I send videos to the model? Is there a web server in the container that's serving the model? Do I make a HTTP POST request?
  3. How do I batch inputs to the model for production?