Closed frankvp11 closed 6 months ago
Sorry about changing the issue so many times- I was indecisive and was changing things from default- changed them all back and got this error - Also why is it showing up with "Misplaced &" - Thats not in my original error? Is that a bug in the dockerfile? Ive tried a multidude of the docker files and they mostly give that error something similar. I also tried going back a release and i still got that error -could it be an issue on my part?
Update: I ended up getting the Dockerfile working on the latest commit of TensorRT however upon entering it and trying to run the command /usr/src/tensorrt/bin/trtexec it gives me the following error: /usr/src/tensorrt/bin/trtexec: error while loading shared libraries: libcudart.so.10.2: cannot open shared object file: No such file or directory
And upon doing sudo find . -iname libcudart.so.10.2 I get the following output ./usr/local/cuda-10.2/targets/aarch64-linux/lib/libcudart.so.10.2
Also, gentle ping: @kevinch-nv
Things I have tried: adding the directory I mentioned above to path- failed tried different dockerfiles - failed reinstalling different version of pytorch - didnt do anything
To give you more information, I'd like this to work so that I can build an engine on it and then run it. Ive already got the model weights and everything that I need I just need the TensorRT OSS to be upgraded.
@kevinch-nv ^ ^
Any updates??? @kevinch-nv
Looking into it - it seems like all we have to do is update the base image here https://github.com/NVIDIA/TensorRT/blob/main/docker/ubuntu-20.04-aarch64.Dockerfile#L19 to 11.6.2. I'm verifying it on my end - can you try to update this locally?
As an aside - I see you've filed this against Jetson TX2, for Jetson platforms you should be following the cross-compilation instructions (https://github.com/NVIDIA/TensorRT#setting-up-the-build-environment), so you should be using the ubuntu-cross-aarch64.Dockerfile
Yeah- thanks for the help. I ended up getting it to work. Turns out I didn't need TensorRT 8.4.x, and all I had to do was follow instructions from https://github.com/NVIDIA-AI-IOT/deepstream_tao_apps/tree/master/TRT-OSS/Jetson and that worked for me. However back to the dockerfile issue, I could try it later after doing some more testing with my current solution.
Also the problem with the cross-aarch dockerfile for me was that it would error out when it copied the jetpack files into the pdk_files. This is because it couldn't find the jetson files
Looking into it - it seems like all we have to do is update the base image here https://github.com/NVIDIA/TensorRT/blob/main/docker/ubuntu-20.04-aarch64.Dockerfile#L19 to 11.6.2. I'm verifying it on my end - can you try to update this locally?
As an aside - I see you've filed this against Jetson TX2, for Jetson platforms you should be following the cross-compilation instructions (https://github.com/NVIDIA/TensorRT#setting-up-the-build-environment), so you should be using the
ubuntu-cross-aarch64.Dockerfile
I don't understand why must to use cross-compile. TX2/Xavier/Orin have tiny ubuntu which can use ubuntu-20.04-aarch64.Dockerfile
build docker?
@kevinch-nv did you end up checking it on your end- and did it work?
JFYI I'm facing the same issue. Replacing the version of cuda to 11.6.2 makes the container build to fail with:
Building container:
> docker build -f docker/ubuntu-20.04-aarch64.Dockerfile --build-arg CUDA_VERSION=11.6.2 --build-arg uid=502 --build-arg gid=80 --tag=tensorrt-aarch64-ubuntu20.04-cuda11.4 .
[+] Building 426.8s (20/25)
=> [internal] load build definition from ubuntu-20.04-aarch64.Dockerfile 0.0s
=> => transferring dockerfile: 3.77kB 0.0s
=> [internal] load .dockerignore 0.0s
=> => transferring context: 34B 0.0s
=> [internal] load metadata for docker.io/nvidia/cuda:11.6.2-devel-ubuntu20.04 1.2s
=> [ 1/21] FROM docker.io/nvidia/cuda:11.6.2-devel-ubuntu20.04@sha256:f9785e8241290c2947fe89eb2b8540e7777337a01c5b6340533e0b6599cdfd83 221.0s
=> => resolve docker.io/nvidia/cuda:11.6.2-devel-ubuntu20.04@sha256:f9785e8241290c2947fe89eb2b8540e7777337a01c5b6340533e0b6599cdfd83 0.0s
=> => sha256:ec0b93578a6e38f0044abb6bc46a5a8ee93c5dc14323d5fcf35b272417c1e2d4 186B / 186B 0.6s
=> => sha256:f9785e8241290c2947fe89eb2b8540e7777337a01c5b6340533e0b6599cdfd83 743B / 743B 0.0s
=> => sha256:3101385b820f9f55ca61e5d2aaf4e912e3c6d5ea7f967e5272a210190ac547c4 2.21kB / 2.21kB 0.0s
=> => sha256:f5e423043b47f33682403b04d1400171e9f7e8d384dd0d4169e3690e2b6d61e6 361.71kB / 361.71kB 0.4s
=> => sha256:9872d9aa75f7a2d051a4d3339dcb1af3616d6484f31577983d81f051668f1fb6 13.45kB / 13.45kB 0.0s
=> => sha256:775bcf4925a33701c1dd9b7bf6ef598a26360e1ced1479fb49cfeb70990915cf 7.76MB / 7.76MB 0.9s
=> => sha256:8941157b58ada869bd12299ba8a56cdc5317923a5ce7df8158c5a3b44ff2fb67 6.43kB / 6.43kB 0.8s
=> => sha256:ff82155d267bda15b08af95fb76ae972b1c837c3c1c703975367e4cc586e2f9e 1.12GB / 1.12GB 177.8s
=> => sha256:b8e91a073034a8ae8c9ab24f8ad39d82b39c0570d10f555d817c3d50c984f4e2 61.66kB / 61.66kB 1.1s
=> => extracting sha256:775bcf4925a33701c1dd9b7bf6ef598a26360e1ced1479fb49cfeb70990915cf 0.2s
=> => sha256:b6e9be7fe06f43c048223a8ae71c3f5664c35ddd11aa686c0173becae8fc15b1 1.41GB / 1.41GB 193.1s
=> => sha256:5087632c6bbd71828eef5d6e1b036264c7278d5f887eecaa6e25560f067e88f9 83.98kB / 83.98kB 1.3s
=> => extracting sha256:f5e423043b47f33682403b04d1400171e9f7e8d384dd0d4169e3690e2b6d61e6 0.0s
=> => extracting sha256:ec0b93578a6e38f0044abb6bc46a5a8ee93c5dc14323d5fcf35b272417c1e2d4 0.0s
=> => extracting sha256:8941157b58ada869bd12299ba8a56cdc5317923a5ce7df8158c5a3b44ff2fb67 0.0s
=> => extracting sha256:ff82155d267bda15b08af95fb76ae972b1c837c3c1c703975367e4cc586e2f9e 18.9s
=> => extracting sha256:b8e91a073034a8ae8c9ab24f8ad39d82b39c0570d10f555d817c3d50c984f4e2 0.0s
=> => extracting sha256:b6e9be7fe06f43c048223a8ae71c3f5664c35ddd11aa686c0173becae8fc15b1 23.9s
=> => extracting sha256:5087632c6bbd71828eef5d6e1b036264c7278d5f887eecaa6e25560f067e88f9 0.0s
=> [internal] load build context 0.0s
=> => transferring context: 38B 0.0s
=> [ 2/21] RUN groupadd -r -f -g 80 trtuser && useradd -o -r -l -u 502 -g 80 -ms /bin/bash trtuser 2.5s
=> [ 3/21] RUN usermod -aG sudo trtuser 0.2s
=> [ 4/21] RUN echo 'trtuser:nvidia' | chpasswd 0.2s
=> [ 5/21] RUN mkdir -p /workspace && chown trtuser /workspace 0.3s
=> [ 6/21] RUN apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2004/sbsa/3bf863cc.pub 0.5s
=> [ 7/21] RUN apt-get update && apt-get install -y software-properties-common 14.1s
=> [ 8/21] RUN add-apt-repository ppa:ubuntu-toolchain-r/test 2.6s
=> [ 9/21] RUN apt-get update && apt-get install -y --no-install-recommends libcurl4-openssl-dev wget git pkg-config sudo ssh libssl-dev pbzip2 pv 10.7s
=> [10/21] RUN apt-get install -y --no-install-recommends python3 python3-pip python3-dev python3-wheel && cd /usr/local/bin && ln -s /usr/bin/python3 python 3.1s
=> [11/21] RUN v="${TRT_VERSION%.*}-1+cuda${CUDA_VERSION%.*}" && apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2004/sbsa/3bf863cc.pub && a 156.7s
=> [12/21] RUN cd /tmp && wget https://github.com/Kitware/CMake/releases/download/v3.21.4/cmake-3.21.4-linux-aarch64.sh && chmod +x cmake-3.21.4-linux-aarch64.sh && ./cmake-3.2 5.1s
=> [13/21] RUN pip3 install --upgrade pip 1.9s
=> [14/21] RUN pip3 install setuptools>=41.0.0 0.7s
=> [15/21] COPY requirements.txt /tmp/requirements.txt 0.0s
=> ERROR [16/21] RUN pip3 install -r /tmp/requirements.txt 5.8s
------
> [16/21] RUN pip3 install -r /tmp/requirements.txt:
#20 0.520 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
#20 0.520 Looking in links: https://download.pytorch.org/whl/cu113/torch_stable.html
#20 0.520 Ignoring onnx: markers 'python_version == "3.10"' don't match your environment
#20 0.520 Ignoring tensorflow-gpu: markers 'platform_machine == "x86_64" and sys_platform == "linux"' don't match your environment
#20 0.520 Ignoring onnxruntime: markers 'python_version == "3.10"' don't match your environment
#20 0.520 Ignoring torch: markers 'python_version == "3.10"' don't match your environment
#20 0.521 Ignoring torchvision: markers 'python_version == "3.10"' don't match your environment
#20 1.484 Collecting onnx==1.10.2
#20 2.265 Downloading onnx-1.10.2-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (12.7 MB)
#20 3.056 ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 12.7/12.7 MB 15.9 MB/s eta 0:00:00
#20 3.995 Collecting onnxruntime==1.8.1
#20 4.339 Downloading onnxruntime-1.8.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (7.4 MB)
#20 4.857 ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 7.4/7.4 MB 14.3 MB/s eta 0:00:00
#20 5.698 ERROR: Could not find a version that satisfies the requirement torch==1.10.2+cu113 (from versions: 1.8.0, 1.8.1, 1.9.0, 1.10.0, 1.10.1, 1.10.2, 1.11.0, 1.12.0, 1.12.1, 1.13.0)
#20 5.698 ERROR: No matching distribution found for torch==1.10.2+cu113
------
executor failed running [/bin/bash -c pip3 install -r /tmp/requirements.txt]: exit code: 1
Hi, if anyone is still facing this issue, there is one more thing you should try.
Just pull the following dockerfile and work with tensorrt like you normally would.
https://catalog.ngc.nvidia.com/orgs/nvidia/containers/l4t-pytorch
This way atleast you won't have to worry about the torch, torchvision and the tensorrt's version.
I will close inactive issues for more than 3 week per our policy, thanks all!
Description
I am trying to use the ubuntu 20.04 aarch64 dockerfile (cuda 11.4 one) it gives me the following error: E: Version '8.4.2-1+cuda11.4' for 'libnvinfer8' was not found E: Version '8.4.2-1+cuda11.4' for 'libnvonnxparsers8' was not found E: Version '8.4.2-1+cuda11.4' for 'libnvparsers8' was not found E: Version '8.4.2-1+cuda11.4' for 'libnvinfer-plugin8' was not found E: Version '8.4.2-1+cuda11.4' for 'libnvinfer-dev' was not found E: Version '8.4.2-1+cuda11.4' for 'libnvonnxparsers-dev' was not found E: Version '8.4.2-1+cuda11.4' for 'libnvparsers-dev' was not found E: Version '8.4.2-1+cuda11.4' for 'libnvinfer-plugin-dev' was not found E: Version '8.4.2-1+cuda11.4' for 'python3-libnvinfer' was not found
Environment
TensorRT Version: Na NVIDIA GPU: Jetson TX2 NVIDIA Driver Version: IDK CUDA Version: na CUDNN Version: na Operating System: na Python Version (if applicable): na Tensorflow Version (if applicable): Na PyTorch Version (if applicable): Na Baremetal or Container (if so, version): Container- ubuntu20.04-aarch64
Relevant Files
The dockerfile(s)
Steps To Reproduce
sudo ./docker/build.sh --file docker/ubuntu-20.04-aarch64.Dockerfile --tag tensorrt-aarch64-ubuntu20.04-cuda11.4
Total Output: Building container: Building container: