Closed davidmartinrius closed 1 year ago
Hi, based on the log below, I suspect it is because the pytorch3d. Can you check the installation of putorch3d? Can you provide details of your GPU?
"/media/pc/Elements/geneface_env/lib/python3.9/site-packages/pytorch3d/renderer/init.py", line 7, in from .blending import ( File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/pytorch3d/renderer/blending.py", line 10, in from pytorch3d import _C ImportError: libtorch_cuda_cu.so: cannot open shared object file: No such file or directory
In the conda environment, with python3.9, I have got this:
(geneface) ➜ conda list | grep pytorch3d pytorch3d 0.7.2 py39_cu113_pyt1120 pytorch3d
I am using a Nvidia RTX 3080, 10GB version.
pytorch3d seems to be installed and the GPU is working well, as I am using it in other AI projects + pytorch + tensorflow (in other environments)
By the way, It passed succcessfully the STEP 3 Verification of the Installation in docs/prepare_env/install_guide.md
It only fails in docs/process_data/process_target_person_video.md STEP 1
Hi, the script is validated in a env with python3.9 with rtx3090. Can you run the sub-steps in the process_data.sh step by step, so we can check where the error is originated?
Hi, I list my environment.yaml as follows:
name: geneface2 channels:
- pytorch3d
- pytorch
- bottler
- iopath
- conda-forge
- defaults dependencies:
- _libgcc_mutex=0.1=main
- _openmp_mutex=5.1=1_gnu
- blas=1.0=mkl
- brotlipy=0.7.0=py39h27cfd23_1003
- ca-certificates=2023.01.10=h06a4308_0
- certifi=2022.12.7=py39h06a4308_0
- cffi=1.15.1=py39h5eee18b_3
- charset-normalizer=2.0.4=pyhd3eb1b0_0
- colorama=0.4.6=pyhd8ed1ab_0
- cryptography=39.0.1=py39h9ce1e76_0
- cudatoolkit=11.3.1=h2bc3f7f_2
- flit-core=3.6.0=pyhd3eb1b0_0
- freetype=2.12.1=h4a9f257_0
- fvcore=0.1.5.post20221221=pyhd8ed1ab_0
- giflib=5.2.1=h5eee18b_3
- idna=3.4=py39h06a4308_0
- intel-openmp=2021.4.0=h06a4308_3561
- iopath=0.1.9=py39
- jpeg=9e=h5eee18b_1
- lcms2=2.12=h3be6417_0
- ld_impl_linux-64=2.38=h1181459_1
- lerc=3.0=h295c915_0
- libdeflate=1.17=h5eee18b_0
- libffi=3.4.2=h6a678d5_6
- libgcc-ng=11.2.0=h1234567_1
- libgomp=11.2.0=h1234567_1
- libpng=1.6.39=h5eee18b_0
- libstdcxx-ng=11.2.0=h1234567_1
- libtiff=4.5.0=h6a678d5_2
- libwebp=1.2.4=h11a3e52_1
- libwebp-base=1.2.4=h5eee18b_1
- lz4-c=1.9.4=h6a678d5_0
- mkl=2021.4.0=h06a4308_640
- mkl-service=2.4.0=py39h7f8727e_0
- mkl_fft=1.3.1=py39hd3c417c_0
- mkl_random=1.2.2=py39h51133e4_0
- ncurses=6.4=h6a678d5_0
- numpy=1.23.5=py39h14f4228_0
- numpy-base=1.23.5=py39h31eccc5_0
- nvidiacub=1.10.0=0
- openssl=1.1.1t=h7f8727e_0
- pillow=9.4.0=py39h6a678d5_0
- pip=23.0.1=py39h06a4308_0
- portalocker=2.7.0=py39hf3d152e_0
- pycparser=2.21=pyhd3eb1b0_0
- pyopenssl=23.0.0=py39h06a4308_0
- pysocks=1.7.1=py39h06a4308_0
- python=3.9.16=h7a1cb2a_2
- python_abi=3.9=2_cp39
- pytorch=1.12.0=py3.9_cuda11.3_cudnn8.3.2_0
- pytorch-mutex=1.0=cuda
- pytorch3d=0.7.2=py39_cu113_pyt1120
- pyyaml=6.0=py39hb9d737c_4
- readline=8.2=h5eee18b_0
- requests=2.28.1=py39h06a4308_0
- setuptools=65.6.3=py39h06a4308_0
- six=1.16.0=pyhd3eb1b0_1
- sqlite=3.40.1=h5082296_0
- tabulate=0.9.0=pyhd8ed1ab_1
- termcolor=2.2.0=pyhd8ed1ab_0
- tk=8.6.12=h1ccaba5_0
- torchvision=0.13.1=cpu_py39h164cc8f_0
- tqdm=4.65.0=pyhd8ed1ab_1
- typing_extensions=4.4.0=py39h06a4308_0
- tzdata=2022g=h04d1e81_0
- urllib3=1.26.14=py39h06a4308_0
- wheel=0.38.4=py39h06a4308_0
- xz=5.2.10=h5eee18b_1
- yacs=0.1.8=pyhd8ed1ab_0
- yaml=0.2.5=h7f98852_2
- zlib=1.2.13=h5eee18b_0
- zstd=1.5.2=ha4553b6_0
- pip:
- absl-py==1.4.0
- astunparse==1.6.3
- audioread==3.0.0
- cachetools==5.3.0
- configargparse==1.5.3
- contourpy==1.0.7
- cycler==0.11.0
- dearpygui==1.8.0
- decorator==4.4.2
- face-alignment==1.3.5
- ffmpeg-python==0.2.0
- filelock==3.9.0
- flatbuffers==23.3.3
- fonttools==4.39.0
- future==0.18.3
- gast==0.4.0
- google-auth==2.16.2
- google-auth-oauthlib==0.4.6
- google-pasta==0.2.0
- grpcio==1.51.3
- h5py==3.8.0
- huggingface-hub==0.13.1
- imageio==2.26.0
- imageio-ffmpeg==0.4.8
- importlib-metadata==6.0.0
- importlib-resources==5.12.0
- joblib==1.2.0
- keras==2.11.0
- kiwisolver==1.4.4
- kornia==0.6.10
- lazy-loader==0.1
- libclang==15.0.6.1
- librosa==0.10.0
- llvmlite==0.39.1
- lpips==0.1.4
- markdown==3.4.1
- markupsafe==2.1.2
- matplotlib==3.7.1
- moviepy==1.0.3
- msgpack==1.0.5
- networkx==3.0
- ninja==1.11.1
- numba==0.56.4
- oauthlib==3.2.2
- opencv-python==4.7.0.72
- opt-einsum==3.3.0
- packaging==23.0
- pandas==1.5.3
- platformdirs==3.1.1
- pooch==1.7.0
- praat-parselmouth==0.4.3
- proglog==0.1.10
- protobuf==3.19.6
- pyasn1==0.4.8
- pyasn1-modules==0.2.8
- pyaudio==0.2.13
- pymcubes==0.1.4
- pyparsing==3.0.9
- python-dateutil==2.8.2
- python-speech-features==0.6
- python-version==0.0.2
- pytz==2022.7.1
- pywavelets==1.4.1
- regex==2022.10.31
- requests-oauthlib==1.3.1
- resampy==0.4.2
- rsa==4.9
- scikit-image==0.20.0
- scikit-learn==1.2.2
- scipy==1.9.1
- soundfile==0.12.1
- soxr==0.3.4
- tensorboard==2.11.2
- tensorboard-data-server==0.6.1
- tensorboard-plugin-wit==1.8.1
- tensorboardx==2.6
- tensorflow==2.11.0
- tensorflow-estimator==2.11.0
- tensorflow-io-gcs-filesystem==0.31.0
- threadpoolctl==3.1.0
- tifffile==2023.2.28
- tokenizers==0.13.2
- transformers==4.26.1
- trimesh==3.20.2
- werkzeug==2.2.3
- wrapt==1.15.0
- zipp==3.15.0 prefix: /home/yezhenhui/anaconda3/envs/geneface2
Hi, the script is validated in a env with python3.9 with rtx3090. Can you run the sub-steps in the process_data.sh step by step, so we can check where the error is originated?
Ok, I will try it and when finished I will tell you. I also will compare your environment packages versions with mine.
Thanks for all 😄
OK, I will stay tuned. If you find any bug in prepare_env, feel free to contact me.
I created a new environment from scratch and reinstalled everything step by step but I got the same error. I am still trying to fix it, it is still not solved. I am just sending this message to let you know that I am still with it. Thank you
Got it.
I had the same problem and solved it. Change export CUDA_VISIBLE_DEVICES from 3 to 0 inside the .sh file you run
I've racked my brain. :rofl: I didn't see it... Thank you!!!! The problem is solved :smile:
Hello,
I installed everything following the instructions. I am working in an Ubuntu 21 with al CUDA dependencies working. I already have installed cuda 11.3 and cuda 11.7 and I already have some working project with cuda.
In this case, when I tried Geneface/docs/process_data/process_target_person_video.md STEP 1 I get the next error:
2023-03-17 21:09:23.506047: W tensorflow/compiler/xla/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libnvinfer.so.7'; dlerror: libnvinfer.so.7: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /media/pc/Elements/geneface_env/lib/python3.9/site-packages/cv2/../../lib64:
Actually, libnvinfer.so.7 does not exist, of course. The think is that I don't know how to install that dependency. I think is related to tensorrt. I tried to downgrade tensorrt but did not work neither. Anyway, I share the full stack trace because it also show another error: "RuntimeError: No CUDA GPUs are available" Actually when I import torch cuda is available. I don't know if the last error is related to libnvinfer.so.7 or not.
Here is the full stack trace:
(/media/pc/Elements/geneface_env) ➜ GeneFace git:(main) ✗ CUDA_VISIBLE_DEVICES=0 data_gen/nerf/process_data.sh $VIDEO_ID
[INFO] ===== extract audio from data/raw/videos/May.mp4 to data/processed/videos/May/aud.wav ===== ffmpeg version 4.4.2-0ubuntu0.22.04.1 Copyright (c) 2000-2021 the FFmpeg developers built with gcc 11 (Ubuntu 11.2.0-19ubuntu1) configuration: --prefix=/usr --extra-version=0ubuntu0.22.04.1 --toolchain=hardened --libdir=/usr/lib/x86_64-linux-gnu --incdir=/usr/include/x86_64-linux-gnu --arch=amd64 --enable-gpl --disable-stripping --enable-gnutls --enable-ladspa --enable-libaom --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libcdio --enable-libcodec2 --enable-libdav1d --enable-libflite --enable-libfontconfig --enable-libfreetype --enable-libfribidi --enable-libgme --enable-libgsm --enable-libjack --enable-libmp3lame --enable-libmysofa --enable-libopenjpeg --enable-libopenmpt --enable-libopus --enable-libpulse --enable-librabbitmq --enable-librubberband --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libspeex --enable-libsrt --enable-libssh --enable-libtheora --enable-libtwolame --enable-libvidstab --enable-libvorbis --enable-libvpx --enable-libwebp --enable-libx265 --enable-libxml2 --enable-libxvid --enable-libzimg --enable-libzmq --enable-libzvbi --enable-lv2 --enable-omx --enable-openal --enable-opencl --enable-opengl --enable-sdl2 --enable-pocketsphinx --enable-librsvg --enable-libmfx --enable-libdc1394 --enable-libdrm --enable-libiec61883 --enable-chromaprint --enable-frei0r --enable-libx264 --enable-shared libavutil 56. 70.100 / 56. 70.100 libavcodec 58.134.100 / 58.134.100 libavformat 58. 76.100 / 58. 76.100 libavdevice 58. 13.100 / 58. 13.100 libavfilter 7.110.100 / 7.110.100 libswscale 5. 9.100 / 5. 9.100 libswresample 3. 9.100 / 3. 9.100 libpostproc 55. 9.100 / 55. 9.100 Input #0, mov,mp4,m4a,3gp,3g2,mj2, from 'data/raw/videos/May.mp4': Metadata: major_brand : mp42 minor_version : 0 compatible_brands: mp42mp41 creation_time : 2021-11-09T10:09:46.000000Z Duration: 00:04:02.97, start: 0.000000, bitrate: 3324 kb/s Stream #0:0(eng): Video: h264 (Main) (avc1 / 0x31637661), yuv420p, 512x512 [SAR 1:1 DAR 1:1], 3004 kb/s, 25 fps, 25 tbr, 25k tbn, 50 tbc (default) Metadata: creation_time : 2021-11-09T10:09:46.000000Z handler_name : ?Mainconcept Video Media Handler vendor_id : [0][0][0][0] encoder : AVC Coding Stream #0:1(eng): Audio: aac (LC) (mp4a / 0x6134706D), 48000 Hz, stereo, fltp, 317 kb/s (default) Metadata: creation_time : 2021-11-09T10:09:46.000000Z handler_name : #Mainconcept MP4 Sound Media Handler vendor_id : [0][0][0][0] File 'data/processed/videos/May/aud.wav' already exists. Overwrite? [y/N] y Stream mapping: Stream #0:1 -> #0:0 (aac (native) -> pcm_s16le (native)) Press [q] to stop, [?] for help Output #0, wav, to 'data/processed/videos/May/aud.wav': Metadata: major_brand : mp42 minor_version : 0 compatible_brands: mp42mp41 ISFT : Lavf58.76.100 Stream #0:0(eng): Audio: pcm_s16le ([1][0][0][0] / 0x0001), 16000 Hz, stereo, s16, 512 kb/s (default) Metadata: creation_time : 2021-11-09T10:09:46.000000Z handler_name : #Mainconcept MP4 Sound Media Handler vendor_id : [0][0][0][0] encoder : Lavc58.134.100 pcm_s16le size= 15183kB time=00:04:02.92 bitrate= 512.0kbits/s speed= 75x
extract_landmarks(ori_imgs_dir)
File "/media/pc/Elements/GeneFace/data_util/process.py", line 55, in extract_landmarks
fa = face_alignment.FaceAlignment(face_alignment.LandmarksType._2D, flip_input=False)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/face_alignment/api.py", line 77, in init
self.face_detector = face_detector_module.FaceDetector(device=device, verbose=verbose, **face_detector_kwargs)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/face_alignment/detection/sfd/sfd_detector.py", line 31, in init
self.face_detector.to(device)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1145, in to
return self._apply(convert)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 797, in _apply
module._apply(fn)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 820, in _apply
param_applied = fn(param)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1143, in convert
return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/cuda/init.py", line 247, in _lazy_init
torch._C._cuda_init()
RuntimeError: No CUDA GPUs are available
[INFO] ===== perform face tracking =====
[INFO] ===== extract semantics from data/processed/videos/May/ori_imgs to data/processed/videos/May/parsing =====
Traceback (most recent call last):
File "/media/pc/Elements/GeneFace/data_util/face_tracking/face_tracker.py", line 8, in
from render_3dmm import Render_3DMM
File "/media/pc/Elements/GeneFace/data_util/face_tracking/render_3dmm.py", line 6, in
from pytorch3d.renderer import (
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/pytorch3d/renderer/init.py", line 7, in
from .blending import (
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/pytorch3d/renderer/blending.py", line 10, in
from pytorch3d import _C
ImportError: libtorch_cuda_cu.so: cannot open shared object file: No such file or directory
[INFO] ===== finished face tracking =====
[INFO] ===== extract background image from data/processed/videos/May/ori_imgs =====
Traceback (most recent call last):
File "/media/pc/Elements/GeneFace/data_util/face_parsing/test.py", line 99, in
evaluate(respth=args.respath, dspth=args.imgpath, cp=args.modelpath)
File "/media/pc/Elements/GeneFace/data_util/face_parsing/test.py", line 59, in evaluate
net.cuda()
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 905, in cuda
return self._apply(lambda t: t.cuda(device))
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 797, in _apply
module._apply(fn)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 797, in _apply
module._apply(fn)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 797, in _apply
module._apply(fn)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 820, in _apply
param_applied = fn(param)
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/nn/modules/module.py", line 905, in
return self._apply(lambda t: t.cuda(device))
File "/media/pc/Elements/geneface_env/lib/python3.9/site-packages/torch/cuda/init.py", line 247, in _lazy_init
torch._C._cudainit()
RuntimeError: No CUDA GPUs are available
[INFO] ===== extracted semantics =====
0%| | 0/304 [00:00<?, ?it/s][ WARN:0@3.438] global loadsave.cpp:244 findDecoder imread('data/processed/videos/May/parsing/0.png'): can't open/read file: check file path/integrity
0%| | 0/304 [00:00<?, ?it/s]
Traceback (most recent call last):
File "/media/pc/Elements/GeneFace/data_util/process.py", line 429, in
extract_background(processed_dir, ori_imgs_dir)
File "/media/pc/Elements/GeneFace/data_util/process.py", line 86, in extract_background
bg = (parse_img[..., 0] == 255) & (parse_img[..., 1] == 255) & (parseimg[..., 2] == 255)
TypeError: 'NoneType' object is not subscriptable
[ WARN:0@0.030] global loadsave.cpp:244 findDecoder imread('data/processed/videos/May/bc.jpg'): can't open/read file: check file path/integrity
[INFO] ===== extract head images for data/processed/videos/May =====
0%| | 0/6073 [00:00<?, ?it/s][ WARN:0@0.072] global loadsave.cpp:244 findDecoder imread_('data/processed/videos/May/parsing/0.png'): can't open/read file: check file path/integrity
0%| | 0/6073 [00:00<?, ?it/s]
Traceback (most recent call last):
File "/media/pc/Elements/GeneFace/data_util/process.py", line 433, in
extract_head(processed_dir)
File "/media/pc/Elements/GeneFace/data_util/process.py", line 138, in extract_head
head_part = (parsing_img[:, :, 0] == 255) & (
TypeError: 'NoneType' object is not subscriptable
[START]
uniuje vizaĉanĝu de fl eĉo morispaas st antl ugehedt di opu ĉiu ni ĉhis ĝkam ant de ŝi evis aj kun sizuol tuanti sevtiin hasen stuo ajbli dajs ofu ĉiu ne ĉiezo kurajĉu da n eva fubuihevn eid amomentas ĉisiĵon an seŝ aŭ seŭ on niu du reĉon adi fto aĝis sekstii mu s la gia juvaŭ ĉi itu da ĉein ĝ disis li ea pisto tmekis hap on ajn a ud refuenĝo bla as ĉiunuvestiva jsi fertoajns an laŭov koosted no sevi bonŝeedz saj impunĝu viru oveŭ ŝi ten zajmuaj prajna ju ĉiuj sed azui feis liop ĉu neceza herbas aŭ ŝe dintro esc dam mpe ŝins kan prinos ĉi gata ui omon da sia buĉo nuĉis sronga da nati s tudeej voj omon te kamtu diris fe ra ce efri bon hasĉ ams de sukcii poj obon sne ŝundris sai fan sekio fraĉj udun ant kv ar ĉoron li sambe ŝun s ju najĉas se zobujoneŭ longa a da fefti ĉiuj p senĝo vaŭĝed li iv antofori aj psendu vaŭĉi ĝu neajn po buan k urei ĉiiunienof pipolundne ŝins udopra ud hisĝuj anto broi t fiuĉa se bonaj seĉ umanon gez sia iŝinĉejpo lenj oro pde ŝia etubi vui zaĉen mojn de nove ĉi raante r gedo roits dio noĝasbo deu sin faŭĉit li iv por f aevlui sen gupa s neniskanĉoj efko s drefremdo m la itpeazomfa d deveŝinsina a kan ĉoj pri neŭzue prosp urin anftaŭ izua not daizu ken isli baadaŭ lon hojm senta ĉ uron duo gradsku fajndas k io ĝoob antdeu suka nod en ŝo r dereso hima kan tribuecuel antdaŭispo hilm eĉ ĝas no t desisti gi ebiniĝ pol da an diis paroj a stas haŭurpi buba k si ki ujn la beto dieloĵ helmpa ornm i buakinpipo dvisaŭt objeĉululi gu naj cirpoĉan n en muĉboj a o junaeti de na sidisumŝipe vdis kurajtne jiŝon junajce ĝe ni op ĉiu neĉesa ĝa eŭpon ĉe o na pipo anĝj nojtis pajda princ ipo dris eŭnlio tal enĝen hor buir k da ĉu dit amen io fiuĉ a aftro ĝis u juna ĝidro ap ipol heva ĉiiv iĉ kraittengs s tro apraŝis juni el ovnei ŝins inglan s kotland ue osdnovan a lend tva junino pipo fron s po s deimstua mfo siespis naciistu ĉamo ĉe s skuustr o hospros anto pavo truagiu ni enov k omiuno ĉeis an fables erkoz zi ĝis ĝast peg gkloibolevanc ertu fajno jea ĝis l upers mufengs e todis av ĝi mapilgie ju sdonĉo fres ĝo bo pajo f est hajm en majpiz iĝe ejo ĉurin sta skulo ka ŭ ofte juniversĉ i ol juviĉoja afto laivĉajno haruerk vi stings laivs majo sĉems ado fing sot pajndas ĉiu evolui a asof anta asti keem pei ĝej koks tiu selĉaĝi li ĉajkenfumezl aosĝie puŝeĝ mi a famo junaiĉe d — anthaf amo lin komen donat uĉ ĉ divacas i ha v gaŭrnop ĉiune ĝi iĝu deamen strueitlad ĉu brinn is konĉuite gata a asnevo de fu sead o ĉ hu evo juao e evol u lev a l polo ĉeeksiko nu mi ensissoa ĉi buakf o jilu norĝaste pri veleĝi fien scuasmi lokehe tu jer voko ĉunzi and jiu nocei leĉmi buŝ iu ĝo famle apisvo l pros prs ant hapi ni ujea
[END]
[INFO] save all feats for training purpose...
[INFO] saved logits to data/processed/videos/May/aud_esperanto.npy
[INFO] ===== extracted esperanto =====
[INFO] ===== extract deepspeech =====
2023-03-17 21:10:08.200192: I tensorflow/core/platform/cpu_feature_guard.cc:193] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 AVX512F AVX512_VNNI FMA
To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.
2023-03-17 21:10:08.298201: I tensorflow/core/util/port.cc:104] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable
video:0kB audio:15183kB subtitle:0kB other streams:0kB global headers:0kB muxing overhead: 0.000502% [INFO] ===== extracted audio ===== [INFO] ===== extract audio labels for data/processed/videos/May/aud.wav ===== [INFO] ===== start extract esperanto ===== [INFO] ===== extract images from data/raw/videos/May.mp4 to data/processed/videos/May/ori_imgs ===== ffmpeg version 4.4.2-0ubuntu0.22.04.1 Copyright (c) 2000-2021 the FFmpeg developers built with gcc 11 (Ubuntu 11.2.0-19ubuntu1) configuration: --prefix=/usr --extra-version=0ubuntu0.22.04.1 --toolchain=hardened --libdir=/usr/lib/x86_64-linux-gnu --incdir=/usr/include/x86_64-linux-gnu --arch=amd64 --enable-gpl --disable-stripping --enable-gnutls --enable-ladspa --enable-libaom --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libcdio --enable-libcodec2 --enable-libdav1d --enable-libflite --enable-libfontconfig --enable-libfreetype --enable-libfribidi --enable-libgme --enable-libgsm --enable-libjack --enable-libmp3lame --enable-libmysofa --enable-libopenjpeg --enable-libopenmpt --enable-libopus --enable-libpulse --enable-librabbitmq --enable-librubberband --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libspeex --enable-libsrt --enable-libssh --enable-libtheora --enable-libtwolame --enable-libvidstab --enable-libvorbis --enable-libvpx --enable-libwebp --enable-libx265 --enable-libxml2 --enable-libxvid --enable-libzimg --enable-libzmq --enable-libzvbi --enable-lv2 --enable-omx --enable-openal --enable-opencl --enable-opengl --enable-sdl2 --enable-pocketsphinx --enable-librsvg --enable-libmfx --enable-libdc1394 --enable-libdrm --enable-libiec61883 --enable-chromaprint --enable-frei0r --enable-libx264 --enable-shared libavutil 56. 70.100 / 56. 70.100 libavcodec 58.134.100 / 58.134.100 libavformat 58. 76.100 / 58. 76.100 libavdevice 58. 13.100 / 58. 13.100 libavfilter 7.110.100 / 7.110.100 libswscale 5. 9.100 / 5. 9.100 libswresample 3. 9.100 / 3. 9.100 libpostproc 55. 9.100 / 55. 9.100 Input #0, mov,mp4,m4a,3gp,3g2,mj2, from 'data/raw/videos/May.mp4': Metadata: major_brand : mp42 minor_version : 0 compatible_brands: mp42mp41 creation_time : 2021-11-09T10:09:46.000000Z Duration: 00:04:02.97, start: 0.000000, bitrate: 3324 kb/s Stream #0:0(eng): Video: h264 (Main) (avc1 / 0x31637661), yuv420p, 512x512 [SAR 1:1 DAR 1:1], 3004 kb/s, 25 fps, 25 tbr, 25k tbn, 50 tbc (default) Metadata: creation_time : 2021-11-09T10:09:46.000000Z handler_name : ?Mainconcept Video Media Handler vendor_id : [0][0][0][0] encoder : AVC Coding Stream #0:1(eng): Audio: aac (LC) (mp4a / 0x6134706D), 48000 Hz, stereo, fltp, 317 kb/s (default) Metadata: creation_time : 2021-11-09T10:09:46.000000Z handler_name : #Mainconcept MP4 Sound Media Handler vendor_id : [0][0][0][0] Stream mapping: Stream #0:0 -> #0:0 (h264 (native) -> mjpeg (native)) Press [q] to stop, [?] for help [swscaler @ 0x5589d57a0740] deprecated pixel format used, make sure you did set range correctly Output #0, image2, to 'data/processed/videos/May/ori_imgs/%d.jpg': Metadata: major_brand : mp42 minor_version : 0 compatible_brands: mp42mp41 encoder : Lavf58.76.100 Stream #0:0(eng): Video: mjpeg, yuvj420p(pc, progressive), 512x512 [SAR 1:1 DAR 1:1], q=2-31, 200 kb/s, 25 fps, 25 tbn (default) Metadata: creation_time : 2021-11-09T10:09:46.000000Z handler_name : ?Mainconcept Video Media Handler vendor_id : [0][0][0][0] encoder : Lavc58.134.100 mjpeg Side data: cpb: bitrate max/min/avg: 0/0/200000 buffer size: 0 vbv_delay: N/A ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.rear6.15x
ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.center_lfe ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM cards.pcm.side ALSA lib pcm_route.c:877:(find_matching_chmap) Found no matching channel map ALSA lib pcm_route.c:877:(find_matching_chmap) Found no matching channel map ALSA lib pcm_route.c:877:(find_matching_chmap) Found no matching channel map ALSA lib pcm_route.c:877:(find_matching_chmap) Found no matching channel map ALSA lib pcm_oss.c:397:(_snd_pcm_oss_open) Cannot open device /dev/dsp ALSA lib pcm_oss.c:397:(_snd_pcm_oss_open) Cannot open device /dev/dsp ALSA lib confmisc.c:160:(snd_config_get_card) Invalid field card ALSA lib pcm_usb_stream.c:482:(_snd_pcm_usb_stream_open) Invalid card 'card' ALSA lib confmisc.c:160:(snd_config_get_card) Invalid field card ALSA lib pcm_usb_stream.c:482:(_snd_pcm_usb_stream_open) Invalid card 'card' [WARN] audio has 2 channels, only use the first. [INFO] loaded audio stream data/processed/videos/May/aud.wav: (3886763,) [INFO] loading ASR model cpierse/wav2vec2-large-xlsr-53-esperanto... /media/pc/Elements/geneface_env/lib/python3.9/site-packages/transformers/configuration_utils.py:379: UserWarning: Passing
gradient_checkpointing
to a config initialization is deprecated and will be removed in v5 Transformers. Usingmodel.gradient_checkpointing_enable()
instead, or if you are using theTrainer
API, passgradient_checkpointing=True
in yourTrainingArguments
. warnings.warn( Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. frame= 6073 fps=174 q=1.0 Lsize=N/A time=00:04:02.92 bitrate=N/A speed=6.97xvideo:287221kB audio:0kB subtitle:0kB other streams:0kB global headers:0kB muxing overhead: unknown [INFO] ===== extracted images ===== [INFO] ===== extract face landmarks from data/processed/videos/May/ori_imgs ===== 2023-03-17 21:09:21.611204: I tensorflow/core/platform/cpu_feature_guard.cc:193] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 AVX512F AVX512_VNNI FMA To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags. 2023-03-17 21:09:21.855254: I tensorflow/core/util/port.cc:104] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable
TF_ENABLE_ONEDNN_OPTS=0
. 2023-03-17 21:09:23.506047: W tensorflow/compiler/xla/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libnvinfer.so.7'; dlerror: libnvinfer.so.7: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /media/pc/Elements/geneface_env/lib/python3.9/site-packages/cv2/../../lib64: 2023-03-17 21:09:23.506404: W tensorflow/compiler/xla/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libnvinfer_plugin.so.7'; dlerror: libnvinfer_plugin.so.7: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /media/pc/Elements/geneface_env/lib/python3.9/site-packages/cv2/../../lib64: 2023-03-17 21:09:23.506411: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Cannot dlopen some TensorRT libraries. If you would like to use Nvidia GPU with TensorRT, please make sure the missing libraries mentioned above are installed properly. Traceback (most recent call last): File "/media/pc/Elements/GeneFace/data_util/process.py", line 438, inTF_ENABLE_ONEDNN_OPTS=0
. 2023-03-17 21:10:09.086860: W tensorflow/compiler/xla/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libnvinfer.so.7'; dlerror: libnvinfer.so.7: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /media/pc/Elements/geneface_env/lib/python3.9/site-packages/cv2/../../lib64: 2023-03-17 21:10:09.087204: W tensorflow/compiler/xla/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libnvinfer_plugin.so.7'; dlerror: libnvinfer_plugin.so.7: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /media/pc/Elements/geneface_env/lib/python3.9/site-packages/cv2/../../lib64: 2023-03-17 21:10:09.087210: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Cannot dlopen some TensorRT libraries. If you would like to use Nvidia GPU with TensorRT, please make sure the missing libraries mentioned above are installed properly. /home/pc/.tensorflow/models/deepspeech-0_1_0-b90017e8.pb 2023-03-17 21:10:11.876266: E tensorflow/compiler/xla/stream_executor/cuda/cuda_driver.cc:267] failed call to cuInit: CUDA_ERROR_NO_DEVICE: no CUDA-capable device is detected 2023-03-17 21:10:11.876287: I tensorflow/compiler/xla/stream_executor/cuda/cuda_diagnostics.cc:169] retrieving CUDA diagnostic information for host: pc 2023-03-17 21:10:11.876291: I tensorflow/compiler/xla/stream_executor/cuda/cuda_diagnostics.cc:176] hostname: pc 2023-03-17 21:10:11.876346: I tensorflow/compiler/xla/stream_executor/cuda/cuda_diagnostics.cc:200] libcuda reported version is: 515.86.1 2023-03-17 21:10:11.876359: I tensorflow/compiler/xla/stream_executor/cuda/cuda_diagnostics.cc:204] kernel reported version is: 515.86.1 2023-03-17 21:10:11.876362: I tensorflow/compiler/xla/stream_executor/cuda/cuda_diagnostics.cc:310] kernel version seems to match DSO: 515.86.1 2023-03-17 21:10:11.879202: I tensorflow/core/platform/cpu_feature_guard.cc:193] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN) to use the following CPU instructions in performance-critical operations: AVX2 AVX512F AVX512_VNNI FMA To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags. tring to extract deepspeech from audio file: data/processed/videos/May/aud.wav The target is: data/processed/videos/May/aud_deepspeech.npy /media/pc/Elements/GeneFace/data_util/deepspeech_features/deepspeech_features.py:50: UserWarning: Audio has multiple channels, the first channel is used warnings.warn( 2023-03-17 21:10:13.438041: I tensorflow/compiler/mlir/mlir_graph_optimization_pass.cc:357] MLIR V1 optimization pass is not enabled