Running python data/face_landmark_detection.py train seems to be very slow (~1s per image).
I added a debug output to double-check if dlib makes use of CUDA:
print (dlib.DLIB_USE_CUDA) for which I get true.
But when checking with nvidia-smi there is nothing running on the GPU but I have one CPU thread running at 100%.
Is there any reason why dlib would not use the GPU here?
Running
python data/face_landmark_detection.py train
seems to be very slow (~1s per image). I added a debug output to double-check if dlib makes use of CUDA:print (dlib.DLIB_USE_CUDA)
for which I get true. But when checking with nvidia-smi there is nothing running on the GPU but I have one CPU thread running at 100%.Is there any reason why dlib would not use the GPU here?