Hi,
I am trying to run posenet project on aarch64 device (raspberry pi 3) on CPU. I am using gstreamer. But on running the pose_camera.py script, no pose is detected. No line drawings can be seen as well. I have a USB camera connected as /dev/video0. Condifence scores for keypoints remains at 0;
When I switch over to using opencv, I get accurate pose detections with non-zero confidence scores.
My input command is:
$python3 pose_camera.py --model model.tflite --videosrc /dev/video0
On running, I see the video on the screen, with live fps and NPoses=1, but i don't see any actual pose detected using the drawn lines.
If i use it using opencv, i get pose outputs, but not using gstreamer. I even installed the posenet_decoder.so
How do I run this project on CPU using the CPU based models?
As a follow- up, how can I port it to run on movenets, using gstreamer? I can run movenet using opencv, but would like to use gstreamer without opencv.
thanks
Click to expand!
### Issue Type
Support
### Operating System
Ubuntu
### Coral Device
Dev Board
### Other Devices
Raspberry Pi 3
### Programming Language
Python 3.9
### Relevant Log Output
_No response_
Description
Hi, I am trying to run posenet project on aarch64 device (raspberry pi 3) on CPU. I am using gstreamer. But on running the pose_camera.py script, no pose is detected. No line drawings can be seen as well. I have a USB camera connected as /dev/video0. Condifence scores for keypoints remains at 0;
When I switch over to using opencv, I get accurate pose detections with non-zero confidence scores.
My input command is: $python3 pose_camera.py --model model.tflite --videosrc /dev/video0
my model.tflite is downloaded from https://raw.githubusercontent.com/google-coral/test_data/master/posenet/posenet_mobilenet_v1_075_353_481_16_quant_decoder.tflite
On running, I see the video on the screen, with live fps and NPoses=1, but i don't see any actual pose detected using the drawn lines. If i use it using opencv, i get pose outputs, but not using gstreamer. I even installed the posenet_decoder.so
How do I run this project on CPU using the CPU based models?
As a follow- up, how can I port it to run on movenets, using gstreamer? I can run movenet using opencv, but would like to use gstreamer without opencv.
thanks
Click to expand!
### Issue Type Support ### Operating System Ubuntu ### Coral Device Dev Board ### Other Devices Raspberry Pi 3 ### Programming Language Python 3.9 ### Relevant Log Output _No response_