Closed AmitMY closed 4 years ago
what do you mean "when the rotation is predicted". It's predicted in both cases. which is evident looking at the bounding boxes. I've changed the anchor selection mechanism a bit to deal with exp overflow, but that's about it.
Ah I expected the title:
HAND TRACKER WITHOUT ROTATION ESTIMATION AND JOINT DETECTION
To be without rotation, but you are right, there obviously is rotation.
What do you think is the reason for the rotation to be predicted differently then?
So if you look here, you'll notice that I pick the detection with the largest bounding box.
In the case of multiple hands I pick the detection with the largest score .
Guess the first option worked better for that one hand, but one would have to test the model on some data with compatible ground truth to know which approach works best generally.
check out the new multihand
branch. It's a bit cleaner than optical_flow. Accuracy could be better in some cases but I believe it's due to simplistic aggregation.
I've opened a clean new issue for this problem.
I used the code in the current master branch to predict the bbox and joints for an image:
Works as well as expected.
Then I used the
hand_tracker2
code in the other branch, which supports multiple hands detection and joints prediction, and got this:Notice that the second hand from right (the one detected in both tests) - seems to get better joints prediction when the rotation is predicted.
I'm not sure if there is any other thing going on here, or if this is just a bad, sample size of 1 example, but should be taken into consideration.
Code: