Open czkaa opened 5 years ago
@kaaaaaaaaaaa Hi,
Just try to use https://raw.githubusercontent.com/AlexeyAB/darknet/master/cfg/yolov3-tiny_3l.cfg
It is required only if the sizes differ more than 2 times, i.e. if automatic data augmentation can't solve this problem.
Yes, you can add borders to the images, so relative sizes will be the same in Training and Testing datasets.
Or for example, if objects on Testing images are 2x smaller, then Train the model with width=832 height=832
, and after training set width=416 height=416
for Detection.
I would like to train a tiny yolo v3 hand detector. It is supposed to detect hands in very specific footage. The camera perspective as well as the size of the hands is always more or less the same in every video:
(sorry for the unnecessary secretiveness but I am not sure about copyright)
So far, I have used the Oxford Dataset for training, where all images look more or less like this:
I got okay results for testing the Oxford Dataset but bad results for testing on the actual footage. Now, I have some questions regarding how to modify the training process for this taks:
1) Is there an equivalence of this hint for tiny yolo?
2) In this repo, I read tha the dataset should include such a set of relative sizes of objects that I want to detect. This is obviously not the case with my dataset. As an improvement, should I e.g. create thick borders aroung the traning images so that the relative size of the hands match the footage-to-detect? Or is it enough if I add some labelled frames from the footage to the dataset? If yes, how many (The training set so far is about 45000 images large.)?
Any help is appreciated!!