Open eng100200 opened 5 years ago
We extend the person bounding box to a fixed ratio 4:3, then crop the person bounding box from the image. Finally we resize the cropped person image patch to 256x192 or 384x288 for training or testing.
@leoxiaobin thank you for reply. It means you have cropped all the persons from the original image and then resize these cropped persons as new seperate single image of sizes 256 x 192 for training, is it? Then it also need to re-label the key points, since the original keypoint labels will be different, is it?
@leoxiaobin resize each of these cropped persons are resized as 256 x 192 and then form new images for training, i.e. each person will be a new single seperate image. is it?
yes
@leoxiaobin morning, how are you? If i directly input the coco images for training and use the same the label data as in coco dataset, then, what will be the affect on training and performance? Do you have any idea? Have you tried it?
@leoxiaobin hello, can i ask you question?
@leoxiaobin i dont understand your code does not compute scale and center for mpii dataset, while, it is provided in json files. Now, further, there are no detection bounding boxes in validation json file, so, do we need to detect the boxes first before testing the validation set?
Hello, may I ask if my data input size is 1280x720, or is the model adaptive to any size of image input? How did you solve this problem?
@leoxiaobin Sorry to bother you, I have a question concerning about the size of the input image, it is mentioned in paper as "We extend the human detection box in height or width to a fixed aspect ratio: height : width = 4 : 3, and then crop the box from the image, which is resized to a fixed size, 256 × 192", what is this mean? You mean if the image in COCO is 300x400, we then extend its width or height to obtain the ratio 4:3 and then crop the extended image to resize to 256x 192? is it? if it is then the corresponding labels are need to be modified. Please reply.