Objectron is a dataset of short, object-centric video clips. In addition, the videos also contain AR session metadata including camera poses, sparse point-clouds and planes. In each video, the camera moves around and above the object and captures it from different views. Each object is annotated with a 3D bounding box. The 3D bounding box describes the object’s position, orientation, and dimensions. The dataset contains about 15K annotated video clips and 4M annotated images in the following categories: bikes, books, bottles, cameras, cereal boxes, chairs, cups, laptops, and shoes
Hi, the shoes in the data does not distinguish left and right, Is that correct? Is there any chance, somewhere can find the label to distinguish left and right? Thanks!
Unfortunately no, we do not distinguish left from the right shoe.
You can use some heuristics to guess the left and right shoe (e.g. the right shoe should be on the right side of the left shoe, if they are both sitting on the same plane) and should work for 95% of the cases.
Hi, the shoes in the data does not distinguish left and right, Is that correct? Is there any chance, somewhere can find the label to distinguish left and right? Thanks!