andyzeng / arc-robot-vision

MIT-Princeton Vision Toolbox for Robotic Pick-and-Place at the Amazon Robotics Challenge 2017 - Robotic Grasping and One-shot Recognition of Novel Objects with Deep Learning.
http://arc.cs.princeton.edu
Apache License 2.0
295 stars 97 forks source link

Non-augmented vs augmented parallel grasping labels are inverted #5

Closed thomascent closed 6 years ago

thomascent commented 6 years ago

Hello, I was just reading through processLabels.m (specifically lines 105-108) and I noticed that in the non-augmented labels, good grasps are set to 128 whereas in the augmented labels they're set to 0.

Is that an error by any chance? It seems like the augmented labels are used for training meaning that positive grasp affordance would be encoded in the first channel of the network's output but in visualize.m they're taken as the second channel.

andyzeng commented 6 years ago

You’re right - the augmented labels should be flipped. Hotfixed with the latest commits. Thanks for the catch!

thomascent commented 6 years ago

No problem! The labels in the preprocessed dataset (this guy: http://vision.princeton.edu/projects/2017/arc/downloads/training-parallel-jaw-grasping-dataset.zip) are also affected so you might like to re-process them