Closed Armandpl closed 10 months ago
is using controlnet actually useful? can't we just train on canny edges from blender generated images? if we do that we won't get the perturbation from the env such as other cars, bridges etc unless we add them in blender which requires hard-coding "scenarios" whereas if we use controlnet we can easily and "organically" add those. If we only train on edges from blender we probably won't generalize very well. Tho maybe we could view adding bridges, cars etc as adding noise and maybe there is a simpler faster way to add noise to images. but it would be a different type of noise i guess
doesn't look like i can add cars and other stuff, its closely follows the canny edges (which are from segmented lane lines) takes 11 seconds to gen 1 image, that's 30 hours for 10k images at 1152x832 vanishing point seems to remain consistent across images
4 steps using turbo = 2s per image, quality seems to be okay but less good?
Make the blender scene:
which focal len to use in blender:
910*6.521/1471.64262 = 4 mm focal len
actually do we really need blender? could just use a perspective projection
looks like it could fly. my main concern now is that there is no car on those images. not sure how to add them. main point of using stable diffusion was to try and distill what it knows about the world. though maybe we could make it work by using random erasing? make it robust to big perturbations? though random black rectangles are a bit different from cars. maybe automatic inpainting? but that would make generating the images even longer.
we also need to add dotted lines, maybe double lines. maybe add curved lines later? though unsure what the vp definition is in those case, let's not worry about it
save gen script config somewhere for repro
Could we go from vp dataset to direction/delta pose dataset this way? if we move the camera forward, does the vanishing point changes? What I'm trying to guess is: is there enough info in the lines to predict the direction of travel? Look at it from above and think about it
lightly annoyed that the vpgnet dataset is/seems to be imbalanced. annoyed i don't have easy access to the DeepVP-1M dataset. Can I make my own dataset?