nv-tlabs / LION

Latent Point Diffusion Models for 3D Shape Generation
Other
735 stars 57 forks source link

Initialization of latent points as target points #67

Closed dqj5182 closed 4 months ago

dqj5182 commented 6 months ago

From the previous issue, it was stated that

" I don't think you need to train it for longer:

in general the longer you train, the worse reconstruction you will get (as shown in the val EMD/CD curve), but smoother latent space (i.e. the latent points closer to N(0,1), this will make training the diffusion model easier). And we need to find a good trade off between them, In the figure you show the latent points is super smooth, I feel like the model can be stopped earlier. "

May I ask which part of the code initializes latent points as GT points? (also wish to know whether the code initializes latent points as GT points even when we resume training code from pre-trained checkpoints)

Looking forward for your reply and always thanks for your kind feedback! @ZENGXH

ZENGXH commented 4 months ago

it's done by 1) skip connection with small weight to the input from the network and 2) adding a offset to the logsigma in the latent space s.t. the variance is small in initilzation.

when resuming from pre-trained, it wont init as GT. In fact, after init from scratch and the training starts, the latent points will be getting away from GT points