Closed kremHabashy closed 1 year ago
Hello,
Just following up on this. Very interested in the model!
Hi, sorry for the late reply. The released code is reorganized and not exactly identical to our experimental code. After careful debugging, I found that the bug of NaN loss is attributed to the computing of KL loss since the bridge_mask
might be all zeros when the bridge_embeds
are all zeros. I fixed the bug by adding length = length.clamp(min=1e-5)
in the function avg_pool()
at the model_color.py
. Please clone our latest commit for your experiments. Thank you!
Thank you!
Hello,
I am going through the pipeline, and have trained the Brownian Bridge. I am however encountering NaN values at the next step of training the planner, as can be seen below.
As you can see, this is happening from the very start, and so I'm not sure where this is coming from. As this step uses the model made in the bridge creation step, I thought that might be the issue, but training there seemed fine. Below is the end of the training for the bridge: