the logprior is calculated as a combination of output of both LMs. Is there a reason you are calculating KL = E{x ~ q(z|x, y)}[log q(z|x, y) - log p(z|y)] instead of KL = E_{x ~ q(z|x)}[log q(z|x) - log p(z)], like in the paper?
Is there a reason only latent y (and not x, from the other domain) is sampled in the data loading process? From equation 3 I presumed we sample from both domains to compare to each lm separately.
Hi. Thank you for such an exciting paper!
I would appreciate it greatly if you could shed some light on these:
In the log_prior calculation, https://github.com/cindyxinyiwang/deep-latent-sequence-model/blob/8a798582b1af5ef7f6ac4ca1f2138fd382a1cb06/src/model.py#L339
the logprior is calculated as a combination of output of both LMs. Is there a reason you are calculating KL = E{x ~ q(z|x, y)}[log q(z|x, y) - log p(z|y)] instead of KL = E_{x ~ q(z|x)}[log q(z|x) - log p(z)], like in the paper?
When loading train data, is there a reason
y
is sampled with1-y_train
? : https://github.com/cindyxinyiwang/deep-latent-sequence-model/blob/8a798582b1af5ef7f6ac4ca1f2138fd382a1cb06/src/data_utils.py#L99Is there a reason only latent y (and not x, from the other domain) is sampled in the data loading process? From equation 3 I presumed we sample from both domains to compare to each lm separately.
Thank you!