Closed peterfeifanchen closed 1 year ago
Hi Peter, I'll reply to some of these questions.
The way I think about training is like this:
TL;DR:
Thanks for the prompt response Albert!
A quick follow up on 2), could you point me to where in the Sashimi code the set up for the unconditioned generation happens? And for clarification, the generation was not conditioned even on a prefix and the loss was maximizing the probability of p(x_t|x<t) ?
Yes to both. The generation script is here: https://github.com/HazyResearch/state-spaces/blob/main/generate.py
Instructions are in the main READMEs which provide the commands for Sashimi generation. I think you can specify the prefix length, i.e. l_prefix=0
means unconditional generation.
I see thank you!
Hi! Like everyone else on here, I would like to first say, thank you very much for putting this all together.
I was looking at the unconditioned audio generation and have a few question about sequence generation specifically and conceptually:
But looking at the code:
` def setup(self):
Learned Parameters (C is complex!)
` It seems Lambda, P, (I guess P=Q here) are declared as params and thus would not optax still optimize them? I am new to flax/jax, so I am wondering how they are been fixed?