Closed jeiglsperger closed 2 years ago
I recognized I had some mistakes in the architecture of my framework (e.g. the creation of the sequences) and the selection of the hyperparameter. After debugging those things it works, therefore I will close this issue.
I have a problem with the dnn_to_bnn() transformation of my lstm network. I define my lstm network the following:
The classes PrepareForlstm, GetOutputZero, and PrepareForDropout are defined as follows:
This network works fine for me. Now I wanted to try out the dnn_to_bnn() API and transform it to a bayesian lstm with:
When executing my code, I get the following traceback:
So the size of the
input
does not match the size ofweight
inout = F.linear(input, weight, bias)
in line 164 inlinear_variational.py
. I tried to trace back how this mats came about, but have not clue why the network does not work anymore. Maybe someone of you has a clue?