Closed NickleDave closed 3 years ago
instead of "ablating" LSTM, run experiments where we vary size of hidden state, i.e. number of hidden units, after fixing #70
this will be one way of getting at "Relative influence of sequential versus local features" as described above in 1.4, although reviewer is asking about bin size v window size
based on email discussions with @yardencsGitHub after further analysis of results in initial submission:
Depending on results we might want to ablate too
correction: after #75 and #76 realize that we used hidden size of 256 for initial submission. 1024 is "hidden size * 4", one dimension of the size of the learnable input-hidden weights i_h -- see LSTM variables in PyTorch docs: https://pytorch.org/docs/stable/generated/torch.nn.LSTM.html
Will need to re-run this