Closed janmbuys closed 6 years ago
Sanity check seems fine. Currently running w/ hidden_dim/clusters == 50/100 and HMM w/ and w/out 1-hot embeddings
Rerunning experiments with step fixed
The val numbers look reasonable, expect for HMM, where I am pretty sure there has to be a bug.
This should approximately replicate the pytorch LM baseline:
python ptb_main.py --type lstm --optim sgd --lr 20 --dropout 0.2 --tie-embeddings --hidden-dim 200 --embed-dim 200 --initrange 0.1
Pushed a possible HMM bug fix.