Closed dimitarsh1 closed 5 years ago
This happens only if the I invoke run_external_eval on the test set. If I don't specify a test set - it works. Until the averaging point (see newly reported issue).
Cheers, Dimitar
which test set are you using? Can you check the maximum length of the sentences in your test set?
Hi.
I managed to run the train and inference without a problem on a GTX 1080Ti (11GB) when I decreased the qe_batch_size and the inference_batch_size to 10 and 1 respectively. The thing is that during inference, I could not set the batch size at all neither by setting the qe_batch_size nor the infer_batch_size and I retrained the qe model. But once that was done - the inference went smoothly.
Thanks for looking into this. I guess during inference it should be possible to set the batch size which in my case didn't work.
Cheers, Dimitar
hparms
has type tf.contrib.training.HParams
. If you want to set a smaller infer batch size without retraining the model, you can add
setattr(hparams, "infer_batch_size", 1)
after https://github.com/lovecambi/qebrain/blob/771519de047279ea25d76b056ed64b78da6cc7c3/qe_model.py#L2170
I think the problem has been solved.
Hello,
I have trained several expert models (different values for the hyperparams) and while that succeeds for each case, I keep on having problems training the qe model.
I have trained models with 120k of vocabulary, 75k vocabulary and 50k vocabulary. Different batch sizes (with the smallest being 20). I always get the following error:
...
And it doesn't matter whether I am using TitanX or 1080Ti.
Any idea?
Thanks. Dimitar