I am trying to train with 10k documents with additional 1k document for eval cycle.
Even for these small number of documents, it is projecting around 4 days of training time on Tesla M60 GPU.
I have changed config to have 10 docs per step with max steps to be 10000 for 10 epochs. It takes around 34 seconds per step, which gives us around 4 days of training time.
I am trying to train with 10k documents with additional 1k document for eval cycle.
Even for these small number of documents, it is projecting around 4 days of training time on Tesla M60 GPU.
I have changed config to have 10 docs per step with max steps to be 10000 for 10 epochs. It takes around 34 seconds per step, which gives us around 4 days of training time.
Am I doing something wrong?