Closed migueljette closed 7 years ago
Hi!
yes, training on GPU-s is not only possible, but strongly recommended. Training speed on AWS with a Tesla K80 GPU should be around 10000 sps.
You will probably see a speedup when you set floatX in ~/.theanorc to float32. Example ~/.theanorc:
[global]
floatX = float32
Another speedup will probably come when you switch to libgpuarray backend as hinted in the warning: https://github.com/Theano/Theano/wiki/Converting-to-the-new-gpu-back-end%28gpuarray%29
Best, Ottokar
Hi!
It worked. I had to do a few things... my installation was not perfect. But now it works:
THEANO_FLAGS=device=cuda0 python main.py ep 256 0.02
Using cuDNN version 5110 on context None
Mapped name None to device cuda0: Tesla K80 (0000:00:1E.0)
256 0.02 Model_ep_h256_lr0.02.pcl
Building model...
Number of parameters is 17912840
Training...
PPL: 1.7279; Speed: 2615.79 sps
PPL: 1.5797; Speed: 4379.89 sps
PPL: 1.4646; Speed: 5649.73 sps
PPL: 1.3994; Speed: 6607.25 sps
PPL: 1.3577; Speed: 7354.74 sps
PPL: 1.3279; Speed: 7954.68 sps
PPL: 1.3055; Speed: 8446.64 sps
PPL: 1.2877; Speed: 8857.51 sps
PPL: 1.2734; Speed: 9200.41 sps
PPL: 1.2614; Speed: 9499.25 sps
PPL: 1.2513; Speed: 9758.59 sps
PPL: 1.2426; Speed: 9985.76 sps
PPL: 1.2350; Speed: 10186.37 sps
Thanks for your help! I look forward to testing this on my data!
Cheers Miguel
Hi there,
Is there a way to train using GPUs? And what is "normal" speed when training the example "ep" model?
On my macbook pro:
On an AWS instance with GPU:
Thanks for your help! Looking forward to training with my own data, but I want to make sure everything is working as expected.
Cheers Miguel