Open kyoguan opened 7 years ago
@kyoguan The loss curve is pretty good. How many training data did you use?
just use the WEB data set, download from https://dl.dropboxusercontent.com/u/42868014/WEB.zip
@kyoguan - Is this the epoch loss or the batch loss?
batch_size = 32*6 and num_gpus = 6(train_multi_gpus.py)
batch_size = 32 (train.py)
it seems the single gpu is mush faster and the training curve is smoother.