Open joapolarbear opened 3 years ago
pre-trained multilayer perceptrons (MLP)
For example, Conv2D, (i) batch size, (ii) number of input and output channels, (iii) kernel size, (iv) padding, (v) stride, and (vi) image size
use the ratio between the peak floating point operations per second (FLOPS) of two GPUS or the ratio between the number of CUDA core on each GPU —> assume that a DNN training workload can exhaust all the computational resources on a GPU.
Reading Notes