PaddlePaddle / Paddle

PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)
http://www.paddlepaddle.org/
Apache License 2.0
22.22k stars 5.58k forks source link

Optimizer Design and related Operator #3655

Closed jacquesqiao closed 7 years ago

jacquesqiao commented 7 years ago
  1. Implement various optimizer operators.
    1. SGD (done) and others listed in https://github.com/PaddlePaddle/Paddle/projects/22
  2. Add optimizer operators into BlockDesc.
  3. Python module

In the new framework with operators, we will use operators to build optimizer, and operators will be add to some block. There are some works to do:

there are three situation to be considered:

  1. single machine single device.
  2. single machine multiple devices.
  3. multiple machine.

Plan:

distribute related issue: https://github.com/PaddlePaddle/Paddle/issues/3656

jacquesqiao commented 7 years ago

For tensorflow: https://github.com/tensorflow/models/blob/master/tutorials/image/cifar10/cifar10_multi_gpu_train.py https://www.tensorflow.org/tutorials/deep_cnn

jacquesqiao commented 7 years ago

close this old one and open a new issue: https://github.com/PaddlePaddle/Paddle/issues/4679