PaddlePaddle / Paddle

PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)
http://www.paddlepaddle.org/
Apache License 2.0
22.07k stars 5.54k forks source link

sevearl questions about sparse training #137

Closed shenhuinuist closed 7 years ago

shenhuinuist commented 7 years ago

Hi, can you provide a demo to show how to use sparse training ? And according to Paddle's documentation , there are four kinds of input : dense_vector , sparse_binary_verctor, sparse_float_vector and integer, do you take sparse_int_vector into account ? I find paddle implement sparse matrix multiplication wihtout calling MKL library. I want to konw wherther the implementation of sparse matrix multiplication in paddle is faster than calling MKL library ?

backyes commented 7 years ago

We will provide sparse train demo with local and cluster mode in few days.

Sparse Train in PaddePaddle is beyond two sparse matrix multiplication, further implies design in Parameter Server Architecture, SGD Optimisation, etc. So it is not directly do performance comparison with only MKL library.

shenhuinuist commented 7 years ago

I mean that just consider the two sparse matrix multiplication, which way is faster? using MKL or using mul function in paddle?

backyes commented 7 years ago

Pull request for Sparse Doc: https://github.com/baidu/Paddle/pull/144, which is under merging. You can preview it here: https://github.com/backyes/Paddle/blob/sparse_train_doc_demo/doc/cluster/opensource/cluster_train.md

We will update you performance data if the fine grained pure sparse matrix performance data is available.

shenhuinuist commented 7 years ago

Thanks!