Closed smalltingting closed 5 years ago
If you are using standard tensorflow or hardware, they can't take the advantage of sparse model to reduce model size and flops. But, will give chance to particular hardware or library to skip the computing with zero weights. let's say i have a GPU with is designed for sparse computing, then the sparse model will run much faster on that than standard GPU. I think there will be more and more platform/hardware can support that in feature.
Thanks for the detailed explanation from @yuanyuanli85 . Some extra information:
If you are using standard tensorflow or hardware, they can't take the advantage of sparse model to reduce model size and flops. But, will give chance to particular hardware or library to skip the computing with zero weights. let's say i have a GPU with is designed for sparse computing, then the sparse model will run much faster on that than standard GPU. I think there will be more and more platform/hardware can support that in feature.
Thanks a lot.
hello,may I ask you two questions?
why weight_sparsification can reduce flops?
and will the size of model reduce after weight_sparsification ?