Closed AllenSun1024 closed 1 year ago
there are several possible reasons.
there are several possible reasons.
- the speed-up is data-dependent, it is not possible to have the same speed-up for all kinds datasets.
- Parameters, like early stopping may result in very different speeds.
- Versions. There are significant changes after the paper, some changes may affect the speed. you can try the early version of LightGBM.
Got it, thank you anyway.
This issue has been automatically locked since there has not been any recent activity since it was closed. To start a new related discussion, open a new issue at https://github.com/microsoft/LightGBM/issues including a reference to this.
Description
With the same parameters except data_sample_strategy, time elapsed for training with parameter
data_sample_strategy = bagging
is 194s while that with parameterdata_sample_strategy = goss
is 1754s, which is 10x slower approximately.As stated in paper <>, GOSS can obtain quite accurate estimation of the information gain with a much smaller data size. However, it doesn't perform well in reality.
Why?