:exclamation: This is a read-only mirror of the CRAN R package repository. lightgbm — Light Gradient Boosting Machine. Homepage: https://github.com/Microsoft/LightGBM Report bugs for this package: https://github.com/Microsoft/LightGBM/issues
Other
1
stars
0
forks
source link
Inconsistent behavior of median regression (objective = "quantile", alpha = 0.5) with early stopping. #1
i'm experiencing unexpected behavior of lightgbm model trainings on a rather small dataset with about 4100 rows.
I'm using earlyStopping with an validation set with 310 rows. It was generated by stratified (with respect to the label y) sampling from the data. I also played a little with the other parameters (like learning_rate etc) but that didnt really help.
The model training will stop after iteration 1 and learn nothing for (objective = "quantile", alpha = 0.5) but will produce meaningful results for the following parameters:
(objective = "regression_l1", alpha = NULL)
(objective = "quantile", alpha = 0.4999)
(objective = "quantile", alpha = 0.5001)
Implications for me:
I avoid using lightgbm with quantile regression in fully automated processes and try to use the more robust "regression_l2" objective if the application context permits.
Here is my code for a minimal working example. Unfortunately i cannot share the data publicly but i think i can provide it to the devs if that's helpful.
R version 4.0.3 (2020-10-10)
Platform: x86_64-w64-mingw32/x64 (64-bit)
Running under: Windows 10 x64 (build 19043)
lightgbm_3.3.1 (CRAN version)
Hi there,
i'm experiencing unexpected behavior of lightgbm model trainings on a rather small dataset with about 4100 rows. I'm using earlyStopping with an validation set with 310 rows. It was generated by stratified (with respect to the label y) sampling from the data. I also played a little with the other parameters (like learning_rate etc) but that didnt really help.
The model training will stop after iteration 1 and learn nothing for (objective = "quantile", alpha = 0.5) but will produce meaningful results for the following parameters:
Implications for me: I avoid using lightgbm with quantile regression in fully automated processes and try to use the more robust "regression_l2" objective if the application context permits.
Here is my code for a minimal working example. Unfortunately i cannot share the data publicly but i think i can provide it to the devs if that's helpful.