Closed aParsecFromFuture closed 1 year ago
I can't reproduce using the latest xgboost. Could you please share a reproducible example?
I can't reproduce using the latest xgboost. Could you please share a reproducible example?
It's from a Kaggle competition. I can't make the notebook public. Do you have any Kaggle account that I can share the notebook with?
I have used the xgboost library a lot, I only face such a problem with this dataset.
I get the same behavior with built-in error metric. (1 - error) is not equal to accuracy score.
Class distribution: (0: 32496, 1: 20523)
Let me get back to this tomorrow. I have an account with the same email address on my github profile.
Let me get back to this tomorrow. I have an account with the same email address on my github profile.
I shared the notebook with trivialfis user. Good luck.
This is a limited-participation competition. Only invited users may participate.
This is a limited-participation competition. Only invited users may participate.
I'm going to open it publicly for today. Let me know when your review is done. Link: https://www.kaggle.com/greysky/temporary-notebook-it-will-be-deleted
thank you for sharing, I can see the notebook, it's the dataset that I can't access. I need to participate in the competition before seeing the dataset, but the competition is invite only.
thank you for sharing, I can see the notebook, it's the dataset that I can't access. I need to participate in the competition before seeing the dataset, but the competition is invite only.
I have created the same notebook with copy-dataset and invited you. Could you try again?
Hi, I think it's possible that this is caused by the verbose=20
in the fit
method, the latest evaluation result is not displayed as the model print only at each 20 iterations.
You can obtain the list of evaluation results by clf.evals_result()
.
NVM, later iteration is changed, still looking.
Hi, I think it's possible that this is caused by the
verbose=20
in thefit
method, the latest evaluation result is not displayed as the model print only at each 20 iterations.You can obtain the list of evaluation results by
clf.evals_result()
.
There is no iteration with output accuracy. I noticed early that even if we do hyper parameter optimization (like Optuna), the output accuracy remains 0.61 as eval_metric output approaches 0.70. I was thinking that this happens because the LogLoss value is constantly increasing. But "disable_default_eval_metric" option didn't prevent that strange behavior.
Thank you for raising the issue and for the helpful assistance! I have reproduced the error using the dataset from the competition.
Accuracy score should be same score of the last iteration but I get 73% and 69%. Validation data has 10k rows. Any idea?