Open crew102 opened 2 years ago
Thanks for the suggestion, @crew102 . We are currently working on a way to replace the 1st step xgboost model for a pre-trained one. Both XGBSEDebiasedBCE and XGBSEStackedWeibull modules will be able to use this feature, which will cover your use case.
I'm wondering if you've considered allowing the user to pass in a separate training sets for the xgboost model vs the survival model?
For example, in XGBSEStackedWeibull, the current state is this:
I'm proposing this:
The rationale for having different datasets used between the models is that it reduces the chance of overfitting. I've found that the risk scores that come out of step 2 are indicating a tighter relationship between risk score and y_train than there actually is, by nature of the fact that we are predicting back on the dataset that the xgboost model was trained on (and then re-relating things to the original outcome variable, y_train).
Thanks for the awesome package