Open kklein opened 3 months ago
FYI @MatthiasLoefflerQC I created a first draft of how the same splits could be used for all base learners, including treatment models. As of now the estimates are still clearly awry, e.g. an RMSE of ~13 compared to ~0.05. This happens for both in-sample and out-of-sample estimation. I currently have no real ideas on what's going wrong; will try to make some progress still
As of now the estimates are still clearly awry, e.g. an RMSE of ~13 compared to ~0.05.
The base models all seem to be doing fine wrt their individual targets at hand. Yet, when I compare pairs of treatment effect model estimates at prediction time, it become blatantly apparent that something is going wrong:
np.mean(tau_hat_control - tau_hat_treatment)
>>> 27.051119307766754
np.mean(tau_hat_control)
>>> 14.104902455634836
np.mean(tau_hat_treatment)
>>> -12.946216852131919
Update: These discrepancies have been substantially reduced by bbfff15. The RMSEs on true cates are still massive when compared to status quo.
TODOs:
cv
are actually not used for training when passingcv
tocross_validate
.CrossFitEstimator
.synchronize_cross_fitting
should be allowed to beFalse
for the X-Learner.Observations
yields the following output:
Checklist
CHANGELOG.rst
entry