Open nknauer opened 2 years ago
Hey, sorry didn't notice this issue, the answer to the first question is here: https://github.com/facebook/prophet/issues/323#issuecomment-1037215919
For the second question, the cross-validation part of this page describes the default behaviour: https://facebook.github.io/prophet/docs/diagnostics.html
performance_metrics
function takes the last 10% of predictions in each horizon and calculates the rmse across those datapoints. I think the reason for this default is that in practice we are usually interested in forecasting out to a particular point in the future, after which we would re-train with the latest data and predict again. So averaging across all future datapoints might not be representative of the "success" of the forecast in practice.rolling_window = 1.0
argument in performance_metrics
function.
@bletham @tcuongd
I am running the prophet model without weekend data in the initial dataset. When creating a future dataframe, I remove weekends.
When doing
cross_validation
function with cutoffs and creating multiple models, will the diagnostic values (RMSE, MSE, etc.) be skewed because we cannot remove weekend data in the future dataframe?My assumption would be yes and I would need to do the cross validation manually (not using the cross_validation function) with cutoffs and creating a future dataframe with weekends removed. Then calculate the diagnostic metrics manually.
(I'm doing this using R by the way)
My second question is that the result currently is by horizon. If I just want one number, per group (specifically RMSE), would you recommend to average all the RMSE values?
Result would therefore be: