thiyangt / seer

:chart_with_upwards_trend: Feature-based Forecast Model Selection (FFORMS) :clock1: :clock130: :clock10: :clock1030: :clock11: :clock1130: :clock12: :clock1230: :clock2: :clock230: :clock3: :clock330: :clock4: :clock430: :clock5: :clock530: :clock6: :clock630: :clock7: :clock730: :clock8: :clock830: :clock9: :clock930:
https://thiyangt.github.io/seer/
79 stars 11 forks source link

Possible improvement to reduce user processing time #4

Closed VicenteYago closed 4 years ago

VicenteYago commented 5 years ago

This is not an issue properly but I dont now where post this.

I have observed that both Professor Hyndman and Pablo Montero and you among others frequently use the M3 / M4 data set to feed the meta-learning classifier.

The result of these incredible tools that you have developed depend on the amount of time series that are used to train the classifier.

If I am right, you have successfully participated in these competitions, so at some point you will have trained the models with all these time series.

My question is: Why do not you provide the configuration of the classifier? In this way users would have a huge advantage when using your software, saving precious computing time training the classifier with the series of M3/M4 competitions.

For example you could provide the classifier already trained or a kind of database with the following format:

What you do is as if someone develops an incredible neural network to detect faces and instead of giving it already trained with all the weights, the user is left to spend hours and hours to train the model with the heavy dataset with which developers have trained the net.

In terms of your terminology, a predetermined offline phase should already be given.

Maybe in the form of subsets of M3/4, for example a classifier for anual series other for hourly/quarterly etc.

thiyangt commented 5 years ago

Yes, the pre-trained classifiers will be uploaded to the package soon.

VicenteYago commented 5 years ago

Great, I'm glad to hear that.