Open sayanb-7c6 opened 3 years ago
Are you sure about the option limit_train_batches=90
in your call to the pytorch lightning trainer? As far as I know it's only for testing purposes.
@georgeblck Thank you, I totally missed that. Will update here, if removing that line improves the accuracy of the model.
Expected behavior
We were trying to reproduce the results from the electricity dataset that the authors of TFT paper made use of. The dataset can be downloaded by following the instructions from here. The specific file we used is called hourly_electricity.csv, it has about 1 million rows.
Actual behavior
We tried 8-10 iterations of the model with different params/hyperparams but were not able to get close. We kept the all the params as close to the ones used in the params, but no luck so far. Some of the optimal params used by the authors can be found here. We also tried overfitting the model on 10% of the training data following the examples from here, but the program crashed because of memory shortage on the gpu. We'll try again with a smaller percentage, but even before crashing the validation loss was in par with our non-overfitting models, so we're not so sure that'll help us.
Here's the relevant code to reproduce the results
The predictions on the validation data looks like this:
We're getting a MAPE of 0.9609 which is pretty bad. Wondering what's the root cause. We've used StandardScaler() from sklearn on the target variable to squish the range and normalise it, but it's not helping.