Closed krsnapaudel closed 3 months ago
@aikepotze: "However, this would need changes to the class structure, as the base nn model cannot re-initialize itself. What do you think?" I don't think the model needs to reinitialize itself. We have saved init_args
for models we create to optimize hyperparameters. That I think is sufficient.
I have validated results from AgML workshop (#261). Now the workshop implementation of LSTMModel
and ExampleLSTM
included in the benchmark have similar results when using workshop data. The architectures are similar. Some differences:
LSTMModel
relies on batch norm for normalizing inputs and intermediate features. ExampleLSTM
normalizes inputs on its own. There is no batch norm after LSTM layer.LSTMModel
does not use LRScheduler
; ExampleLSTM
does.The NRMSE on test years [2012, 2018] is approximately 25% for both models.
Tried to simplify nn-models code.
settings
,args
andkwargs
.predict_batch
andpredict
.