Closed m-ali-awan closed 2 years ago
I have further dived into code and came to know that outputs from TemporalAttentionLayer are nan always, and from FeatureAttentionLayer some come out as nans.
If you haven't already, I suggest you ensure that
Regards, Axel
I have further dived into code and came to know that outputs from TemporalAttentionLayer are nan always, and from FeatureAttentionLayer some come out as nans.
Have you solved this problem yet? I also have this problem, During training, the losses are all nan,and there are many zeros in the data. And I don't understand why the num_values in labeled_anomalies.csv and the shape in the .npy file in the train folder are different. For example, C-1 in labeled_anomalies.csv is 2264, but C-1.npy is 2158. '2264' and '2158' don't match.
Same problem with the SMD dataset, using the default hyperparameters.
Setting use_gatv2 to False can produce normal loss.
Check out my answer on #13, I believe it is due to uninitialized bias parameters.
Check out my answer on #13, I believe it is due to uninitialized bias parameters.
It works for me now, great!
Hi, hope you are fine. Thanks for this wonderful work. I tried training with MSL, and SMD, and my losses are always nan. Moreover, I also tried GDN repo, and I found that there is a difference in MSL data as compared to this repo. Thanks for any help.
Regards, Ali