yuqinie98 / PatchTST

An offical implementation of PatchTST: "A Time Series is Worth 64 Words: Long-term Forecasting with Transformers." (ICLR 2023) https://arxiv.org/abs/2211.14730
Apache License 2.0
1.37k stars 248 forks source link

about res attention #81

Open Hannibal046 opened 8 months ago

Hannibal046 commented 8 months ago

Hi, thanks for the work. After digging into the code, I found that res_attention parameter is set to False in PatchTST_self_supervised setting while True in PatchTST_supervised setting. Could you please share some insights behind this? Thanks! image

bb6a3510baa6d1454646485e84038fc