Closed Wty1122 closed 1 year ago
Hi! This is a very good questions. Some of the discussion about Transformer decoder may lead to performance degrading can been found in another recent paper (https://arxiv.org/pdf/2212.02789.pdf). But we haven't done serious analysis on that.
Hi, it seems that only the encoder part of the transformer is used in the model. However, both Autoformer and FEDformer use the structure of encoder + decoder. Is it better to use the encoder than the full structure (encoder + decoder) on the time series forecasting task? Could you provide some literature or experimental support?