if self._generate_PE is not None:
positional_encoding = self._generate_PE(K, self._d_model)
positional_encoding = positionalencoding.to(decoding.device)
decoding.add(positional_encoding)
for layer in self.layers_decoding:
decoding = layer(decoding, encoding)
Why are both input parameters of the decoder the output of the encoder? Shouldn't one of the inputs be a future sequence?
tst/transformer.py Line 138
Decoding stack
decoding = encoding
Add position encoding
if self._generate_PE is not None: positional_encoding = self._generate_PE(K, self._d_model) positional_encoding = positionalencoding.to(decoding.device) decoding.add(positional_encoding)
for layer in self.layers_decoding: decoding = layer(decoding, encoding)
Why are both input parameters of the decoder the output of the encoder? Shouldn't one of the inputs be a future sequence?