in your Wayformer implementation, the intra-class attention modules (tf_tl, tf_map and tf_other) are nn.ModuleLists.
Hence, for loops are required to use each module. I changed your code accordingly.
(Not an issue in your default config, but as soon as you increase n_layer_tf)
Hi,
in your Wayformer implementation, the intra-class attention modules (
tf_tl
,tf_map
andtf_other
) are nn.ModuleLists. Hence, for loops are required to use each module. I changed your code accordingly. (Not an issue in your default config, but as soon as you increasen_layer_tf
)