Closed timmeinhardt closed 1 year ago
Hi! We don't use it, we just forgot to remove it from the code.
Did you ablate whether adding/removing the level encoding has an effect on the performance?
In our early experiments, adding it didn't change the performance but we didn't check in the final model.
Your
MaskedTransformerDecoder
never applies the level encoding during the forward pass. Was this intentionally?