Closed vieting closed 3 years ago
This actually doesn't fix anything, except that it adds multi_head_attention_forward
. So all worked fine now? (After the previous fixes.)
Yes, all the fixes that I needed to get this to work were handled in separate PRs before. Not sure what you mean with all worked fine now, but at least the test here works. It's a rather simple case though, I didn't test all the possibilities like e.g. using bias_k
, bias_v
, add_zero_attn
.
but at least the test here works
Yes, that's what I mean. I assume it reflects what you actually need to go forward with Wav2Vec2 or other things.
Adds
F.multi_head_attention_forward
and a test case which is a simplified version of theMultiheadAttention
module in fairseq.