byu-dml / d3m-dynamic-neural-architecture

1 stars 1 forks source link

Make (dag) attention models more user friendly #162

Closed bjschoenfeld closed 5 years ago

bjschoenfeld commented 5 years ago

The attention models require that attention_in_features be divisible by n_heads. This is problematic for the genetic tuning as it does not always choose these args to satisfy that requirement. It would be much nicer if all reasonable arg settings we valid. attentention_in_features should be replaced with another argument that is used to compute attentention_in_features by multiplying by n_heads.