CUNY-CL / yoyodyne

Small-vocabulary sequence-to-sequence generation with optional feature conditioning
Apache License 2.0
25 stars 15 forks source link

Generalization of `expert` `teacher_forcing` and `monotonicity` across model architectures #198

Open bonham79 opened 2 weeks ago

bonham79 commented 2 weeks ago

Something I've been thinking about with expansion of library: a decent amount of the work we've been using involves application of inductive biases and teacher-prompted training to model architecture. Currently we have:

One thing I would like to do with the next overhaul is modularize these beyond their respective models (like we're trying to do with https://github.com/CUNY-CL/yoyodyne/issues/77 for teacher forcing) so that they can be 'dropped in' wherever. This would allow 'fun' combinations such as:

A lot of these things won't necessarily click, but I believe adding this new modularity layer would allow easier use of curricular learning and exploration scheduling that isn't easy to implement in other libraries. Expanding utility.

(This is down the roads thought. Post-beta.)

Adamits commented 2 weeks ago

Without thinking through how these combinations would work too much, this sounds exciting and like a good idea! I am on board.

kylebgorman commented 2 weeks ago

Yeah that sounds like a Johns Hopkins PhD dissertation ;)

bonham79 commented 2 weeks ago

Yeah that sounds like a Johns Hopkins PhD dissertation ;)

Am I missing a reference for the JHU?

kylebgorman commented 2 weeks ago

Am I missing a reference for the JHU?

no it just used to be the home of this sort of thing