namiyousef / multi-task-learning

Repository for multi task learning
2 stars 2 forks source link

Re-design lambda's to make sense w.r.t OEQ #8

Closed namiyousef closed 2 years ago

namiyousef commented 2 years ago

Currently, the way the weights are specified is arbitrary. We can choose any lambda values of our choice, without any constraints on them. For example, one configuration has lambdas = [1, 1/100, 1/10000]

This can make sense, e.g. if the loss functions we choose are of O(1), O(100) and O(10000) respectively. However, in my opinion this choice is extremely arbitrary (note it isn't invalid per se). We need to choose losses such that they are all of the same order, then we can apply lambda's such that sum(lambdas) = 1 (thereby constraining the problem). Without this, I'm not sure if the OEQ makes sense?

After a quick search during our meeting today I found these:

Both seem to go for arbitrary losses. I would prefer going for a solution where the losses are of the same scale, allowing us to apply the constraint above. It would make more sense from an OEQ pov.

What are your thoughts?

namiyousef commented 2 years ago

Closing issue. Seems like scaling is arbitrary. The final code will have separate inputs for scaling and weight constraints separately.