yaringal / multi-task-learning-example

A multi-task learning example for the paper https://arxiv.org/abs/1705.07115
MIT License
838 stars 206 forks source link

Sounds like a lucky result comes from a wrong formula deduction #9

Open edfall opened 4 years ago

edfall commented 4 years ago

I read the paper carefully, the formula in paper is fundamentally wrong.

yaringal commented 4 years ago

"Under the formula (2) and (3), the probility output has a gaussian distribution."

"However, the probility can't be a gaussian distribution as it distributed in [0,1] rather than (-infty, +infty)."

"if we just look at the first line in formula (7), if independent assumption is established, -log p(y1, y2|f(w,x)) = -log p(y1|f(w,x)) - log(y2|f(w,x)); which is just a sum of cross-entropy loss over different tasks."

"This is apparently contradicted with the result under additional gaussian assumption."

"Somehow, the paper repalce the cross entrophy loss with mse"

"which finally reach the result that higher loss task should have higher theta weights."

jiazhiguan commented 3 years ago

fabulous!