Closed gaopeng-eugene closed 6 years ago
Impeccable reasoning! Joint optimization is wrong because, otherwise, they wouldn't have won the award!
In the iccv paper supplement, they prove alternating optimization for pi is convex and guarantee converge to a unique optimal by several iterations. The info Gan is very similiar with random forest and decision tree. However, when you use joint optimization nothing exist. Unless you can give a prove. Please go to author’s homepage and read the provement.
Another reason I found this paper is interesting is dynamic routing in capsule net and follow-up paper EM optimization. It seems that this two papers have some link.
Another to prove joint optimization is wrong is when you visulize the distribution in the leaf. There is no patterns because it’s only a feature representation which do not contain any semantic meaning. However, when you train by EM. Same class will follow similiar routing path. The leaf distribution is more interpretable just because past experience show decision tree leaf is very easy to understand.
Seems like you just read the analysis but yet understood what are the disadvantages of EM and why it requires convexity analysis.
First, EM algorithm, or alternating optimization in general, suffers from slow convergence. This is because the other variables that are not optimized slow down the convergence of the other variable.
Second, alternating optimization is not guaranteed to converge to the optimum. This is why more analysis is required for alternating optimization.
I explained the same thing in the readme. So please read before you make random comments that do not make much sense.
Thank you so much for your answer. Have you ever compared joint optimisation with alternating optimisation ? If joint optimisation is the right answer, why the author chose the much more complex and difficult to implement alternating optimisation?
The author respond me just now saying that joint optimization is OK. But they only report EM method because its more effective.
Okay, I'm not working on this project. Why don't you implement their method and compare? You just have to remove the softmax variables from the optimization and optimize them separately using the update rule they proposed. It wouldn't be that difficult and this way, you can actually contribute to the society. (:
Sure, I will test it on ROI region with joint optimisation. Again, big thanks for the implementation. Looking forward to see your interesting work. 2D-R2N2 is very good 😊
Do you know any paper proposing boost differentiable decision forests? Basically, the newly added tree will focus on misclassified sample ?
In the ICCV paper, the author only proposed alternating optimisation trained with resilient SGD. Joint optimization is definitely not the right solution. If they author proposed joint optimization in the ICCV paper, the contribution is minor and will not be ICCV best paper.