yrlu / irl-imitation

Implementation of Inverse Reinforcement Learning (IRL) algorithms in Python/Tensorflow. Deep MaxEnt, MaxEnt, LPIRL
587 stars 146 forks source link

LPIRL: Redundant Constraints #4

Open aaronsnoswell opened 6 years ago

aaronsnoswell commented 6 years ago

Hi! Thank you for this great reference implementation - it is very helpful.

I was going over the LPIRL implementation and I think you have some redundant constraints in your LP matrices - see line 59 in lp_irl.py - this loop does the same thing as the previous loop on line 55, resulting in a redundant set of constraints.

Thanks again,

MithunNallana commented 5 years ago

@aaronsnoswell

Yeah. I too think the same way. In line 60, It should be -1.0 instead of 1.0. If I am correct, constraint in line 55 and line 59 should provide constraints that proxy l1 norm. Can you please check this @yrlu .