OpenRobotLab / HIMLoco

Learning-based locomotion control from OpenRobotLab, including Hybrid Internal Model & H-Infinity Locomotion Control
https://junfeng-long.github.io/HIMLoco/
Other
250 stars 27 forks source link

Code release time #2

Closed lingxiao-guo closed 5 months ago

lingxiao-guo commented 6 months ago

Hi, I wonder how soon will you release the code, since this wonderful work has already accepted. Thanks!

pengzhenghao commented 6 months ago

+1

JL-Brain commented 5 months ago

+2

jnaliu commented 5 months ago

+1

jnaliu commented 5 months ago

Hi, I hope this message finds you well.

In Figure 2 of your paper, you illustrate the components of the policy network's input, which include a hybrid internal embedding derived from the robot's historical observations. I am intrigued by the distinction between the explicit velocity ( \hat{v}_t ) and the actual velocity ( v_t ) as depicted in the figure. The paper describes ( \hat{v}_t ) as part of the hybrid internal embedding optimized through contrastive learning to simulate the robot's response. However, the precise definition and role of ( v_t ) within the context of the proprioceptive information are not explicitly detailed in the text I have access to.

Could you please clarify the following:

  1. Regarding ( v_t ), how does it relate to the explicit velocity ( \hat{v}_t ) provided by the proprioceptive sensors, and how is it used within the framework of your hybrid internal model?

Your insights would be invaluable for my understanding of the nuanced aspects of the HIM and its application in agile legged locomotion. I appreciate your time and expertise in elucidating these concepts.

Thank you very much.

Best regards

Junfeng-Long commented 5 months ago

v_t is the target of \hat{v}_t, the hybrid internal model gives an estimation of v_t, which is \hat{v}_t. \hat{v}_t is used as a part of policy input to estimate disturbances.

Junfeng-Long commented 5 months ago

Sorry for the late. Code is released.