Open efelem opened 5 years ago
Hi Efelem,
We conceptually assume the target pose in the paper as a modified reference pose. The network learns how we modify the reference motion according to the dynamic states of the character. To compute the actual target pose, we add the network output and the pose from the reference motion.
Hi,
After learning a walking gait for several days and observing the results the gait seems ok. However I noticed that the BVH file is still needed to run the network after optimization. I am a bit confused because in the paper there is a claim that says :
When looking further I noticed that the call to
GetActivationFromNN
in theWindow::Step()
methods makes use of the target positions. Indeed in this method there is a call toEnvironment::GetDesiredTorques()
. When looking at this method we can see that themTargetPositions
is used and the output of the mimicking networkmAction
is added to the desired positions provided by mTargetPositions.The
mTargetPositions
is set in theEnvironment::SetAction
method which looks like this :I noticed two things : 1) The
mTargetPositions
comes from theCharacter::GetTargetPosAndVel
which explicitly use the BVH file. 2) the0.1
factor is multiplying the output of the mimicking controller.Now my question is, how do you proceed to get rid of this TargetPosition after learning ? Is there a missing part of the code that implements a sort of shrinkage of mTargetPositions in favor of the of output of the mimicking controller. Or did you use an other technique to achieve this ? Or did I mistunderstood some parts of the code ?
Thanks !