Closed patxikuku closed 4 years ago
Gradients are computed and used in L-BFGS-B
. This is the whole point of using TensorFlow
. Perhaps this is not immediately obvious when examining the code, because the gradient computation is handled via GPflow
.
Hello, Does it mean using numerical calculated gradient, not analytically calculated gradient?
It’s neither, it’s via automatic differentiation.
I thought the minimize() automatically calculate the gradient using the finite-difference method. (In case of scipy.optimize.minimize; "If None or False, the gradient will be estimated using 2-point finite difference estimation with an absolute step size.") https://docs.scipy.org/doc/scipy/reference/generated/scipy.optimize.minimize.html
Anyway, I'll study TensorFlow and GPflow. Thanks!
Hello,
if I understood correctly, the authors of PILCO uses a gradient based method for optimising the policy. In the current implementation it doesn't seem to the case, you use L-BFGS-B without giving the computation of the jacobian.
Did you make any experiments using a gradient based method ?