Open takafusui opened 3 years ago
No, primarily because there weren't any good implementations of nonlinear conjugate gradients last time we looked.
Is there a reason you need this specifically over something like LBFGS?
Thank you @jacobrgardner for your answer. Not really. I usually use either adam or PyTorch-LBFGS to train GP. I want to try SCG and compare it with the other two optimizers.
@takafusui we're open to a PR, if you'd like to implement it! 😉
Hi,
Are there any implementations where the scaled conjugate gradient (: SCG) method is used to optimize the model's hyperparameters, instead of LBFGS or the stochastic gradient descent?