Closed cherepanovic closed 4 years ago
@cherepanovic Sorry for the long delay on responding here -- both Geoff and I were at NeurIPS all last week.
The parameters involved in variational inference include:
10 x 784
matrix you see is probably the mixing parameters in the softmax likelihood.num_features x num_inducing
mean and a num_features x num_inducing x num_inducing
covariance matrix, so that e.g. mu[0]
represents the variational mean for the first GP.@jacobrgardner glad to see you and thanks a lot for your response!
what you are seeing is a num_features x num_inducing mean
it is graphical interpretable?
Is there generally a way to interpret the learning/learned density/boundaries in lower dimensions graphically?
@cherepanovic - "graphically interpretable" is a bit subjective. However, each of the features output by the GP are independent. What I would do is i would make num_features
plots, each plotting inducing_mean
as a function variational_strategy.inducing_points
for each one of the num_features
.
Is there generally a way to interpret the learning/learned density/boundaries in lower dimensions graphically?
This might be difficult because the output is a bit high dimensional. Something like TSNE might be your best bet.
(To answer your question - as far as I know there is no established practice, but these are ideas for where I would start.)
@jacobrgardner Hello, I just see the value of variational covariance, but it is a lower triangular matrix. So, is it a covariance matrix or just a cholesky factor (L) ? If I would like to get covariance matrix, should I use LL^T ?
in following constellation Dense(784,784) -> GP (see snippets below) I have following parameters for the optimization
The first two lines are parameters of the dense layer.
The rest is supposing to be from the GP
Could you give a short explanation what the parameters are and whether it makes sense to visualize some of them during training?
code snippets
GP layer
DLK snippet
optimizer