Closed jan-matthis closed 8 months ago
Yes, that's a very good idea!
I have code for SBC by Talts et al.. This would require some hyperparameter decision which we should discuss. But the coding should be straight forward.
A review of the Bayesian workflow, which will likely be very relevant for this endeavour: Bayesian workflow (Gelman et al. 2020)
I also just noticed that our logging is very rudimentary, e.g., we only log the number of epochs trained, and the single best validation performance. It would be much better to at least log the (best) validation log prob for every epoch to see how it evolved.
For visual analysis, posterior predictive, and others, it could be interesting to consider ArViZ.
Can you please check the link to the publication by Hermans et al? The link appears not to work.
How about the coverage property? See Prangle D, Blum MGBB, Popovic G, Sisson SA. Diagnostic tools for approximate Bayesian computation using the coverage property. Aust N Z J Stat. 2014;56: 309–329. doi:10.1111/anzs.12087
Most the of the methods are implemented by now: SBC, ArviZ, coverage.
In discussion with @ppjgoncalves, we thought it would be very useful to add diagnostics to
sbi
to help users decide whether inference is working as it should. I created this issue to collect ideas for diagnostics to implement and for discussion on the topic more generally.Note that while some diagnostics might be easy to write as functions, others require manual steps, so eventually we might want to make a new tutorial or incorporate them into existing ones.
Diagnostics