jasp-stats / jasp-issues

This repository is solely meant for reporting of bugs, feature requests and other issues in JASP.
59 stars 29 forks source link

Filtering data changes the output (still) #794

Closed K1NDRED closed 4 years ago

K1NDRED commented 4 years ago

JASP Issue dataset.zip

Steps to reproduce:

  1. Go to the data window
  2. Click on ''Training_group" column
  3. Filter out the NB group by X-ing it (see pic 1)
  4. Run a Bayesian RM Anova (see pic 2 & 2a), note the values of the output
  5. Delete the Bayesian RM Anova from the output
  6. Go back to the data window & Click on ''Training_group" column
  7. Click the X to remove the filter so that both training groups are available
  8. Filter out the NB group again by X-ing it
  9. Run the Bayesian RM Anova again (see pic 2 & 2a), note the values of the output are different.
  10. Refreshing the results gives different outputs constantly
pic 1 pic 2 pic 2a
boutinb commented 4 years ago

@JohnnyDoorn Could you look at this issue?

JohnnyDoorn commented 4 years ago

Hi @K1NDRED ,

Could you give some screenshots of the different output? How big are these changes? The Bayesian ANOVA's use MCMC-sampling, so it is expected that there are small differences between subsequent runs, but these differences typically should not lead to qualitatively different conclusions.

Another solution to your problem could be to try this with the latest version of JASP. I cannot seem to reproduce this issue.

Kind regards, Johnny

K1NDRED commented 4 years ago

Hi @K1NDRED ,

Could you give some screenshots of the different output? How big are these changes? The Bayesian ANOVA's use MCMC-sampling, so it is expected that there are small differences between subsequent runs, but these differences typically should not lead to qualitatively different conclusions.

Another solution to your problem could be to try this with the latest version of JASP. I cannot seem to reproduce this issue.

Kind regards, Johnny

Hello,

Thank you for your response. Upon further testing, the differences aren't large enough to change the qualitative conclusions.

This might not be the place but I'm not sure I understand why using the MCMC-sampling causes these variations.

Cheers

JohnnyDoorn commented 4 years ago

Hi @K1NDRED ,

MCMC-sampling is a computational technique that is at the core of many Bayesian analyses. This technique uses random number sampling in order to approximate the posterior distribution, which is what Bayesian analyses are based on. Because of the sampling, every time you run an analysis the numbers will vary slightly (but should not lead to qualitatively different conclusions).

I will close this issue now, please reopen if anything is still unclear!

Kind regards Johnny