Closed WeiFoo closed 9 years ago
cant see those results
and do you mean RQ3... which is about discrete values... or RQ4... which is about the numeric distributions of fig4.
and if you meanf fig4, are you saying that statistically, there is no difference in the curves shown in fig4, as determined by KS?
In 4.2 section, the next to last paragraph. I added a sentence about KS test done on table 4, and 5
what about the fig4 nums. are they insignificantly different?
in fig4, if D value got from KS test greater than 1.36 x sqrt(34/(17*17)) = 0.46, we can reject that two distributions are from the same distribution.i.e. they're different distributions.
According to these D's, we will say, they're all not significantly different from each other!
what is the threshold for "different"?
0.46
Precc
CART | RF | WHERE | |
---|---|---|---|
CART | 0.29 | 0.42 | |
RF | 0.18 | ||
WHERE |
F
CART | RF | WHERE | |
---|---|---|---|
CART | 0.24 | 0.24 | |
RF | 0.29 | ||
WHERE |
can you add the above tables to the paper
Sure, I will do it.
added! right below fig.4
This observation is supported by the KS results of Table 7.
At a 95% confidence, the KS thresh- old is 1.3634/(17 ∗ 17) = 0.46
which is greater than the values in Figure 4.
That is, no result in Figure 4 is significantly different to any
other– which is to say that there is no evidence that np=10
is a poor choice of search space size.
Here we do KS on improvements of different learners, our conclusion is no result in Figure 4 is significantly different to any other
Do we need to do KS test on results of each learner with np = 10 and recommended np? then the conclusion would be for each learner, tuning with np = 10 is no significantly different to any recommended np's?
To me, the above two conclusions are slightly different.
i thought your treatments were
np=10 vs np=EvertthingElse
if so, our current results are fine.
but please advise
I added the KS results into RQ3, but what I did was pass those numbers from tuned CART and Tuned RF columns to do KS test. is it right?