DP-3T / bt-measurements

4 stars 2 forks source link

Statistical treatment of the data #4

Closed pdehaye closed 4 years ago

pdehaye commented 4 years ago

I had a look in more details to the data, specifically the office and train data. Thanks to @ s___m__ on Twitter for getting started with some of the necessary work.

I was very surprised at first to see the following jumps between 1.5m and 3m thresholds:

train

train_confusion_15 train_confusion_3

office

office_confusion15 office_confusion_3

In both cases it seems things go so much better with the 3m threshold than the 1.5m one, which is counter-intuitive. I then checked the histograms (train, then office): train_histogram office_histogram

Of course it now makes sense: you can't misclassify many signals at range >3m if you don't take many such measurements.

On the other hand, it is kind of alarming, for two different reasons:

In the end, it looks like you just summed the sampled data for all scenarios to get a big dataset, that was then used to pick thresholds.

Have you checked what you were doing with a statistician?

gannimo commented 4 years ago

We aggregated the measurements from the different scenarios in experiment 34 for the PR curves. We looked into the effects of different distributions of distances and packet loss at distances and how this could change the precision / recall curves. We further looked into other uncertainties in distance calculation like those in phone calibration, phone position on the person, different materials in the environment etc. All of these considerations, including the feedback from the system and information from other researchers led to the determination of the thresholds.