Closed pdehaye closed 4 years ago
We aggregated the measurements from the different scenarios in experiment 34 for the PR curves. We looked into the effects of different distributions of distances and packet loss at distances and how this could change the precision / recall curves. We further looked into other uncertainties in distance calculation like those in phone calibration, phone position on the person, different materials in the environment etc. All of these considerations, including the feedback from the system and information from other researchers led to the determination of the thresholds.
I had a look in more details to the data, specifically the office and train data. Thanks to @ s___m__ on Twitter for getting started with some of the necessary work.
I was very surprised at first to see the following jumps between 1.5m and 3m thresholds:
train
office
In both cases it seems things go so much better with the 3m threshold than the 1.5m one, which is counter-intuitive. I then checked the histograms (train, then office):
Of course it now makes sense: you can't misclassify many signals at range >3m if you don't take many such measurements.
On the other hand, it is kind of alarming, for two different reasons:
<=r
is expected to grow quadratically inr
(for open floorplans). You are very much undersampling there.In the end, it looks like you just summed the sampled data for all scenarios to get a big dataset, that was then used to pick thresholds.
Have you checked what you were doing with a statistician?