Closed christophM closed 1 year ago
Hello @christophM,
I think this behavior is actually normal, let me explain:
Please tell me if it is clear enough for you
Vincent
Hi Vincent
It makes sense to me now.
I really appreciate that you responded so quickly and took the time to explain it. I learned something new today.
Would it make sense to add a sentence to the documentation of MapieClassifier.predict
:
Options False
and True
can result in coverages larger than $1-\alpha$, see [1]
[1] Angelopoulos, Anastasios, et al. "Uncertainty sets for image classifiers using conformal prediction." arXiv preprint arXiv:2009.14193 (2020).
Again, thanks a lot, I really appreciate not only your answer but all of the team's effort to build and maintain MAPIE.
Best, Christoph
Description
The classification coverage for method="cumulated_score" doesn't match $1 - \alpha$ for options include_last_label=False and include_last_label=True.
To Reproduce Steps to reproduce the behavior: Got to this COLAB notebook or see code here:
For the 'randomized' option the method produces around the expected coverage: 0.526
But for include_last_label=True:
The coverage is much higher 0.752
And also for include_last_label=False:
The result coverage is 0.653
Expected behavior I'd expect the coverage for "include_last_label=randomized" to be bounded between "include_last_label=False" and "include_last_label=True". Instead, False/True options have a way too high coverage.
So I'd expect for the coverages: False < randomized < True But the result was 0.653 0.526 0.752
Additional context