Closed ThebTron closed 2 years ago
Merging #81 (8aac10f) into main (3e2f9e4) will increase coverage by
0.00%
. The diff coverage is100.00%
.
@@ Coverage Diff @@
## main #81 +/- ##
=======================================
Coverage 94.18% 94.19%
=======================================
Files 24 24
Lines 1239 1240 +1
Branches 192 192
=======================================
+ Hits 1167 1168 +1
Misses 42 42
Partials 30 30
Impacted Files | Coverage Δ | |
---|---|---|
src/omlt/gbt/gbt_formulation.py | 93.65% <100.00%> (+0.05%) |
:arrow_up: |
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact)
,ø = not affected
,? = missing data
Powered by Codecov. Last update 3e2f9e4...8aac10f. Read the comment docs.
Legal Acknowledgement By contributing to this software project, I agree my contributions are submitted under the BSD license. I represent I am authorized to make the contributions and grant the license. If my employer has rights to intellectual property that includes these contributions, I represent that I have received permission to make contributions and grant the required license on behalf of that employer.
This PR addresses #80. GBT models imported from scikit-learn have an additional constant base score that is added to the prediction of the model. This doesn't apply for lightgbm and the base score defaults to zero. I also added a note to the jupyter notebook that only lightgbm and scikit-learn tree model inputs are officially supported.