Open amritbhanu opened 7 years ago
so if we use naive bayes, pdf never gets worse and other measures improve, a little
and if we other learners the get LARGE reductions in pf and recall?
Not all learners for large reductions in pf and recall. We only see that in Dec trees and rbf svm.
Prof. here we can use tuning to find the cliff percentage, alpha, beta, about how much to share and get the best performance.
Amritanshu Agrawal PhD Student CS @ NC State aagrawa8@ncsu.edu http://amritag.wixsite.com/amrit
On Thu, May 4, 2017 at 5:51 PM, Tim Menzies notifications@github.com wrote:
so if we use naive bayes, pdf never gets worse and other measures improve, a little
and if we other learners the get LARGE reductions in pf and recall?
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/ai-se/LAS-Phishing/issues/18#issuecomment-299319461, or mute the thread https://github.com/notifications/unsubscribe-auth/ADTGCEp0JiATcnyTWM2aDq_P4m860FbCks5r2khhgaJpZM4NRPkz .
need a monthly report
Baseline:
Experiment:
Baseline:
Results:
AOL
Apple
Facebook
Google
Paypal
Yahoo