ai-se / ResourcesDataDrivenSBSE

Other
4 stars 0 forks source link

what are the standard papers discussing methodologies for sbse? #11

Closed timm closed 6 years ago

timm commented 6 years ago

here are two. how many others?

minkull commented 6 years ago

I believe that Arcuri and Briand's is the most popular paper in the SBSE community.

For ML (not SBSE), we have the following:

And then more specific SA ones:

timm commented 6 years ago

fyi: personal bias

e.g. see the median and IQR in the following charts? and how scott knott divided them into 3 sane divisions? me likey.

image

and here's a more elaborate example. all those numbers, and there is only 6 divisions. sane

image

timm commented 6 years ago

whish isn't to say we dont list demsar. just note that many folks in MSR used Demsar for a few years then moved to SK

minkull commented 6 years ago

Hi Tim,

One of my concerns with Scott-Knott is that it's parametric. So, it is kind of against some of the things that the SE community is normally keen on, such as the use of median instead of mean. Did you ever get any trouble with reviewers when using this test? Or, maybe the reviewers don't know the test very well yet, and then don't complain about this point?

The main reason for me to normally use Friedman instead of Scott-Knott is that Friedman is non-parametric. In terms of simplicity to read, it is actually possible to generate some nice plots for Friedman's post hoc tests, showing which approaches perform similar to or different from the top ranked approach.

Best, Leandro

-- Dr. Leandro L. Minku Lecturer (Assistant Professor) in Computer Science Department of Informatics University of Leicester, UK

On 19 Jan 2018, at 17:22, Tim Menzies notifications@github.com<mailto:notifications@github.com> wrote:

fyi: personal bias

e.g. see the median and IQR in the following charts? and how scott knott divided them into 3 sane divisions? me likey.

[image]https://user-images.githubusercontent.com/29195/35162908-e69bea94-fd12-11e7-9ddd-8b3d67c3bac9.png

and here's a more elaborate example. all those numbers, and there is only 6 divisions. sane

[image]https://user-images.githubusercontent.com/29195/35162959-1ce6b17e-fd13-11e7-9727-6c2c52e11744.png

— You are receiving this because you were assigned. Reply to this email directly, view it on GitHubhttps://github.com/ai-se/resourcesDataDriveSSBSE/issues/11#issuecomment-359033183, or mute the threadhttps://github.com/notifications/unsubscribe-auth/AFTL7V5CT88hzO9W5c6pdulKdW7-xyfAks5tMM9PgaJpZM4Rj6o6.

timm commented 6 years ago

mother told me never to argue about stats. maybe we should focus on what we do agree on (evolutionary methods)

pause

oh what the hell

scott-knott is not parametric. it is a clustering method and the operators used to assess the merits of combining clusters are a domain decision

now, usually, it is used with an ANOVA operator but that is a design choice. i use non-parametric boostrapping and cliffs delta for the top-down version

for a bottom's up version, as near as i can tell, the canada people use cliffs delta (again, non-parametric) to see if cluster i,j can be replaced by k. if yes, then repeat for k,l. else, try again with j,k.

as to kinky friedmann (http://www.kinkyfriedman.com/, just kidding), those plots show so much overlap that i dont know what is going on. and those rank values hide the true eval numbers... never a good thing in my book

as to ease of reading, i dig the plot shown above. or one of wei's charts:

image

markuswagnergithub commented 6 years ago

Slightly newer than the Arcuri 2011 is the Arcuri 2013 ("only" 170 citations instead of 430), which is the official extended journal version of the ICSE paper (see footnote on first page) http://orbilu.uni.lu/handle/10993/1071