Closed behrica closed 3 years ago
A complete different way to "select" a contribution for you, would be that my colleges start by selecting manually the "projects of interest / relevance / representative", and making sure manually that they indeed contain a "level" which was an abstract screening. (and probably look at the concrete questions asked in the SR and decide based on ......) to share or not to share with you
This would very likely result in a far smaller number, maybe 10 (compared to > 126 of my "take all" approach I described before)
Maybe even both contributions can be usefull....
Please provide me with your comments.
In the "take all" scenario we would talk about at least: ~ 150000 references of which ~ 130000 where excluded.
I started to have a look at our full SR database.
Maybe I start to describe what I have here, and then we can iteratively think about what would be worth to include. Maybe we can have as well an other call.
We have in total 299 "projects" in Distiller. Quite some of them are "tests" or other garbage. Hard to say how many, but by looking at the project names, there might be at least 100 project which should not be looked at at all. so 200 projects remaining.
Each of it has at least "one level", in which a level could mean different things:
"title screening" "abstract screening" "title + abstract screening" "full text screening" "data extraction" "abstract screening 1" vs "abstract screening 2". ...... .....
There is no "clear nomenclature" or metadata on this, but often we use the word "abstract" in the name of the level to indicate "abstract screening"
The number of "levels" in total (including garbage projects) is: 1226
So in total we have 1226 times , that "humans have decided to exclude x papers out of y"
(sometimes x or y or x and y are 0)
I filtered the levels by the ones which have "abstract" in the "level name". These SHOULD be all about abstract screening, but we might have more.
This leaves then 126 rows.
I just pasted here for you information, some of the "statistics" I get for these.
We can see that the first row is:
related to an EFSA question: EFSA-Q-2012-00234 and was about "Leishmaniosis" . (from this you could find the EFSA output: https://efsa.onlinelibrary.wiley.com/doi/epdf/10.2903/sp.efsa.2014.EN-466 where on page 19 you find a summary of the SR. Sometimes we publish as well the concrete references included / excluded, but not always
we look at a level/phase of the systematic review called "Title and abstract screening - Study eligibility form: Title and abstract screening" (so I think, this is a "real" abstract screening, probably worth to add to your database or to be used in a simulation)
it started with 961 references
we excluded 877 and included 84
So one contribution to you could be the (at least 126) abstract screenings from our database, including its meta-data:
Some might be "half done", but that you could see from the numbers of "total papers", "included", "excluded" , "conflict". I would say "nearly all" are complete.
I have automated all extractions, so the "volume of SRs" does not make any difference for me.