[x] Be sure to restart your R Session to apply these updates
[x] Get the latest version of the dataset from osf (delete your raw_data, so that the script automatically downloads the data)
[x] Run the import script
[x] Does it run into issues due to missing libraries? during restructuring, import statements for libraries like janitor might have been lost in some datasets - re-add them if necessary
[x] Does the validator complain about the processed data? Complain to Adrian (or fix the listed issues if you feel like it)
Common issues to check:
Trials
[x] Are trials now unique between administrations?
[x] Is exclusion info handled correctly? (look closely at how exclusion information is marked)
Trial Types
[x] Check if the trial type IDs are created independently of administrations/subjects
[x] Is vanilla_trial coded appropriately?
Stimuli
[ ] If the images are on osf, make sure the image path contains the image path on osf
[x] Make sure each row represents a label-image association
[x] the labels should be the words that the participants hear. For example, "apple" is okay, "red_apple_little" is wrong and was probably erroneously extracted from the file name
[x] Are there items in the imported dataset not mentioned in the paper?
[x] Are distractors represented correctly?
Special explanation for distractors: If an item only ever appeared in distractor position, it still gets its own row. The label is typically the label given to the image in the experiment description (e.g., "the distractor was an image of a chair"). If there is no obvious label provided in the experiment design, leave label blank.
Subjects
[x] Does CDI data follow the new aux_data format?
[x] Age rounded correctly? (decision: we do no rounding)
General
[x] Double-check the citation and update it in the dataset table and make sure it’s consistent with the peekbank datasets google sheet: peekbank datasets
[x] Are there any TODOs left in the code - resolve/double check
[x] Make sure any TO-DOs or other decision points in the comments of the code are documented in the ReadMe AND removed from the code to prevent ambiguity
[x] General data sanity-checking (summary output helps here)
[x] is there are the general numbers (e.g. # of participants, # of stimuli, average trials per administratoin) in the summary consistent with the paper? aoi_timepoints are hard to gague, but a super small number is probably bad
[x] is the subject summary (age, sex distribution) approximately consistent with the paper? (note that it is not surprising if it is not identical - often we have a slightly different dataset and are not trying reproduce the exact numbers)
[x] is the target side distribution skewed towards one side?
[x] any weird trial durations?
[x] do the cdi rawscore numbers match the instrument and measure?
[x] is the exclusion % and the exclusion reasons sensible? (bearing in mind that we only have exclusion info for some datasets)
[x] Inspect the timecourse and accuracy plots/output at the end of the import:
[x] Compare timecourse patterns with paper (as best as possible)
[x] Does the timing seem right? (accuracy spike later than PoD might be sensible, earlier is suspicious)
[x] (if multiple conditions) Does the number conditions make sense in the context of the paper?
[x] (if multiple conditions) Are the overall accuracies for conditions vastly different in a way not explained by the paper?
[x] Any odd item-level patterns?
[x] Any odd subject-level patterns?
[x] Any large (unexpected) discrepancies between data reported in paper vs. data in the imported dataset?
[x] After checking everything and rerunning the script: Upload the output to osf
Checklist for code review v2024
To start:
Common issues to check:
Trials
Trial Types
Stimuli
Subjects
General