Open mih opened 5 years ago
@cmaumet @jbpoline which of the metadata-related bits (last section) should appear on the poster, or rather, how to tweak them to become relevant?
@dkeator NIDM-E would be next on my TODO list, in order to bring proper metadata on the analysis inputs into the picture ... likely post-OHBM
Target: Do a full GLM analysis with all the great machinery in the cmdline. Preferably without intermediate scriting.
Bottom line: All actually tricky parts of the workflow work well, but quite a few minor issue make it needlessly hard to succeed within a reasonable time frame. Here is the protocol with comment and links to issues:
study raw data
Fresh dataset
Import DICOM (will create subdatasets inside)
Add non-DICOM study data
Configure a "converter" for the stimulation protocol, to be used for BIDSification
BIDSification
BIDS-compliant dataset is a fresh dataset
that has the study raw dataset linked
Convert everything to BIDS by selection all desired studyspecs.
analysis
Any analysis is also a fresh dataset
that has the BIDSified data linked as input
Build an FSL analysis container based on a recipe (note: takes a fraction of the download time from singularity hub). Analysis in a container == better chances to reproduce.
Get the container build specifications.
Build the actual container (include a permission fix to be able to get the image file tracked in the dataset).
Add the remaining pieces to be able to compute a GLM analysis (a script to convert the stimulation log into FSL EV3 files, and a template analysis configuration).
Convert the stimulus log into EV3 files and build an actual analysis configuration for
sub-001
from the template.Compute the GLM analysis
Inform the dataset that it carries FSL FEAT, and
run
provenance metadata. And pull it out.metadata
Here are a few pointers re where we stand in terms of metadata availability at the analysis end. Here is a dump for
thresh_zstat1.nii.gz
, i.e. this metadata is immediately bound to the path, and accessible without query.(in DataLad's internal data structure)
It know that this file was generated by
14bd68c839c5254764044d88893abef4d9e5e716
, and this is what we know about it:Ah this guy:
What else did he damage:
Needless to say that this is not how query should work in the end, but the information is in the beast, and it is consolidated across three different metadata sources (datalad-core, PROV on datalad's own command capture, and NIDM-R). By enriching the information in each one of them alone, one could achieve substantial information retrieval capabilties.