Centre-IRM-INT / GT-MVPA-nilearn

GT MVPA nilearn from Marseille
3 stars 3 forks source link

Your take on demeaning & normalizing data before MVPA analyses #22

Open JeanneCaronGuyon opened 2 years ago

JeanneCaronGuyon commented 2 years ago

Hi all

I've seen in some papers, and in the lab I'm in now (Collignon's), authors "demeaning" the data prior to running MVPA analyses, to basically take out the univariate differences (more activation in condition A than B) and only consider patterns differences. For instance Rezk et al., 2020 : "Before running the decoding analysis in each ROI, the patterns of each block and each condition (motion and static) was demeaned individually to minimize the univariate activation level differences and to mathematically equate the mean activity of each condition and each block."

To me, that makes sense especially for cross modal decoding as in areas previously defined as "unisensory" (eg. for modality A) we will probably get way less activation for modality B, and that could influence the decoding (see for instance Smith et al. 2011). However some have warned about the potential negative effect of these methods on the interpretation of the data (I'm thinking of Ramirez's paper for instance - who talks more about RSA though).

Then, there's also the z-scoring of the patterns, again for instance Rezk et al., 2020 : "In each cross-validation fold, the training data (5 blocks per condition for the visual motion localizer, 12 blocks per condition for the auditory localizer) were normalized (z-scored) across conditions and the classifier was trained to discriminate the motion and static conditions."

We (Anne, Caro, Jean-Luc and I) have been looking at procedures to perform MVPA analyses in a more relevant way, especially as we use two different sensory modalities (either Auditory & Tactile or Visual & Tactile) and also do cross modal decoding, for which such methods could be actually pertinent.

Have you been using such methods and in which context ? Or theoretically, what's your take on their use and their impact on subsequent interpretation of the decoding performances ?

Thanks !! Jeanne

JeanneCaronGuyon commented 2 years ago

A little follow up on that, I've been even told that we should always normalize our data across conditions during the training phase... (and the extra step of demeaning the patterns would then just be used for cross-modal decoding) I see how that can help, but I really don't know if it's the classic/"right" (fair?) way to go... I'm super interested in having your input, if you have any opinion on that !! Maybe @SylvainTakerkart ? Thanks :)

SylvainTakerkart commented 2 years ago

Hi,

Here is y take on all this:

Well, this was a true "réponse de normand"! My conclusion is: go for what you believe in from a neuroscience point of view, at least you'll be able to defend this in front of a reviewer! As long as you don't touch the test set in your training phase, you're good from a methodological point of view!

JeanneCaronGuyon commented 2 years ago

Thanks Sylvain, I definitely see what you mean... I do believe that going with what makes sense in a neuroscience point of view feels best, and that will most likely help settle on analyses we trust, we'll see what we go for but thank you for your input on these matters !