TY - JOUR
T1 - Cross-validation and permutations in MVPA
T2 - validity of permutation strategies and power of cross-validation schemes
AU - Valente, Giancarlo
AU - Castellanos, Agustin Lage
AU - Hausfeld, Lars
AU - De Martino, Federico
AU - Formisano, Elia
N1 - Copyright © 2021. Published by Elsevier Inc.
PY - 2021/9
Y1 - 2021/9
N2 - Multi-Voxel Pattern Analysis (MVPA) is a well established tool to disclose weak, distributed effects in brain activity patterns. The generalization ability is assessed by testing the learning model on new, unseen data. However, when limited data is available, the decoding success is estimated using cross-validation. There is general consensus on assessing statistical significance of cross-validated accuracy with non-parametric permutation tests. In this work we focus on the false positive control of different permutation strategies and on the statistical power of different cross-validation schemes. With simulations, we show that estimating the entire cross-validation error on each permuted dataset is the only statistically valid permutation strategy. Furthermore, using both simulations and real data from the HCP WU-Minn 3T fMRI dataset, we show that, among the different cross-validation schemes, a repeated split-half cross-validation is the most powerful, despite achieving slightly lower classification accuracy, when compared to other schemes. Our findings provide additional insights into the optimization of the experimental design for MVPA, highlighting the benefits of having many short runs.
AB - Multi-Voxel Pattern Analysis (MVPA) is a well established tool to disclose weak, distributed effects in brain activity patterns. The generalization ability is assessed by testing the learning model on new, unseen data. However, when limited data is available, the decoding success is estimated using cross-validation. There is general consensus on assessing statistical significance of cross-validated accuracy with non-parametric permutation tests. In this work we focus on the false positive control of different permutation strategies and on the statistical power of different cross-validation schemes. With simulations, we show that estimating the entire cross-validation error on each permuted dataset is the only statistically valid permutation strategy. Furthermore, using both simulations and real data from the HCP WU-Minn 3T fMRI dataset, we show that, among the different cross-validation schemes, a repeated split-half cross-validation is the most powerful, despite achieving slightly lower classification accuracy, when compared to other schemes. Our findings provide additional insights into the optimization of the experimental design for MVPA, highlighting the benefits of having many short runs.
KW - MVPA
KW - Cross-validation
KW - Permutation test
KW - Statistical validity
KW - Statistical power
KW - CHANCE LEVEL
KW - FMRI
KW - CLASSIFICATION
KW - TESTS
U2 - 10.1016/j.neuroimage.2021.118145
DO - 10.1016/j.neuroimage.2021.118145
M3 - Article
C2 - 33961999
SN - 1053-8119
VL - 238
SP - 1
EP - 38
JO - Neuroimage
JF - Neuroimage
M1 - 118145
ER -