Comparing within-subject classification and regularization methods in fMRI for large and small sample sizes

In recent years, a variety of multivariate classifier models have been applied to fMRI, with different modeling assumptions. When classifying high‐dimensional fMRI data, we must also regularize to improve model stability, and the interactions between classifier and regularization techniques are stil...

Full description

Saved in:
Bibliographic Details
Published inHuman brain mapping Vol. 35; no. 9; pp. 4499 - 4517
Main Authors Churchill, Nathan W., Yourganov, Grigori, Strother, Stephen C.
Format Journal Article
LanguageEnglish
Published New York, NY Blackwell Publishing Ltd 01.09.2014
Wiley-Liss
John Wiley & Sons, Inc
John Wiley and Sons Inc
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:In recent years, a variety of multivariate classifier models have been applied to fMRI, with different modeling assumptions. When classifying high‐dimensional fMRI data, we must also regularize to improve model stability, and the interactions between classifier and regularization techniques are still being investigated. Classifiers are usually compared on large, multisubject fMRI datasets. However, it is unclear how classifier/regularizer models perform for within‐subject analyses, as a function of signal strength and sample size. We compare four standard classifiers: Linear and Quadratic Discriminants, Logistic Regression and Support Vector Machines. Classification was performed on data in the linear kernel (covariance) feature space, and classifiers are tuned with four commonly‐used regularizers: Principal Component and Independent Component Analysis, and penalization of kernel features using L1 and L2 norms. We evaluated prediction accuracy (P) and spatial reproducibility (R) of all classifier/regularizer combinations on single‐subject analyses, over a range of three different block task contrasts and sample sizes for a BOLD fMRI experiment. We show that the classifier model has a small impact on signal detection, compared to the choice of regularizer. PCA maximizes reproducibility and global SNR, whereas Lp‐norms tend to maximize prediction. ICA produces low reproducibility, and prediction accuracy is classifier‐dependent. However, trade‐offs in (P,R) depend partly on the optimization criterion, and PCA‐based models are able to explore the widest range of (P,R) values. These trends are consistent across task contrasts and data sizes (training samples range from 6 to 96 scans). In addition, the trends in classifier performance are consistent for ROI‐based classifier analyses. Hum Brain Mapp 35:4499–4517, 2014. © 2014 Wiley Periodicals, Inc.
Bibliography:istex:44E5DE7EEFFCFD2829B6BCE1D374AAC44FCE8708
Brain, Mind and Behaviour Grant from the James S. McDonnell Foundation and Heart and Stroke Foundation of Ontario, through the Centre for Stroke Recovery
Bridging CIHR - No. IAO123872
CIHR - No. MOP84483
ark:/67375/WNG-V5FFV0Q2-9
ArticleID:HBM22490
ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ObjectType-Article-2
ObjectType-Feature-1
content type line 23
ISSN:1065-9471
1097-0193
1097-0193
DOI:10.1002/hbm.22490