Back to all papers

Deep convolutional neural networks outperform vanilla machine learning when predicting language outcomes after stroke.

Authors

Hope TMH,Bowman H,Leff AP,Price CJ

Affiliations (3)

  • Department of Imaging Neuroscience, Institute of Neurology, University College London, 12 Queen Square, London WC1N 3AR, the United Kingdom of Great Britain and Northern Ireland; Department of Psychological and Social Sciences, John Cabot University, Via della Lungara 233, 00165, Rome, Italy. Electronic address: [email protected].
  • School of Psychology University of Birmingham Edgbaston Birmingham B15 2TT the United Kingdom of Great Britain and Northern Ireland.
  • Department of Imaging Neuroscience, Institute of Neurology, University College London, 12 Queen Square, London WC1N 3AR, the United Kingdom of Great Britain and Northern Ireland.

Abstract

Current medicine cannot confidently predict patients' language skills after stroke. In recent years, researchers have sought to bridge this gap with machine learning. These models appear to benefit from access to features describing where and how much brain damage these patients have suffered. Given the very high dimensionality of structural brain imaging data, those brain lesion features are typically post-processed from the images themselves into tabular features. With the introduction of deep Convolutional Neural Networks (CNN), which appear to be much more robust to high dimensional data, it is natural to hope that much of this image post-processing might be unnecessary. But prior attempts to demonstrate this (in the area of post-stroke prognostics) have so far yielded only equivocal results - perhaps because the datasets that those studies could deploy were too small to properly constrain CNNs, which are famously 'data-hungry'. The study draws on a much larger dataset than has been employed in previous work like this, referring to patients whose language outcomes were assessed once during the chronic phase post-stroke, on or around the same days as they underwent high resolution MRI brain scans. Following the model of our own and others' past work, we use state of the art 'vanilla' machine learning models (boosted ensembles) to predict a variety of language and cognitive outcomes scores. These models employ both demographic variables and features derived from the brain imaging data, which represent where brain damage has occurred. These are our baseline models. Next, we use deep CNNs to predict the same language scores for the same patients, drawing on both the demographic variables, and post-processed brain lesion images: i.e., multi-input models with one input for tabular features and another for 3-dimensional images. We compare the models using 5 × 2-fold cross-validation, with consistent folds. The CNN models consistently outperform the vanilla machine learning models, in this domain. Deep CNNs offer state of the art performance when predicting language outcomes after stroke, outperforming vanilla machine learning and obviating the need to post-process lesion images into lesion features.

Topics

Journal Article

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.