Published on Tue Jun 29 2021

Transfer learning improves outcome predictions for ASD from gene expression in blood

Robasky, K., Kim, R., Yi, H., Xu, H., Bao, B., Chang, A. W. T., Courchesne, E., Lewis, N. E.

The number of variables (genes) is often much larger than the number of observations (human subject tissue samples) We investigated means for improving model performance on the types of under-constrained problems that are typical in human genetics. We showed that a non-pre-trained, naive model performance can be improved by 7.8% by transferring learning from a highly performant model trained on GTEx data.

1
0
0
Abstract

BackgroundPredicting outcomes on human genetic studies is difficult because the number of variables (genes) is often much larger than the number of observations (human subject tissue samples). We investigated means for improving model performance on the types of under-constrained problems that are typical in human genetics, where the number of strongly correlated genes (features) may exceed 10,000, and the number of study participants (observations) may be limited to under 1,000. MethodsWe created train, validate and test datasets from 240 microarray observations from 127 subjects diagnosed with autism spectrum disorder (ASD) and 113 typically developing (TD) subjects. We trained a neural network model (a.k.a., the naive model) on 10,422 genes using the train dataset, composed of 70 ASD and 65 TD subjects, and we restricted the model to one, fully-connected hidden layer to minimize the number of trainable parameters, including a dropout layer to help prevent overfitting. We experimented with alternative network architectures and tuned the hyperparameters using the validate dataset, and performed a single, final evaluation using the holdout test dataset. Next, we trained a neural network model using the identical architecture and identical genes to predict tissue type in GTEx data. We transferred that learning by replacing the top layer of the GTEx model with a layer to predict ASD outcome and we retrained the new layer on the ASD dataset, again using the identical 10,422 genes. FindingsThe naive neural network model had AUROC=0.58 for the task of predicting ASD outcomes, which saw a statistically significant 7.8% improvement from transfer learning. InterpretationWe demonstrated that neural network learning could be transferred from models trained on large RNA-Seq gene expression to a model trained on a small, microarray gene expression dataset with clinical utility for mitigating over-training on small sample sizes. Incidentally, we built a highly accurate classifier of tissue type with which to perform the transfer learning. FundingThis work was supported in part by NIMH R01-MH110558 (E.C., N.E.L.) Author SummaryImage recognition and natural language processing have enjoyed great success in reusing the computational efforts and data sources to overcome the problem of over-training a neural network on a limited dataset. Other domains using deep learning, including genomics and clinical applications, have been slower to benefit from transfer learning. Here we demonstrate data preparation and modeling techniques that allow genomics researchers to take advantage of transfer learning in order to increase the utility of limited clinical datasets. We show that a non-pre-trained, naive model performance can be improved by 7.8% by transferring learning from a highly performant model trained on GTEx data to solve a similar problem.