A deep learning model for prediction of autism status using whole-exome sequencing data | PLOS Computational Biology
Skip to main content
Advertisement
  • Loading metrics

A deep learning model for prediction of autism status using whole-exome sequencing data

  • Qing Wu,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliations Department of Molecular Biology, Cell Biology and Biochemistry, Brown University, Providence, Rhode Island, United States of America, Center for Translational Neuroscience, Robert J. and Nancy D. Carney Institute for Brain Science and Brown Institute for Translational Science, Brown University, Providence, Rhode Island, United States of America, Center for Computational Molecular Biology, Brown University, Providence, Rhode Island, United States of America

  • Eric M. Morrow,

    Roles Conceptualization, Funding acquisition, Supervision, Writing – original draft, Writing – review & editing

    Affiliations Department of Molecular Biology, Cell Biology and Biochemistry, Brown University, Providence, Rhode Island, United States of America, Center for Translational Neuroscience, Robert J. and Nancy D. Carney Institute for Brain Science and Brown Institute for Translational Science, Brown University, Providence, Rhode Island, United States of America, Center for Computational Molecular Biology, Brown University, Providence, Rhode Island, United States of America, Developmental Disorders Genetics Research Program, Department of Psychiatry and Human Behavior, Emma Pendleton Bradley Hospital, East Providence, Rhode Island, United States of America

  • Ece D. Gamsiz Uzun

    Roles Conceptualization, Investigation, Supervision, Writing – original draft, Writing – review & editing

    dilber_gamsiz@brown.edu

    Affiliations Center for Translational Neuroscience, Robert J. and Nancy D. Carney Institute for Brain Science and Brown Institute for Translational Science, Brown University, Providence, Rhode Island, United States of America, Center for Computational Molecular Biology, Brown University, Providence, Rhode Island, United States of America, Department of Pathology and Laboratory Medicine, Warren Alpert Medical School of Brown University, Providence, Rhode Island, United States of America, Department of Pathology and Laboratory Medicine, Rhode Island Hospital, Providence, Rhode Island, United States of America

Abstract

Autism is a developmental disability. Research demonstrated that children with autism benefit from early diagnosis and early intervention. Genetic factors are considered major contributors to the development of autism. Machine learning (ML), including deep learning (DL), has been evaluated in phenotype prediction, but this method has been limited in its application to autism. We developed a DL model, the Separate Translated Autism Research Neural Network (STAR-NN) model to predict autism status. The model was trained and tested using whole exome sequencing data from 43,203 individuals (16,809 individuals with autism and 26,394 non-autistic controls). Polygenic scores from common variants and the aggregated count of rare variants on genes were used as input. In STAR-NN, protein truncating variants, possibly damaging missense variants and mild effect missense variants on the same gene were separated at the input level and merged to one gene node. In this way, rare variants with different level of pathogenic effects were treated separately. We further validated the performance of STAR-NN using an independent dataset, including 13,827 individuals with autism and 14,052 non-autistic controls. STAR-NN achieved a modest ROC-AUC of 0.7319 on the testing dataset and 0.7302 on the independent dataset. STAR-NN outperformed other traditional ML models. Gene Ontology analysis on the selected gene features showed an enrichment for potentially informative pathways including calcium ion transport.

Author summary

Autism is a developmental disability. Genetic factors are considered as major contributors to the development of autism. Here, we present a deep neural network model, Separate Translated Autism Research Neural Network (STAR-NN), to predict autism status using WES data. STAR-NN showed a modest performance and was validated using an independent dataset. However, ML based models may have advantages in autism status prediction using genomic data and should be further studied.

Introduction

Autism is a developmental disability that affects social and communication skills. In recent years, autism has been identified among 1% of children in the world population [1]. Studies have found that early diagnosis of autism was significantly associated with improved social characteristics [2]. To date, there is evidence supporting the importance of early diagnosis, followed by early intervention for having long-term positive impact [36]. While autism is highly heterogeneous, genetic variations are major contributors to the development of autism [711]. Rare genetic variants have garnered substantial focus [10,12]. Hundreds of loci and genes associated with autism have been identified in recent years with the advancement of next generation sequencing (NGS) [7,8,10,1215]. Clinical genetic testing, including whole exome sequencing (WES), to identify rare variants is recommended [58].

Machine learning (ML) methods, including advanced deep learning (DL) models, offer the advantage of capturing subtle patterns among input features, making them efficient tools for phenotype prediction and feature weight inference [16]. Liu et al. developed a DL model to identify structural variations associated with several brain conditions, including ADHD, depression, and autism using whole genome sequencing (WGS) datasets as input [17]. DL methods have been used in cancer research for various purposes, ranging from predicting disease status (primary or metastatic) to prognosis prediction [1822]. Both ML models and a limited number of DL models have been tested for autism prediction and the identification of autism subgroups [2328]. As the diagnosis of autism is based on the observation of phenotypic presentations, DL models that used medically observed features, such as those from electronic health records (EHR), achieved a good prediction rate. Onishchenko et al. trained a DL model using EHR data from 30 million children under 6-years old to assess the comorbidity of autism [23]. Lin et al. trained a random forest and support vector machine classifier on gene expression profile from 31 children with autism to identify subgroups [25]. Using WES data from 598 schizophrenia families and 2392 autism families, researchers used the Extreme Gradient Boosting (XGBoost) model to classify individuals with autism and individuals with schizophrenia achieving a prediction accuracy above 80% [26]. However, these ML studies which used genetic data as an input lack an independent dataset to demonstrate their predictive ability and to further test the applicability of their method beyond the scope of training and testing data. Previous studies also found that small datasets might exhibit ascertainment bias in data collection. Successful biomedical ML models trained and tested on small datasets often failed on large heterogeneous datasets [29,30].

In this study, we developed a DL model to predict autism status using WES and microarray data. The model was trained and tested using a WES dataset, which includes 43,203 individuals (16,809 individuals with autism and 26,394 non-autistic controls). We further validated the model performance using an extra independent dataset from 27,879 individuals (13,827 individuals with autism and 14,052 non-autistic controls). Our model took into account the effects of rare and common variants [10,12,14]. A total of 1489 features, including a binary gender status, polygenic score (PGS) and 1487 genes, were pre-selected as the key features for our model.

Results

Autism status prediction model with selected features

We developed a DL model to predict autism status based on the individuals’ genetic data (Fig 1). We named our model “Separate Translated Autism Research–Neural Network (STAR-NN)” as it separates variants based on their functional effects on each gene. The impact of both common and rare variants was considered in our model. Rare protein truncating variants (PTVs) include nonsense, frameshift and canonical splicing variants. Missense variants were annotated by Missense Badness, Polyphen-2, Constraint (MPC) score [31] and separated into three groups: MisA, MisB and MisC (Methods). A higher MPC score indicates a higher likelihood of damaging effect of a missense variant. Missense variants with MPC score above 1, predicted as possibly damaging missense variants, were grouped into MisA and MisB forming MisAB. MisC represents the possibly benign missense variants. Due to the specificity of model structure, STAR-NN learned the impact of different types of variants in the same gene separately. Gender, PGS calculated from significantly associated common variants (MAF > 1%), together with three types of rare variants, PTVs, MisAB, and MisC, were used as input to predict a binary autism status: individuals with autism or non-autistic controls. We trained and tested the model using the combined WES1 and WES2 datasets (WES12) from SPARK dataset (16,809 individuals with autism and 26,394 non-autistic controls). 80% of the samples were used for training, 10% for validation and 10% for testing.

thumbnail
Fig 1. The workflow and framework of STAR-NN.

After quality control, rare variants (minor allele frequency, MAF < 1%) identified from whole exome sequencing data were separated into four categories based on their function effect: protein truncating variants (PTVs), MisA (Missense variants with MPC > 2), MisB (Missense variants with 1 < MPC < 2) and MisC (Missense variants with 0 < MPC < 1). MisA and MisB were then combined as MisAB. Three types of rare exonic variants were used as input for STAR-NN model. In addition, polygenic score (PGS) generated from common variants (MAF > 1%) from microarray data were also used as input for STAR-NN. STAR-NN uses a three-to-one mapping strategy to learn different types of variants on the same gene separately. G represents gene node, S with grey color represents the option to add gene sets node before final output (shaded circle). *, Quality control on WES1 and WES2 used the same standards, further details provided in Materials and Methods. #, numbers in brackets showing (the count of variants, in the count of individuals) in the dataset.

https://doi.org/10.1371/journal.pcbi.1012468.g001

While selecting features is not a necessary step in DL models, the DL model might not efficiently learn the importance of certain rare features due to the sparsity of the rare variants. For that reason, we used an automated ML model, Tree-based Pipeline Optimization Tool (TPOT) [32], to select the best features for the model (Methods, S1 Fig). A final list of 1489 selected features, including 1487 genes, PGS and gender were used. Of the 1487 pre-selected genes (S1 Table), 115 genes have previously been found in “SFARI Gene”, a database of genes implicated in autism susceptibility [33] (hypergeometric test, p-value = 3.057e-5. SFARI genes list released on 1/11/2022).

Our model outperformed traditional ML models, including decision trees, random forest, XGBoost, L1 logistic regression, L2 logistic regression, linear support vector and a basic DNN model that does not separate the rare variants by their functional effect, further supporting the importance of separation of variants based on their functional effect. Receiver Operating Characteristic Area Under the Curve (ROC-AUC) for STAR-NN was 0.7317 (Fig 2A and S2 Table). Our model demonstrated a faster training process than logistic regression with L2 regularization (L2LR) which exhibited the highest performance among the traditional ML models. STAR-NN took an average of 179.4 seconds for each training process whereas L2LR took an average of 690 seconds for each training (S3 Table). We used 1 CPU, 32 core and 256G memory for the training of each model on high performance computing cluster, Oscar maintained and supported by Center for Computation and Visualization at Brown University.

thumbnail
Fig 2. Performance of STAR-NN.

A. ROC-AUC plot, showing STAR-NN outperformed six traditional machine learning model and a basic deep neural network (DNN) model. Variants of different type was not separated in traditional machine learning model and the basic DNN. B. ROC-AUC plot, showing STAR-NN with selected gene features outperformed the model using other gene sets as input. C. The density plot of PGS for individuals with autism and non-autistic controls. D. The distribution plot of score generated from STAR-NN for individuals with autism and non-autistic controls.

https://doi.org/10.1371/journal.pcbi.1012468.g002

To test the predictive performance of selected features, we compared the model performance using 4 different groups of gene features as input, including 1487 selected gene features, 1031 SFARI genes, a combination of 2405 selected features and SFARI genes and the 19117 full gene set. We found that our model, using selected features has the highest performance (ROC-AUC = 0.7317) (Fig 2B and S4 Table).

We generated PGS using common variants for each individual (Fig 2C). Compared to the PGS, which exhibits a small difference between individuals with autism and non-autistic controls (mean PGS: 0.00156 for individuals with autism and -0.00152 for non-autistic controls), our model significantly separated the two groups (mean scores: 0.5784 for individuals with autism and 0.4271 non-autistic controls, respectively; Mann-Whitney-Wilcoxon Test, p<2.2e-16, Fig 2D). While a significant difference was observed between PGS of males with autism and non-autistic controls, PGS of females with autism and non-autistic controls shows no significant difference (Fig 3A and 3B). We also tested the STAR-NN model without PGS as an input and ROC-AUC was 0.7281 with gender, PTVs, MisAB and MisC as input (S5 Table). Compared to PGS, the score generated from STAR-NN has a significant distinction between individuals with autism and non-autistic controls for both males and females (Fig 3C and 3D).

thumbnail
Fig 3. Score from STAR-NN in male and female population.

The density plot of PGS for individuals with autism and non-autistic controls in females (A) and males(B). The density plot of autism score generated from STAR-NN in females (C) and males (D). The dashed line shows the mean value for each distribution.

https://doi.org/10.1371/journal.pcbi.1012468.g003

We tested the individual effect of PTV, MisAB and MisC on the prediction of autism status. Basic DNN model using PGS, biological sex and aggregated count of PTVs per gene as input resulted in an ROC-AUC of 0.7080. The ROC-AUC of 0.7015 was generated using PGS, biological sex and aggregated count of MisAB per gene as input. The basic DNN model using PGS, biological sex and aggregated count of MisC per gene as input had a lower performance, (ROC-AUC of 0.6982) compared with using aggregated count of PTVs per gene or using aggregated count of MisAB per gene as input (S6 Table). We also tested STAR-NN with a 2-to-1 mapping structure to assess the necessity of including MisC as input. We used gender, PGS, aggregated counts of PTVs and MisAB as input and obtained an ROC-AUC of 0.7157 for the 2-to-1 STAR-NN model (S7 Table). Meanwhile, we tested STAR-NN model performance of a 4-to-1 mapping structure by separating PTVs, MisAB, MisC and synonymous variants on the same gene at the input level and merged into one. We found that including synonymous variants slightly decreased model performance (S8 Table). Each ROC-AUC value mentioned above were based on 10 random repeats. The results showed that PTV, MisAB and MisC are all contributing to the prediction of autism status. STAR-NN, incorporating the combined effect from PTVs, MisAB and MisC, had a slightly better performance than basic DNN model with individual effects of PTV, MisAB and MisC as input. This suggests the importance of 3-to-1 mapping structure of STAR-NN to the prediction of autism status.

Validation with an independent dataset

To evaluate the predictive ability of STAR-NN model, we used an extra independent dataset which was also generated by SPARK. 27,879 participants were included in the most recent release of iWES1_v1 from SPARK (13,827 individuals with autism and 14,052 non-autistic controls). STAR-NN achieved similar performance with selected features and SFARI genes (ROC-AUC = 0.7302 for selected features, and ROC-AUC = 0.7319 for SFARI genes, S9 Table). The model performance using the independent dataset was similar to that using the original dataset.

Input features additively contributed to the autism status prediction

In STAR-NN, we used sigmoid function in the output layer for autism status prediction. Sigmoid ensures a non-linear relationship between input features and the prediction results. Subsequently, we tested if there is a non-additive effect across input features. Deep neural network (DNN) model with multiple hidden layers captures the higher interactions across input features [20,34]. Therefore, we compared the performance of STAR-NN with no hidden layer, which models an additive contribution of gene features, to STAR-NN with one hidden layer, which models a higher interaction of biological pathways across input features. As an increase in model complexity might cause overfitting problems and thereby reduce model performance, we also tested the STAR-NN with one hidden layer using “relu” and “linear” as the activation function, separately. Both showed similar performance in prediction accuracy, precision and recall rate (ROC-AUC = 0.7285 for “relu” as activation function, ROC-AUC = 0.7284 for “linear” activation function) (S10 Table) [35]. Therefore, we kept the structure of STAR-NN with no hidden layer between gene layer and the output layer, indicating the gene features are additively contributed to the autism status prediction.

Genes related to developmental disorders were enriched in selected genes

To investigate the function of selected genes, we performed a disease ontology pathway enrichment analysis. We found that the selected genes are enriched with genes related to developmental disorders, intellectual disability and autism (S2 Fig). Then, we performed a gene ontology (GO) term analysis to study the biological function of selected genes. GO terms associated with stimulus, including “detection to stimulus” and “response to stimulus” were significantly enriched. Synapse and dendrite related pathways, including “postsynapse”, “dendrite” and “glutamatergic synapse” were significantly enriched. “Ion binding” and “ion transport” related GO terms were also significantly enriched (S12 Table). The counts of variants in selected features are provided in supplementary tables (S11 Table). Previous findings showed that pathways related to ion channel, including calcium ion channel, potassium ion channel and sodium ion channel are significantly associated with autism [3638].

Discussion

Autism is a developmental disability with variability in phenotypic features and genetic characteristics. Autism is highly heritable, from 64% to 91% [39]. Rare genetic variants have been identified as major contributors [10,12]. In a clinical setting, patients are advised to undergo genetic testing to evaluate the presence of rare variants [4043]. With the advances in NGS technologies and bioinformatics tools, we need better tools to interpret complex genetic test results. Applications of ML methods have been evaluated in precision medicine research. SVMs and RF were the top two most commonly used ML algorithms in studies on phenotype prediction and factor identification [25,30,4447]. Compared to traditional ML methods, DL models as an advanced part of ML with more flexible model construction to capture non-linear and additive effect across input features have been studied in clinical phenotype and subtype classification, such as primary and metastatic cancer classification [1921]. To date, limited studies have used either traditional ML or DL models in autism status prediction [23,25,26,28]. The constraints for usage of genetic data for autism status prediction include 1) lack of large datasets to support the effectiveness of developed model in a heterogeneous population, and 2) lack of an external dataset to prove the predictive ability of developed models. Small datasets might have the potential risk of ascertainment bias, which increases the uncertainty of model performance in a heterogeneous setting [30]. An external dataset is needed to test the reproducibility of the model results outside the scope of training and testing dataset.

Recent studies showed that variants with different level of damaging effects on the same gene contribute differently to the development of autism [7,8,48]. PTVs and possibly damaging missense variants (MisAB) were found to be significantly enriched in individuals with autism and individuals with developmental delay (DD), while mild effect missense variants (MisC) were found to be enriched in DDs only [7,48,49]. In STAR-NN, by constructing a 3-to-1 sparse connection between variants layer and gene layer, the model learned the effect of different level of damaging variants on the same gene separately. Besides, STAR-NN also considered PGS which generated from common variants [8]. In this way, STAR-NN predicted autism status with a modest ROC-AUC of 0.73 and a F1 score of 0.69.

With the expectation to improve the prediction accuracy by capturing higher interaction, the non-additive effect, across gene features, we also tested the model performance by adding hidden layers between the gene layer and the output layer. However, the best performance was still obtained from the model with single gene layer, suggesting the pre-selected gene features, PGS and gender worked in a linear additive way to the prediction of autism status with current sample size. This finding is aligned with the results in a recent study [49], in which the higher interaction across sex, PGS and rare variants score were tested using pairwise linear model and only found the additive effect across input features. As the performance of deep neural network is affected by sample sizes, more samples are needed to test the existence of higher interactions across input features.

In STAR-NN, we used aggregated count of variants on genes as input, considering the impact of variants on a gene level to the prediction of autism status. A similar inference method was also used in P-NET and BANNs, both of which are biologically interpretated neural network models [20,50]. In an alternative approach, Jiao et al. used aggregated variants detected by WGS per 1M bp bin as input for their DL model to classify primary and metastatic cancer [21]. In our study, we selected 1487 gene features. SHANK3 is a well-known autism gene and in SFARI database. GOPC gene was previously found among the top significantly differentially expressed genes in a Parkinson’s disease iPSC model and was identified as a potential target in the treatment of Parkinson [51,52]. Despite autism is a neurodevelopmental disorder, studies also found some parkinsonism features, such as motor issues, in older adults with autism [53,54]. TYW3 gene has been previously found associated with amyotrophic lateral sclerosis (ALS) and tic spectrum disorder (TSD) [55,56].

Copy number variants (CNVs) are genetic factors important for the etiology of autism. To date, 71 autism loci have been identified from CNV and WES data [12]. Gene expression profiles generated from a small dataset with autism patients have previously been used to predict subgroups in which RF and SVM were applied [25]. Besides mutation data, CNVs, gene expression and DNA methylation should be considered in ML models, as previous study suggested a connection between impaired methylation and the etiology of autism [57,58]. To further improve the prediction accuracy in the future, these data types should be included when available. Larger sample size is also needed to better detect the higher interactions across genetic features. Another limitation in STAR-NN is the population used for training and testing the model. Around 70% of the individuals in WES12 are from European population. More samples from diverse populations are needed to further train and test the model. Overall, STAR-NN showed a modest performance to predict autism status. Further studies are needed to assess the viability of ML based models to predict autism status using genomic data.

Ethics approval and consent to participate

The research performed in the study is approved by institutional review board (IRB) of Brown University.

Materials and methods

SPARK whole exome sequencing (WES) data

The sequencing was performed by Simons Foundation Powering Autism Research for Knowledge (SPARK) [59,60]. Variants were called by GATK v4.2.1.0 [61], weCall v2.0.0 [62] and DeepVariant v1.2.0 [63,64] separately and stored in the SPARK genomic dataset. While both GATK and weCall reported a higher number of variants than DeepVariant, DeepVariant has been documented as more precise compared to other callers [63,65]. Studies showed that the number of false positive variants called by GATK was higher compared to those called by DeepVariant [65,66]. Thus, in this study we utilized vcf files produced by DeepVariant. The WES1 and WES2 datasets were released in September 2019 and June 2020, respectively. The WES1 and WES2 cohorts are 77.12% and 73.86% European, respectively. The population distributions are shown in a histogram in S3 Fig. From the 27,281 participants in WES1, 4,681,309 variants were called. From the 15,995 participants in WES2, 4,212,994 variants were called.

An additional independent dataset was also generated by SPARK. This dataset was released in February 2022, two years after the release of the WES1 and WES2 datasets. A total of 27,879 participants were included comprising13,827 individuals with autism and 14,052 non-autistic controls. In this dataset, 8,336,937 variants were called by DeepVariant. However, the SPARK datasets may contain related individuals and therefore kinship is not accounted for in the model.

Quality control

Hail (version 0.2.99) was used to perform quality control on 4,681,309 variants from 27,281 participants in WES1 and 4,212,994 variants from 15,995 participants in WES2. Initially, variants on low complexity regions were removed. Variants labeled as “MONOALLELIC” were also removed. The remaining number of variants were 4,537,198 and 4,085,509 in WES1 and WES2, respectively. Subsequently, we excluded genotypes with the following criteria: 1) genotype calls on chromosome Y for female participants; 2) genotypes with read depth lower than 10 or higher than 1000.; 3) for homozygous reference calls, genotype with the allele balance higher than 0.1 with genotype quality lower than 25; 4) for homozygous alternative calls, genotype with allele balance lower than 0.9 with the phred-scaled likelihood (PL) lower than 25 for homozygous reference; 5) for heterozygous or hemizygous calls, genotypes with allele balance lower than 0.25 and PL for heterozygous reference lower than 25. Samples with a call rate lower than 90%, average genotype quality lower than 20 and average genotype depth lower than 10 were excluded. In WES1 and WES2, 27,266 and 15,986 samples remained, respectively. Regarding variant QC, variants with a call rate above 0.1 and Hardy-Weinberg equilibrium p-value greater than 1e-12 were retained. This resulted in 4,276,376 variants from the 27,266 participants in WES1 and 3,875,705 variants from the 15,986 participants in WES2. We merged the data from WES1 and WES2 incorporating a total of 6,281,625 variants from 43,203 participants in our study (excluding 91 duplicated samples).

Rare variants and variant annotation

To obtain high confidence rare variants, variants with genotype quality lower than 25 or call rate lower than 90% were removed. Rare variants were defined as minor allele frequency (MAF) less than 1% among samples in SPARK dataset and MAF less than 1% in non-neuronal population in gnomAD (exome v2.1.1.). A total of 3,857,525 variants remained. To make our model work efficiently, synonymous variants were excluded, due to the CADD score of synonymous variants is relatively low compared with nonsense variants and missense variants (S4 Fig). 1,469,036 rare variants from 19,117 genes were kept for the model development. The above variants were annotated by dbNSFP4.2a using Annovar (version: 2018Apr16) [67]. We categorized missense variants by their Missense badness, Polyphen2, Constraint (MPC) score [31]. We utilized the three-tier classification of missense variants from the study by Satterstorm et al [7]. MisA represents missense variants with MPC score above 2, indicating the probably damaging missense variants group; MisB represents missense variants with MPC score between 1 and 2, indicating the possibly damaging missense variants; MisC represents missense variants with MPC score below 1, indicating mild impact missense variants.

SPARK genotype data and common variants

Genotype data from 69,592 samples in the SPARK dataset were downloaded. PLINK 1.9 [68,69] was used to generate confidence common variant calls from 650,206 variants. 1,276 samples with more than 10% missing genotypes were excluded. 121,896 variants that failed the Hardy-Weinberg equilibrium exact test with a p-value<1e-06 were excluded. Common variants were defined as having a MAF above 1%. 100,254 variants were excluded due to a lower MAF. 373,475 common variants from 68,316 samples were retained. Subsequently, common variants for individuals from WES1 and WES2 rare variants data were matched using the unique individual identifier (SP_ID, provided in the data).

Generation of gene by sample matrix

The genotype of rare variants in 43,227 samples have been converted to 0,1,2 representing homozygous reference, heterozygous and homozygous alternative, respectively. The matrix table of variants-by-sample was sparse (greater than 99%). We converted the variants-by-sample matrix into gene-by-sample matrix using: Gij = ∑k Vkj, where k represents the number of variants in gene i. Vkj = 0 if sample j does not carry the variant, Vkj = 1 if sample j carries the variant. Therefore, Gij represents the aggregated count of variants in gene i of sample j. In STAR-NN model, where PTVs, MisAB and MisC on the same gene were separated, the Gij represents the aggregated count of single type of variant in gene i of sample j.

Generation of PGS

Effect sizes of common variants were calculated using SBayesR [70]. Summary statistics, including variant ID, rsID (or the genomic location of variant when rsID is not provided), effect allele, alternative allele, odds ratio of effect allele, standard error of odds ratio, p-value, allele frequency of effect allele and per-variant sample size, generated by Grove et al. were used as input for SBayesR–gwas-summary parameter [8]. The banded LD matrix (https://cnsgenomics.com/software/gctb/#LDmatrices) provided by SBayesR were used as input for–mldm parameter. We kept the other parameters as default and excluded MHC region. The output of SBayesR were used as input of–score parameter for PGS calculation by PLINK 1.9.

Feature selection

We employed an automated ML method, Tree-based Pipeline Optimization Tool (TPOT) [32], for the feature selection and the hyperparameter tuning for each feature selection method. Four feature selection methods from sklearn python package were used in TPOT, including “removing features with low variance”, “univariate feature selection”, “recursive feature elimination” and “feature selection using SelectFromModel”. Simultaneously, we generated four gene-by-sample matrices based on variant type (PTV, MisA, MisB and MisC) for feature selection. The value in the matrix is the aggregated count of variants per genes per sample in the training data, which comprises 80% of the samples. Feature selection in TPOT is performed on training data only. The differences across four matrices are the different combination of variants (S1 Fig).

Features selected using methods “SelectFwe” and “SelectionPercentile” from sklearn python package, had the highest ROC-AUC for each sample by gene matrix. In total, 954 genes from 19,117 genes were selected from sample-by-gene matrix covering aggregated count of PTVs and missense variants per gene. 350 genes from 16,832 genes were selected from sample-by-gene matrix covering aggregated count of PTVs per gene. 348 genes from 17,483 genes were selected from sample-by-gene matrix covering aggregated count of PTVs and deleterious missense variants (MisA) per gene. 365 genes were selected from 18,284 genes using sample-by gene-matrix covering aggregated count of PTVs, deleterious missense variants (MisA) and possibly deleterious missense variants (MisB) per gene (S1 Table). We combined selected genes from four matrices and removed overlaps resulting in 1487 selected genes.

Gene ontology (GO) and disease ontology (DO) enrichment analysis

GO and DO enrichment analysis were performed on selected genes using gprofiler2/v0.2.1 R package [71] and DOSE/v3.22.1 R package [72], respectively.

STAR-NN model

STAR-NN is a feedforward neural network with nodes representing variants and genes, and edges representing location relationship between variants and genes. In STAR-NN, rare variants were pre-annotated and separated into three categories: PTVs, possibly damaging missense variants (MisAB, a combination of MisA and MisB) and possibly benign missense variants (MisC). The input layer is the variant layer with each node representing the aggregated count of one type of variants in one gene (57,375 input nodes for 3 variant types from 19,125 genes). The second layer is the gene layer with each node representing a gene. The connection between variant layer and gene layer is sparse and follows a three-to-one pattern, in which variants are only connected to their located gene. For genes that do not carry certain type of variants, we used 0 to represent the blank. The connections between gene layer and output layer were fully connected. The output of each layer follows f(x) = g(wx+b), where w is the weight, x is input of each layer and b is bias, g() represents the “tanh” function, , in sparse layer, “relu” function, , in gene layer and “sigmoid” function, , in the output layer. The learning rate was initially set as 3*10−5 with a learning rate decay follows * , where lr represents initial learning rate. Adam optimizer was used to reduce binary cross entropy loss between the label of a training sample and the predicted probability output of the sample [73]. The model makes an early stop if the loss for validation dataset does not further decrease in 50 epochs and retrieves the model parameter from the best performance according to the minimum loss. The model outputs a probability score between 0 and 1. The prediction is made based on a 0.5 threshold, where score above 0.5 indicates autism and score below 0.5 indicates non-autistic control. The prediction performance is measured using metrics of average F1, ROC-AUC, binary accuracy, precision, and recall rate over 10 random repeats on one random split. The performance of model on 10 random repeats on 10 random splits showed a similar performance and the results are saved in S13 Table.

For the basic DNN model used as a baseline comparison, we did not separate variants in each gene by their functional effect. Therefore, there is no variant layer in the basic DNN model. The aggregated count of variants in genes are used directly as input gene layer. Then, the gene layer is fully connected to the output layer. Gene layer has the following structure: f(x) = g(wx+b), where g() represents “sigmoid” function, the same as STAR-NN. The difference between the basic DNN model and the STAR-NN model is an additional input variant layer. The gene layer in STAR-NN takes the weighted aggregated effect of three types of variants per gene whereas the gene layer in the basic DNN model takes the unweighted aggregated effect of three types of variants per gene.

We also compared the performance of STAR-NN with six traditional ML models, including three tree-based models; decision trees, random forest, XGBoost, three linear models; L1 and L2 logistic regression, linear support vector classifier (LinearSVC). Same as the basic DNN model, aggregated variants in each gene were used as input for the traditional ML models. For traditional ML models, we also separated the training and testing data into 80% and 20% using the same random seed as used in the STAR-NN. We used default parameters settings for DecisionTreeClassifier and RandomForestClassifier from sklearn python package [74] and the default parameter settings of XGBClassifier from xgboost package [75]. For LogisticRegression with L1 regularization and LogisticRegression with L2 regularization from sklearn python package, we set C = 0.001 with “saga” solver and max_iter = 2000. For LinearSVC from sklearn python package, we used C = 0.001 and penalty = “l2”.

Supporting information

S1 Fig. Workflow of selected features using TPOT from four gene sets.

https://doi.org/10.1371/journal.pcbi.1012468.s001

(EPS)

S2 Fig. Enriched disease ontology (DO) term from selected gene features.

developmental disorder, intellectual disability, and autism are among the top enriched DO terms.

https://doi.org/10.1371/journal.pcbi.1012468.s002

(EPS)

S3 Fig. Population distribution on individuals from WES1 and WES2 datasets.

https://doi.org/10.1371/journal.pcbi.1012468.s003

(EPS)

S4 Fig.

CADD score distribution on synonymous variants (A), exonic variants (B) and splicing variants (C). CADD score above 20 indicates top 1% of most deleterious effect. Majority of synonymous variants have a CADD score below 20, a density peak at score of 1 and score of 8. Majority of splicing variants have a CADD score above 20 and a density peak above score of 30.

https://doi.org/10.1371/journal.pcbi.1012468.s004

(EPS)

S1 Table. The list of 1489 selected features generated by TPOT.

https://doi.org/10.1371/journal.pcbi.1012468.s005

(CSV)

S2 Table. Traditional machine learning model performance using prediction accuracy, precision, recall, ROC-AUC and F1.

https://doi.org/10.1371/journal.pcbi.1012468.s006

(CSV)

S3 Table. The training time for STAR-NN model and the logistic regression model with L2 regularization (LRL2).

https://doi.org/10.1371/journal.pcbi.1012468.s007

(CSV)

S4 Table. STAR-NN performance comparison on four gene sets in testing dataset.

Four gene sets include selected features, SFARI genes, combination of SFARI genes and selected features and full gene list.

https://doi.org/10.1371/journal.pcbi.1012468.s008

(CSV)

S5 Table. STAR-NN performance on input without PGS.

https://doi.org/10.1371/journal.pcbi.1012468.s009

(CSV)

S6 Table. STAR-NN performance on input from single type of variants.

https://doi.org/10.1371/journal.pcbi.1012468.s010

(CSV)

S7 Table. STAR-NN performance on input without mild effect missense variants.

https://doi.org/10.1371/journal.pcbi.1012468.s011

(CSV)

S8 Table. STAR-NN performance on input with additional synonymous variants.

https://doi.org/10.1371/journal.pcbi.1012468.s012

(CSV)

S9 Table. STAR-NN performance comparison on four gene sets on extra independent dataset.

https://doi.org/10.1371/journal.pcbi.1012468.s013

(CSV)

S10 Table.

STAR-NN model performance comparison on 1) no additional hidden layer, 2) one additional hidden layer between the gene layer and the output layer using relu activation function and 3) one additional hidden layer between the gene layer and the output layer using linear activation function.

https://doi.org/10.1371/journal.pcbi.1012468.s014

(CSV)

S11 Table. Count and frequency of variants in four populations, including male cases, male controls, female cases, and female controls.

Counts of variants represents the aggregated variants in each population group. Variant frequency is the count of variant divided by total number of individuals in each population group. Proportion of variants between cases and controls in male and female are also calculated. Proportion is calculated based on variant frequency in each population group.

https://doi.org/10.1371/journal.pcbi.1012468.s015

(XLSX)

S12 Table. Enriched GO terms on 1487 selected genes.

https://doi.org/10.1371/journal.pcbi.1012468.s016

(CSV)

S13 Table. STAR-NN performance on 10 random splits on 10 random repeats.

https://doi.org/10.1371/journal.pcbi.1012468.s017

(CSV)

Acknowledgments

We thank families participated in the studies of SPARK. We thank Simons Foundation for initiating the study, data collection and data sharing. We thank Jacob Grove for sharing the summary statistics on GWAS study. We thank the Center for Computation and Visualization (CCV) at Brown University for the maintenance of computing resources. We thank Gabrielle Masse, MSHI, ACRP-CP for reviewing this manuscript and her valuable feedback.

References

  1. 1. Chiarotti F, Venerosi A. Epidemiology of Autism Spectrum Disorders: A Review of Worldwide Prevalence Estimates Since 2014. Brain Sci. 2020;10(5). Epub 2020/05/07. pmid:32370097; PubMed Central PMCID: PMC7288022.
  2. 2. Gabbay-Dizdar N, Ilan M, Meiri G, Faroy M, Michaelovski A, Flusser H, et al. Early diagnosis of autism in the community is associated with marked improvement in social symptoms within 1–2 years. Autism. 2021:13623613211049011. Epub 2021/10/09. pmid:34623179.
  3. 3. Fuller EA, Kaiser AP. The Effects of Early Intervention on Social Communication Outcomes for Children with Autism Spectrum Disorder: A Meta-analysis. J Autism Dev Disord. 2020;50(5):1683–700. Epub 2019/02/26. pmid:30805766; PubMed Central PMCID: PMC7350882.
  4. 4. Hyman SL, Levy SE, Myers SM, Council On Children With Disabilities SOD, Behavioral P. Identification, Evaluation, and Management of Children With Autism Spectrum Disorder. Pediatrics. 2020;145(1). Epub 2019/12/18. pmid:31843864.
  5. 5. Kodak T, Bergmann S. Autism Spectrum Disorder: Characteristics, Associated Behaviors, and Early Intervention. Pediatr Clin North Am. 2020;67(3):525–35. Epub 2020/05/24. pmid:32443991.
  6. 6. Rollins PR, De Froy AM. Reexamining Pathways Early Autism Intervention in Children Before and After the Third Birthday: A Randomized Control Trial. J Autism Dev Disord. 2022. Epub 2022/05/22. pmid:35596830; PubMed Central PMCID: PMC9123830.
  7. 7. Satterstrom FK, Kosmicki JA, Wang J, Breen MS, De Rubeis S, An JY, et al. Large-Scale Exome Sequencing Study Implicates Both Developmental and Functional Changes in the Neurobiology of Autism. Cell. 2020;180(3):568–84 e23. Epub 2020/01/26. pmid:31981491; PubMed Central PMCID: PMC7250485.
  8. 8. Grove J, Ripke S, Als TD, Mattheisen M, Walters RK, Won H, et al. Identification of common genetic risk variants for autism spectrum disorder. Nat Genet. 2019;51(3):431–44. Epub 2019/02/26. pmid:30804558; PubMed Central PMCID: PMC6454898.
  9. 9. Krumm N, Turner TN, Baker C, Vives L, Mohajeri K, Witherspoon K, et al. Excess of rare, inherited truncating mutations in autism. Nat Genet. 2015;47(6):582–8. Epub 2015/05/12. pmid:25961944; PubMed Central PMCID: PMC4449286.
  10. 10. Iossifov I, O’Roak BJ, Sanders SJ, Ronemus M, Krumm N, Levy D, et al. The contribution of de novo coding mutations to autism spectrum disorder. Nature. 2014;515(7526):216–21. Epub 2014/11/05. pmid:25363768; PubMed Central PMCID: PMC4313871.
  11. 11. Gaugler T, Klei L, Sanders SJ, Bodea CA, Goldberg AP, Lee AB, et al. Most genetic risk for autism resides with common variation. Nat Genet. 2014;46(8):881–5. Epub 2014/07/21. pmid:25038753; PubMed Central PMCID: PMC4137411.
  12. 12. Sanders SJ, He X, Willsey AJ, Ercan-Sencicek AG, Samocha KE, Cicek AE, et al. Insights into Autism Spectrum Disorder Genomic Architecture and Biology from 71 Risk Loci. Neuron. 2015;87(6):1215–33. Epub 2015/09/25. pmid:26402605; PubMed Central PMCID: PMC4624267.
  13. 13. Leppa VM, Kravitz SN, Martin CL, Andrieux J, Le Caignec C, Martin-Coignard D, et al. Rare Inherited and De Novo CNVs Reveal Complex Contributions to ASD Risk in Multiplex Families. Am J Hum Genet. 2016;99(3):540–54. Epub 2016/08/30. pmid:27569545; PubMed Central PMCID: PMC5011063.
  14. 14. De Rubeis S, He X, Goldberg AP, Poultney CS, Samocha K, Cicek AE, et al. Synaptic, transcriptional and chromatin genes disrupted in autism. Nature. 2014;515(7526):209–15. Epub 2014/11/05. pmid:25363760; PubMed Central PMCID: PMC4402723.
  15. 15. Ruzzo EK, Perez-Cano L, Jung JY, Wang LK, Kashef-Haghighi D, Hartl C, et al. Inherited and De Novo Genetic Risk for Autism Impacts Shared Networks. Cell. 2019;178(4):850–66 e26. Epub 2019/08/10. pmid:31398340; PubMed Central PMCID: PMC7102900.
  16. 16. Kohane IS. Finding a new balance between a genetics-first or phenotype-first approach to the study of disease. Neuron. 2021;109(14):2216–9. Epub 2021/07/23. pmid:34293292.
  17. 17. Liu Y, Qu HQ, Mentch FD, Qu J, Chang X, Nguyen K, et al. Application of deep learning algorithm on whole genome sequencing data uncovers structural variants associated with multiple mental disorders in African American patients. Mol Psychiatry. 2022;27(3):1469–78. Epub 2022/01/09. pmid:34997195; PubMed Central PMCID: PMC9095459.
  18. 18. Lello L, Raben TG, Yong SY, Tellier L, Hsu SDH. Genomic Prediction of 16 Complex Disease Risks Including Heart Attack, Diabetes, Breast and Prostate Cancer. Sci Rep. 2019;9(1):15286. Epub 2019/10/28. pmid:31653892; PubMed Central PMCID: PMC6814833.
  19. 19. Mohammed M, Mwambi H, Mboya IB, Elbashir MK, Omolo B. A stacking ensemble deep learning approach to cancer type classification based on TCGA data. Sci Rep. 2021;11(1):15626. Epub 2021/08/04. pmid:34341396; PubMed Central PMCID: PMC8329290.
  20. 20. Elmarakeby HA, Hwang J, Arafeh R, Crowdis J, Gang S, Liu D, et al. Biologically informed deep neural network for prostate cancer discovery. Nature. 2021;598(7880):348–52. Epub 2021/09/24. pmid:34552244; PubMed Central PMCID: PMC8514339.
  21. 21. Jiao W, Atwal G, Polak P, Karlic R, Cuppen E, Subtypes PT, et al. A deep learning system accurately classifies primary and metastatic cancers using passenger mutation patterns. Nat Commun. 2020;11(1):728. Epub 2020/02/07. pmid:32024849; PubMed Central PMCID: PMC7002586.
  22. 22. Poirion OB, Jing Z, Chaudhary K, Huang S, Garmire LX. DeepProg: an ensemble of deep-learning and machine-learning models for prognosis prediction using multi-omics data. Genome Med. 2021;13(1):112. Epub 2021/07/16. pmid:34261540; PubMed Central PMCID: PMC8281595.
  23. 23. Onishchenko D, Huang Y, van Horne J, Smith PJ, Msall ME, Chattopadhyay I. Reduced false positives in autism screening via digital biomarkers inferred from deep comorbidity patterns. Sci Adv. 2021;7(41):eabf0354. Epub 2021/10/07. pmid:34613766; PubMed Central PMCID: PMC8494294.
  24. 24. Rahman R, Kodesh A, Levine SZ, Sandin S, Reichenberg A, Schlessinger A. Identification of newborns at risk for autism using electronic medical records and machine learning. Eur Psychiatry. 2020;63(1):e22. Epub 2020/02/27. pmid:32100657; PubMed Central PMCID: PMC7315872.
  25. 25. Lin PI, Moni MA, Gau SS, Eapen V. Identifying Subgroups of Patients With Autism by Gene Expression Profiles Using Machine Learning Algorithms. Front Psychiatry. 2021;12:637022. Epub 2021/06/01. pmid:34054599; PubMed Central PMCID: PMC8149626.
  26. 26. Sardaar S, Qi B, Dionne-Laporte A, Rouleau GA, Rabbany R, Trakadis YJ. Machine learning analysis of exome trios to contrast the genomic architecture of autism and schizophrenia. BMC Psychiatry. 2020;20(1):92. Epub 2020/03/01. pmid:32111185; PubMed Central PMCID: PMC7049199.
  27. 27. Wang H, Avillach P. Diagnostic Classification and Prognostic Prediction Using Common Genetic Variants in Autism Spectrum Disorder: Genotype-Based Deep Learning. JMIR Med Inform. 2021;9(4):e24754. Epub 2021/03/15. pmid:33714937; PubMed Central PMCID: PMC8060867.
  28. 28. Dingemans AJM, Hinne M, Jansen S, van Reeuwijk J, de Leeuw N, Pfundt R, et al. Phenotype based prediction of exome sequencing outcome using machine learning for neurodevelopmental disorders. Genet Med. 2022;24(3):645–53. Epub 2021/12/16. pmid:34906484.
  29. 29. Vabalas A, Gowen E, Poliakoff E, Casson AJ. Machine learning algorithm validation with a limited sample size. PLoS One. 2019;14(11):e0224365. Epub 2019/11/08. pmid:31697686; PubMed Central PMCID: PMC6837442.
  30. 30. Vadapalli S, Abdelhalim H, Zeeshan S, Ahmed Z. Artificial intelligence and machine learning approaches using gene expression and variant data for personalized medicine. Brief Bioinform. 2022. Epub 2022/05/21. pmid:35595537.
  31. 31. Kaitlin E. Samocha JAK, Konrad J., O’Donnell-Luria Anne H., Pierce-Hoffman Emma, MacArthur Daniel G., Neale Benjamin M., Daly Mark J. Regional missense constraint improves variant deleteriousness prediction. bioRxiv. 2017.
  32. 32. Le TT, Fu W, Moore JH. Scaling tree-based automated machine learning to biomedical big data with a feature set selector. Bioinformatics. 2020;36(1):250–6. Epub 2019/06/06. pmid:31165141; PubMed Central PMCID: PMC6956793.
  33. 33. SFARI. SFARI Gene 3.0 (Human Gene Module) 2022. Available from: https://gene.sfari.org/database/human-gene/.
  34. 34. Demetci P, Cheng W, Darnell G, Zhou X, Ramachandran S, Crawford L. Multi-scale inference of genetic trait architecture using biologically annotated neural networks. PLoS Genet. 2021;17(8):e1009754. Epub 2021/08/20. pmid:34411094; PubMed Central PMCID: PMC8407593.
  35. 35. Reed R, MarksIi RJ. Neural Smithing: Supervised Learning in Feedforward Artificial Neural Networks. Cambridge: MIT Press; 1999.
  36. 36. Galan-Vidal J, Socuellamos PG, Baena-Nuevo M, Contreras L, Gonzalez T, Perez-Poyato MS, et al. A novel loss-of-function mutation of the voltage-gated potassium channel Kv10.2 involved in epilepsy and autism. Orphanet J Rare Dis. 2022;17(1):345. Epub 2022/09/07. pmid:36068614; PubMed Central PMCID: PMC9446776.
  37. 37. Lee H, Lin MC, Kornblum HI, Papazian DM, Nelson SF. Exome sequencing identifies de novo gain of function missense mutation in KCND2 in identical twins with autism and seizures that slows potassium channel inactivation. Hum Mol Genet. 2014;23(13):3481–9. Epub 2014/02/07. pmid:24501278; PubMed Central PMCID: PMC4049306.
  38. 38. Daghsni M, Rima M, Fajloun Z, Ronjat M, Bruses JL, M’Rad R, et al. Autism throughout genetics: Perusal of the implication of ion channels. Brain Behav. 2018;8(8):e00978. Epub 2018/06/24. pmid:29934975; PubMed Central PMCID: PMC6085908.
  39. 39. Tick B, Bolton P, Happe F, Rutter M, Rijsdijk F. Heritability of autism spectrum disorders: a meta-analysis of twin studies. J Child Psychol Psychiatry. 2016;57(5):585–95. Epub 2015/12/29. pmid:26709141; PubMed Central PMCID: PMC4996332.
  40. 40. Moreno-De-Luca D, Kavanaugh BC, Best CR, Sheinkopf SJ, Phornphutkul C, Morrow EM. Clinical Genetic Testing in Autism Spectrum Disorder in a Large Community-Based Population Sample. JAMA Psychiatry. 2020;77(9):979–81. Epub 2020/05/14. pmid:32401282; PubMed Central PMCID: PMC7221847.
  41. 41. Srivastava S, Love-Nichols JA, Dies KA, Ledbetter DH, Martin CL, Chung WK, et al. Meta-analysis and multidisciplinary consensus statement: exome sequencing is a first-tier clinical diagnostic test for individuals with neurodevelopmental disorders. Genet Med. 2019;21(11):2413–21. Epub 2019/06/12. pmid:31182824; PubMed Central PMCID: PMC6831729.
  42. 42. Tammimies K, Marshall CR, Walker S, Kaur G, Thiruvahindrapuram B, Lionel AC, et al. Molecular Diagnostic Yield of Chromosomal Microarray Analysis and Whole-Exome Sequencing in Children With Autism Spectrum Disorder. JAMA. 2015;314(9):895–903. Epub 2015/09/02. pmid:26325558.
  43. 43. Chen LS, Xu L, Huang TY, Dhar SU. Autism genetic testing: a qualitative study of awareness, attitudes, and experiences among parents of children with autism spectrum disorders. Genet Med. 2013;15(4):274–81. Epub 2013/01/05. pmid:23288207.
  44. 44. Zhao S, Bao Z, Zhao X, Xu M, Li MD, Yang Z. Identification of Diagnostic Markers for Major Depressive Disorder Using Machine Learning Methods. Front Neurosci. 2021;15:645998. Epub 2021/07/06. pmid:34220416; PubMed Central PMCID: PMC8249859.
  45. 45. Wang Y, Miller M, Astrakhan Y, Petersen BS, Schreiber S, Franke A, et al. Identifying Crohn’s disease signal from variome analysis. Genome Med. 2019;11(1):59. Epub 2019/10/01. pmid:31564248; PubMed Central PMCID: PMC6767648.
  46. 46. Trakadis YJ, Sardaar S, Chen A, Fulginiti V, Krishnan A. Machine learning in schizophrenia genomics, a case-control study using 5,090 exomes. Am J Med Genet B Neuropsychiatr Genet. 2019;180(2):103–12. Epub 2018/04/29. pmid:29704323.
  47. 47. Held E, Cape J, Tintle N. Comparing machine learning and logistic regression methods for predicting hypertension using a combination of gene expression and next-generation sequencing data. BMC Proc. 2016;10(Suppl 7):141–5. Epub 2016/12/17. pmid:27980626; PubMed Central PMCID: PMC5133520.
  48. 48. Fu JM, Satterstrom FK, Peng M, Brand H, Collins RL, Dong S, et al. Rare coding variation provides insight into the genetic architecture and phenotypic context of autism. Nat Genet. 2022. Epub 2022/08/19. pmid:35982160.
  49. 49. Antaki D, Guevara J, Maihofer AX, Klein M, Gujral M, Grove J, et al. A phenotypic spectrum of autism is attributable to the combined effects of rare variants, polygenic risk and sex. Nat Genet. 2022. Epub 2022/06/03. pmid:35654974.
  50. 50. Demetci P, Cheng W, Darnell G, Zhou X, Ramachandran S, Crawford L. Multi-scale Inference of Genetic Trait Architecture using Biologically Annotated Neural Networks. bioRxiv. 2020:2020.07.02.184465.
  51. 51. Luo WY, Xing SQ, Zhu P, Zhang CG, Yang HM, Van Halm-Lutterodt N, et al. PDZ Scaffold Protein CAL Couples with Metabotropic Glutamate Receptor 5 to Protect Against Cell Apoptosis and Is a Potential Target in the Treatment of Parkinson’s Disease. Neurotherapeutics. 2019;16(3):761–83. Epub 2019/05/11. pmid:31073978; PubMed Central PMCID: PMC6694344.
  52. 52. Novak G, Kyriakis D, Grzyb K, Bernini M, Rodius S, Dittmar G, et al. Single-cell transcriptomics of human iPSC differentiation dynamics reveal a core molecular network of Parkinson’s disease. Commun Biol. 2022;5(1):49. Epub 2022/01/15. pmid:35027645; PubMed Central PMCID: PMC8758783.
  53. 53. Starkstein S, Gellar S, Parlier M, Payne L, Piven J. High rates of parkinsonism in adults with autism. J Neurodev Disord. 2015;7(1):29. Epub 2015/09/01. pmid:26322138; PubMed Central PMCID: PMC4553212.
  54. 54. Geurts HM, McQuaid GA, Begeer S, Wallace GL. Self-reported parkinsonism features in older autistic adults: A descriptive study. Autism. 2022;26(1):217–29. Epub 2021/06/23. pmid:34154378; PubMed Central PMCID: PMC8750133.
  55. 55. Wei L, Tian Y, Chen Y, Wei Q, Chen F, Cao B, et al. Identification of TYW3/CRYZ and FGD4 as susceptibility genes for amyotrophic lateral sclerosis. Neurol Genet. 2019;5(6):e375. Epub 2019/12/25. pmid:31872054; PubMed Central PMCID: PMC6878836.
  56. 56. Hildonen M, Levy AM, Hansen CS, Bybjerg-Grauholm J, Skytthe A, Debes NM, et al. EWAS of Monozygotic Twins Implicate a Role of mTOR Pathway in Pathogenesis of Tic Spectrum Disorder. Genes (Basel). 2021;12(10). Epub 2021/10/24. pmid:34680906; PubMed Central PMCID: PMC8535383.
  57. 57. Tremblay MW, Jiang YH. DNA Methylation and Susceptibility to Autism Spectrum Disorder. Annu Rev Med. 2019;70:151–66. Epub 2019/01/30. pmid:30691368; PubMed Central PMCID: PMC6597259.
  58. 58. Nardone S, Sams DS, Reuveni E, Getselter D, Oron O, Karpuj M, et al. DNA methylation analysis of the autistic brain reveals multiple dysregulated biological pathways. Transl Psychiatry. 2014;4(9):e433. Epub 2014/09/03. pmid:25180572; PubMed Central PMCID: PMC4203003.
  59. 59. SPARK_Consortium. SPARK: A US Cohort of 50,000 Families to Accelerate Autism Research. Neuron. 2018;97(3):488–93. Epub 2018/02/09. pmid:29420931; PubMed Central PMCID: PMC7444276.
  60. 60. Feliciano P, Zhou X, Astrovskaya I, Turner TN, Wang T, Brueggeman L, et al. Exome sequencing of 457 autism families recruited online provides evidence for autism risk genes. NPJ Genom Med. 2019;4:19. Epub 2019/08/28. pmid:31452935; PubMed Central PMCID: PMC6707204.
  61. 61. Auwera Gvd, O’Connor BD. Genomics in the cloud: using Docker, GATK, and WDL in Terra. First edition. ed. Sebastopol, CA: O’Reilly Media; 2020. xxiv, 467 pages p.
  62. 62. weCall. weCall. Available from: https://github.com/Genomicsplc/wecall.
  63. 63. Poplin R, Chang PC, Alexander D, Schwartz S, Colthurst T, Ku A, et al. A universal SNP and small-indel variant caller using deep neural networks. Nat Biotechnol. 2018;36(10):983–7. Epub 2018/09/25. pmid:30247488.
  64. 64. Yun T, Li H, Chang PC, Lin MF, Carroll A, McLean CY. Accurate, scalable cohort variant calls using DeepVariant and GLnexus. Bioinformatics. 2021;36(24):5582–9. pmid:33399819; PubMed Central PMCID: PMC8023681.
  65. 65. Yun T, Li H, Chang PC, Lin MF, Carroll A, McLean CY. Accurate, scalable cohort variant calls using DeepVariant and GLnexus. Bioinformatics. 2021. Epub 2021/01/06. pmid:33399819; PubMed Central PMCID: PMC8023681.
  66. 66. Pedersen BS, Brown JM, Dashnow H, Wallace AD, Velinder M, Tristani-Firouzi M, et al. Effective variant filtering and expected candidate variant yield in studies of rare human disease. NPJ Genom Med. 2021;6(1):60. Epub 2021/07/17. pmid:34267211; PubMed Central PMCID: PMC8282602.
  67. 67. Wang K, Li M, Hakonarson H. ANNOVAR: functional annotation of genetic variants from high-throughput sequencing data. Nucleic Acids Res. 2010;38(16):e164. Epub 2010/07/06. pmid:20601685; PubMed Central PMCID: PMC2938201.
  68. 68. Purcell S, Chang C. PLINK [1.9]. Available from: www.cog-genomics.org/plink/1.9/.
  69. 69. Chang CC, Chow CC, Tellier LC, Vattikuti S, Purcell SM, Lee JJ. Second-generation PLINK: rising to the challenge of larger and richer datasets. Gigascience. 2015;4:7. Epub 2015/02/28. pmid:25722852; PubMed Central PMCID: PMC4342193.
  70. 70. Lloyd-Jones LR, Zeng J, Sidorenko J, Yengo L, Moser G, Kemper KE, et al. Improved polygenic prediction by Bayesian multiple regression on summary statistics. Nat Commun. 2019;10(1):5086. Epub 2019/11/11. pmid:31704910; PubMed Central PMCID: PMC6841727.
  71. 71. Raudvere U, Kolberg L, Kuzmin I, Arak T, Adler P, Peterson H, et al. g:Profiler: a web server for functional enrichment analysis and conversions of gene lists (2019 update). Nucleic Acids Res. 2019;47(W1):W191–W8. pmid:31066453; PubMed Central PMCID: PMC6602461.
  72. 72. Yu G, Wang LG, Yan GR, He QY. DOSE: an R/Bioconductor package for disease ontology semantic and enrichment analysis. Bioinformatics. 2015;31(4):608–9. Epub 20141017. pmid:25677125.
  73. 73. Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. 2017. https://arxiv.org/abs/1412.6980.
  74. 74. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research. 2011;12:2825–30.
  75. 75. Chen T, Guestrin C. XGBoost: A Scalable Tree Boosting System. CoRR. 2016;abs/1603.02754. https://doi.org/10.1145/2939672.2939785.