 Research
 Open Access
 Published:
Supervised learning methods in modeling of CD4+ T cell heterogeneity
BioData Mining volume 8, Article number: 27 (2015)
Abstract
Background
Modeling of the immune system – a highly nonlinear and complex system – requires practical and efficient data analytic approaches. The immune system is composed of heterogeneous cell populations and hundreds of cell types, such as neutrophils, eosinophils, macrophages, dendritic cells, T cells, and B cells. Each cell type is highly diverse and can be further differentiated into subsets with unique and overlapping functions. For example, CD4+ T cells can be differentiated into Th1, Th2, Th17, Th9, Th22, Treg, Tfh, as well as Tr1. Each subset plays different roles in the immune system. To study molecular mechanisms of cell differentiation, computational systems biology approaches can be used to represent these processes; however, the latter often requires building complex intracellular signaling models with a large number of equations to accurately represent intracellular pathways and biochemical reactions. Furthermore, studying the immune system entails integration of complex processes which occur at different time and space scales.
Methods
This study presents and compares four supervised learning methods for modeling CD4+ T cell differentiation: Artificial Neural Networks (ANN), Random Forest (RF), Support Vector Machines (SVM), and Linear Regression (LR). Application of supervised learning methods could reduce the complexity of Ordinary Differential Equations (ODEs)based intracellular models by only focusing on the input and output cytokine concentrations. In addition, this modeling framework can be efficiently integrated into multiscale models.
Results
Our results demonstrate that ANN and RF outperform the other two methods. Furthermore, ANN and RF have comparable performance when applied to in silico data with and without added noise. The trained models were also able to reproduce dynamic behavior when applied to experimental data; in four out of five cases, model predictions based on ANN and RF correctly predicted the outcome of the system. Finally, the running time of different methods was compared, which confirms that ANN is considerably faster than RF.
Conclusions
Using machine learning as opposed to ODEbased method reduces the computational complexity of the system and allows one to gain a deeper understanding of the complex interplay between the different related entities.
Background
Immune cell differentiation and modeling
The process of immune cell differentiation plays a central role in orchestrating immune responses. This process is based on the differentiation of naïve immune cells that, upon activation of their transcriptional machinery through a variety of signaling cascades, become phenotypically and functionally different entities capable of responding to a wide range of viruses, bacteria, parasites, or cancer cells. Functionally, immune cells have been classified in either regulatory or effector cell subsets. The cell differentiation process involves a series of sequential and complex biochemical reactions within the intracellular compartment of each cell. The Systems Biology Markup Language (SBML) is an XMLbased format widely used to represent as well as store models of biological processes. SBML allows the encoding of biological process including their dynamics. This information can be unambiguously converted into a system of Ordinary Differential Equations (ODEs). Of note, ODE models are extensively used to model biological processes such as cell differentiation, immune responses towards specific pathogens, autoimmune processes or intracellular activation of specific cellular pathways [1–3]. Several equations are usually required to adequately represent these complex immunological processes, being either at the level of the whole organism, tissue, cells or molecules
In one of our previous studies, Carbo et. al. published the first comprehensive ODE model of CD4+ T cell differentiation that encompassed both effector T helper (Th1, Th2, Th17) and regulatory Treg cell phenotypes [3]. CD4+ T cells play an important role in regulating adaptive immune functions as well as orchestrating other subsets to maintain homeostasis [4]. These cells interact with other immune cells by releasing cytokines that could further promote, suppress or regulate immune responses. CD4+ T cells are essential in B cell antibody class switching, in the activation and growth of CD8+ cytotoxic T cells, and in maximizing bactericidal activity of phagocytes such as macrophages. Mature T helper cells express the surface protein CD4, for which this subset is referred as CD4+ T cells. Upon antigen presentation, naïve CD4+ T cells become activated and undergo a differentiation process controlled by the cytokine milieu in the tissue environment. The cytokine environmental composition therefore represents a critical factor in CD4+ T cell differentiation. As an example, a naïve CD4+ T cell in an environment rich in IFNγ or IL12 will differentiate into Th1. In contrast, an environment rich in IL4 will induce a Th2 phenotype. Some other phenotypes are also balanced by each other: Th17 cells, induced by IL6, IL1β and TGFβ, are closely balanced by regulatory T cells (induced by TGFβ only) [5]. Furthermore, competition for cytokines by competing clones of CD4+ T cells within an expanding cell population (proliferation), cell death and expression of other selective activation factors such as the T cell receptor, OX40, CD28, ICOS and PD1 are key steps that influence CD4+ T cell differentiation.
Computational approaches allow concurrent multiparametric analysis of biological processes. Computational algorithms and models have become powerful and widely used tools to improve the efficiency and reduce cost of the knowledge discovery process. Systems modeling approaches combined with experimental immunology studies can integrate existing knowledge and provide novel insights on rising trends and behaviors in biological processes such as CD4+ T cell differentiation and function. The CD4+ T cell differentiation model was built upon the current paradigms of molecular interactions that occur in CD4+ T cells, which consists of 60 ODEs, 53 reactions, and 94 species. The mathematical model ensures proper modulation of intracellular pathways and cell phenotypes via external cytokines representing the cytokine milieu. Two types of kinetic equations were employed to mathematically compute dynamic biological processes in the CD4+ T cell model: 1) mass action and 2) Hill equation kinetics. Despite their simplicity, mass action kinetics are widely accepted and extensively validated in biological systems due to their inherent ability to accurately represent elementary reactions and species degradation [6]. Mass action rates are also extremely reliable for stochastic modeling simulations. In the CD4+ T cell model, the natural loss of model species due to mRNA and protein decay was fit using mass action rate laws. On the other hand, sigmoidal Hill equations were used to model more complex molecular processes that behave via “on/off” switch mechanisms including protein phosphorylation, cytokinereceptor binding and transcription. Extensive studies have demonstrated the benefits of the Hill equation for studying combinatorial regulation, especially in sigmoidal Hill equations [7], and thus this equation set captures complexities arising when a particular model species can be modified by more than one input. Results from modeling the pleiotropic and highly dynamic regulation of CD4+ T cell differentiation has guided experimentation to elucidate underlying regulatory mechanisms, identify novel putative CD4+ T cell subsets or potential targets, and enrich our understanding of the dynamics of the process [8, 9].
ODEbased modeling approaches require detailed knowledge about kinetic parameters, some of which can be estimated from literature and some from in silico experiments. However, models that are based on a large parameter set will be subject to higher level of inaccuracies. Thus, the use of novel modeling approaches applicable to the immune system and specifically to the CD4+ cell differentiation has a high value for investigation.
Multiscale modeling and model reduction
Current biomedical research involves performing experiments and developing hypotheses that link different scales of biological systems such as: intracellular signaling or transcriptional interactions, cellular behavior and cell population behavior, as well as tissue and organismlevel events. Computational modeling efforts exploring multiscale systems have to incorporate an array of techniques due to the different time and space scales involved. In one of our previous studies, Mei et. al. presented Enteric Immunity Simulator (ENISI), an agentbased simulator for modeling mucosa immune responses to enteric pathogens [10]. ENISI uses a rulebased approach and can simulate cells, cytokines, cell movement and cellcell interactions. To be able to model finegrained intracellular behaviors, a multiscale modeling approach that embeds intracellular models into the intercellular tissue level models is needed. Indeed, the multiscale modeling approach includes four scales: Intracellular, Chemokine/Cytokine diffusion (intercellular), Cellular, and Tissue. The current version of ENISI incorporates Cellular Scale, Chemokine Scale and Tissue Scale. The cellular scale represents how the cells interact with nearby cells and incorporates the plasticity of a cell based on stochastic and temporal rules. The chemokine scale represents the chemokine concentration and diffusion process. Finally, the tissue scale represents the spatial and compartmental information (Fig. 1).
Finegrained ODE models of intracellular pathways controlling immune cell differentiation are adequate for studying mechanisms of cell differentiation. However, they can be highly complex and expensive from a computational standpoint, especially when embedded within largescale agentbased simulations. ENISI Visual models a large number of cells and microbes in the gastrointestinal mucosa. If each agent is represented by 60 ODEs, as an example, the simulation will be hardly scalable. Therefore, to be able to develop efficient agentbased multiscale models, model reduction needs to be performed. In addition, multiscale models usually do not require all the internal details of intracellular scales to have predictive value. In essence, novel model reduction strategies could be used to address the multiscale scalability requirements to reduce molecular models before integrating them into largescale agentbased tissuelevel models.
Supervised learning methods and their applications
Supervised machine learning methods use training data to learn the structure of a system and utilize that knowledge to predict the outcome for an unseen condition. Supervised learning methods have been applied in multiple areas, such as bioinformatics, cheminformatics, database marketing, spam detection, and pattern recognition in general [11]. Artificial Neural Network (ANN), Linear Regression (LR), Support Vector Machines (SVM) and Random Forest (RF) are examples of supervised machine learning methods.
Artificial neural networks algorithms, inspired by the biological neural systems, are powerful in modeling and data mining tools based upon the theory of connectionism [12]. In biological systems, neurons are connected to each other through synapses. A neuron receives inputs from multiple neurons and outputs a value based upon the activation function. Perceptron is one of the easiest data structures for the study of neural networking. The perceptron models neuron’s behavior in the following way: First the perceptron receives several input values. The connection for each of the inputs has a weight in the range of 0 to 1. The threshold unit then sums the inputs, and if the sum exceeds the threshold value a signal is sent to the output node, otherwise no signal is sent. The perceptron can learn by adjusting the weights to approach the desired output [13].
Building on the algorithm of the simple perceptron, the multilayer perceptron (MLP) model not only gives a perceptron structure for representing more than two classes, it also defines a learning rule for this kind of networks. The MLP is divided into three layers: the input layer, the hidden layer and the output layer, where each layer in this order processes inputs and deliver outputs to next layer [13]. The extra layers give the structure needed to recognize nonlinearly separable classes (Fig. 2). The network structures and the parameters of the activation function are important factors when developing neural network models. Feedforward neural networks are frequently used structures in modeling. There are effective learning algorithms for the parameters once the structures are set in the feedforward ANNs.
Artificial neural network algorithms are widely used for data mining tasks such as classification and pattern recognition. Neural network algorithms are especially effective in modeling nonlinear relationships which makes them ideal candidates for differentiation processes. Importantly, this process is scalable. However, there are also some practical challenges. It is not possible to know in advance the ideal network topology. Therefore, ANNbased methods require testing several network settings or topologies in order to find the best solution. This technical challenge triggers an extended training period. Our initial pilot study was the first to apply neural network algorithms into studying the immune cell differentiation [14]. Based on the initial success, the study was systematized and expanded.
Linear regression model are attractive because of their simplicity and reduced computational complexity. Linear regression is an approach for modeling the relationship between a scalar dependent variable and explanatory variables [15]. In linear regression, data are modeled using linear predictor functions, while unknown model parameters are estimated from the data. Such models are called linear models [16]. Linear regression is a regression analysis that is studied rigorously, and used widely in practical applications. Linear regression is extensively used in biological [17], behavioral and social sciences to describe possible relationships between variables.
Support Vector Machines is another widelyused supervised learning algorithm for classification and regression problems. SVM contains all the main features that characterize maximum margin algorithm: a nonlinear function is leaned by linear learning machine mapping into high dimensional kernel induced feature space [18]. Given a set of training examples with each marked as belonging to one category, an SVM training algorithm builds a model that could assign new examples into one category [19]. SVMs are helpful in text and hypertext categorization as their application can significantly reduce the need for labeled training instances in both the standard inductive and transductive settings. SVMs are also useful in medical science to predict survival in breast cancer [20].
Finally bagging of classification trees is one wellknown ensemble learning method. In bagging, each successive tree is independently constructed using a bootstrap sample of the dataset. A simple majority vote is then taken for prediction [21]. Based on bagging theory, Breiman proposed the Random Forest algorithm, which adds an additional layer of randomness to bagging [22]. In addition to constructing each tree using a different bootstrap data sample, random forests change how the classification and regression trees are constructed. In standard trees, each node is split using the best split among all variables, while in a random forest, each node is split using the best among a subset of predictors randomly chosen at that node. In addition, RF algorithm has only two parameters (the number of variables in the random subset at each node and the number of trees in the forest) and is usually not very sensitive to their values [23]. RF method is based on the aggregation of a large number of decision trees. Specifically, it is an ensemble of trees constructed from a training dataset and internally validated to yield a prediction of the response given the predictors for future observations. An important feature of RF is its outofbag (OOB) error [24]. Each observation is an OOB observation for some of the trees. The OOB error of the RF is the average error frequency obtained when the observations from the dataset are predicted using the trees for which they are OOB. Through this internal validation, the error estimation is less optimistic and usually considered as a good estimator of the expected error for independent data (Fig. 3). For instance, Random Forest models have been successfully used in recent years to explore metabolic syndrome serum profiling [25] and predict avian influenza H5N1 outbreaks [26]. However, to the best of our knowledge, this study is the first to apply RF algorithms into studying the immune cell differentiation.
Related work
Modeling the CD4+ T cell differentiation is challenging because of the complexity of the immune system, plasticity between phenotypes, feedback loops involved in regulation and combinatorial effects of cytokines. The immune system protects the human body from pathogens by recognizing, containing, and destroying nonself or foreign antigens [27, 28]. At the highest level, the immune system can be divided into innate and adaptive branches. The innate immune system, involving cells such as macrophages, epithelial cells, neutrophils, and dendritic cells, responds quickly but nonspecifically to stimuli [29]. On the contrary, the adaptive immune system involving T cells and B cells responds more specifically to antigens [30]. Immune cells are activated and differentiated into evergrowing numbers of cell subsets such as CD4+ T cells and macrophages [31–33]. These cells are regulated by different cytokines in their microenvironment. Using CD4+ T cells as an example, Th1 cells stably express IFNγ, whereas Th2 cells express IL4. The discovery and investigation of two other CD4+ T cell subsets, induced regulatory T (iTreg) cells and Th17 cells, has led to a rethinking of the notion that helper T cell subsets represent irreversibly differentiated endpoints. Mounting evidence supports the tissue environmentdependent plasticity of CD4+ T cell subsets and suggests the emergence of new phenotypes. When both TGFβ and IL6 are present in the environment, naïve CD4+ T cells differentiate into Th17 [34, 35]. When TGFβ alone presents in the environment, CD4+ T Cells differentiate into Treg [14]. When IFNγ and IL12 are present, T cells differentiates into Th1 [36].
Systems biology has become an important paradigm in immunology research, using mathematical and computational models to synthesize and mine exiting knowledge, and discover new knowledge from big data [37]. Biological systems and processes can be modeled using a variety of methods [38–40]. In some instances, biological processes can be mapped to networks where nodes and edges represent biological agents such as cells and their interactions [41]. Furthermore, mathematical or computational dynamics can be applied to the network models so that in silico simulations can be performed [1, 42]. SBML is a XMLbased file format used to represent computational models of biological processes [43]. There are many types of models used for modeling biological processes such as Bayesian networks, ODE, and agentbased models [44]. For metabolic and signaling networks, the biochemical reactions can be represented by firstorder ODEs [45].
In line with our systems and translational immunology efforts under Modeling Immunity to Enteric Pathogens (www.modelingimmunity.org) of computational model building, calibration, refinement and validation, Carbo et. al. published the first ODE model of CD4+ T Cell differentiation, which comprises of 60 ODEs [2]. The model as shown in Fig. 4 represents the intracellular pathways that are critical for CD4+ T cell differentiation. The hypotheses generated by this model were fully validated using in vivo animal models of inflammatory bowel disease (IBD). Computational modeling and mouse adoptive transfer studies were combined to gain a better mechanistic understanding of the modulation of CD4+ T cell differentiation and plasticity at the intestinal mucosa of mice. Sensitivity analyses highlighted the importance of PPARγ in the regulation of Th17 to iTreg plasticity. Indeed, validation experiments demonstrated that PPARγ is required for the plasticity of Th17 promoting a functional shift towards an iTreg phenotype. More specifically, PPARγ activation is associated with upregulation of FOXP3 and suppression of IL17A and RORγt expression in colonic lamina propria CD4+ T cells. Conversely, the loss of PPARγ in T cells results in colonic immunopathology driven by Th17 cells in adoptive transfer studies.
In another study Mei et. al. presented ENISI Visual, an agentbased simulator for modeling enteric immunity [10]. ENISI Visual provides high quality visualizations for simulating gut immunity to enteric pathogens and is capable of simulating gut immunity, including pathogen invasion, proinflammatory immune responses, pathogen elimination, regulatory immune responses, and restoring homeostasis. ENISI Visual can also help immunologists test novel hypotheses and design biological experiments accordingly. Undoubtedly a holistic model of the immune response could provide even more valuable insights; however, it needs to take into consideration complexities at the different layers: intracellular, cellular, intercellular, tissue and whole organism. Modeling a complex system at four levels of magnitude – multiscale modeling (MSM) – poses new set of challenges. Multiscale modeling requires considering different spatial and temporal scales, ranging from nanometers to meters and nanoseconds to years. Therefore, different technologies have to be integrated to provide most accurate predictions. Multiscale modeling frameworks have been recently developed and attempted to address some of these challenges [46–48]. In our recent work, we have developed ENISI MSM, a multiscale modeling platform driven by highperformance computing and designed specifically for computational immunology, which integrates agent based modeling (ABM), ODEs and partial differential equations (PDEs) [48]. Our ENISI MSM platform is calibrated with experimental data and tested for the CD4+ T cell differentiation model which is able to perform a variety of in silico experimentation for generating new hypothesis. However, running simulations on the MSM platform that requires COPASI [49] to solve complex ODEs is computationally expensive and time consuming. Replacing the ODEbased steps in the MSM by machine learning methods would significantly improve its computational performance and allow researchers to perform broad and comprehensive in silico experimentation that uncovers emerging properties of the immune system and results in new nonintuitive computational hypotheses about immune responses.
Machine learning methods – supervised learning methods in particular – are key in building predictive models from observations, therefore facilitating knowledge discovery for complex systems. Neural network algorithm is a supervised learning approach and has been widely used in data mining tasks [50, 51] as well as medical applications [52, 53]. Snow et al. developed neural networks for prostate cancer diagnosis and prognosis [54]. Lek et al. introduced neural networks in ecological modeling [55]. Brusic et al. used neural networks for predicting major histocompatibility complex (MHC) binding peptides [56]. Learning is an important research topic in neural networks. White presented neural network learning algorithms from the statistical perspective [57]. Hagan et al. presented an effective learning algorithm called backpropagation for training feedforward networks [58]. In addition to modeling and predictions, neural network algorithm has also been used for solving ordinary and partial differential equations [59].
Our initial work [14] presented ANN as an alternative to solving ODEs using in silico data; in that study ANN was compared with LR model and it was shown to outperform the latter. In the present work, we compare four different learning methods: ANN, LR, SVM and RF. We optimize the parameters of the models and apply them to in silico data with and without added noise. We corroborate our findings with experimental data and demonstrate that both ANN and RF are capable of predicting the dynamic behavior of the output cytokines in four out of five cases. Finally, we also evaluated the methods based on their computational performance.
Methods
To model cell differentiation we first define the problem and make the following assumptions. There are m input cytokines that regulate immune cell differentiation: C_{i1}, C_{i2}, …, C_{im}. There are also n output cytokines secreted by immune cells: C_{o1}, C_{o2}, …, C_{on}. The cytokine concentrations are positive continuous values.
The problem of modeling immune cell differentiation is to develop one model for the following functional relationship:
The model is designed to predict the output cytokine concentrations given concentrations of input cytokines.
T cell differentiation process as a use case
This study focuses on the T cell differentiation. However, the techniques and algorithms developed herein can be applied to differentiations of other types of immune cells, such as macrophages, dendritic cells, B cells, etc. The input cytokines are internalized by the naïve T cells and regulate the T cell differentiation process. The output cytokines are externalized and secreted.
Data for training and testing models
The data for modeling the relationship from the input and output cytokines can be derived from the T Cell differentiation ODE model [2] which was calibrated using data from biological experiments. By changing the concentrations of the input cytokines, the steady state of the ODE model is calculated. The steady state results provide a measure of the output cytokines that can be used in the model. Creation of datasets was achieved by using the parameter scan task of COPASI tool [49]. COPASI is a software application for simulation and analysis of biochemical networks and their dynamics, which supports models in the SBML standard and can simulate their behavior using ODEs. A fivedimensional scan was performed, where five output cytokines were independently measured. All the data is normalized to the range of [0, 1]. The method used to create datasets is equaldistance sampling. For each input cytokine, five values were chosen (0, 0.25, 0.5, 0.75, and 1). Since there is a total of five input cytokines, 625 data points were created by the parameter scan process. One hundred of the data points were selected randomly for training and the remaining 525 data points were used for testing. Additionally, uniformly distributed noise was added to the output for a quantitative analysis. Table 1 shows an example of data points used in the study.
Supervised learning methods
Artificial neural networks
ANN models can be used to model nonlinear relationships. We developed the ANN model for T cell differentiation using a package in R named neuralnet [60]. The learning algorithm used is backpropagation. The function neuralnet is used for training neural networks, which provides the opportunity to define the required number of hidden layers and hidden neurons. The most important arguments of neuralnet function include formula (a symbolic description of the model to be fitted), data (a data frame containing the variables specified in formula), and a hidden vector (specifying the number of hidden layers and hidden neurons in each layer) [60]. To optimize the performance of the ANN model, we tested different sizes of hidden layers, including 1, 2, 4, 5, 6, 7, 8, 10, and 11 hidden neurons. By comparing the average absolute difference between the model predictions and real outputs from the test data, the neural network model with seven hidden neurons was identified to perform best (Table 2). Size of hidden layers is a critical model parameter. If the number of layers is too small underlearning can occur whereas a size too large can cause overlearning or over fitting. In this study, our results demonstrated that with the network of four inputs and five outputs, seven hidden neurons were necessary to best model the complex nonlinear system of cell differentiation using backpropagation (Fig. 5).
Linear regression
LR model was tested for its simplicity. R has a linear regression module lm that was adapted and used in this study. The lm function is used to fit linear models, which can be used to carry out regression, single stratum analysis of variance and analysis of covariance [61].
Support vector machine
A model was created using SVM algorithm. The R Package, e1071 [62], was applied to build the support vector machine models using the same training data and test data as used by our previous modeling approaches. To optimize the performance of the SVM model, we tested different width of radial kernel, including baseline (0.25), 1, 0.1, 0.01, and 0.001. By comparing the average absolute difference between the model predictions and real outputs from the test data, the model with a kernel size of 0.25 (baseline) was identified to perform best (Table 3).
Random forest
A RF model was created using the randomForest package in R [23]. The function randomForest is used for building trees, which provides the opportunity to define the number of trees to grow and the number of variables randomly sampled as candidates at each split. For each output cytokine, a Random Forest model was built. In essence, for five outputs, IL17, RORgt INFγ, Tbet, and FOXP3, five Random Forest models were created. To optimize the performance of the RF model, two main variables – mtry and ntree – were optimized (see Fig. 6). By comparing the average absolute difference between the model predictions and real outputs from the test data, the random forest model with 1000 trees and 4 variables randomly sampled as candidates at each split was identified to perform best.
Capability of the models to analyze data with noise
Stochasticity is an inherent component of biological processes and an important aspect in modeling such systems [43–46]. Thus, we incorporated noise to the output data points. A uniformly distributed noise in range of [−0.5 %, 0.5 %] and [−1 %, 1 %] was add to all five output data points independently in order to assess whether the learning methods could be used to model the system with same level of accuracy. The level of noise that was applied is relatively low because there is no indication of any species in the model to be subject to a low copy number and therefore to significantly higher levels of fluctuation and noise. In a similar manner, 100 data points were selected randomly as the training dataset and the remaining 525 data were used for testing.
Model validation
Testing the models using in vitro data
In vitro data were obtained from recent publications, which were used to further validate neural network model and random forest model. The first publication [63] shows that TGFβ and IL6 are required for the lineage commitment of pathogenic Th17 cells. TGFβ and IL6 drive the production of IL17 by T cells and restrain Th17 cellmediated pathology, such as production of IFNγ. The second publication [64] shows that TGFβ and IL6 together induces the differentiation of pathogenic Th17 cells from naïve T cells. Meanwhile, IL6 as an acute phase protein induced during inflammation, completely inhibits the generation of Foxp3+ Treg cells induced by TGFβ. The third publication [65] shows that Tbet upregulation and subsequent IL12 stimulation are essential for induction of Th1 mediated immunopathology in Crohn’s disease. Also, augmentation of IFNγ production by IL12/IL18 was independent of Tbet expression. The experimental condition (cytokine concentrations) was extracted from publications as inputs of supervised learning models. The outputs of models were compared with experimental data to identify whether the dynamics behaviors of cells could be predicted correctly.
10fold cross validation
Crossvalidation is a model validation technique for assessing how the results of a statistical analysis could generalize to an independent dataset. The goal of cross validation is to define a dataset to test the model in the training phase, in order to limit overfitting and give an insight on how the model will generalize to an independent dataset. 10fold crossvalidation approach was used in our studies. Firstly, the whole dataset was randomly divided into 10 subsets. Since there are totally 625 data samples, five subsets contains 63 data respectively and the other five subsets have 62 data per set. Then, 10 rounds of learning were performed for each machine learning approaches; on each round one subset of the data was used as a test set and the remaining 9 data subsets are used as training data. The average of meansquared error on the predictions of each machine learning approach was used as an estimate of accuracy.
Running time comparison among different supervised learning methods
The running time of different supervised learning methods was calculated using proc.time function of R, which determines how much real and CPU time the currently running R process has already taken.
Results
LR model was tested for its simplicity. R has a linear regression module lm that was adapted and used in this study. The lm function is used to fit linear models, which can be used to carry out regression, single stratum analysis of variance and analysis of covariance [61]. The result of the linear regression model can be summarized as linear transformation from the input cytokines to the output cytokines as shown by the Eq. 2. The transformation matrix, M_{Tran}, (Eq. 3) summarizes the relationship between input and output cytokine concentrations.
where rows represent IFNγ, IL12, IL6, and TGFβ respectively.
The prediction error, the average absolute difference between the model predictions and real outputs from the test data, of the linear model is shown in Table 3. Considering the data are normalized within [0, 1], the prediction error of linear regression model is larger than that of neural network model. This corroborates that the T cell differentiation process is highly nonlinear and linear regression will not be an appropriate method for this highly complex and nonlinear process. By calculating the prediction error, it is concluded that the performance of support vector machine model is better than linear regression model, but worse than neural network model (Table 4). The prediction error of the RF model (Table 4) is better than linear regression and support vector machine. RF’s performance is comparable with ANN method.
The Table 5 shows that the ANN model and RF model still outperform the linear regression model and the support vector machine model when noise is added to the data. However, the performance of these two models deteriorates slightly when compared to data without added noise.
Testing the best performers – ANN and RF – using in vitro data is important in order to assess the predictive quality of the models. We gathered three publications that provide experimental data for our input and output cytokines. We tested the methods and in four out of five cases both models were able to predict the dynamic behavior of the system. For instance in the first study, when IL6 and TGFβ are 100 and 10 respectively the level of IL17 is measured to be 4875 (baseline level is 188), corresponding to an upregulation of IL17 with respect to baseline. The model predictions for the same input values are 0.99 versus 0.14 for ANN and 0.769 versus 0.128 for RF: and upregulation with respect to the baseline. Four out of five experimental data are reproduced with the correct dynamic behavior. It is essential to note that even though it is important to test the system with experimental data, there is discrepancy between different experimental conditions and the final results. For instance, the third and fourth studies measure the same cytokines; however, the results demonstrate different dynamic behavior. In the third study the authors observe an upregulation of Foxp3; while, in the fourth study authors observe a downregulation of the same output. The difference can be due to experimental conditions as well as input range (Table 6).
In order to control overfitting and give an insight on how the models will generalize to an independent dataset crossvalidation was performed. Crossvalidation is a model validation technique for assessing how the results of a statistical analysis would generalize to an independent dataset in a practical setting. The goal of crossvalidation is to define a dataset to test the model in the training phase. 10fold crossvalidation approach was used to evaluate models created by different machine learning, including ANN, LR, SVM, and RF. The average of meansquared error on the predictions of each model was used as an estimate of accuracy, which is shown in Table 7. ANN and RF are still the best performers following this analysis.
The R function, proc.time, was used to determine how much real and CPU time (in seconds) the training and testing processes of each supervised learning methods have already taken (Table 8). proc.time returns five elements for backwards compatibility, but its print method prints a named vector of length 3. The first two entries are the total user and system CPU times of the current R process and any child processes on which it has waited, and the third entry is the ‘real’ elapsed time since the process was started. The system specification is Intel® Core(TM) i74500 CPU @ 1.80 GHz 2.40 GHz and 4.00 GB RAM. The comparison between ANN and RF shows that ANN is faster on both real and CPU time.
Discussion
In this study, we presented four different supervised learning methods – ANN, LR, SVM, and RF – to model the CD4+ T Cell differentiation. Immune cell differentiation is an important immunological process that is not fully characterized. Based upon our previous studies on the ODE model of CD4+ T cell differentiation and agentbased modeling for enteric immunity, it is concluded that developing multiscale models requires significant reduction of the intracellular ODE model before integrating them into the intercellular agentbased models. However, since immune cell differentiation is a highly nonlinear process, the linear regression model was not capable of fitting the data well. Linear regression models provide a simplistic approach that is very well scalable and was shown to outperform neural network models in a recent study [66]. ANN and RF were shown to be best performers with in silico data with and without added noise.
In particular a feedforward neural network model has been developed, focusing on modeling the relationship between the input external cytokines regulating the cell differentiation and the output cytokines secreted and externalized by the immune cell subsets. After training using back propagation algorithm, this neural network model predicts the concentrations of the output cytokines with an average prediction error of 0.0379 for the five output cytokines concentrations. The neural network model significantly reduces the ODE model complexity by focusing on the needs of multiscale models. This approach is scalable and can be integrated into future multiscale modeling efforts.
In our analysis, we also explored SVM as potential candidates for the modeling of T cell differentiation. SVMs provide a number of advantages over ANN. For instance, ANN algorithm is more prone to overfitting as compared to SVMs [67]. In additions, unlike ANN, computational complexities of SVM do not depend on the dimensionality of the input space [68] and therefore it could provide a more scalable framework. Finally, solution to SVM is global [69], where ANN could suffer from multiple local minima [70]. However, in our analysis the ANN significantly outperformed SVM.
Random Forest has also outperformed SVM and Linear regression model in our experiment. RF and ANN have comparative performance. Inherited from classification and regression trees, random forest algorithm has the following advantages. It handles categorical predictors, highly nonlinear interactions, and missing values. It is computationally simple even for large problems. Furthermore, RF does not require formal distributional assumptions (nonparametric) and provides an automatic variable selection process [22, 71]. In addition, it also overcomes disadvantages of classification and regression trees. For example, on accuracy aspect, random forest turns out to perform very well compared to many other classifiers, including discriminant analysis, support vector machines and neural networks [23], and is robust against overfitting [22]. Random forest methods have been widely applied in bioinformatics and computational biology. A major field of application of RF method is largescale genetic association studies. The response is typically a phenotype of interest, while the predictors are genetic markers, often SNPs that can be seen as predictors with two or three categories. RFs yield both a prediction tool and a ranking of the SNPs with respect to their classification ability [72]. Other applications of random forest include prediction of patient outcome from high dimensional gene expression data, where patients are instances and their outcome is the response to be predicted [73]. Another class of applications is the prediction of molecule properties based on sequence information, such as the prediction of replication capacity based on HIV1 sequence variation [74]. However, RF methods have not been heavily used in immunology studies. To the best of our knowledge this study is the first one applying the random forest algorithm for immune cell differentiation.
Furthermore, analysis of data with noise is an important step, as biological systems are stochastic processes in general. As we have shown the performance of the ANN system deteriorates but only marginally. RF outperforms all the other methods when low level noise is added to the in silico data, while ANN performs best using date with high level noise. Therefore, the constructed modeling framework is stable and robust to slight variations.
Finally, by testing the best performers – ANN and RF – using in vitro experimental data, the predictive accuracy of the models was accessed. In four out of five cases both models were able to predict the dynamic behavior of the system, which demonstrated that our models are capable of predicting dynamic behaviors of cell differentiation system with high accuracy. There are two contradictory data on FOXP3. One data showed that Foxp3 did not substantially change (slightly increase) [63], while another data showed that addition of TGFβ plus IL6 to T cells during differentiation completely abrogated the expression of Foxp3 [64]. This contradictory observation could be due to different background of mice used in these two studies or different initial expressing level of Foxp3 in T cells [75]. In addition to prediction accuracy, the running time of each supervised learning method was measured. It is concluded that ANN performs more efficiently than RF with similar accuracy.
Conclusions
This is the first study using neural networks as well as random forest to model immune cell differentiation. We have shown that the proposed modeling framework is robust to noise, and outperforms two other widely used methods – LRM and SVM. Furthermore, ANN and RF models represent ideal candidates for integration into the agentbased models that we have developed using ENISI MSM to study the immunological processes comprehensively and systematically. Using machine learning as opposed to ODEbased methods will reduce the computational complexity of the system and allow us to gain a deeper understanding of the complex interplay between the molecules, cells and tissues of the immune system to advance the development of safer and more efficacious therapeutics.
Abbreviations
 ANN:

Artificial neural networks
 ENISI:

Enteric immunity simulator
 iTreg:

Induced regulatory T
 LR:

Linear regression
 MHC:

Major histocompatibility complex
 MLP:

Multi layer perceptron
 MSM:

Multi scale modeling
 ODE:

Ordinary differential equation
 OOB:

Outofbag
 RF:

Random forest
 Treg:

Regulatory T cells
 SVM:

Support vector machines
 SBML:

Systems biology markup language
 Th:

T helper
References
 1.
Carbo A, Hontecillas R, Andrew T, Eden K, Mei Y, Hoops S, et al. Computational modeling of heterogeneity and function of CD4+ T cells. Front Dev Biol. 2014;2:31.
 2.
Carbo A, Hontecillas R, Kronsteiner B, Viladomiu M, Pedragosa M, Lu P, et al. Systems Modeling of Molecular Mechanisms Controlling Cytokinedriven CD4+ T Cell Differentiation and Phenotype Plasticity. PLoS Comput Biol. 2013;9:e1003027.
 3.
Carbo A, OlivaresVillagomez D, Hontecillas R, BassaganyaRiera J, Chaturvedi R, Piazuelo MB, et al. Systems modeling of the role of interleukin21 in the maintenance of effector CD4+ T cell responses during chronic Helicobacter pylori infection. mBio. 2014;5:e01243–01214.
 4.
Zhu J, Paul WE. Peripheral CD4 T cell differentiation regulated by networks of cytokines and transcription factors. Immunol Rev. 2010;238:247–62.
 5.
Eisenstein EM, Williams CB. The T(reg)/Th17 cell balance: a new paradigm for autoimmunity. Pediatr Res. 2009;65:26R–31.
 6.
Goldbeter A. A minimal cascade model for the mitotic oscillator involving cyclin and cdc2 kinase. Proc Natl Acad Sci. 1991;88:9107–11.
 7.
Mangan S, Alon U. Structure and function of the feedforward loop network motif. Proc Natl Acad Sci U S A. 2003;100:11980–5.
 8.
Kidd BA, Peters LA, Schadt EE, Dudley JT. Unifying immunology with informatics and multiscale biology. Nat Immunol. 2014;15:118–27.
 9.
Yosef N, Shalek AK, Gaublomme JT, Jin H, Lee Y, Awasthi A, et al. Dynamic regulatory network controlling TH17 cell differentiation. Nature. 2013;496:461–8.
 10.
Mei Y, Hontecillas R, Zhang X, Bisset K, Eubank S, Hoops S, et al. ENISI Visual, an agentbased simulator for modeling gut immunity. In: Bioinformatics and Biomedicine (BIBM), 2012 IEEE International Conference on. IEEE. 2012. p. 1–5.
 11.
Mohri M, Rostamizadeh A, Talwalkar A: Foundations of machine learning. Cambridge, MA: MIT press; 2012.
 12.
Yegnanarayana B: Artificial neural networks. Delhi, India: PHI Learning Pvt. Ltd.; 2009.
 13.
Nielsen F. Neural Networks algorithms and applications, Neil’ s Brock Business College. 2001.
 14.
Mei Y, Hontecillas R, Zhang X, Carbo A, BassaganyaRiera J. Neural network models for classifying immune cell subsets. In: Bioinformatics and Biomedicine (BIBM), 2013 IEEE International Conference on. 2013. p. 5–11.
 15.
Freedman D: Statistical models: theory and practice. New York, NY: Cambridge University Press; 2009.
 16.
Seal HL. Studies in the History of Probability and Statistics. XV The historical development of the Gauss linear model. Biometrika. 1967;54:1–24.
 17.
Laurila K, Autio R, Kong L, Närvä E, Hussein S, Otonkoski T, et al. Integrative genomics and transcriptomics analysis of human embryonic and induced pluripotent stem cells. BioData Mining. 2014;7:32.
 18.
Smola AJ, Schölkopf B. A tutorial on support vector regression. Stat Comput. 2004;14:199–222.
 19.
Cortes C, Vapnik V. Supportvector networks. Mach Learn. 1995;20:273–97.
 20.
Vanneschi L, Farinaccio A, Mauri G, Antoniotti M, Provero P, Giacobini M. A comparison of machine learning techniques for survival prediction in breast cancer. BioData mining. 2011;4:12.
 21.
Breiman L. Bagging predictors. Mach Learn. 1996;24:123–40.
 22.
Breiman L. Random forests. Mach Learn. 2001;45:5–32.
 23.
Liaw A, Wiener M. Classification and Regression by randomForest. R News. 2002;2:18–22.
 24.
Boulesteix AL, Janitza S, Kruppa J, König IR. Overview of random forest methodology and practical guidance with emphasis on computational biology and bioinformatics. Wiley Interdisciplinary Reviews: Data Mining Knowledge Discovery. 2012;2:493–507.
 25.
Lin Z, Vicente Gonçalves CM, Dai L, Lu HM, Huang JH, Ji H, et al. Exploring metabolic syndrome serum profiling based on gas chromatography mass spectrometry and random forest models. Anal Chim Acta. 2014;827:22–7.
 26.
Kane MJ, Price N, Scotch M, Rabinowitz P. Comparison of ARIMA and Random Forest time series models for prediction of avian influenza H5N1 outbreaks. BMC Bioinformatics. 2014;15:276.
 27.
Boyd WC: Fundamentals of immunology. Fundamentals Immunol 1946.
 28.
O’Shea JJ, Paul WE. Mechanisms underlying lineage commitment and plasticity of helper CD4+ T cells. Science. 2010;327:1098–102.
 29.
Akira S, Uematsu S, Takeuchi O. Pathogen recognition and innate immunity. Cell. 2006;124:783–801.
 30.
Alberts B. Molecular biology of the cell. 4th ed. New York: Garland Science; 2002.
 31.
Mosmann TR, Sad S. The expanding universe of Tcell subsets: Th1, Th2 and more. Immunol Today. 1996;17:138–46.
 32.
Groux H, O’Garra A, Bigler M, Rouleau M, Antonenko S, de Vries JE, et al. A CD4+ Tcell subset inhibits antigenspecific Tcell responses and prevents colitis. Nature. 1997;389:737–42.
 33.
Murray PJ, Wynn TA. Protective and pathogenic functions of macrophage subsets. Nat Rev Immunol. 2011;11:723–37.
 34.
Mangan PR, Harrington LE, O’Quinn DB, Helms WS, Bullard DC, Elson CO, et al. Transforming growth factorβ induces development of the TH17 lineage. Nature. 2006;441:231–4.
 35.
Korn T, Mitsdoerffer M, Croxford AL, Awasthi A, Dardalhon VA, Galileos G, et al. IL6 controls Th17 immunity in vivo by inhibiting the conversion of conventional T cells into Foxp3+ regulatory T cells. Proc Natl Acad Sci. 2008;105:18460–5.
 36.
Kohno K, Kataoka J, Ohtsuki T, Suemoto Y, Okamoto I, Usui M, et al. IFNgammainducing factor (IGIF) is a costimulatory factor on the activation of Th1 but not Th2 cells and exerts its effect independently of IL12. J Immunol. 1997;158:1541–50.
 37.
Kitano H. Computational systems biology. Nature. 2002;420:206–10.
 38.
Kreeger PK, Lauffenburger DA. Cancer systems biology: a network modeling perspective. Carcinogenesis. 2010;31:2–8.
 39.
Noble D. Modeling the heartfrom genes to cells to the whole organ. Science. 2002;295:1678–82.
 40.
Kitano H. Systems biology: a brief overview. Science. 2002;295:1662–4.
 41.
Kitano H, Funahashi A, Matsuoka Y, Oda K. Using process diagrams for the graphical representation of biological networks. Nat Biotechnol. 2005;23:961–6.
 42.
Foster I, Kesselman C: The Grid 2: Blueprint for a new computing infrastructure. San Fransisco, CA: Elsevier; 2003.
 43.
Hucka M, Finney A, Sauro HM, Bolouri H, Doyle JC, Kitano H, et al. The systems biology markup language (SBML): a medium for representation and exchange of biochemical network models. Bioinformatics. 2003;19:524–31.
 44.
Machado D, Costa RS, Rocha M, Ferreira EC, Tidor B, Rocha I. Modeling formalisms in systems biology. AMB Express. 2011;1:1–14.
 45.
Gillespie DT. A rigorous derivation of the chemical master equation. Physica A: Statistical Mechanics and its Applications. 1992;188:404–25.
 46.
Mancuso JJ, Cheng J, Yin Z, Gilliam JC, Xia X, Li X, et al. Integration of multiscale dendritic spine structure and function data into systems biology models. Front Neuroanat. 2014;8:130.
 47.
Buganza Tepole A, Kuhl E. Computational modeling of chemobiomechanical coupling: a systemsbiology approach toward wound healing. Comput Methods Biomech Biomed Engin. 2014;24:1–18.
 48.
Mei Y, Carbo A, Hontecillas R, Hoops S, Liles N, Lu P, et al. ENISI MSM: A novel multiscale modeling platform for computational immunology. In: Bioinformatics and Biomedicine (BIBM), 2014 IEEE International Conference on. 2014.
 49.
Hoops S, Sahle S, Gauges R, Lee C, Pahle J, Simus N, et al. COPASIa COmplex PAthway SImulator. Bioinformatics. 2006;22:3067–74.
 50.
Craven MW, Shavlik JW. Using neural networks for data mining. Futur Gener Comput Syst. 1997;13:211–29.
 51.
Lu H, Setiono R, Liu H. Effective data mining using neural networks. Knowledge Data Eng IEEE Trans. 1996;8:957–61.
 52.
Dayhoff JE, DeLeo JM. Artificial neural networks. Cancer. 2001;91:1615–35.
 53.
Ling H, Samarasinghe S, Kulasiri D. Novel recurrent neural network for modelling biological networks: oscillatory p53 interaction dynamics. Biosystems. 2013;114:191–205.
 54.
Snow PB, Smith DS, Catalona WJ. Artificial neural networks in the diagnosis and prognosis of prostate cancer: a pilot study. J Urol. 1994;152:1923–6.
 55.
Lek S, Guégan JF. Artificial neural networks as a tool in ecological modelling, an introduction. Ecol Model. 1999;120:65–73.
 56.
Brusic V, Rudy G, Harrison LC: Prediction of MHC binding peptides using artificial neural networks. Complex systems: mechanism of adaptation 1994:253–260
 57.
White H. Learning in artificial neural networks: A statistical perspective. Neural Comput. 1989;1:425–64.
 58.
Hagan MT, Menhaj MB. Training feedforward networks with the Marquardt algorithm. Neural Networks, IEEE Trans. 1994;5:989–93.
 59.
Lagaris IE, Likas A, Fotiadis DI. Artificial neural networks for solving ordinary and partial differential equations. Neural Networks, IEEE Trans. 1998;9:987–1000.
 60.
Günther F, Fritsch S. Neuralnet: Training of neural networks. R J. 2010;2:30–8.
 61.
Ihaka R, Gentleman R. R: a language for data analysis and graphics. J Comput Graph Stat. 1996;5:299–314.
 62.
Dimitriadou E, Hornik K, Leisch F, Meyer D, Weingessel A. Misc functions of the Department of Statistics (e1071), TU Wien. R Package. 2008;1:5–24.
 63.
McGeachy MJ, BakJensen KS, Chen Y, Tato CM, Blumenschein W, McClanahan T, et al. TGFbeta and IL6 drive the production of IL17 and IL10 by T cells and restrain T(H)17 cellmediated pathology. Nat Immunol. 2007;8:1390–7.
 64.
Bettelli E, Carrier Y, Gao W, Korn T, Strom TB, Oukka M, et al. Reciprocal developmental pathways for the generation of pathogenic effector TH17 and regulatory T cells. Nature. 2006;441:235–8.
 65.
Matsuoka K, Inoue N, Sato T, Okamoto S, Hisamatsu T, Kishi Y, et al. Tbet upregulation and subsequent interleukin 12 stimulation are essential for induction of Th1 mediated immunopathology in Crohn’s disease. Gut. 2004;53:1303–8.
 66.
Liu X, Li NS, Lv LS, Huang JH, Tang H, Chen JX, et al. A comparison of the performances of an artificial neural network and a regression model for GFR estimation. Am J Kidney Dis. 2013;62:1109–15.
 67.
Panchal G, Ganatra A, Shah P, Panchal D. Determination of overlearning and overfitting problem in back propagation neural network. Int J Soft Comput. 2011;2:40–51.
 68.
Patil RA, Gupta G, Sahula V, Mandal A. Power Aware Hardware Prototyping of Multiclass SVM Classifier Through Reconfiguration. In: VLSI Design (VLSID), 2012 25th International Conference on. IEEE. 2012. p. 62–7.
 69.
Burges CC. A Tutorial on Support Vector Machines for Pattern Recognition. Data Min Knowl Disc. 1998;2:121–67.
 70.
Olson DL, Delen D: Advanced data mining techniques. Berlin and Heidelberg, Germany: Springer; 2008.
 71.
Hastie T, Tibshirani R, Friedman J, Hastie T, Friedman J, Tibshirani R: The elements of statistical learning. Berlin and Heidelberg, Germany: Springer; 2009.
 72.
Cabras S, Castellanos ME, Biino G, Persico I, Sassu A, Casula L, et al. A strategy analysis for genetic association studies with known inbreeding. BMC Genet. 2011;12:63.
 73.
DíazUriarte R, De Andres SA. Gene selection and classification of microarray data using random forest. BMC Bioinformatics. 2006;7:3.
 74.
Segal MR, Barbour JD, Grant RM. Relating HIV1 sequence variation to replication capacity via trees and forests. Stat Appl Genet Mol Biol. 2004;3:Article2.
 75.
Li Y, Zheng SG. The secret of FOXP3 downregulation in the inflammation condition. Int J Clin Exper Pathol. 2012;5:624–5.
Acknowledgements
This work was supported in part by NIAID Contract No. HHSN272201000056C to JBR and funds from the Nutritional Immunology and Molecular Medicine Laboratory (URL: www.nimml.org).
Author information
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
JBR, RH, PL, VA, YM, and AC planned the experiments. PL, YM and VA performed the experiments. RH, SH, and JBR supervised the project. JBR, RH, PL, VA, and YM wrote the manuscript. All authors read edited and approved the final manuscript.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Lu, P., Abedi, V., Mei, Y. et al. Supervised learning methods in modeling of CD4+ T cell heterogeneity. BioData Mining 8, 27 (2015). https://doi.org/10.1186/s1304001500606
Received:
Accepted:
Published:
Keywords
 Support Vector Machine
 Artificial Neural Network
 Random Forest
 Neural Network Model
 Artificial Neural Network Model