Research Article Volume 9 Issue 3
Target classification using machine learning approaches with applications to clinical studies
Chen Qian,1,2 Jayesh P. Rai,1 Jianmin Pan,1 Aruni Bhatnagar,3 Craig J. McClain,3,4,5,6 Shesh N. Rai1,2,5,6
Regret for the inconvenience: we are taking measures to prevent fraudulent form submissions by extractors and page crawlers. Please type the correct Captcha word to see email ID.
1Biostatistics and Bioinformatics Facility, James Graham Brown Cancer Center, University of Louisville, USA
2Department of Biostatistics and Bioinformatics, University of Louisville, USA
3Department of Medicine, University of Louisville, Louisville, USA
4Robley Rex Louisville VAMC, USA
5University of Louisville Alcohol Research Center, University of Louisville, USA
6University of Louisville Hepatobiology & Toxicology Center, University of Louisville, USA
Correspondence: Dr. Shesh N. Rai, Biostatistics and Bioinformatics Facility, James Graham Brown Cancer Center, University of Louisville, Louisville, Kentucky, 40202, USA
Received: May 09, 2020 | Published: June 2, 2020
Citation: Qian C, Rai JP, Pan J, et al. Target classification using machine learning approaches with applications to clinical studies. Biom Biostat Int J. 2020;9(3):91-95. DOI: 10.15406/bbij.2020.09.00305
Download PDF
Abstract
Machine learning has been a trending topic for which almost every research area would like to incorporate some of the technique in their studies. In this paper, we demonstrate several machine learning models using two different data sets. One data set is the thermograms time series data on a cancer study that was conducted at the University of Louisville Hospital, and the other set is from the world-renowned Framingham Heart Study.
Thermograms can be used to determine a patient’s health status, yet the difficulty of analyzing such a high-dimensional dataset makes it rarely applied, especially in cancer research. Previously, Rai et al.1 proposed an approach for data reduction along with comparison between parametric method, non-parametric method (KNN), and semiparametric method (DTW-KNN) for group classification. They concluded that the performance of two-group classification is better than the three-group classification. In addition, the classifications between types of cancer are somewhat challenging.
The Framingham Heart Study is a famous longitudinal dataset which includes risk factors that could potentially lead to the heart disease. Previously, Weng et al.2 and Alaa et al.3 concluded that machine learning could significantly improve the accuracy of cardiovascular risk prediction. Since the original Framingham data have been thoroughly analyzed, it would be interesting to see how machine learning models could improve prediction.
In this manuscript, we further analyze both the thermogram and the Framingham Heart Study datasets with several learning models such as gradient boosting, neural network, and random forest by using SAS Visual Data Mining and Machine Learning on SAS Viya. Each method is briefly discussed along with a model comparison. Based on the Youden’s index and misclassification rate, we select the best learning model. For big data inference, SAS Visual Data Mining and Machine Learning on SAS Viya, a cloud computing and structured statistical solution, may become a choice of computing.
Keywords: machine learning, misclassification, target classification, thermo gram, time series
Introduction
Currently, artificial intelligence (AI) has been discussed in a host of subjects. Similar to data mining and deep learning, machine learning serves as a core branch within the AI, and can be used widely in business, medicine, and statistics, etc. The differences between machine learning and deep learning are often confusing, and some would even think they are the same thing. In fact, they have overlaps, but are actually two different branches. Here, we will discuss and utilize machine learning. Machine learning is a modeling approach that learns and identifies patterns from data, and then makes predictions from those data, with minimal human intervention.
The interpretability of the model is trivial in machine learning, as the main focus is on predictive accuracy.4 To achieve a high predictive accuracy, a model with high complexity would be expected, but that does not mean that more the better. The appropriate complexity must to be chosen in order to construct a good fitting model with the best generalizability. For example, in regression models, more variables do not mean a better model fitting, as the problem of overfitting will likely exist. In contrast, if a model lacks sufficient information to show the true association, then it is underfitted.
In machine learning, data usually would be split into training data and validation data (test data is optional). Training data are used for building models in which the algorithm learns from this set of data. On the other hand, validation data are used to adjust the model which built from the training data for better generalizability, but the model does not learn from this set of data. The model that has the best performance on the validation data will be selected. In our data application examples, both data sets were split into usually 70% training and 30% validation by using random sampling when partitioning. All three methods (gradient boosting, neural network, and random forest) were performed independently for each data set. Models within each method were built based on the training data, and the best model among each method was selected based on the evaluation of the validation data. Comparison between three methods was carried out at the end based on misclassification rates. In addition, auto-tuning was not used in either application, as we only wanted to see the initial assessment of models on classification.
Since the purpose of this manuscript is to show the accuracy of three machine learning models on two data sets, specific algorithms and concepts regarding each machine learning model are not discussed in detail. Algorithms are adopted from the book The Elements of Statistical Learning: Data Mining, Inference, and Prediction and are just for reference5 purposes in this manuscript. Readers can refer to this book for further detail.
Data
Thermogram data
The University of Louisville Institutional Review Board have approved the study protocol and patient consent procedures (IRB# 08.0108, 08.0636, 608.03, 08.0388). All participating patients gave written informed consent for their tissues and blood to be entered a tissue repository (IRB# 608.03, 08.0388) and utilized for research purposes. The IRB specifically approved the use of plasma specimens from the biorepository for use in this study without the need for further consent (IRB# 08.0108, 08.0636).
Plasma samples from 100 healthy individuals with known demographic characteristics were purchased from Innovative Research (Southfield, MI). Cervical cancer specimens were obtained from women with invasive cervical carcinoma attending the clinics of the Division of Gynecologic Oncology. Lung cancer specimens were obtained from patients attending the clinics of the Division of Thoracic Oncology.
The thermogram time series data contained 186 samples, of which 35 were cervical cancer, 54 were lung cancer, and 97 were normal subjects. Details regarding the collection of differential scanning calorimetry (DSC) data can be found in Rai et al.1 For each sample, measurement of heat capacity (HC) values (cal/ °C.g) was made at every temperature point, from 45 degrees to 90 degrees, with 0.1 degree intervals. Negative values, often caused by machine reading errors among low HC values, were not compatible with models, and have been imputed with the next closest positive value to create a continuous curve. For example, if the HC values at 50, 50.1 and 50.2 degrees are 0.02, -0.01, 0.03 respectively, then we impute the negative value at 50.1 degrees with 0.03. The goal of our analysis was to make a three-group classification based on the thermogram data. Other variables were also used in the model including age, gender, and ethnicity.
Framingham heart study data
The Framingham Heart Study data is a very large set, with 5209 samples in total. In this study, our goal was to discern between heart disease and non-heart disease. We only took data from subjects who were dead and had a known cause of death. We treated both cerebral vascular disease and coronary heart disease as “heart disease”, and combined “other” and “cancer” as a second category. After initial preparation, we had 1463 samples in total. Next, we calculated the main artery pressure (MAP) by using variables of diastolic and systolic pressure. With a high MAP (>100), the patient was deemed to have hypertension. Thus, we categorized all samples into two parts: 434 dead with heart problem, and 1029 dead without heart disease. To identify a heart problem, the subject should have both hypertension and have died with heart disease. Other variables used in the model included age, medical record weight (MRW), blood pressure status, cholesterol status, and smoking status.
Methods
Gradient boosting
To understand gradient boosting, one should also know about the AdaBoost (adaptive boosting) algorithm. It is an additive model that starts with a decision tree in which every observation is given the same weight, and after the initial evaluation, the weights change depending on the difficulty to classify, and that all leads to the second tree. The new model is the combination of both trees. The algorithm of AdaBoost can be briefly written as follows:
Let Y be output variable, X be a vector of predictor variables, and G(X) be a classifier that produces a prediction.
- Initialize the observation weights
- For iteration
to M
- Fit a classifier
to the training data using weights
.
- Compute weighted error rate:
- Compute
given to
in producing the final classifier:
- Update the individual weights of each observation for the next iteration.
Set
- Output
The AdaBoost is equivalent to forward stagewise additive modeling, in which the algorithm can be written as:
- Initialize
- For
- Compute
- Set
where
are the expansion coefficients, and
are basis functions of the multivariate argument x, characterized by a set of parameters
to the current expansion
. The squared-error loss function can be further expressed as
where
can be written as
to represent the residual of the current model on the ith observation.
Going forward, a weak learner is introduced to compensate for the shortcomings of the existing weak learners at each stage, and the process is repeated for certain iterations. The final model thus contains the weighted sum of the predictions of all tree models. In gradient boosting, the algorithm identifies the shortcoming of the weak learners by gradient, whereas AdaBoost identifies shortcomings by high-weight data points. The gradient comes from the loss function which is a function that measures how well the predictive model fits when classifying targets. Therefore, the goal of gradient boosting is to add the learner that can maximize the correlation with the negative gradient of the loss function. The algorithm of gradient boosting for K-class classification can be expressed as follow:
Let targets yik coded as 1 if observation i is in class k, and zero otherwise.
- Initialize
- For
- Set the class conditional probabilities
- For
- Compute residual
- Fit a regression tree to the targets
giving terminal regions
.
- Compute
- Update
- Output
.
Neural networks
In neural networks, there are at least three layers: input layer, hidden layer, and target layer. The hidden layer can be adjusted to increase accuracy and improve model fitness, but in our case, we used only one hidden layer. Within each layer, there are neurons that are connected to other neurons in other layers. The size of the neuron indicates the absolute value of the estimated weights, which shows the importance of a certain variable to the target classification. In a model for K-class classification, there are total of K units at the top of the network diagram, with the kth unit modeling the probability of class k. There are K target measurements
, each being coded as a 0 or 1 variable for the kth class.
is defined as the hidden unit in the neural network, and
is called the target that could be modeled as a function of linear combinations of
.
can be calculated from linear combinations of the inputs. These can be expressed in the following way:
where
, and
. The neural network makes a prediction based on the input variables, as it is acting more like a regression model. The activation function
is usually chosen to be the
The output function,
allows a final transformation of the vector of outputs, T. For K-class classification, the softmax function,
, is used. The neural network algorithm works well on almost everything regardless of the relation between inputs and outputs. Once trained, it is a great way of handling large scale of computation.
Random forests
A forest model is built up with a huge number of individual decision trees that can be treated as an ensemble. The final prediction is a combination of the predictions of the ensemble. But unfortunately, decision trees are not stable by themselves. However, with the forest model, the overall performance of the tree is still stable.6 Those trees are uncorrelated, which is a huge advantage when eliminating errors. In other words, when running a large number of trees, the correct trees always outperform the wrong trees. A group of trees can always produce better prediction than a single tree. The workflow of random forests can be written as follow:
- For tree b=1 to B:
- Draw a bootstrap sample
of size N from the training data.
- Grow a random forest tree
to the bootstrapped data, by recursively repeating the following steps for each terminal node of the tree, until the minimum node size is
reached.
- Select m variables at random from the p variables.
- Pick the best variable/split-point among the m.
- Split the node into two daughter nodes.
- Output the ensemble of trees
.
To make a prediction at a new point x, let
be the class prediction of the bth random forest (rf) tree. Then
.
Results
First data application--thermogram
Among the three supervising models, the Neural Network is the best model for this dataset (Table 1). The misclassification rate is 0.0545 which is slightly better than the gradient boosting model (0.0909).
Methods
|
Misclassification Rate
|
Accuracy (1-MR)
|
KS Youden
|
Gradient Boosting
|
0.0909
|
0.9091
|
0.8349
|
Neural Network
|
0.0545
|
0.9455
|
0.8862
|
Random Forests
|
0.2
|
0.8
|
0.6554
|
Table 1 Model comparison for the first data application
In Gradient Boosting, variables with relatively higher importance are age, temperature at 50.1 degrees, temperature at 49.8 degrees, and temperature at 50.3 degrees (Figure 1). Those temperature points are the key variables when differentiating groups. The cutoff value for the receiver operating characteristic (ROC) curve is at 0.07. Similarly, in the Neural Network, the most important variables are temperatures at 49.8 degrees, 49.2 degrees, and 50.1 degrees (Figure 2). Results indicate that temperatures around 50 degrees are more relevant to the types of cancers. The cutoff value is at 0.25. The random forest model shows relatively important temperature points at 50.1 degrees, 49.8 degrees, and 50.3 degrees (Figure 3). This mostly matches the results from the other two models. The cutoff value is at 0.33 for the ROC curve.
Figure 1 Gradient boosting variable importance and ROC plots.
Figure 2 Neural network relative importance and ROC plots.
Figure 3 Random forests variable importance and ROC plots.
Second data application--framingham
Results were generated using SAS Visual Data Mining and Machine Learning on SAS Viya. To determine the best algorithm that fits the data, the misclassification rate and the Kolmogorov-Smirnov (Youden) statistic are used. The lower the misclassification rate, the better the model fits the data. One minus the misclassification rate yields the accuracy. It shows how many samples were correctly classified. On the other hand, the K-S Youden statistic is a goodness-of-fit measurement that represents the maximum distance between the ROC of the model and the ROC of the baseline. Based on the misclassification rate, Gradient Boosting is the best algorithm to fit the dataset (Table 2). The accuracy is 0.8486, which is very good. Neural Network and Random Forests also perform very well, but not as well as Gradient Boosting.
Methods
|
Misclassification Rate
|
Accuracy (1-MR)
|
KS Youden
|
Gradient Boosting
|
0.1514
|
0.8486
|
0.6898
|
Neural Network
|
0.2754
|
0.7246
|
0.5825
|
Random Forests
|
0.2246
|
0.7754
|
0.605
|
Table 2 Model comparison for the second data application
We take a closer look at each of the algorithms in detail. In Gradient Boosting, the highest impact variable is blood pressure status, which is not surprising since blood pressure is highly associated with heart disease (Figure 4). The area under the ROC curve represents the classification accuracy. The bigger the area, the better the accuracy. Dotted lines indicate K-S Youden statistics. In this model, the cutoff value is at 0.02. Neural network also confirms that blood pressure has the highest relative importance (Figure 5). The ROC curve shows that the cutoff value is at 0.34. In the random forest model, blood pressure is again the most important variable (Figure 6). The cutoff value for the ROC curve is at 0.08.
Figure 4 Gradient boosting variable importance and ROC plots.
Figure 5 Neural network relative importance and ROC plots.
Figure 6 Random forests variable importance and ROC plots.
Conclusion & discussion
From both data demonstrations, all three models have proven their abilities to make high accuracy predictions. When comparing the accuracy of the first dataset to results reported in Rai et al.,1 the difference is minimal. In fact, the Neural Network model performs better than DTW-KNN method (Table 3). However, it should be noted that the way the data were pre-processed was slightly different.
Method
|
Accuracy
|
KS Youden
|
Rai Proposed Method
|
0.65
|
N/A
|
KNN
|
0.80
|
N/A
|
DTW-KNN
|
0.80
|
N/A
|
Gradient Boosting
|
0.90
|
0.83
|
Neural Network
|
0.94
|
0.88
|
Random Forests
|
0.80
|
0.65
|
Table 3 Comparison between six classification methods in terms of accuracy in thermogram time series data
This paper only demonstrates three commonly used machine learning models on two-group and three-group classification. Future research could further extend these results to other supervising and non-supervising models on multi-group classification. In our examples, we did not use auto-tuning or use higher tree numbers to achieve a better accuracy. As always, the machine learning process is time consuming.
When making a classification, it is always used on a large-scale data set. In both of our samples, though we have lots of variables, the sample size in each group is still small.
Acknowledgments
C. Qian was supported by the National Institute of Health grant 5P50 AA024337 (CJM) and the University of Louisville Fellowship.
S. N. Rai was partly supported with Wendell Cherry Chair in Clinical Trial Research Fund and NIH grants P20GM113226 and P50AA024337 (CJM).
Disclosure
The authors report no conflicts of interest in this work.
References
- Rai SN, Srivastava S, Pan J, et al. Multi-group diagnostic classification of high-dimensional data using differential scanning calorimetry plasma thermograms. PLoS ONE. 2019;14(8):e0220765.
- Weng SF, Reps J, Kai J, et al. Can machine-learning improve cardiovascular risk prediction using routine clinical data? PLoS ONE. 2017;12(4):e0174944.
- Alaa AM, Bolton T, Di Angelantonio E, et al. Cardiovascular disease risk prediction using automated machine learning: A prospective study of 423,604 UK Biobank participants. PLoS ONE. 2019;14(5):e0213653.
- Breiman L. Random Forests. Machine Learning. 2001;45:5–32.
- Hastie Trevor, Jerome Friedman, Robert Tisbshirani. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. New York: Springer, 2017.
- Breiman L, Friedman J, Olshen R, et al. Classification and Regression Trees. Wadsworth, New York.1984.
©2020 Qian, et al. This is an open access article distributed under the terms of the,
which
permits unrestricted use, distribution, and build upon your work non-commercially.