Healthcare, Machine Learning, Clinical Data, Sensor Data, Omics Data
Health care is a wide term that concerns to a system that involves improvement of medical services in order to serve the medical demands of the people. In healthcare, efforts are made by patients, physicians, vendors, health companies and IT companies for maintaining and restoring health records. Over the past decade, Indian health care is known as one of the fast-rising industry in the world. Healthcare analysis is handling various types of diseases including cancer, diabetes, strokes and so on using machine learning. Cancer is one of the deadliest diseases. Different types of cancer are present in this human word including lung cancer, breast cancer, prostate cancer, stomach cancer and so on. Around 12% cases of lung cancer come every year in which 10% cases died from it. Similarly, for breast cancer, 11% cases come in which 9% dies from breast cancer. This happens in each type of cancer. Cancer prevalence of year 2018 is taken and is shown in Figure 1 below. It describes the total cases and death cases for each cancer type. For handling cancer in health care analysis, there is a need to generate correct and quality data. In this competitive world, healthcare must need to use the data in such a manner that there will always be rise in quality of health care and decline in the cost needed for the treatment purpose . From past years, heath care research with Machine Learning (ML) has been increasing steadily. Due to variety of medical data including clinical data, omics data or EHR data, it is difficult for humans to infer the data and to make decisions. Accordingly, ML has been proposed in health care for better understanding of data and for better decision-making process .
Figure 1. Evolution of the effective thermal conductivity of some monolithic, silica aerogels.
Machine Learning was originated by Samuel in 1950 to play strategic games like chess. It is the mechanism of making machines to learn automatically without being explicitly programmed. The main focus of Machine Learning is to develop a computer program which can access the data and use this data for learning purpose. It is the ability of machine to make use of statistical techniques and advanced algorithms to make more powerful prediction and making the data driven system more powerful by replacing the rule-based system. The main component of machine learning is data which is the backbone for any model. The more relevant data is the more accurate predictions are. After data, we need to select the algorithm based on the problem for more accurate predictions. Machine Learning can be used in many fields such as finance, retail, health care and social data .
Types of machine learning algorithms
Machine learning can be used for different purpose. Machine learning algorithms are basically classified into three categories based on their objective which varies from each other. It includes supervised learning, unsupervised learning and reinforcement learning.
Supervised learning: Supervised learning involves training the model on the labeled data and uses this trained model to make predictions on the new data. It involves splitting of data into two sets including training set and testing set. First the model is trained on training set and afterwards the performance is tested on the testing set. The performance of the model can be evaluated using performance metrics . Supervised learning can be classification problem or regression problem. In supervised classification, the labeled value is a discrete value. The algorithms in this are used to classify to which class or category the problem belongs. On the other side, the models are used to predict the outcome based on continuous (numeric) data is supervised regression learning . For the classification of raw data, first the data is selected and then preprocessing is performed in which all NA values are removed. Then the data is normalized using z-score or min max normalization. Once the normalization is performed feature selection procedure is applied to select the best features. After the features are selected, some supervised ML algorithms includes K Nearest neighbor, Decision trees, Support Vector Machines, Naïve Bays Classifier, Neural Network and Ensemble methods  are used for classification of raw data as shown in Figure 2.
Figure 2. Types of Machine Learning algorithm.
Unsupervised learning: Unsupervised Learning also involves training of the data except for the fact that the labeled value or target value is not known. In this, machine try to cluster the similar type of the data by finding the hidden pattern. Rather than making prediction, the main aim of unsupervised learning is to discover the patterns. The performance of the model in unsupervised learning cannot be evaluated as the label value is absent or unknown. The algorithms involved in unsupervised learning are K-mean clustering, Association Rule Mining, Topic Modeling and Dimensionality Reduction Techniques .
Semi-supervised learning: As supervised learning works on labeled data and unsupervised learning on unlabeled data, then a lot of information is lost from labeled data which can be obtained from unlabeled data. So, in this case semi-supervised learning comes to mind. It is a mixture of supervised and unsupervised learning in which it takes both the unlabeled and labeled data. Labeled data should be of shorter length as compared to unlabeled data. The idea behind semi-supervised learning is that there is a considerable change in performance when both labeled and unlabeled data is used in conjunction. The training set used is of shorter length. It is normally used to detect outliers.
Reinforcement learning: Reinforcement Learning works by developing a system which improves its performance by taking feedback from the environment and taking possible steps to improve them. It is an act of learning from environment by interacting with it without any help from humans. It is an iterative process.
The different types of machine learning algorithms and their applications are shown in Figure 2 above.
As healthcare is emerging now days, researchers are focusing on types of data used for prediction. For Example, Ajay et al. focus on clinical and genomic data and used machine learning algorithms to analyze them. But other data types are also present to work upon including sensor and Omics data. The prime motive of our survey is to include all types of data and analyze them using machine learning. It is described in the Table 1 as follows.
|Ajay Kumar et al.||✓||✓||__||__||__|
Table 1: Comparison with other surveys.
Section 2 presents different type of data used by authors for diagnosis and prevention of certain kind of disease and their work done for achieving it. Section 3 shows conclusion achieved from the related surveys.
Healthcare analysis using ML
As in healthcare sector, there is enormous information about the patient health. So it is impossible for humans to process it. Consequently, ML provides a technique to recognize patterns from the massive data and use algorithms to predict future outcome of the patients. ML in healthcare helps users to perceive understanding about the potency of existing programs and identify the treatment that provides best result for patients according to their condition.
Types of healthcare data
Different types of data have come into view in healthcare now days including clinical data, sensor data, Omics data and so on. This type of data includes different mining methods to extract the more relevant features and then different algorithms needs to be trained for better future prediction.
Clinical data: Clinical data is the data which is collected during the ongoing treatment of the patient including the Electronic Health Record (EHR) data which is comprised of laboratory tests, radiology images, allergies and so on (Figure 3). The work on clinical is applied by following authors.
Figure 3. Types of dataset.
Wengert et al.  proposed ML algorithms for early prediction of pathological complete response (pcr) to neoadjuvant chemotherapy and survival outcome of breast cancer patients using Multiparametric Magnetic Resonance Imaging (mpMRI) data. Samples of 38 women with breast cancer were taken and eight classifiers including linear support vector machine, linear discriminant analysis, logistic regression, random forests, stochastic gradient descent, adaptive boosting and Extreme Gradient Boosting (XGBoost) were applied to rank the features for pcr including residual cancer burden (RCB), Recurrence Free Survival (RFS) and disease-specific survival DSS. Area under Curve value was extracted for each feature of pcr. From the experimental results, XGBoost produces the best result with higher accuracy for RCB and DSS and logistic regression for RSS as compared to other classifiers. Dagli et al.  defined multilevel perception model for two year survival prediction of non-small cell lung cancer patients. Samples of 559 patients were taken and attributes were ranked with RelifF feature selection method. From results, Multilayer Neural Network was found as the best prediction model with area under curve value of 0.75. Kayal et al.  proposed new improved classification approach for survival prediction of Hepatocellular Carcinoma (HCC) patients. Samples of 165 patients were taken from which authors defined that out of 49 risk factors, 15 risk factors were responsible for HCC. The outcome of the experiment proved that the accuracy obtained by Deep Neural Network is significantly higher than Cox models (SVM) and Unsupervised model (KNN).
Zheng et al.  proposed a framework to identify Type- 2 Diabetes Mellitus (T2DM) patients using Electronic Health Record (EHR) data. A total of 300 patient samples were taken and 114 features were extracted on which different machine learning algorithms including k-Nearest Neighbor (kNN), Random Forest (RF), Decision Tree (DT), naïve bayes, Support Vector Machine (SVM) and logistic regression were applied. From results, SVM produces the best result with accuracy of 96%. Sumei et al.  developed computer assistant classification method by combining convolution MRI and profusion MRI data for diagnosis of different type of brain tumor and for grading of gliomas. Samples of 102 brain tumor patient were taken and Support vector machine recursive feature elimination (SVM-RFE), k-nearest neighbor and linear discriminant analysis were applied to them. The result showed that SVM RFE produced the best result with accuracy of 85% for classification of tumor and 88% for grading the gliomas. Kristin et al.  defined different machine learning algorithms including penalized logistic regression, random forest models, and extreme gradient boosted decision trees for identification of high-risk surgical patients. Authors trained the algorithms on Pythia data containing electronic health records having 194 clinical features including patient demographics, smoking status, medications, comorbidities, procedure information, and proxies for surgical patients. The experimental results show that the best result was produced by penalized logistic regression model with AUC value of 0.924. Andrew et al.  investigated five machine learning algorithms comprising penalized logistic regression, gradient boosting machine, artificial neural network with a single hidden layer, linear support vector machine and random forest for delirium risk prediction based on electronic health record data. A total of 18223 patient samples were taken and experiment was performed. From results, it was proved that gradient boosting algorithm produced the best result with AUC value of 0.855. Fatemeh et al.  proposed machine learning models for first emergency admission prediction based on EHR data. Authors applied Cox model on a sample of 4.6 million patient samples for prediction of risk for first emergency admission and then random forest and gradient boosting algorithm were used. Authors identified that gbm model performed best with AUC value of 0.779. Maryam et al.  investigated Seattle heart failure model for the prediction of heart failure by using EHR data. Samples of 5044 patient were taken and features were extracted to calculate the survival score. Authors first calculated the survival score of heart patients with Cox proportional regression model who survived for one, two or five years and then the patient who died after five years were excluded and different machine learning models comprising random forest, logistic regression, support vector regression, decision tree and ada boost were applied on the remaining patients. From experiment results, logistic regression performed best with 11% improvement in AUV curve value. Stephen H.
Weng et al.  defined machine learning algorithms including random forest, logistic regression, gradient boosting machines and neural networks on samples of 378,256 patients for the prediction of cardiovascular risk. After data was prepared and features were extracted. Authors applied the different machine learning algorithms and identified that neural network performed best with AUC value of 0.72 as shown in Table 2.
|Type of health
|Wengert et al.||Clinical (mpMRI)||Support vector machine,linear discriminant analysis,logostic regression,random forests,stochastic gradient descent,adaptive boosting,extreme gradient boosting (XG Boost)||Area under curve||XG Boost produced the best result with AUC value of 0.94 for RCB and 0.92 for DSSwith AUC value of 0.83||The dataset used in this way very small. In this some features are extracted which effect the imaging features as well as the prediction of RCB.DSS. So this can be covered in future.|
|Yash Dagali etal. ||Clinical||Multilayer Neural Network,Logistic Regression,Single Perception neural network||Area under curve,95% confidence interval,Misclassification rate,True positive rate,false positive rate,accuracy and precision||Multilayer Neural Network produced the best result with AUC value of 0.75,confidence value of 0.693-0.806,true positive rate of 0.68,false positive rate of, accuracy of 0.76 and precision value of 0.72|
|Chayan Kuma Karvey et al.||Clinical||Deep Neural Network, Support Vector Machine, K-Nearest Neighbor||Accuracy, Precision, Recall, Fmeasure||Deep neural network produced higher accuracy of 78% and precision,Recall and Fmeasure value of 83.58, 81.25 and 80%||In future,focus should be on appropriate feature selection method for efficient survival prediction|
|Tao Zheng et al.||Clinical (EHR)||Support vector machine,
random forest decision
tree, naïve bayes
|SVM produced best result
with accuracy 96%,
96%,precision 91% and
AUC value of 0.96.
|Dataset used was very
small. Large sample
can be used for better
|Sumei Waang et al.||Clinical (MRI)||Support vector machine recursive feature
|SVM RFE produced best result for both classification
of tumor and grading of
gliomas with accuracy,
sensitivity and specificity
value of 85%, 87%, 79% and
88%, 85% and 96%.
The Nf value is 20, entropy
and sd is 0.82 and 0.92
|Dataset used was very small. Large sample
can be used for better
|Kristin M. Korey et al.
|Clinical (HER)||Penalized logistic regression, random forest
gradient boosted decision
trees basis function
|Penalized logistic regression produced best result with
and ppv value of 95%, 76%,
76%, 0.924, 0.174 and 0.390
|The effect of Pythia risk calculator needed
to be evaluated for
|Andrew Wong et al. ||Clinical (EHR)||Penalized logistic regression,Gradient boosting machine,
Artificial neural network
with a single hidden
layer,Linear support vector machine and random forest
Area under curve
|Gradient boosting machine produced best result with
sensitivity,specificity and AUC value of 59.7%, 23.1%
|Fatemeh Rahimian et al. ||Clinical (EHR)||Cox model, Gradient boosting, Random forest||Area under curve, confidence interval||Gradient boosting machine produced best result with
AUC and 95% CI value of
0.779 and 0.847
|Better techniques can be used for risk
|Maryam et al. ||Cox model,Gradient boosting, Random forest Support vector
regression, Decision tree,
|Area under curve||Logistic regression performed best with an
improvement of 11% in AUC
|StephenH Weng et al. ||Clinical (EHR)||Random forest, Logistic regression,Gradient boosting machines and Neural networks||Area under curve, confidence interval, positive predictive and negative predictive value||Neural networks produced best result with AUC,CI, PPV and NPV value of 0.728, 0.75-0.76, 18.4% and 95.70%||Complexity may lead to overfitting. Best models needed for better results|
Table 2: Summarized clinical data analysis using Machine Learning.
Sensor data: Data elements produced by sensors including time series signals which is an ordered sequence of pairs is sensor data. These data elements are processed by computing devices and can be simple numerical or categorical value or can be more complex data. The work on sensor data is applied by following authors. Luca et al.  proposed machine learning algorithms to detect Parkinson’s disease (PD) by using data streams collected from wearable sensors.
Experiment was performed on 20 individuals and movement of individual was recorded by 6 wearable sensors. Total 13 tasks were performed by individuals and experiment was conducted on one day and was repeated 2 weeks later. From this a total of 41,802 data clips were used. After the data was trained using convolutional neural networks and random forest classifier for the detection of bradykinesia and tremor. Results proved that random forest classifier performed better with AUROC value of 0.73 for the detection of bradykinesia and 0.79 for the detection of tremor.
David et al.  defined machine learning classification algorithms for detection of risk of developmental arrays (AD) and Typical Development (TD) in infants. Long day inertial movement of infants were recorded using Opal sensors fixed on the ankle of the infant and data was divided into two sets, 0 to 6 months and 6 to 12 months. A total of 19 movement features including movement count, duration, average acceleration and peak acceleration from two sets were extracted using univariate feature selection methods which were Recursive Feature Elimination (RFE), and stepwise feature selection. Authors used three machine learning algorithms support vector machine, logistic regression and adaboost for prediction and the outcome of the result proved that SVM performed best for 0-6 month infants with accuracy of 90% and adaboost for 6-12 month infant with accuracy of 83%. Sota et al.  proposed shoe-type pressure sensor and single inertial measurement unit attached to the trunk for detection of assistance motion with different foot. A total of 8 Flexiforce sensors were attached to the sole. 5 people were asked to perform the experiment and were asked to perform in two variations including short step and long step. Features were extracted from the data obtained and were trained using classification method. Experimental results show that proposed system performed best with accuracy of 90%. Prabhjot et al.  investigated hybrid approach comprising Bayesian network and heuristic technique in neural network for stress detection using mobile phone sensing mechanism by measuring the Blood Pressure Management (BPM) and Heart Rate (HR) value. Data was collected using sensors embedded in mobile phones and hybrid approach was applied to detect stress using BPM values and HR values as shown in Table 3. From result, hybrid approach performed well with accuracy of 92.86% for BPM and 85.71% for HR. Shamsul et al.  proposed Deep-belief network for recognition of human activity using data from body sensors. Sensor data was collected and important feature were extracted using Kernel Principle Component Analysis (KPCA) and Linear Discriminant Analysis (LDA). Then, the model was trained using deep-belief network with 40 hidden layers. From results, it is cleared that deep belief network performed best for activity detection with an accuracy of 97.5%. Elimen et al.  defined hybrid approach comprising Convolutional Neural Network and Long Short-term Memory Recurrent Neural Network (CNN-LSTM) for emotion detection using smart phones and wearable sensor devices data. Sample of 40 female patients were taken from which 550,432 sensor data values were collected comprising of on-body data, environmental data and self-report emotion level data captured using mobile phone app. Then data was preprocessed and trained using hybrid CNN-LSTM model for emotion detection. The outcome of the result proved that the proposed hybrid approach performed best with accuracy of 95%. Diana et al.  investigated four machine learning classifier including decision trees, ensemble, logistic regression and Deepnets for the detection of fall in elderly people using 3D-axis accelerometer fitted in 6lowPAN wearable device. The accelerometer reading was collected and feature was extracted with sliding window technique. Fall was detected using machine learning classifiers and the outcome of result proved that ensemble algorithm performs best with accuracy of 94%. Jessica et al.  proposed 90 second fear induction task to measure the motion of participant using a wearable sensor for the detection of anxiety and depression among young children. Samples of 64 children were taken and they were subjected to 20 second potential threat phase. Data was collected after 20 second threat phase and features were extracted from sensor data. Authors then subjected the data to k-nearest neighbor model and proved that the proposed model produced best result with an accuracy of 75%.
|Type of health care data||ML algorithms||Performance parameters||Results||Future scope|
|Luca Lonini et al ||Wearable sensors data from sensors attached to arms, hands and thighs||Random Forest, Convolutional neural network||Area under ROC curve, confidence interval, Standard Deviation||Random forest classifier performed better with AUROC, CI and SD value of 0.73, o.68-0.77 and 21.7% for the detection of bradykinesia and 0.79, 0.74-0.84 and 11.5% for the detection of tremor.||Due to limited number of individuals, it was difficult to detect dyskinesia. So this can be a research area in future.|
|David Goodfello w et al ||Wearable sensors data collected from sensors attached to ankle of infrant||Support vector machine, Logistic regression, Adaboost||Accuracy precision Recall, F1 score||SVM performed best with accuracy, precision, recall and F1 score value of 90%, 92%, 90% and 90% for 0-6 month infant and adaboost performed best for 6-12 month infant with accuracy, precision, recall and F1 score value of 83%, 83%, 84%, 83%.||Future studies were required to determine the robustness of classifier and for better prediction of typical development|
|Sota Nocano et al. ||Shoe type pressure sensor data||Classification methods||Accuracy||Proposed system performed best with 80% accuracy.||For better results, appropriate foot position is necessary. So this can be done in future.|
|Prabhjot Kaur et al. ||Wireless sensor using mobile phone||Hybrid approach including Bayesian network and heuristic technique in neural network||Accuracy||Hybrid approach performed well with accuracy of 92.86% for BPM and 85.71% for HR.||Detection of stress using real time sensor data can be a topic for future study.|
|Shamsul Huda et al. ||Body Sensors||Support vector machine, Deep belief network||Accuracy||Deep belief network performed best with an accuracy of 97.5%||Parallelism was not considered in this approach. So this can be a future research.|
|Elimen Kanjo et al. ||Physiological, environmental and location data using mobile phones and wearable devices||Convolutional neural network, Long short-term memory recurrent neural network||Accuracy, Precision, Recall, F1 score, RMSE Error rate||Hybrid CNN-LSTM performed best with Accuracy, Precision, Recall, F1 score, and RMSE value of 92%, 95%, 94%, 94% and 29%.|
|Diana et al. ||Data collected using 3D axis accelerometer fitted in 6lowPAN wearable device||Logistic regression, Ensemble, Deepnet, Decision tree||Accuracy, Precision, Sensitivity, specificity, Gain||Ensemble performed best with accuracy, Precision, Sensitivity, specificity and gain value of 92%, 92%, 92% and 67% avg gain.||For more benefits, integration of more sensors and developments of new services related to health can be considered as research topic.|
|Jessica Hruschak et al. ||Data collected from wearable sensor||K nearest neighbour||Accuracy||KNN performed best with 75% accuracy.||Future studies should consider additional feature selection techniques to reduce the number of features.|
Table 3: Summarized sensor data analysis using Machine Learning.
Omics data: Omics data is collection of huge amount of complex and high dimensional data consisting of genomic, transcriptomic and proteomics data. Handling this type data required various techniques including machine learning algorithms.
Genomic data: Genomic data is collection of gene expression, copy number variation, sequence number and DNA data and is used in bioinformatics. The work on genomic data is applied by following authors. Patrick et al.  proposed machine learning algorithms for improving hazard characterization in microbial risk assessment. Because of high dimensionality of genomics data, authors defined ML based predictive risk modelling for risk assessment. Dataset related to DNA isolation and sequencing were collected and feature extraction was performed to extract the relevant features. Machine Learning classifiers including random forest, support vector machine, logic boost were applied and results were evaluated. From results, it was proved that logic boost performed best with an accuracy of 75%. Yaron et al.  proposed DeepGestalt, a deep learning framework for identification of facial phenotypes of genetic disorders. Samples of 17000 patients with 200 syndromes were taken. Features were extracted and DeepGestalt was applied in which face detection was done using deep convolutional neural network (DCNN) and then image is normalized and cropped into different segments which is then converted to grey scale. After Gestalt model was trained and predict the syndrome with 91% accuracy. Marcus et al.  investigated machine learning algorithm XGBoost for prediction of minimum antimicrobial concentration among patients. Samples of 5278 non-typhoidal Salmonella genomes were collected. Short read sequenced data was collected for each strain with genome assembled service and XGBoost was applied which used gradient boosting ensemble method to reduce the error. The outcome of the result proved that XGBoost produced best result with accuracy of 95%. Kumardeep et al.  defined deep learning model and six machine learning algorithms comprising random forest, support vector machine, linear discriminant analysis, prediction analysis for microarrays, recursive partitioning and regression trees and generalized boosting model for prediction of estrogen receptor status in breast cancer patients based on metabolomics data. Samples of 271 patients were taken in which 204 patients are with positive estrogen receptor and 67 with negative receptor. K-nearest neighbor was used for normalization of data. The normalized data was trained using machine learning and deep learning algorithm. From experimental results, it was proved that deep learning algorithm performed best with AUC value of 0.93.
Transcriptomic data: Transcriptomic data is a collection of multiple mRNA transcripts data within a biological sample. These samples are analyzed and extracted to generate different datasets. The work on transcriptomic data is applied by following authors. Carly et al.  proposed a framework to integrate multiple gene expression datasets to identify gene signatures for the diagnosis of tuberculosis. Samples of 1164 patients were taken by integrating 4 datasets. Features were extracted and machine learning algorithms including random forest, support vector machine with polynomial kernel and Partial least square discriminant analysis applied and results were evaluated. From results, it was proved that random forest performed best with an accuracy of 95%. Suhas et al.  proposed a hybrid approach comprising a deep unsupervised single cell clustering which integrates the feature generated by deep learning model for profiling of single-cell RNAsequencing data. Samples were taken and features were extracted. Model was trained and the proposed model performed the best result with accuracy of 96%. Marin et al.  investigated machine learning algorithm for tracking age related changes of human muscle skeleton on transcriptomic data. Gene-expression profiles of donor were analyzed to compare signatures of old and young donors. Machine learning algorithm comprising neural network was applied on signature data which built a biomarker for aging. The outcome of the result proved that proposed technique produced best result with accuracy of 80%.
Proteomic data: Proteomic data is a collection of proteins expressed in the form cell, tissue or an organism. It is the representation of actual functional molecules in the cell. The work on proteomic data is applied by following authors. Christine et al.  proposed deep learning algorithms for the analysis of FLT3-ITD in acute leukemia patients. Samples of 191 patients with protein data were taken which have serum level of 231 patients. Deep learning with stacked auto- encoders was used and dimensionality reduction reduces the proteins from 291 to 20. From results, it was proved that the proposed model performed best with accuracy of 97% as shown in Table 4. Babita et al.  proposed a hybrid space for the prediction of protein structure class. A hybrid approach including SkipGram based word2hovac and Atchleys space II, III, IV for electron ion interaction were applied for amino acid sequence representation . For feature extraction of time and frequency domain, Stockwell transformation was applied. It was applied on six datasets including small sized samples comprising 498, 277 and 204 and large sized samples comprising PDB25, 640 and FC699. Deep recurrent neural network was used for classification. The result proved that proposed approach performed best with accuracies of 95.9%, 94.9%, 85.36%, 84.2%, 94.3% and 93.1% for both small sized and large sized datasets.
|Type of health care data||ML algorithms||Performance parameters||Results||Future scope|
|Patrick Murigu et al.||Genomic data (Sequencing and DNA isolation)||Support vector, machine, Random forest, Logic boost||Accuracy, sensitivity, specificity, confidence interval, positive and negative predictive value||Logic Boos produced the best result with accuracy, sensitivity, specificity, PPV, NPV and CI value of 75%, 60%, 86%,0.25, 0.95 and 0.60-.80.|
|Yaron Gurovich et al. ||Genomic||Deep Gestalt||Accuracy||DeepGestalt correctly identified syndrome with 91% accuracy.||Dataset used was very work small work on large dates was required for better predictions|
|XGBoost produced the best
Result accuracy and
quartile bound value of 95%
|In future, subtle
Genomic changes must
identified that result in different MIC.
|random forest, support
vector machine, linear
prediction analysis for
partitioning and regression
tree sand generalized
boosting model, deep
|Deep learning produced best
result with AUC value of
identification of DL
methods was not there.
This can be a research
topic in future.
|random forest, support
vector machine with
polynomial kernel and
Partial least square
|RF produced best result with accuracy, sensitivity, specificity and AUC value of
95%, 89%, 97%, 97%.
|Additional dataset may be included for better
et al. 
|Single-cellRNA sequencing data||Deep unsupervised single
|Accuracy||Proposed technique produced best result with accuracy of
vector machine, Elastic net, Deep feature selection
coefficient of determination mean average
|Deep feature selection
produced best result with
accuracy of 95,0.96,0.92 and
|Proposed technique can
be applied to further
|Christine A. Liang
|Protein Data||Deep neural network with
|Proposed technique produced
best result with accuracy, sensitivity and specificity value of 97%, 90% and 98%.
|This technique can be used in further research to determine chemotherapy response.|
|Babita Majhi et
|Small sized and large sized
|Recurrent neural network||Accuracy,
recall, f1 score
performed best with accuracies of 95.9%, 94.9%,
85.36%,84.2%,94.3%and 93.1% for both small sized
and large sized datasets.
|Another factors of
Atchleys may be
considered for better prediction. This is work of future studies.
Table 4: Summarized Omics data analysis using Machine Learning.
A different type of data is present in healthcare. To analyze this variety of data various Machine learning algorithms such as supervised, unsupervised and reinforced algorithms are used to improve prediction which can be analyzed using various performance parameters like accuracy, sensitivity, specificity, precision, F1 score, and Area under Curve. In this paper, machine learning algorithms are defined and use of machine learning algorithms for analyzing different types of healthcare data like clinical, omics and sensor data is done. From the survey, it is concluded that for analyzing different types of data in healthcare, various machine learning algorithms and feature extraction techniques are proposed by various authors for survival prediction of cancer patients.
I am thankful to Dr. Ashima Singh for her assistance in preparing article
Arwinder Dhillon designed the study, contributed to the study, contributed in analyzing the data and also wrote the paper. Dr. Ashima Singh read, help in required changes and approved the final manuscript.
The authors declare no potential conflicts of interests with respect to the authorship and/or publication of this paper.
Received Date: Jun 27, 2019 / Accepted Date: Jul 25, 2019 /
Copyright: © 2019 Arwinder Dhillon et al. This is an open access paper distributed under the Creative Commons Attribution License. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Competing interests: The authors have declared that no competing interests exist.