TSP Cmes 43921

Download as pdf or txt
Download as pdf or txt
You are on page 1of 34

Computer Modeling in

Tech Science Press


Engineering & Sciences

DOI: 10.32604/cmes.2021.016728

REVIEW

Multi-Disease Prediction Based on Deep Learning: A Survey

Shuxuan Xie, Zengchen Yu and Zhihan Lv*

College of Data Science Software Engineering, Qingdao University, Qingdao, 2660712, China
*
Corresponding Author: Zhihan Lv. Email: lvzhihan@qdu.edu.cn
Received: 20 March 2021 Accepted: 06 May 2021

ABSTRACT
In recent years, the development of artificial intelligence (AI) and the gradual beginning of AI’s research in the
medical field have allowed people to see the excellent prospects of the integration of AI and healthcare. Among
them, the hot deep learning field has shown greater potential in applications such as disease prediction and drug
response prediction. From the initial logistic regression model to the machine learning model, and then to the
deep learning model today, the accuracy of medical disease prediction has been continuously improved, and the
performance in all aspects has also been significantly improved. This article introduces some basic deep learning
frameworks and some common diseases, and summarizes the deep learning prediction methods corresponding
to different diseases. Point out a series of problems in the current disease prediction, and make a prospect for the
future development. It aims to clarify the effectiveness of deep learning in disease prediction, and demonstrates the
high correlation between deep learning and the medical field in future development. The unique feature extraction
methods of deep learning methods can still play an important role in future medical research.

KEYWORDS
Deep learning; disease prediction; Internet of Things; COVID-19; precision medicine

1 Introduction
With the progress of society, people’s living habits and environmental conditions are gradually
changing, which invisibly increases people’s hidden dangers of various diseases. Major diseases
such as Brain Health, Cardiovascular Diseases, Vision Impairment, Diabetes, Cancer, etc., have
a serious impact all over the world. Looking at diabetes alone, there are 422 million people in
the world who are troubled, and Type 2 diabetes patients account for more than 90% [1]. With
the increase of age, the heart senescence and loss of function make the risk of heart diseases
increase. Heart disease deaths account for more than 30% of global deaths [2]. These major
diseases seriously affect human health and reduce personal productivity. At the same time, it will
aggravate social pressure and increase medical and health care expenditures. The main purpose
of disease prediction is to predict the risk probability of an individual suffering from a certain
disease in the future. There are a large number of influencing factors that need to be taken into
account for diverse diseases in different populations. Including a set of features with extremely
wide dimensions that need to be detected, as well as complex and variable individual and disease

This work is licensed under a Creative Commons Attribution 4.0 International License,
which permits unrestricted use, distribution, and reproduction in any medium, provided the
original work is properly cited.
490 CMES, 2021, vol.128, no.2

differences need to be distinguished. Simply wanting to accomplish these tasks manually is not
only thorny, but also consumes a lot of human and financial resources.
The development of society has promoted the progress of technology. It makes the acquisition
of medical data sets easier and easier. Medical institutions all over the world are collecting
various health-related data, and the scale is constantly expanding. The types of medical data
include basic patient information, electronic health records (EHR) [3], electronic medical records
(EMR) [4], image data, medical instrument data, etc. Although they belong to one type of
basic data, they are extremely vulnerable to impact. For example, incomplete or redundant data
caused by its timeliness, or personal security reasons make the data untrue and privacy exposure
issues. In addition, medical data is different from ordinary data in complexity, high-dimensionality,
heterogeneity, irregularity, and involves more unstructured data, which is difficult to achieve by
simply handling these data manually. Moreover, medical diagnosis in reality is still limited by many
factors, such as medical conditions, the level of the diagnostician, and the differences between
different patients. Based on the above problems, people urgently need some auxiliary methods to
help predict and diagnose diseases.
Among the many methods of studying complex data, machine learning has developed rapidly,
and the deep learning part of it has suddenly emerged and has become the most eye-catching
branch in the field of machine learning. The ability of deep learning to process complex data,
the method of extracting the main features of multi-dimensional data, the efficient response to
unstructured data, and the classification strategy with higher accuracy are all superior to the
previous technical methods. This also allows more people to begin to understand and step into
the field of deep learning. In recent years, the development of deep learning technology has
become more and more mature. It has brought disruptive changes in many fields such as image
recognition, speech recognition, and natural language processing. At the same time, the application
of deep learning to disease prediction has gradually attracted people’s attention and has achieved
many impressive results [5]. The characteristics of medical data are highly compatible with the
deep learning model, which can better complete various tasks.
The foundation of deep learning is artificial neural network (ANN), which is a structure
composed of multiple neurons according to a certain level, also known as a perceptron. With
the changes in people’s demand for models, the expansion of data scale and the improvement
of computing power, the neural network has gradually evolved from a shallow model to a deep
model, and structurally becomes a multilayer perceptron, it also called deep neural network
(DNN). A deep neural network includes an input layer, a hidden layer, and an output layer. The
core of the traditional multi-layer perceptron neural network is the back propagation algorithm
during training, as shown in Fig. 1.
It includes the procedure of forward propagation of information and the process of optimiz-
ing loss function through back propagation. Visually speaking, it is like letting a computer build
a series of programs similar to the human brain, exercising the computer has human intelligence,
so that it can simulate the principles of human vision, hearing, and other intelligent behaviors,
and evolve in the process of continuous practice. Similarly, when predicting diseases, the computer
can be used to imitate medical experts to diagnose diseases and accumulate prior experience in
the continuous practice process to improve the accuracy of prediction, thereby making the model
more robust. For a variety of medical data and problems, DNN has gradually evolved a variety
of network structure models to deal with [6], which will be explained in the next section.
CMES, 2021, vol.128, no.2 491

Figure 1: Deep neural network


In the follow-up content of this article, the second part first introduces some popular deep
learning models and methods, and briefly introduces some improved model frameworks, leading
to some basic medical problem solving strategies. For different diseases, the third part briefly
demonstrates the concept of diseases, overviews the deep learning methods used in predicting
diseases, and summarizes the experiences and improvements. The fourth part will point out some
of the problems and challenges that exist when the current deep learning methods are combined
with disease prediction. The fifth part introduces the extension of the intelligent fusion of the two
under the current technological development conditions, and prospects for the future application
prospects. The sixth part summarizes the content of the full text.

2 Deep Learning Model


2.1 CNN
When a neural network containing only a fully connected layer reaches a deeper level, the
amount of parameters will gradually increase, which will increase the memory footprint and
increase the computational cost. In order to overcome the problems caused by some fully con-
nected layers and promote the development of neural networks to a deeper level, a solution has
been gradually explored. By using the idea of local correlation and weight sharing, a convolutional
neural network (CNN) is designed, which not only greatly reduces the number of parameters,
but also greatly improves the training efficiency. The proposal of LeNet in 1998 marked the
birth of CNN and defined the basic framework of CNN. Fig. 2 shows the basic structure,
including convolutional layer, pooling layer, fully connected layer. And the architectural changes
after convolution and pooling, at the end, get output after fc.
However, due to the influence of computer computing power and data volume at that time,
it has not been developed to a greater extent. Until 2012, the overcoming of some objective
factors, coupled with the excellent results of CNN in some international competitions, CNN
quickly became the new favorite method in the field of computer vision. After that, improved
CNN models and CNN variant models starting occurrence, such as VGG net [7], ResNet [8],
DenseNet [9], Xception [10], etc. As shown in Tab. 1, the basic information of several frameworks
is introduced. Although CNN was originally invented for computer vision, in subsequent studies
it has gradually proved that CNN is also quite effective in other fields [11,12], and the prominent
492 CMES, 2021, vol.128, no.2

directions include medicine. Imaging classification, disease prediction [13,14], medical image visu-
alization [15], etc. CNN can process many types of data, the most prominent is the processing of
pictures with spatial local correlation data, such as medical field chest X-ray, CT and other image
data. Regardless of whether it is a single image for disease diagnosis or a combination of multiple
images for auxiliary prediction [16], the CNN model can realize the processing of medical images
through the feature extraction and fusion technology of complex data. Then through the fully
connected layer to achieve disease classification, its performance is better than traditional machine
learning models [17].

Figure 2: Convolutional neural networks (CNN)

Table 1: Commonly used CNN model and architecture


Model name Proposal time General structure
VGG (16, 19) 2014 Use the smaller 3 ∗ 3 convolution kernel to
stack, alternately use the largest pooling layer,
then connect the fully connected layer, and
finally classify by Softmax function.
ResNet 2015 The residual module is used, and skip
connection is added between input and output.
While the number of network layers is
deepened, the defects of gradient dispersion
are avoided.
Xception 2016 A new concept of depth separable convolution
operation is used. It can ensure that large-scale
data can be processed more effectively while
the amount of parameters is unchanged.
DenseNet 2017 The improvement on the basis of ResNet, the
difference is that when the connection is
established, it realizes the dense connection of
all the front layers and the back layers.

2.2 RNN
Time series forecasting, task-based dialogue, etc. all require models to learn features from the
sequence. The neurons in each layer of the neural network including the fully connected layer
CMES, 2021, vol.128, no.2 493

cannot be connected to each other, which leads to the problem of not being able to effectively
manage the time series. CNN can solve some problems in the spatial dimension, but it cannot
handle data in the temporal dimension. In response to some of the problems mentioned above,
recurrent neural networks (RNN) came into being.
RNN is composed of neurons and feedback loops. For scenarios where the previous input
and the next input have a dependency relationship, RNN shows unique advantages [18]. RNN
performs better than most machine learning algorithms when processing big data with delays
and abnormal noise. However, RNN faces the possibility of gradient explosion and gradient
disappearance during the training process, as well as the short-term memory defects of RNN [19],
which derived improved models LSTM and GRU. Compared with the basic RNN, LSTM adds
a state vector and a gate mechanism. As shown in Fig. 3, the main structure of LSTM includes:
Input Gate, Forget Gate, Output Gate. The GRU only contains Reset Gate and Update Gate.
Through the gating mechanism to achieve the control of information refresh and forgetting. In
this figure, xt represents the input data, ht is the state vector, ht−1 to ht represent the change of the
state vector after gate control. σ , tanh represents the activation function. For some information
with a long timeline, LSTM shows excellent performance that is different from general RNN. It
can receive a variety of data, including images, images, and audio. However, the computational
cost of LSTM is high and the model parameters are large, so a simpler GRU model is derived.
The gate control structure of GRU is reduced to two parts: reset gate and update gate. Although
GRU is relatively simple in structure, it is as effective as LSTM in many instances [20]. Both
models have good performance when dealing with data similar to vertical time stamps. Especially
in the prediction of some diseases, it is necessary to consider the past medical records, or to
extract effective information from multiple time spans and combine them [21]. RNN is very useful
in such scenarios [22]. In practical applications, RNN is more often combined with CNN or other
models to solve the double complexity of medical data in space and time [23–26].

Figure 3: Improved RNN structures. (a) LSTM structure diagram (b) GRU structure diagram
494 CMES, 2021, vol.128, no.2

2.3 Autoencoder (AE)


The existence of a large number of unlabeled data in reality. It promotes the creation of
unsupervised learning algorithms. It is difficult to label some exact labels on the data collected in
actual medical scenarios. This requires some technology to automatically classify or aggregate the
data to discover more further potential value of the data. Furthermore, there are many ways to
obtain medical data, and it is related to the actual medical environment. Some data will be lost
occasionally, and the information will be incomplete or distorted [27]. As a neural network model,
autoencoder perform efficient feature extraction on unlabeled data through unsupervised learning,
and perform well in feature dimensionality reduction, noise removal and restoration of original
data [28].
The classic AE model consists of two modules: encoding and decoding, as shown in Fig. 4.
Since the input data itself has no label information, only the input data x itself, AE uses x as the
output at the same time, constructing a mapping of f :x−>x . In practice, the network consists of
two parts, the left side is the encoding process f 1:x−>z, and the right side is the decoding process
f :z−>x . The dimensionality reduction is achieved during encoding, and the lower-dimensional
variable z is obtained, and then the z-dimensionality is increased through the decoding process to
obtain a new x . The core problem to be solved in this process is to ensure that the input x and
the final output x are as the same as possible, thereby obtaining the optimized objective function:
Minimie L = dist(x, x ). The functions f1 and f2 are trained through deep neural networks and
combined with other deep learning models to better obtain representation data. Compared with
the machine learning model, the original input x can be recovered better. The AE model has
a high degree of freedom and can be combined with a variety of models [29]. In practical
applications, various AE architectures can be constructed based on changeable medical data,
such as Sparse Autoencoder [30], Variational Autoencoder (VAE) [31], etc., to achieve efficient
processing of low-recognition medical data from many aspects.

Figure 4: Autoencoder architecture

2.4 Generative Adversarial Networks (GAN)


GAN is a kind of model which is used to train synthetic image based on game theory. The
idea of game theory is mainly embodied in the way of network training. Two sub-networks are
CMES, 2021, vol.128, no.2 495

set up respectively: generation network and discriminant network. They use a process similar to
a competition, then apply the backpropagation algorithm to continuously optimize the generation
network and the discriminant network. After that, obtain mutual improvement in turn. The
generating network is mainly responsible for learning the true distribution of samples, and the
discriminating network is mainly responsible for correctly distinguishing the samples generated
by the generating network from the real samples. The essence of the generative network is a
deconvolutional neural network, and the discrimination network is a convolutional neural network.
Since Gan was first proposed in 2014 to the present, regardless of the image size or image
fidelity, the performance of GAN has been constantly improving and breakthrough. The main
application of GAN is in image processing, including image-to-image translation, using GAN
for image restoration [32,33], improving image resolution [34], medical image fusion [35,36]
etc. Recently, the application of Gan in medical imaging has gradually increased and achieved
remarkable results [37]. Traditional Gan often encounters instability problems such as gradient
disappearance and mode collapse during training. Most of the improved methods introduced for
these problems have good performance [38–40]. Although GAN is still in the early stages of
medical imaging, it has already demonstrated its advantages in the medical field. With the further
resolution of these deficiencies, GAN will certainly provide effective help to the medical field to
the greatest extent.

2.5 Other Models and Methods


Except for the models mentioned above, Deep Belief Networks (DBN) has also appeared in
medical-related research many times in recent years. For example, when processing medical images,
it helps to achieve better image fusion [41]. Through DBN processing of voice information, help
Parkinson’s disease diagnosis [42] and so on. DBN can be used in both supervised learning and
unsupervised learning, it can play the role of classifier in supervised learning, and it have the
same effects as AE on unsupervised learning, assist doctor process medical information more
comprehensively.
In handling practical applications, research methods that combine multiple models perform
better. Different models are used to deal with problems from different angles. The method of
combining CNN and RNN proposed by [43] uses CNN to process horizontal data, RNN solves
vertical problems, and then combines with Bagging, the accuracy is significantly better than
independent models.
Besides, transfer learning has become more and more popular for the past few years, and
the trend of combining with deep learning has become more obvious. One of the drawbacks
of deep learning is that data dependence is relatively serious and requires the help of a large
amount of data. Transfer learning can be completed with a small amount of data, which is
similar to a “polymorphic” idea, which provides ideas for model reuse, which not only saves time,
it is also conducive to the further improvement of the model. More transfer learning should
be combined with a variety of DL models [44]. Talo et al. [45] use deep transfer learning to
recognize magnetic resonance images, which can effectively assist radiologists in their diagnostic
work. Transfer learning is applied to medical research, and new problems are dealt with through
the trained model, and the initial model is slightly modified to be used in cancer image lesion
recognition, all have good performance [46]. With the maturity of medical imaging technology,
transfer learning can help the medical field achieve more far-reaching research.
496 CMES, 2021, vol.128, no.2

3 Disease Classification and Related Prediction


3.1 Pneumonia
Pneumonia is an inflammation of multiple organs in the lungs caused by a variety of factors
and pathogens. Bacterial pneumonia is the most common. Common symptoms include difficulty
breathing, fever, and chest pain, which have a greater impact on the health of children and the
elderly. According to statistics, the ratio of pneumonia patients to deaths is huge across the world.
Although antibiotics and vaccines currently exist, there are still major hidden dangers to health
and even life threatening. COVID-19, which is raging around the world, is a mild to moderate
respiratory disease. It is essentially pneumonia caused by a coronavirus [47]. COVID-19 not only
causes damage to the lungs, but also affects other human systems. The harm including cardio-
vascular system, cranial nervous system [48,49], etc. Since February 19, new coronary pneumonia
has gradually developed into a global health problem, which is extremely important for the early
diagnosis and prediction of pneumonia.
Obstruction of the respiratory tract caused by pneumonia is more common, and patients with
severe lung damage can have symptoms of breathing difficulties. Fan et al. [50] designed a portable
respiration measurement device, combined with the algorithm of BP neural network, which effec-
tively improved the performance of existing ventilators. The detection of pneumonia disease is
generally through X-ray image or chest radiography, and the efficient processing of images by
deep learning can help radiologists to screen abnormal images of the lungs. Voroshnina et al. [51]
used X-ray images to test the CNN model. The data set used contains 5863 X images, with clear
image classification of diseased lungs and normal lungs. The author determined the most suitable
CNN model by setting different layers, optimizers, training parameters and epochs, and finally
came up with a 4-layer alternating Conv2D layer and the structure of the MaxPooling2D layer
has the highest pneumonia detection accuracy rate of 89.24%. Rajpurkar et al. [52] improved on
the traditional CNN model and constructed the CheXNet network, which contains 121 layers.
Also tested on the chest X-ray image data set, the index exceeded the previous optimal model, and
the results of detecting pneumonia were better than the level of professional radiologists. Qjidaa’s
team [53] applied the pre-trained and improved VGG16 network model to pneumonia prediction.
The advantage of this network is that it is simple in design and can extract better features in the
image. The article shows that this model has the same performance as radiologists in the location
of lesions. When using this model to predict chest radiographs, the external (different data sets)
AUC obtained is as high as 95% and the accuracy rate is 87.5%. In order to better monitor
COVID-19, [54] evaluated three pre-trained models Xception, VGG16, and InceptionResnet-V2.
In the end, the CNN model using Xception has the highest accuracy, with an accuracy rate of
97.19% higher than the others model’s 95.42% and 93.87%.
At the same time, the application of transfer learning can provide new ideas for predicting
pneumonia [55]. First build a deep learning framework, and then use transfer learning to apply
the parameters that have been trained in other data sets to your own model, which can avoid
the situation where the optimal parameters cannot be obtained due to insufficient data. Liang
et al. [56] proposed a novel way to predict pneumonia. The underlying framework is a convo-
lutional network containing a residual structure, where the residual structure reduces overfitting,
and the convolutional layer avoids the loss of feature information caused by the increase in
model depth. Combined with transfer learning, the final accuracy rate is 90.5%, which is better
than Xception’s 87.8%. Other indicators are also better than the existing CNN model; Chouhan
et al. [57] also combines transfer learning and deep learning. It mainly uses five pre-trained deep
learning models to extract features and classify them to get the output, and then use an ensemble
CMES, 2021, vol.128, no.2 497

model to integrate the five pre-trained models. The resulting model combines the output of five
pre-trained models, and its performance is better than that of a single model.
In addition to the traditional standalone model, a variety of fusion methods have also been
derived. Saul et al. [58] designed a model combining convolutional neural network and residual
network to classify images. Through the analysis of X-ray images, an accuracy rate of 78.73%
was finally obtained, which is higher than the optimal previous independent model 76.8%; Wu
et al. [59] creatively proposed a multi-view fusion model, as shown in Fig. 5. Its bottom layer uses
the ResNet50 architecture. Based on the original model, the classic RGB channel image is changed
to the three lungs view image as input. Compared with the single-view model, the use of medical
images from a more comprehensive perspective reduces some one-sided problems caused by single
factors, while avoiding overfitting and having better performance. The data set used is 495 high-
resolution CT images to diagnose COVID-19, showing the great potential of deep learning to
train models through CT images to achieve efficient screening of Convid-19. COVID-19 involves
a large amount of patient information and involves many security and privacy issues. In terms
of ensuring the security of medical data, Hossain et al. [60] proposed the B5G framework to
detect some influencing factors of COVID-19 in daily life, and cleverly combine with blockchain
technology to improve data security.

Figure 5: Multi-view deep learning fusion model [59]

Under the modern background of COVID-19, the epidemic has strongly impacted the global
economy and occupies most of the medical resources in society. The relationship between economy
and quality has become a problem that must be faced in reality. Reference [61] proposed an
auction mechanism and decision-making model, which can assist product selection and effectively
promote a win-win situation in economy and quality. The use of deep learning models can
effectively alleviate the shortage of medical resources, reduce the pressure on doctors, and achieve
rapid diagnosis as well accurate prediction of the individualized prognosis of COVID-19, which is
essential for the management of COVID-19 [62]. In the research of the future, we should further
combine DL with CT imaging, X-ray image, and other clinical information. More consideration
498 CMES, 2021, vol.128, no.2

can be given to comprehensive diagnosis, it means that predict at the same time through the
patient’s CT images, chest radiography and other information to avoid errors caused by single
information.

3.2 Cardiovascular Disease (CVD)


Cardiovascular diseases have many different types, such as common coronary heart disease,
hypertension, arrhythmia, and congenital heart disease. According to statistics, nearly 18 million
people die from cardiovascular disease every year, accounting for 31% of the global deaths, and
it is the world’s rank first fatal disease [63]. Especially in the fast-paced society, excessive work
pressure and irregular diets have made cardiovascular diseases more common. There are many
factors that cause cardiovascular diseases, such as overt indicators like blood pressure and blood
lipids, as well as recessive factors consist of genes and heredity. As the most common arrhythmia
in clinical work, atrial fibrillation often causes the formation of brain thrombosis or embolization
of other organs. The general detection is to analyze the changes in heart rhythm, but the actual
clinical abnormality of heart rate is not obvious, and it is difficult to detect [64]. Deep learning has
a unique processing method for high-dimensional complex features, and it can identify abnormal
features among many influencing factors. In general, heart diseases are assisted in the analysis
by Electrocardiograph (ECG), which can detect various abnormalities from arrhythmia to acute
coronary syndrome [65]. Hannun et al. [65] used DNN to process ECG to classify the heart
rhythm category, which exceeded the sensitivity of cardiologists, and could effectively reduce the
misjudgment of abnormal heart rhythm diagnosis. DL can not only shorten the diagnosis time
when helping CVD prediction, but also improve the accuracy, which provides many conveniences
for the prevention and treatment of CVD.
During the past research, there are many practices using machine learning to predict CVD.
With the development of deep learning, more and more DL methods have surpassed the tradi-
tional ML methods in many aspects. Ayon et al. [66] used seven models (including six comparative
models) to predict coronary heart disease. The author utilized the Statlog data set and Cleveland
data set to tests, and used five-fold and ten-fold cross-validation methods to make full use of
the data. As shown in Fig. 6, only the indicators under the five-fold cross-validation of the
Statlog data set are extracted here. Compared with the commonly used ML models, the accuracy,
sensitivity and NPV of DNN reached 98.15%, 98.67% and 98.32%, all of which were better than
the optimal ML model SVM. It is almost the same as SVM in Specificity and Precision and F1
score. In ten-fold cross-validation, DNN has the same excellent performance. And it can be seen
from the figure that the performance of DNN on various evaluation indicators is very stable,
and the classification effect in practical applications is also better. From this, it can be seen that
when dealing with disease data with a wider dimension, the performance of DL is better than the
general ML models.
Beyond that, clinical data of coronary heart disease often have unbalanced categories, and
the quality of data collection is not reach the standard. Based on this problem, Dutta et al. [67]
proposed an efficient neural network with convolutional layers model, which uses sub sampling
methods to solve the problem of abnormal data categories, and compares it with a variety of
latest machine learning models. Although it is equal to the optimal machine learning model in
predicting the recall value (0.77), the accuracy rate reaches 82% when testing negative cases, which
is higher than the optimal model SVM 76%, and the classification performance is also better
than commonly used machine learning models. MIT-BIH arrhythmia database is the first set of
more general evaluation materials for arrhythmia detection [68]. Shi et al. [69] used the improved
CMES, 2021, vol.128, no.2 499

XGBoost for the classification of abnormal heartbeat detection, combined with the MIT-BIH
atrial fibrillation database (AFDB), and merged the threshold classifier on the hierarchical classi-
fier to achieve multi-level classification of heartbeat conditions. Wang et al. [70] proposed a neural
network only contains the full connection layer to detect abnormal heartbeat in accordance with
AAMI standard (classifying heartbeat types into five categories: N, S, V, F and Q). The basic
structure contains two layers of neural networks, and the results are further optimized through
two classifications, and the effectiveness is verified by multiple data sets. Jin et al. [71] designed
a residual network and combined domain adaptation methods in transfer learning. By learning
and training from the source data domain, and then applying it to the target data domain, the
average accuracy rate can be increased by 4.5%. Effectively solved the problem of unbalanced and
unlabeled data sets.

Figure 6: Evaluation metrics of the heart disease prediction (ten-fold) system using Statlog data
set by Ayon et al. [66]

RNN also outperforms most ML algorithms when processing big data with delays and
anomalous noise. The prediction of CVD generally needs to consider the patient’s historical medi-
cal records to help decision-making. For data distributed along the timeline, RNN can be used to
process. On the other hand, RNN’s unique gate control structure can realize real-time monitoring
of patients’ medical records and predict physical conditions at any time. Anandajayam et al. [72]
Use the RNN model to predict heart disease based on the past medical records of patients with
a history of heart disease, comparing the results with logistic regression, the accuracy of RNN
is higher than 90%, which is better than logistic regression 85%. The result shows the advantages
of RNN in processing large data set. Kuang et al. [73] created a dynamic LSTM model by
improving the traditional LSTM. The unified processing of input data at different times is better
than traditional LSTM in the final classification result. For the abnormal heartbeat phenomena
that existed in the past, there are comparatively large individual differences. Jin et al. [74] proposed
the hybrid LSTM model in his research, utilizes the advantages of active learning to automatically
extract features and can study ECG fragments in a wider time span. It better solved the problems
that arise due to individual differences.
500 CMES, 2021, vol.128, no.2

What’s more, using the combination of RNN and CNN to build a hybrid model can deal
with more diverse disease prediction scenarios. And different model combination methods can
further improve the performance of the hybrid model. Shi et al. [75] proposed a network structure
that includes both CNN and LSTM, by combining several input layers for feature input. While
realizing automatic feature extraction, it also makes full use of manual features to achieve a
subject accuracy rate of 94.20%. Jin et al. [76] created a hybrid model called twin-attentional
convolutional-LSTM, which uses CNN and LSTM as the bottom layer. It makes full use of
LSTM’s ability to capture time-domain information, and effectively makes up for the shortcomings
of CNN in processing time-sensitive information. An accuracy rate of 98.51% is obtained on
the AFDB data set. Ali et al. [77] constructed a hybrid model combining CNN and GRU to
predict heart disease. Combining LDA and PCA feature algorithms for feature extraction, then
improve accuracy through k-fold cross-validation, the final CNN-GRU accuracy rate is 94.5%,
which is higher than 92% of LSTM and 93.7% of CNN-LSTM. Deep belief networks can also
be applied to disease prediction, and it can be used to both supervised learning and unsupervised
learning. Lu et al. [78] proposed an improved DBN model, combined with unsupervised training
and supervised optimization, the model has excellent stability and prediction accuracy. In the final
examination, the model achieved a classification accuracy of 91.26%. At the same time, the use
of deep learning can help identify disease risk factors and find abnormal indicators of disease
from multiple indicators. Poplin’s et al. [79] used the monitoring of retinal characteristics to non-
invasively detect the characteristics of blood vessels. Starting from the retinal image, combined
with CNN to predict and verify the influencing factors of cardiovascular diseases, they successfully
achieved the classification of multiple influencing factors.
Taking into account the characteristics of CVD, RNN can be used more to achieve early
prediction, and its unique processing method for time series data can more comprehensively
investigate the patient’s medical history. At present, the popular use of Electronic health record
(EHR) provides an opportunity for RNN to develop. In the future, EHR and RNN should be
more combined to realize multi-faceted disease monitoring.

3.3 Alzheimer’s Disease (AD)


Alzheimer’s disease is a syndrome, mainly manifested as a decline in cognitive and behavioral
abilities, and is a degenerative disease of the nervous system. Approximately 50 million people
worldwide suffer from AD. Although the initial mortality rate is not high, it has a serious impact
on the burden of individuals and families [80]. The initial period of the patient’s onset is not
easy to detect, but will gradually get worse over time. Therefore, early AD is easily overlooked,
and there is no cure for AD in the current medical field. Once suffering from AD, the medical
community is still at a loss and can only be relieved by early diagnosis. The early prediction and
prevention of AD is a global issue and a major challenge for global technology. Reference [81]
pointed out the general development process of AD patients, from the initial normal state to
mild cognitive impairment (MCI), and eventually may develop into AD. It is discovered in time
that MCI can effectively alleviate and avoid deterioration to AD. The article also pointed out
difficulties of AD prediction. Although there is still no comprehensive method to predict AD, deep
learning technology has received more attention in the early detection and automatic classification
of AD.
Magnetic resonance image (MRI) and Positron emission tomography (PET) are common
medical image data used to diagnose AD. These two images have high-dimensional, multi-modal
and other complex characteristics [82]. The previous processing methods are time-consuming and
CMES, 2021, vol.128, no.2 501

poor in performance. With the development of deep learning in image processing, more studies
have begun to use DL models to process these brain medical images. Jo et al. [82] discussed the
performance difference between the DL method and the hybrid method, as well as the impact of
different AD image data, finally obtained the hybrid method with better performance, and the
test indicators obtained when multiple images were mixed more acurrate. Ji’s team [83] proposed
a 3D-CNN model that can not only extract features from the two dimensions of time and space,
but also capture more complete spatial features than the 2D model, which is more suitable for the
requirements of medical imaging analysis. Cheng’s team [84] based on the characteristics of MRI
in the spatial and temporal dimensions, verify the effectiveness of 3D-CNN in diagnosing AD by
MRI. The Alzheimer’s Disease Neuroimaging Initiative (ADNI) database has extremely abundant
sample types and enormous data size, including MRI, PET, Genetic Data, Biospecimen Data and
other AD diagnostic data. Most of the influencing factors of AD are taken into consideration,
and it is one of the most common databases for AD research. Wang et al. [85] improved on
the basis of 3D-CNN, constructed the 3D-densely CNN model in Fig. 7. The author uses the
ANDI dataset and performed classification tests on AD/MCI/Normal, the accuracy was all in
More than 93%. The 18F-FDG PET image has a unique distribution in the low-metabolic area of
the cortex, which can help identify MCI and AD [86]. By comparing with the MRI of AD/MCI
patients, 18F-FDG PET has higher sensitivity in helping to detect AD [81]. Using the CNN model
of InceptionV3, combined with 18F-FDG PET images, proved that the DL model can correctly
classify lesions.

Figure 7: Wang proposed ensemble 3D-DenseNet framework for AD and MCI diagnosis [85]

In order to get rid of some limitations of identifying and classifying brain abnormalities, Cui
et al. [87] proposed a framework that combines RNN and CNN. CNN learns spatial features,
RNN extracts longitudinal features, and the results are tested on the ANDI dataset. The result is
better than the CNN average method. Reference [88] pointed out that the bilateral hippocampus
502 CMES, 2021, vol.128, no.2

of AD/MCI patients will show lesion features, which can be identified by MRI images. Li
et al. [89] used a hybrid convolutional and recurrent neural network to diagnose AD by identifying
the lesions in the hippocampus in MRI, the consequence can get 91.0% AUC when classifying
AD with normal controls.
Synthesize these papers, the deep learning model considers a more comprehensive perspective
when processing brain medical images. Compared with traditional medical methods and medical
experts, it can dig deeper brain abnormalities. However, the test results obtained through the
model cannot provide interpretable reasons for decision-making. In practical applications, the
image quality requirements are also tough to reach, and the performance in different image sets
is not stable enough. There are many brain pathological factors, and the changes are extremely
rapid. There is still a lot of room for the prediction of AD. The medical field should also
further strengthen research on the prevention and treatment of AD, which can better provide basic
knowledge for AI to provide medical services. It can also better to tap and develop DL technical
capabilities.

3.4 Cancer
Cancer is a general term for a large group of diseases. Mainly refers to the irregular growth
of cells under the influence of disease-inducing factors, which become tissue masses. Cancer has
a high mortality rate and ranks second in the global cause of death [90]. Cancer, also known
as tumor, is generally divided into benign and malignant. Benign tumors are easy to clear and
have a low recurrence rate; malignant tumors are more stubborn, easy to recur, and easy to
metastasize, which can damage other tissues and organs. Although malignant tumors are difficult
to cure, if timely prevention and detection and treatment can prevent the tumor from developing
in the direction of deterioration [91,92]. According to statistics, 30%–50% of cancer deaths can
be prevented, early diagnosis and screening are essential. Because of there are many types of
cancerous cells and large differences between cells, the comprehensiveness of prediction technology
is required. In recent years, deep learning has shown its unique advantages in the classification of
cancer lesions by virtue of the ability to automatically extract key data from the original data [93].
The traditional single-omics method can no longer meet the needs when studying complex
diseases [94]. Cancer is a disease with a long development cycle and the possibility of multiple
gene mutations. More data and more related genes must be considered to obtain more effective
information. The multi-omics joint method can obtain a more comprehensive flow of disease
information, and fundamentally promote the accurate diagnosis of cancer diseases, it will become
a new trend in cancer research [95]. The UCSC Xena database has a wealth of tumor data
sets. The Cancer Genome Atlas (TCGA) contains rich multi-omics data resources. It is conducive
to multi-omics research on cancer diseases. Tan et al. [96] proposed a MOSAE (Multi-omics
Supervised Autoencoder) model for genomics research, through specific AE to achieve the pro-
cessing and fusion of multiple omics structures, use TCGA Pan-Cancer data to compare with
six commonly used machine learning models, the ROC and AUC obtained are better than other
models. Tong et al. [97] developed two different AEs based on the original AE to achieve different
functions. ConcatAE was used to fuse modal information, and CrossAE was used to perform
multi-modal denoising. These two models are applied to breast cancer prediction, and combined
with multi-omics data to test the performance of these models, both are better than the single
model. Asada et al. [98] combined RNA, miRNA expression and clinical information as assembled
dataset, tested the types of genes related to the prognosis of lung cancer through ML and DL,
and successfully discovered new genes related to patient survival. Takahashi’s team [99] used AE
CMES, 2021, vol.128, no.2 503

to process six multi-omics data sets, reduce dimensionality and unify features, then utilize ML
to establish a lung cancer prognosis prediction model. The model shows advantages in multiple
indicators. After searching for keywords in multiple omics, the current research on the combination
of deep learning and multiple omics for cancer diagnosis prediction is still relatively scarce. The
development of existing genomics technology is more conducive to the acquisition of cancer
genomics data. In the future, we should make full use of the advantages of DL technology
for multi-data processing, and focus on multi-omics integration research. For the dynamic and
unstable development of cancer, the condition of cancerous cells can be detected from more
angles, and at the same time, it can help to discover new factors for cancer monitoring or
evaluation, achieving accurate cancer prediction.
DL model can better extract potential features when processing diseased cell data, and then
enhance the data through AE or GAN, solve unlabeled and irregular data, optimize the noise
data, which further benefit detect the disease of cancer cells. Khamparia et al. [100] uses CNN and
variational autoencoder for the diagnosis and classification of cervical cancer. The main structure
is CNN to extract high-dimensional features, and then through AE to reduce the dimensionality,
while enhancing the data, finally handed over to the dense layer and Softmax layer for classifica-
tion. After setting different filters and comparing with the existing method, the accuracy of 99.4%
when the size is 3 × 3 is obtained and it is better than the comparison method. Iizuka et al. [101]
uses InceptionV3 network combined with RNN to help classify and diagnose adenocarcinoma,
adenoma, and non-tumor. Through the classification test in the WSI image dataset, the AUC
is above 96.0%. Khan et al. [102] uses the ResNet101 model at the bottom layer, which can
extract deeper features than the traditional CNN model, and at the same time, with the help of
transfer learning, it reduces the data demand of the model. After applying the improved model to
the WCE image database (including four different types of stomach abnormalities), through the
optimization algorithm and the fusion method of maximum correlation, the accuracy performance
increased from the initial 87.45% to 99.46%. By comparing with other neural networks, it has
obvious advantages in accuracy, which is better than all previous methods.
Apart from predicting cancer lesions, DL can also be used for the detection of anti-cancer
drugs. Baptista et al. [103] used the DL method to help predict the sensitivity of anti-cancer
drugs, which greatly reduced the limitations of drug testing in clinical use. Mencattini et al. [104]
studied cancer drug response at the cellular level, it can be evaluated in vitro, combined with the
Alex-NET framework for drug response classification, using the basic circuit shown in Fig. 8, and
the final classification effect is obvious. Kuenzi et al. [105] also study the reaction of drugs in the
process of carcinogenesis from the cellular level. Realize the clinical research of anti-cancer drugs
from a new perspective.
Advances in the medical field and the rise of a new generation of computers will promote the
emergence of new methods of cancer treatment. Traditional cancer diagnosis methods are done
manually by radiologists. A medical image has several million pixels, but the cancerous part may
only occupy more than a dozen pixels. Manual diagnosis is time-consuming, and there are certain
misjudgments. Moreover, the requirements for radiologists are higher. The use of deep learning
can identify deeper cancer cell lesions, while considering horizontal and vertical characteristics
to provide a more comprehensive diagnosis basis. To a greater extent, the pressure on medical
diagnosis can be reduced, and while improving the defects of artificial care, it can achieve earlier
and more accurate cancer cell diagnosis [106].
504 CMES, 2021, vol.128, no.2

Figure 8: Schematic diagram of the method proposed by Mencattini et al. [104]

3.5 Diabetes Mellitus


Diabetes mellitus is mainly manifested as an abnormal increase in blood glucose levels,
exceptional insulin secretion or unusual functional absorption, which is a metabolic disease. If it
cannot be diagnosed and treated in time, diabetes can affect multiple organs, and complications
can seriously threaten human health. Diabetes is divided into Type 1 and Type 2. Both types show
a certain genetic correlation, so it is essential for genetic research to prevent the development of
diabetes from the root cause. Deep learning’s screening of diabetes-related features can not only
help diagnose diabetes, but also achieve accurate disease classification.
In medical diagnosis, blood glucose levels are often used to test for diabetes, and ECG signals
can also help in diabetes monitoring. Besides, the abnormality of Heart Rate Variability (HRV)
is used as a voucher to identify diabetic lesions [107]. Ayon et al. [108] used DNN for diabetes
feature monitoring, combined with two cross-validation method tests, and compared with a variety
of machine learning methods. The accuracy of 98.35% in five-fold cross-validation is obtained,
which is higher than 97.47% of the optimal SVM. Padmapritha [109] uses LSTM to help test
blood glucose (BG) levels. Li et al. [110] proposed a CRNN model. The bottom layer includes
CNN and an improved LSTM structure, which can quickly identify blood glucose abnormalities
and has high sensitivity in predicting BG levels. Swapna et al. [107] also used the method of
combining CNN and LSTM to diagnose with the help of HRV data, which can achieve an
accuracy of 95.7%. Pavithra’s team [111] used Deep GRU and LSTM to test on the diabetes
CMES, 2021, vol.128, no.2 505

dataset and found that the accuracy of GRU is higher than that of LSTM. Perdomo et al. [112]
creatively proposed a combination of obstructive sleep apnea and Type 2 diabetes to diagnose,
combining DBNN and CNN to build a training network, and the fusion model is better than
independent DBN and CNN with 98.80% accuracy.
Diabetes data monitoring in daily medical treatment relies on diabetes monitoring equipment.
In China, the treatment rate and control rate of diabetes have not reached 50%. Whether it is a
disease analysis platform or basic equipment, there are some gaps in the real-time monitoring of
diabetes. The huge demand provides basic conditions for the practical application of the Internet
of Things (IoT) and promotes the development of the global IoT. To a certain extent, it has also
promoted the advancement of medical treatment, combined with the advanced processing methods
of deep learning, to provide more intelligent measures for future disease diagnosis. Diabetes
monitoring can be found from various angles, tiny retinal cracks [112], and facial features to
identify the disease. In actual applications, a richer diagnosis model should be established and
applied to more scenarios. AI help is the only way to diagnose diabetes in the future.

3.6 Other Disease


Over the above categories of diseases, deep learning technology is also widely used in other
diseases, the following is a brief summary.

3.6.1 Hepatitis
Hepatitis belongs to a type of acute and chronic inflammation. There are five classifications
in medicine. Different types of hepatitis have different manifestations and alternate phenomena.
Most people have no obvious symptoms at the initial stage, and more methods are needed to
help predict. Failure to diagnose hepatitis in time may lead to liver cirrhosis and even worsen
liver cancer. Timely discovery and vaccination can achieve a greater degree of prevention and
treatment. Jyoti et al. [113] used DNN to predict hepatitis and tested different activation functions,
through the result obtained an accuracy of 92.3% when using the swish function, which is higher
than other methods. Yang’s team [114] proposed a 7-layer AlexNet CNN structure, combined with
transfer learning to apply the pre-trained model to the detection of liver lesions. While realizing
the algorithm automation, the accuracy is no less than the previous ML method and traditional
neural network. Xiao et al. [115] creatively studied a variety of hepatobiliary diseases from eye
feature images, combined with the ResNet-101 network to help identify hepatitis, and proved
the correlation between abnormal eye function and hepatobiliary diseases. Frid-Adar et al. [116]
used GAN to enhance medical data, synthesized high-quality liver lesion images, combined with
CNN to complete lesion classification, obtained 85.7% sensitivity and 92.4% specificity, which are
better than the comparison method 78.6% and 88.4%. Lu et al. [117] used Raman spectroscopy to
build an MSCIR model. The standard architecture was combined with CNN and RNN, and the
data processing was strengthened from multiple levels. The accuracy rate was 96.15%, higher than
the 91.53% of traditional NN. Wang et al. [118] also identified patients with type B pneumonia
from Raman spectroscopy, used PCA to reduce dimensionality, and then used LSTM to process
long-span data, finally reaching an accuracy of 97.32%. Both papers use PCA for dimensionality
reduction. If you change to AE or Gan, you may get better results.

3.6.2 Infectious Disease


Generally speaking, it is infection with multiple pathogens such as bacteria and viruses,
causing serious damage to multiple organs of the body. There are many ways of transmission and
there is no specific trend. Infectious shows high fulminant, once spread, it is difficult to control.
506 CMES, 2021, vol.128, no.2

The source of infectious diseases must be cut off early and the characteristics of the pathogen
must be explored. COVID-19, which still plagues the world, is a viral infectious disease. As a
thorny global medical problem, many teams currently use DL to help predict and diagnose. Zhu
et al. [119] used BiPathCNN combined with virus sequences to predict the possible source of the
virus and proved that COVID-19 has the same infectivity as other coronaviruses. Fu et al. [120]
used ResNet-50, and combined with transfer learning to identify viral pneumonia from a variety
of lung diseases through CT images, with an accuracy rate of 98.8%.
AI contributes to the discovery and high-risk factors timely control. It also helps effectively
improve the prevention and control of infectious diseases [121]. Using DNN and LSTM to
predict infectious diseases, and testing with optimal control parameters, the average performance
of the ARIMA model is improved by 24% and 19%, respectively. The DNN is still stable when
the infectious disease spreads, and the average performance very well, LSTM is more accurate.
In recent years, the development of big data has played a significant role in the prevention
of infectious diseases. Regarding the variability and unpredictability of infectious diseases, big
data has realized deeper mining to solve difficult problems. The existence of big data must have
corresponding processing technology, and deep learning can effectively solve this demand [122].

3.6.3 Kidney Disease


Kidney diseases include renal functional damage and tissue inflammation. The typical one
is chronic kidney disease (CKD). If early diagnosis and treatment are ignored, it may lead to
kidney cancer. In the latest study, Alnazer et al. [123] evaluated the medical imaging methods
used for CKD detection, including MRI, CT, PET, etc. These complex medical images can more
comprehensively analyze kidney disease conditions and provide a data basis for DL’s disease
research. Wu et al. [124] created an automated architecture, a multifunctional network (MF-Net)
mainly composed of CNN, for kidney abnormality detection, and obtained the best average clas-
sification accuracy of 94.67% compared with multiple fusion models rate. Navaneeth et al. [125]
improved the traditional convolutional neural network and introduced dynamic processing in the
pooling layer. And proposed to identify chronic kidney disease by the level of urea in saliva,
then use SVM as a classifier to obtain 97.67% classification accuracy, which is better than the
96.12% of traditional CNN. Bhaskar et al. [126] also used saliva urea abnormalities to detect
diseases, developed an ammonia sensor, and obtained data from it, and then used the CNN-SVM
model for classification, which can achieve an accuracy of 98.04%. Khamparia et al. [127] use
stacked AutoEncoder for feature enhancement, then use the Softmax function as the classifier,
and even reached 100% accuracy when performing CKD classification. Ma’s team [128] proposed
a Heterogeneous Modified ANN model. This model combines multiple classifiers to achieve the
prediction of chronic kidney disease. During the experiment, the data integration is carried out
through the medical Internet of Things, which ensures the reliability of the data and reduces the
calculation time. Chen et al. [129] used an Adaptive hybridized deep CNN structure to detect
kidney disease, combined with the medical Internet of Things, and conveniently realized model
optimization and prediction, which was significantly better than other NN models in AUC.

4 Existing Problems and Technical Obstacles


4.1 Insufficient Interpretability
Although the deep learning model has excellent performance in feature extraction and classifi-
cation capabilities, its poor model interpretability has become a criticism that is hard to ignore. As
two independent branches of science, deep learning and medicine will inevitably be heterogeneous
in all aspects in research. Some existing disease prediction methods separate medical domain
CMES, 2021, vol.128, no.2 507

knowledge from deep learning methods. How to properly integrate these two kinds of subjects is
very important, which requires attention to the interpretability of the model. The interpretability
of the model indicates the extent to which humans can understand the reasons for the decision. In
general, the model is required to answer the “why” question, that is, how the model predicts. Some
traditional statistical methods are manually selected features based on medical knowledge, this
approach is closely integrated with medical knowledge. The deep learning method is completely
based on data, and rarely considered the previous domain knowledge or experience and some
known risk factors.
Deep neural network is like a black box model. After the model is fully trained, new data
is input to the model and the prediction results are output. The model only lets people get the
prediction result by the corresponding input data, but does not show people how to predict,
which inevitably leads to some lack of trust. In order to apply the deep learning method to the
real disease prediction and healthcare system, realize the intelligence of disease prediction, it is
necessary to solve the basic credibility problem and make the model convincing. The credibility of
deep learning models needs to be based on certain model interpretability. Only when the problem
of model interpretability is solved, deep learning can be more widely applied to the actual scenes
in the medical field and achieve the purpose of combining with the real application. In conclusion,
more attention should be paid to the study of model interpretability in future research. Whether
it is the criteria, scope or role of interpretability, appropriate methods should be selected and
combined with realistic requirements to make the model “humanized”. Only in this way, it can
help elevate the believability while using the deep learning ideas. In this way not only provide
rapid and accurate diagnosis for patients, but also provide explainable diagnosis basis.

4.2 Clinical Implementation Issues


Most of the current deep learning methods combined with disease prediction are still in
the theoretical stage and have not been applied to the clinic. There are several reasons for this
situation.

4.2.1 Generalization Ability


The poor generalization ability of the model has led to doubts about the practicality and
accuracy of the model in actual use. For example, in some areas with poor medical equipment,
high-quality medical images cannot be extracted, and it is difficult to meet the basic requirements
of the model for input data. Reference [72] pointed out that although DNN has obvious advan-
tages in various indicators, it will show instability when applying different test sets. The main
reason is that DL has higher data requirements. For this situation, on the one hand, CNN or
other deep learning models with denoising function can be designed [130]. On the other hand,
image enhancement technology can also be applied, using appropriate filters, to enhance the
overall or local features of the image, amplify key features, and improve the applicability of the
image. At the same time, make full use of the existing excellent image fusion technology, and
integrate various image data to enrich the diagnosis basis [131]. For medical data, the dataset is
small, and there are still many problems of class imbalance. In the future, more attention should
be paid to the research of data availability to improve the generalization ability of the model.

4.2.2 Stability
Whether the deep learning model is stable in practical applications is still unknown. To truly
apply deep learning to clinical applications, a high degree of stability is one of the necessary
conditions. In reality, medical data are spotty and vary widely. When the neural network model
508 CMES, 2021, vol.128, no.2

is applied to the actual medical situation, it may face the situation that the training sample
does not match the real sample. When the model refined by training samples is processing real
samples, there may be some abnormal features, insufficient classification accuracy or underfitting,
and the model is not robust enough. If the stability of the model cannot be guaranteed, it will
not only lead to a decrease in the performance and efficiency of the algorithm, inaccurate disease
prediction, but also some clinical safety problems may occur, and even the life safety of medical
patients may be threatened. Therefore, the transition to clinical application should be realized
under the premise of ensuring the stability of the model.

4.2.3 Privacy Security


When using deep learning methods to predict diseases, the medical records and personal data
of disease patients are generally involved, so the protection of data and the privacy of patients
should naturally be fully considered. How to protect the privacy of patients and prevent the
leakage of personal information is particularly important. Lv et al. [132] proposed the use of
differential privacy technology to process medical data, which can effectively ensure data security.
Reference [60] used surveillance systems combined with blockchain technology enables reliable
protection of data in telemedicine. The so-called technology-generated problems should be solved
by technology, and the problems generated by the application of deep learning technology should
also be supported by corresponding technologies. Although distributed technology has made some
achievements in data protection in recent years, and the goal of storing patient data in the
original medical institution or original storage device has been achieved [133]. However, distributed
technology is still relatively independent, and the research on integration with deep learning
technology is not enough. In future research, more attention should be paid to the protection of
doctors and patients’ privacy [88]. There are many types of medical data, with large differences,
and the quality of the data is uneven due to the influence of medical conditions. Therefore, while
using existing technologies, we should develop application technologies that are more suitable for
medical data and can ensure privacy and security. Let patients no longer worry about the exposure
of personal information, so as to promote the realization of clinical applications more quickly.
In addition to the three points mentioned above, there are also some influencing factors.
For example, in disease prediction, the input data of the model is manually labeled data. It is
inevitable that there will be some personal biases, which may cause ethical problems. This requires
the establishment of a realistic ethical framework to avoid the influence of personal emotions,
and it is obviously not easy to establish a global consensus on ethics. In addition, the application
threshold of deep learning technology is relatively high, most areas still do not have the basic
conditions, and there is a shortage of talents in related fields. Therefore, there is still a long way
to go if really want to apply deep learning technology to actual clinical practice.

4.3 The Challenge of Model Training


The deep learning models mentioned above still have some shortcomings, such as AE or
GAN. The problems include the credibility of synthetic data. Because of the lack of some neces-
sary associations and mappings, it is not enough to be convincing. Instability and the uncertainty
of evaluation indicators are both problems that need to be solved urgently. Only when these
problems and deficiencies are resolved can it become a more trustworthy technology in the field
of healthcare. In addition, under the characteristics of medical data, there are more over-fitting
and under-fitting phenomena during model training [134]. The lack or redundancy of the training
data set will cause the above problems. It can be solved by appropriately increasing or decreasing
the size of the dataset, starting from the model itself, and changing the model structure [135].
CMES, 2021, vol.128, no.2 509

What’s more, data quality is still the biggest challenge in model training. The excellent
performance of DL models in disease prediction relies on high-quality medical data. Although
it is easy to obtain medical data under existing conditions, the data quality is still low. Many
medical data need to have a wealth of experience in medical experts to give a proprietary label.
The processing of image features is particularly important [136], and medical data sets involve
many privacy issues, and they are stored in independent institutions. The data sets are closed,
not open, and a large number of datasets cannot be used in actual research. The inability to get
adequate training [137] also hinders the development of many innovative models.

5 Future Work and Prospect


5.1 Deep Learning Drives Precision Medicine (PM)
Precision medicine is extremely significant in future research. No matter in patient diagnosis,
treatment, or prognosis management, precision medicine has achieved many breakthroughs under
the support of the development of big data. Compared with traditional medical treatment, it can
provide patients with more effective and timely medical services. Since the introduction of preci-
sion medicine in 2015 to the present, it has always been the key to the future global healthcare
and is a golden target for development plans in many countries around the world [138]. Its unique
diagnostic paradigm and basic concepts have opened up a new idea for human health and medical
care [139–141]. Afzal et al. [138] discussed the development process and future direction of PM
from the perspective of informatics, and proposed the PM framework shown in Fig. 9, which fully
displays the source of data, related processing, practical applications and services object. It can
be seen that precision medicine will be combined with multi-domain applications in the next era
development. A set of relatively complete operation methods can realize the systematic operation
of medical diagnosis and promote the development of medical treatment to more intelligent.
Just as tailored, precision medicine pays more attention to the study of individual dif-
ferences [142], and refines the impact of each individual’s personal factors on disease [143].
Evaluation of personal health from genomics, living environment and other levels, coupled with
clinical data analysis, has a higher resolution. For example, Panayides et al. [144] proposed that
starting from the method of radio genomics, combined with precision medicine, abnormal diseases
can be found faster when dealing with complex diseases. PM can be summarized in three correct
terms, namely, the right time, the right patient, and the right medicine to achieve timely diagnosis
of the patient. And precision medicine also has a relatively good performance in the prevention of
malignant diseases, especially when dealing with cancer [145,146], tumor [147] and other related
diseases. Its unique diagnosis and treatment methods have promoted disease prediction. And
treatment is developing towards the era of precision medicine [148].
At present, precision medicine in Western countries has made many achievements, while
the development of the Asia-Pacific region is still in its infancy. On the one hand, the urgent
problem to be solved at the moment is to ensure the diversity and high quality of gene collection.
On the other hand, the crux needs a firm promise that the extraction is consistent with the
genetic characteristics of the population in the Asia-Pacific region. Reference [138] shows the
close relationship between precision medicine and artificial intelligence. The rapid development
of AI and the emergence of related platforms have provided many conveniences for precision
medicine [149]. With the advent of the era of big data, combined with deep learning as the engine
of the era, it fundamentally promotes the development of precision medicine [150].
510 CMES, 2021, vol.128, no.2

Figure 9: A holistic integrated precision medicine framework [138]

Regarding the remaining constraints between precision medicine and big data [151], it is
necessary to fundamentally handle the relationship between PM, big data and deep learning. Deep
learning not only provides advanced technical means, but also arouses people’s understanding of
the future medical multi-angle thinking. As a part of precision medicine, accurate prediction of
diseases reflects the advantages of future medical care and plays a key role in promoting the
development of modern medical technology. The current precision medicine research is still in the
exploratory development stage [152–154], the research status of different diseases is quite different,
and the application of deep learning technology is relatively narrow. In the future, in the process
of exploring the medical field, AI researchers should focus more on precision medicine. Combining
the research on pathological changes caused by gene mutations in the medical field, construct a
deep learning model that is more in line with the requirements of precision medicine. Applying
deeper and more changeable models to it. While promoting the development of precision medicine,
it also drives the multi-faceted extension of deep learning, making medical care conform to the
needs of the society in the future.

5.2 Combine with the Internet of Things


The maturity of the Internet of Things technology has strengthened the development of
intelligent society. Especially the arrival of the current 5G era has brought qualitative changes to
the application of IoT technology. Broader coverage of the Internet of Things, low-latency and
high-throughput performance accelerate the spread of data, and the high demand for health and
medical modernization, all of which promote the integration of the Internet of Things, medical
and DL fields. E-healthcare is an important application combining medical care and the Internet
of Things in recent years. It can not only provide “interaction” of medical information, but also
provide a system and platform for telemedicine services. Reference [155] shows that in the current
CMES, 2021, vol.128, no.2 511

raging COVID-19, the healthcare system has played a great role, providing important help for
better tracking patients and monitoring disease symptoms. The combination of the Internet of
Things and AI has further improved the previous single disease prediction method. The cloud-
based electronic physical localization system proposed by Hossain [156] can effectively process
speech and EEG signals. Bisio et al. [157] proposed an eHealth system based on the Internet
of Things (IoT) to help remotely monitor the rehabilitation of stroke patients. Compared with
traditional medical systems, it facilitates remote monitoring. Once promoted, the DL method can
be used to efficiently process data, which can provide a wider range of remote end-to-end services
for patient rehabilitation.
The Internet of Medical Things (IoMT) has developed rapidly in the past few years [158].
The infrastructure such as common life wearable devices (bracelet, belt, diagnosis and treatment
glasses, intelligent underwear, etc.) [159], intelligent home devices [160], has become an important
source of medical data acquisition. Reference [161] created a medical system that can detect and
perceive emotions, and obtain patient status by monitoring emotions. The daily monitoring of
personal activities can be realized, the operation is simple, and the system provides dynamic data.
On this basis, through the processing of dynamic data with deep learning, real-time monitoring
of patient health can be achieved [162]. Reference [163] proposes a new type of intelligent medical
equipment, and integrates deep learning methods in disease monitoring, realizes physical condition
detection by acquiring all-round information of patients, which greatly improves the performance
of the detection system. In addition, combined with the use of sensors, it is possible to track
key physical indicators during the postoperative recovery phase of the patient. Another combined
with the DL method, the longitudinal timeline data stream is processed to remotely monitor the
patient’s recovery status. Realize the symbiotic development with Deep Learning through these
medical IoT devices. As shown in Fig. 10, personal medical data is first obtained at the terminal,
and then transmitted through various networks. The data and personal information are handed
over to the IoMT platform. IoMT uses DL to process massive amounts of data information,
thereby realizing all the data on the IoMT platform. Required medical operations, including
interactions with medical service systems and hospitals. This describes the complete framework of
the combination of the medical Internet of Things and DL.
In prospective work, more intelligent processing of data should be realized through the
Internet of Things to provide higher-quality basic data for deep learning models. At the same
time, it is necessary to develop more IoT medical system platforms, produce more smart hardware,
and build a truly smart city. Smart medical care is an important link in a smart city, and a highly
interactive medical platform is even more critical. The medical monitoring framework proposed
by [164] can simultaneously receive voice signals and EGG signals, and achieves 93% accuracy
in voice case monitoring. It greatly improves the shortcomings of difficulty in interaction. On
the basis of ensuring the security of the Internet of Things platform, it further broadens the
application coverage and more comprehensively serves the user population [165]. Intelligence is
the key word for future medical health and urban development. To truly realize comprehensive
medical intelligence, better integration of medical Internet of Things and deep learning technology
is essential. A new technological model is used to help disease diagnosis and reduce pressure on
medical resources.

5.3 Specialized Model Improvement


A very important feature of the deep learning model is that the architecture is very change-
able. Different models have many features that can be changed. They can be changed at will
512 CMES, 2021, vol.128, no.2

according to the needs. Regardless of the data or the scene, the general actual needs can be based
on the general model to make a slight change to help implement it. This feature is very compatible
with the medical field, and can meet the needs of different diseases, individual differences in
different patients, and complex medical conditions.

Figure 10: Integrated architecture diagram of medical IoT platform

Specifically, for example, when dealing with complex structures and high-dimensional features,
DL models tend to have deeper levels, which leads to the possibility of gradient disappearance
during the gradient propagation process. Based on the characteristics of medical data, a DL
model that is more conform to the characteristics of medical data should be designed [166–168].
Comprehensive consideration of various influencing factors, including different training sets, opti-
mization methods, training times, parameter effects, etc. Improve the utilization of important data
features in medical data, reduce unnecessary parameters, avoid over-fitting, and fundamentally
optimize the efficiency of the classifier. In addition, according to different types of data used in
medical diagnosis, such as electronic medical records arranged in a time series, the advantages
of transfer learning can be fully utilized here to design a framework that combines multiple
models [169]. Combined with the existing results, as much as possible to reduce unnecessary
parameter training process, maximize the expansion of the functions of existing models. While
making good use of transfer learning to achieve specific model optimization, it can also be
combined with multi-domain applications, such as Internet of Things related equipment [170], and
a big data foundation to form a more complete integrated framework.
For different diseases, it will be affected by many factors when predicting. Moreover, the
actual situation in medical diagnosis varies greatly, and there is no fixed objective law. In addition,
CMES, 2021, vol.128, no.2 513

sometimes, there is a certain correlation between different diseases, and the complications of
one disease may cause the production of other diseases. This adds many limitations to practical
applications. According to the characteristics of different diseases, the corresponding DL model
should be designed according to the unique usage data and prediction methods, so as to facilitate
the rapid expansion of the model according to different situations in practical applications. To
ensure that model research develops towards diversification, while establishing standard treatment
frameworks corresponding to different diseases.

6 Conclusion
The development of medical healthcare in the future has more modern prospects under the
background of the era of big data and the coverage of the Internet of Things. In the face of the
high dimension and instability of medical data, deep learning has become the main driving force
for the future development with its unique feature processing method and variable model structure.
This article introduces a series of popular deep learning prediction methods and frameworks for
different disease, and introduces the optimal selection method based on commonly used medical
diagnosis data. Pointed out some existing medical practical problems, as well as the limitations
of deep learning itself. In the outlook, it is proposed to combine with precision medicine to
conduct individual disease research on the genetic level; combine with the Internet of Things
to develop intelligent medical platforms and equipment; at the same time, it proposes model
improvement ideas for the particularity of medical data. In the future, we believe that disease
diagnosis and deep learning will develop in a more diversified direction, and will certainly extend
from today’s auxiliary diagnosis to decision-making diagnosis. For different diseases, more and
more DL models have emerged. The models are intertwined and learn from each other, resulting in
a richer DL system network. This corresponds to the complex medical system, which will facilitate
the development of medical diagnosis and clinical application as well as promote the development
of the medical field.

Funding Statement: This work was supported in part by the National Natural Science Foundation
of China (Nos. 61902203, 61976242) and Key Research and Development Plan-Major Scientific
and Technological Innovation Projects of Shandong Province (2019JZZY020101).
Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding
the present study.

References
1. Cho, N., Shaw, J. E., Karuranga, S., Huang, Y., da Rocha Fernandes, J. D. et al. (2018). IDF diabetes atlas:
Global estimates of diabetes prevalence for 2017 and projections for 2045. Diabetes Research and Clinical
Practice, 138, 271–281. DOI 10.1016/j.diabres.2018.02.023.
2. Steenman, M., Lande, G. (2017). Cardiac aging and heart disease in humans. Biophysical Reviews, 9(2),
131–137. DOI 10.1007/s12551-017-0255-9.
3. Shickel, B., Tighe, P. J., Bihorac, A., Rashidi, P. (2017). Deep EHR: A survey of recent advances in deep
learning techniques for electronic health record (EHR) analysis. IEEE Journal of Biomedical and Health
Informatics, 22(5), 1589–1604. DOI 10.1109/JBHI.2017.2767063.
4. Sun, W., Cai, Z., Li, Y., Liu, F., Fang, S. et al. (2018). Data processing and text mining technologies on elec-
tronic medical records: A review. Journal of Healthcare Engineering, 2018, 9. DOI 10.1155/2018/4302425.
5. Cao, C., Liu, F., Tan, H., Song, D., Shu, W. et al. (2018). Deep learning and its applications in biomedicine.
Genomics, Proteomics & Bioinformatics, 16(1), 17–32. DOI 10.1016/j.gpb.2017.07.003.
514 CMES, 2021, vol.128, no.2

6. Shrestha, A., Mahmood, A. (2019). Review of deep learning algorithms and architectures. IEEE Access, 7,
53040–53065. DOI 10.1109/ACCESS.2019.2912200.
7. Dumoulin, V., Visin, F. (2016). A guide to convolution arithmetic for deep learning. arXiv preprint
arXiv:1603.07285v2.
8. Kaur, T., Gandhi, T. K. (2019). Automated brain image classification based on VGG-16 and transfer
learning. 2019 International Conference on Information Technology, pp. 94–98. Bhubaneswar, India, IEEE.
9. Huang, G., Liu, Z., van der Maaten, L., Weinberger, K. Q. (2017). Densely connected convolutional
networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700–4708.
Honolulu, HI, USA.
10. Kassani, S. H., Kassani, P. H., Khazaeinezhad, R., Wesolowski, M. J., Schneider, K. A. et al. (2019). Diabetic
retinopathy classification using a modified xception architecture. 2019 IEEE International Symposium on
Signal Processing and Information Technology (ISSPIT), pp. 1–6. Ajman, United Arab Emirates, IEEE.
11. Tang, S., Yuan, S., Zhu, Y. (2020). Convolutional neural network in intelligent fault diagnosis toward
rotatory machinery. IEEE Access, 8, 86510–86519. DOI 10.1109/ACCESS.2020.2992692.
12. Khare, S. K., Bajaj, V. (2020). Time-frequency representation and convolutional neural network-
based emotion recognition. IEEE Transactions on Neural Networks and Learning Systems, 1–9. DOI
10.1109/TNNLS.2020.3008938.
13. Siddiqi, R. (2019). Automated pneumonia diagnosis using a customized sequential convolutional neural
network. Proceedings of the 2019 3rd International Conference on Deep Learning Technologies (ICDLT
2019), pp. 64–70. New York, NY, USA.
14. Zhang, N., Cai, Y. X., Wang, Y. Y., Tian, Y. T., Wang, X. L. et al. (2020). Skin cancer diagnosis
based on optimized convolutional neural network. Artificial Intelligence in Medicine, 102, 101756. DOI
10.1016/j.artmed.2019.101756.
15. Hossain, M. S., Amin, S. U., Alsulaiman, M., Muhammad, G. (2019). Applying deep learning for
epilepsy seizure detection and brain mapping visualization. ACM Transactions on Multimedia Computing,
Communications, and Applications (TOMM), 15(1s), 1–17. DOI 10.1145/3241056.
16. Zhang, Y., Liu, Y., Sun, P., Yan, H., Zhao, X., Zhang, L. (2020). IFCNN: A general image fusion framework
based on convolutional neural network. Information Fusion, 54, 99–118. DOI 10.1016/j.inffus.2019.07.011.
17. Sharma, H., Jain, J. S., Bansal, P., Gupta, S. (2020). Feature extraction and classification of chest X-ray
images using cnn to detect pneumonia. 2020 10th International Conference on Cloud Computing, Data
Science & Engineering (Confluence), pp. 227–231. Noida, India, IEEE.
18. Sherstinsky, A. (2020). Fundamentals of recurrent neural network (RNN) and long short-term memory
(LSTM) network. Physica D: Nonlinear Phenomena, 404, 132306. DOI 10.1016/j.physd.2019.132306.
19. Long, D., Zhang, R., Mao, Y. (2019). Recurrent neural networks with finite memory length. IEEE Access,
7, 12511–12520. DOI 10.1109/ACCESS.2018.2890297.
20. Lynn, H. M., Pan, S. B., Kim, P. (2019). A deep bidirectional GRU network model for biometric elec-
trocardiogram classification based on recurrent neural networks. IEEE Access, 7, 145395–145405. DOI
10.1109/ACCESS.2019.2939947.
21. Cui, Y., Zhao, S., Wang, H., Xie, L., Chen, Y. et al. (2018). Identifying brain networks at multiple time scales
via deep recurrent neural network. IEEE Journal of Biomedical and Health Informatics, 23(6), 2515–2525.
DOI 10.1109/JBHI.2018.2882885.
22. Choi, E., Bahadori, M. T., Schuetz, A., Stewart, W. F., Sun, J. (2016). Doctor ai: Predicting clinical events
via recurrent neural networks. Machine Learning for Healthcare Conference, pp. 301–318. Los Angeles, CA,
PMLR 56.
23. Banerjee, I., Ling, Y., Chen, M. C., Hasan, S. A., Langlotz, C. P. et al. (2019). Comparative effectiveness of
convolutional neural network (CNN) and recurrent neural network (RNN) architectures for radiology text
report classification. Artificial Intelligence in Medicine, 97, 79–88. DOI 10.1016/j.artmed.2018.11.004.
24. Zhang, P., Wang, X., Zhang, W., Chen, J. (2018). Learning spatial–spectral–temporal EEG features with
recurrent 3D convolutional neural networks for cross-task mental workload assessment. IEEE Transactions
on Neural Systems and Rehabilitation Engineering, 27(1), 31–42. DOI 10.1109/TNSRE.2018.2884641.
CMES, 2021, vol.128, no.2 515

25. Qin, C., Schlemper, J., Caballero, J., Price, A. N., Hajnal, J. V. et al. (2018). Convolutional recurrent neural
networks for dynamic MR image reconstruction. IEEE Transactions on Medical Imaging, 38(1), 280–290.
DOI 10.1109/TMI.2018.2863670.
26. Liang, G., Hong, H., Xie, W., Zheng, L. (2018). Combining convolutional neural network with
recursive neural network for blood cell image classification. IEEE Access, 6, 36188–36197. DOI
10.1109/ACCESS.2018.2846685.
27. Kim, J. C., Chung, K. (2020). Multi-modal stacked denoising autoencoder for handling missing data in
healthcare big data. IEEE Access, 8, 104933–104943. DOI 10.1109/Access.6287639.
28. Chen, M., Shi, X., Zhang, Y., Wu, D., Guizani, M. (2017). Deep features learning for medical image
analysis with convolutional autoencoder neural network. IEEE Transactions on Big Data, 1, 1. DOI
10.1109/TBDATA.2017.2717439.
29. Azarang, A., Manoochehri, H. E., Kehtarnavaz, N. (2019). Convolutional autoencoder-based multispectral
image fusion. IEEE Access, 7, 35673–35683. DOI 10.1109/ACCESS.2019.2905511.
30. Zhang, J., Lin, F., Xiong, P., Du, H., Zhang, H. et al. (2019). Automated detection and localization of
myocardial infarction with staked sparse autoencoder and treebagger. IEEE Access, 7, 70634–70642. DOI
10.1109/ACCESS.2019.2919068.
31. Wei, R., Mahmood, A. (2021). Recent advances in variational autoencoders with representation learning
for biomedical informatics: A survey. IEEE Access, 9, 4939–4956. DOI 10.1109/ACCESS.2020.3048309.
32. Yeh, R. A., Chen, C., Yian Lim, T., Schwing, A. G., Hasegawa-Johnson, M. et al. (2017). Semantic image
inpainting with deep generative models. Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 5485–5493. Honolulu, HI, USA.
33. Zhou, Z., Wang, Y., Guo, Y., Qi, Y., Yu, J. (2019). Image quality improvement of hand-held ultrasound
devices with a two-stage generative adversarial network. IEEE Transactions on Biomedical Engineering,
67(1), 298–311. DOI 10.1109/TBME.2019.2912986.
34. Bing, X., Zhang, W., Zheng, L., Zhang, Y. (2019). Medical image super resolution using improved generative
adversarial networks. IEEE Access, 7, 145030–145038. DOI 10.1109/ACCESS.2019.2944862.
35. Huang, J., Le, Z., Ma, Y., Fan, F., Zhang, H. et al. (2020). MGMDcGAN: Medical image fusion using multi-
generator multi-discriminator conditional generative adversarial network. IEEE Access, 8, 55145–55157.
DOI 10.1109/ACCESS.2020.2982016.
36. Li, Y., Zhao, J., Lv, Z., Li, J. (2021). Medical image fusion method by deep learning. International Journal
of Cognitive Computing in Engineering, 2, 21–29. DOI 10.1016/j.ijcce.2020.12.004.
37. Yi, X., Walia, E., Babyn, P. (2019). Generative adversarial network in medical imaging: A review. Medical
Image Analysis, 58, 101552. DOI 10.1016/j.media.2019.101552.
38. Yang, Z., Chen, Y., Le, Z., Fan, F., Pan, E. (2019). Multi-source medical image fusion based on wasserstein
generative adversarial networks. IEEE Access, 7, 175947–175958. DOI 10.1109/ACCESS.2019.2955382.
39. Hong, H., Li, X., Wang, M. (2019). Gane: A generative adversarial network embedding. IEEE Transactions
on Neural Networks and Learning Systems, 31(7), 2325–2335. DOI 10.1109/TNNLS.2019.2921841.
40. Wang, C., Xu, C., Yao, X., Tao, D. (2019). Evolutionary generative adversarial networks. IEEE Transactions
on Evolutionary Computation, 23(6), 921–934. DOI 10.1109/TEVC.2019.2895748.
41. Kaur, M., Singh, D. (2020). Fusion of medical images using deep belief networks. Cluster Computing, 23(2),
1439–1453. DOI 10.1007/s10586-019-02999-x.
42. Al-Fatlawi, A. H., Jabardi, M. H., Ling, S. H. (2016). Efficient diagnosis system for Parkinson’s disease using
deep belief network. 2016 IEEE Congress on Evolutionary Computation, pp. 1324–1330. IEEE.
43. Dua, M., Makhija, D., Manasa, P. Y. L., Mishra, P. (2020). A CNN–RNN–LSTM based amalgamation
for Alzheimer’s disease detection. Journal of Medical and Biological Engineering, 40(5), 688–706. DOI
10.1007/s40846-020-00556-1.
44. Khan, S., Islam, N., Jan, Z., Din, I. U., Rodrigues, J. J. C. (2019). A novel deep learning based framework for
the detection and classification of breast cancer using transfer learning. Pattern Recognition Letters, 125,
1–6. DOI 10.1016/j.patrec.2019.03.022.
516 CMES, 2021, vol.128, no.2

45. Talo, M., Baloglu, U. B., Yıldırım, Ö., Acharya, U. R. (2019). Application of deep transfer learning for
automated brain abnormality classification using MR images. Cognitive Systems Research, 54, 176–188.
DOI 10.1016/j.cogsys.2018.12.007.
46. Hu, Z., Tang, J., Wang, Z., Zhang, K., Zhang, L. et al. (2018). Deep learning for image-based cancer
detection and diagnosis−A survey. Pattern Recognition, 83, 134–149. DOI 10.1016/j.patcog.2018.05.014.
47. Zhu, N., Zhang, D., Wang, W., Li, X., Yang, B. et al. (2020). A novel coronavirus from patients with pneu-
monia in China, 2019. New England Journal of Medicine, 382, 727–733. DOI 10.1056/NEJMoa2001017.
48. Bansal, M. (2020). Cardiovascular disease and COVID-19. Diabetes & Metabolic Syndrome: Clinical
Research & Reviews, 14(3), 247–250. DOI 10.1016/j.dsx.2020.03.013.
49. Nishiga, M., Wang, D. W., Han, Y., Lewis, D. B., Wu, J. C. (2020). COVID-19 and cardiovascular
disease: From basic mechanisms to clinical perspectives. Nature Review Cardiology, 17, 543–558. DOI
10.1038/s41569-020-0413-9.
50. Fan, D., Yang, J., Zhang, J., Lv, Z., Huang, H. et al. (2018). Effectively measuring respiratory flow with
portable pressure data using back propagation neural network. IEEE Journal of Translational Engineering
in Health and Medicine, 6, 1–12. DOI 10.1109/JTEHM.2017.2688458.
51. Voroshnina, N., Shabley, A., Vaulin, S. (2020). Neural network approach for prediction of pneumonia. 2020
International Multi-Conference on Industrial Engineering and Modern Technologies (FarEastCon), pp. 1–5.
Vladivostok, Russia, IEEE.
52. Rajpurkar, P., Irvin, J., Zhu, K., Yang, B., Mehta, H. et al. (2017). Chexnet: Radiologist-level pneumonia
detection on chest X-rays with deep learning. arXiv preprint. arXiv: 1711.05225.
53. Qjidaa, M., Ben-Fares, A., Mechbal, Y., Amakdouf, H., Maaroufi, M. et al. (2020). Development of a
clinical decision support system for the early detection of COVID-19 using deep learning based on chest
radiographic images. 2020 International Conference on Intelligent Systems and Computer Vision (ISCV), pp.
1–6. Fez, Morocco, IEEE.
54. Phankokkruad, M. (2020). COVID-19 pneumonia detection in chest X-ray images using transfer learning
of convolutional neural networks. Proceedings of the 3rd International Conference on Data Science and
Information Technology, pp. 147–152. Xiamen, China.
55. Raghu, M., Zhang, C., Kleinberg, J., Bengio, S. (2019). Transfusion: Understanding transfer learning for
medical imaging. arXiv preprint. arXiv: 1902.07208.
56. Liang, G., Zheng, L. (2020). A transfer learning method with deep residual network for pediatric pneumonia
diagnosis. Computer Methods and Programs in Biomedicine, 187, 104964. DOI 10.1016/j.cmpb.2019.06.023.
57. Chouhan, V., Singh, S. K., Khamparia, A., Gupta, D., Tiwari, P. et al. (2020). A novel transfer learn-
ing based approach for pneumonia detection in chest X-ray images. Applied Sciences, 10(2), 559. DOI
10.3390/app10020559.
58. Saul, C. J., Urey, D. Y., Taktakoglu, C. D. (2019). Early diagnosis of pneumonia with deep learning. arXiv
preprint. arXiv: 1904.00937.
59. Wu, X., Hui, H., Niu, M., Li, L., Wang, L. et al. (2020). Deep learning-based multi-view fusion model
for screening 2019 novel coronavirus pneumonia: A multicentre study. European Journal of Radiology, 128,
109041. DOI 10.1016/j.ejrad.2020.109041.
60. Hossain, M. S., Muhammad, G., Guizani, N. (2020). Explainable AI and mass surveillance system-
based healthcare framework to combat COVID-19 like pandemics. IEEE Network, 34(4), 126–132. DOI
10.1109/MNET.011.2000458.
61. Abdulsalam, Y., Hossain, M. S. (2020). COVID-19 networking demand: An auction-based mechanism for
automated selection of edge computing services. IEEE Transactions on Network Science and Engineering,
1–1. DOI 10.1109/TNSE.2020.3026637.
62. Chowdhury, M. E., Rahman, T., Khandakar, A., Mazhar, R., Kadir, M. A. et al. (2020). Can
AI help in screening viral and COVID-19 pneumonia? IEEE Access, 8, 132665–132676. DOI
10.1109/ACCESS.2020.3010287.
63. Timmis, A., Townsend, N., Gale, C. P., Torbica, A., Lettino, M. et al. (2020). European society
of cardiology: Cardiovascular disease statistics 2019. European Heart Journal, 41(1), 12–85. DOI
10.1093/eurheartj/ehz859.
CMES, 2021, vol.128, no.2 517

64. Shi, H., Wang, H., Qin, C., Zhao, L., Liu, C. (2020). An incremental learning system for atrial fibrillation
detection based on transfer learning and active learning. Computer Methods and Programs in Biomedicine,
187, 105219. DOI 10.1016/j.cmpb.2019.105219.
65. Hannun, A. Y., Rajpurkar, P., Haghpanahi, M., Tison, G. H., Bourn, C. et al. (2019). Cardiologist-level
arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network.
Nature Medicine, 25(1), 65–69. DOI 10.1038/s41591-018-0268-3.
66. Ayon, S. I., Islam, M. M., Hossain, M. R. (2020). Coronary artery heart disease prediction: A com-
parative study of computational intelligence techniques. IETE Journal of Research, 66, 1–20. DOI
10.1080/03772063.2020.1713916.
67. Dutta, A., Batabyal, T., Basu, M., Acton, S. T. (2020). An efficient convolutional neural net-
work for coronary heart disease prediction. Expert Systems with Applications, 159, 113408. DOI
10.1016/j.eswa.2020.113408.
68. Moody, G. B., Mark, R. G. (2001). The impact of the MIT-bIH arrhythmia database. IEEE Engineering in
Medicine and Biology Magazine, 20(3), 45–50. DOI 10.1109/51.932724.
69. Shi, H., Wang, H., Huang, Y., Zhao, L., Qin, C. et al. (2019). A hierarchical method based on weighted
extreme gradient boosting in ECG heartbeat classification. Computer Methods and Programs in Biomedicine,
171, 1–10. DOI 10.1016/j.cmpb.2019.02.005.
70. Wang, H., Shi, H., Lin, K., Qin, C., Zhao, L. et al. (2020). A high-precision arrhythmia classification method
based on dual fully connected neural network. Biomedical Signal Processing and Control, 58, 101874. DOI
10.1016/j.bspc.2020.101874.
71. Jin, Y., Qin, C., Liu, J., Lin, K., Shi, H. et al. (2020). A novel domain adaptive residual network for automatic
atrial fibrillation detection. Knowledge-Based Systems, 203, 106122. DOI 10.1016/j.knosys.2020.106122.
72. Anandajayam, P., Krishnakoumar, C., Vikneshvaran, S., Suryanaraynan, B. (2019). Coronary heart disease
predictive decision scheme using big data and RNN. 2019 IEEE International Conference on System,
Computation, Automation and Networking (ICSCAN), pp. 1–6. Pondicherry, India, IEEE.
73. Kuang, J. W., Yang, H. Z., Liu, J. J., Yan, Z. J. (2019). Dynamic prediction of cardiovascular disease using
improved LSTM. International Journal of Crowd Science, 3(1), 14–25. DOI 10.1108/IJCS-01-2019-0002.
74. Jin, Y., Qin, C., Liu, J., Li, Z., Shi, H. et al. (2021). A novel incremental and interactive method for actual
heartbeat classification with limited additional labeled samples. IEEE Transactions on Instrumentation and
Measurement, 70, 1–12. DOI 10.1109/TIM.2021.3069021.
75. Shi, H., Qin, C., Xiao, D., Zhao, L., Liu, C. (2020). Automated heartbeat classification based
on deep neural network with multiple input layers. Knowledge-Based Systems, 188, 105036. DOI
10.1016/j.knosys.2019.105036.
76. Jin, Y., Qin, C., Huang, Y., Zhao, W., Liu, C. (2020). Multi-domain modeling of atrial fibrillation detection
with twin attentional convolutional long short-term memory neural networks. Knowledge-Based Systems,
193, 105460. DOI 10.1016/j.knosys.2019.105460.
77. Ali, A. A., Hassan, H. S., Anwar, E. M. (2020). Heart diseases diagnosis based on a novel convolution neural
network and gate recurrent unit technique. 2020 12th International Conference on Electrical Engineering, pp.
145–150. Cairo, Egypt, IEEE.
78. Lu, P., Guo, S., Zhang, H., Li, Q., Wang, Y. et al. (2018). Research on improved depth belief
network-based prediction of cardiovascular diseases. Journal of Healthcare Engineering, 2018, 1–9. DOI
10.1155/2018/8954878.
79. Poplin, R., Varadarajan, A. V., Blumer, K., Liu, Y., McConnell, M. V. et al. (2018). Prediction of cardiovas-
cular risk factors from retinal fundus photographs via deep learning. Nature Biomedical Engineering, 2(3),
158–164. DOI 10.1038/s41551-018-0195-0.
80. Lyu, M. J., Yuan, S. M. (2020). Cloud-based smart dog music therapy and pneumonia detection system
for reducing the difficulty of caring for patients with dementia. IEEE Access, 8, 20977–20990. DOI
10.1109/ACCESS.2020.2969482.
81. Ding, Y., Sohn, J. H., Kawczynski, M. G., Trivedi, H., Harnish, R. et al. (2019). A deep learning model to
predict a diagnosis of Alzheimer disease by using 18F-fDG PET of the brain. Radiology, 290(2), 456–464.
DOI 10.1148/radiol.2018180958.
518 CMES, 2021, vol.128, no.2

82. Jo, T., Nho, K., Saykin, A. J. (2019). Deep learning in Alzheimer’s disease: Diagnostic classification
and prognostic prediction using neuroimaging data. Frontiers in Aging Neuroscience, 11, 220. DOI
10.3389/fnagi.2019.00220.
83. Ji, S., Xu, W., Yang, M., Yu, K. (2012). 3D convolutional neural networks for human action
recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(1), 221–231. DOI
10.1109/TPAMI.2012.59.
84. Cheng, D., Liu, M., Fu, J., Wang, Y. (2017). Classification of MR brain images by combination of multi-
cNNs for AD diagnosis. Ninth International Conference on Digital Image Processing (ICDIP 2017), vol.
10420, pp. 1042042. Hong Kong, China, International Society for Optics and Photonics.
85. Wang, H., Shen, Y., Wang, S., Xiao, T., Deng, L. et al. (2019). Ensemble of 3D densely connected convolu-
tional network for diagnosis of mild cognitive impairment and Alzheimer’s disease. Neurocomputing, 333,
145–156. DOI 10.1016/j.neucom.2018.12.018.
86. Riederer, I., Bohn, K. P., Preibisch, C., Wiedemann, E., Zimmer, C. et al. (2018). Alzheimer disease and mild
cognitive impairment: Integrated pulsed arterial spin-labeling MRI and 18F-fDG PET. Radiology, 288(1),
198–206. DOI 10.1148/radiol.2018170575.
87. Cui, R., Liu, M. (2019). RNN-Based longitudinal analysis for diagnosis of Alzheimer’s disease. Computer-
ized Medical Imaging and Graphics, 73, 1–10. DOI 10.1016/j.compmedimag.2019.01.005.
88. Ahmed, O. B., Benois-Pineau, J., Allard, M., Amar, C. B., Catheline, G. (2015). Classification of Alzheimer’s
disease subjects from MRI using hippocampal visual features. Multimedia Tools and Applications, 74(4),
1249–1266. DOI 10.1007/s11042-014-2123-y.
89. Li, F., Liu, M., Alzheimer’s Disease Neuroimaging Initiative. (2019). A hybrid convolutional and recurrent
neural network for hippocampus analysis in Alzheimer’s disease. Journal of Neuroscience Methods, 323,
108–118. DOI 10.1016/j.jneumeth.2019.05.006.
90. Siegel, R. L., Miller, K. D., Jemal, A. (2020). Cancer statistics. CA: A Cancer Journal for Clinicians, 70(1),
7–30. DOI 10.3322/caac.21590.
91. Tiwari, A., Srivastava, S., Pant, M. (2020). Brain tumor segmentation and classification from magnetic
resonance images: Review of selected methods from 2014 to 2019. Pattern Recognition Letters, 131, 244–260.
DOI 10.1016/j.patrec.2019.11.020.
92. Tsherniak, A., Vazquez, F., Montgomery, P. G., Weir, B. A., Kryukov, G. et al. (2017). Defining a cancer
dependency map. Cell, 170(3), 564–576. DOI 10.1016/j.cell.2017.06.010.
93. Sultan, H. H., Salem, N. M., Al-Atabany, W. (2019). Multi-classification of brain tumor images using deep
neural network. IEEE Access, 7, 69215–69225. DOI 10.1109/ACCESS.2019.2919122.
94. de Anda-Jáuregui, G., Hernández-Lemus, E. (2020). Computational oncology in the multi-omics era: State
of the art. Frontiers in Oncology, 10, 423. DOI 10.3389/fonc.2020.00423.
95. Hasin, Y., Seldin, M., Lusis, A. (2017). Multi-omics approaches to disease. Genome Biol, 18, 83. DOI
10.1186/s13059-017-1215-1.
96. Tan, K., Huang, W., Hu, J., Dong, S. (2020). A multi-omics supervised autoencoder for pan-cancer
clinical outcome endpoints prediction. BMC Medical Informatics and Decision Making, 20(3), 1–9. DOI
10.1186/s12911-020-1114-3.
97. Tong, L., Mitchel, J., Chatlin, K., Wang, M. D. (2020). Deep learning based feature-level integration
of multi-omics data for breast cancer patients survival analysis. BMC Medical Informatics and Decision
Making, 20, 225. DOI 10.1186/s12911-020-01225-8.
98. Asada, K., Kobayashi, K., Joutard, S., Tubaki, M., Takahashi, S. et al. (2020). Uncovering prognosis-
related genes and pathways by multi-omics analysis in lung cancer. Biomolecules, 10(4), 524. DOI
10.3390/biom10040524.
99. Takahashi, S., Asada, K., Takasawa, K., Shimoyama, R., Sakai, A. et al. (2020). Predicting deep learning
based multi-omics parallel integration survival subtypes in lung cancer using reverse phase protein array
data. Biomolecules, 10(10), 1460. DOI 10.3390/biom10101460.
100. Khamparia, A., Gupta, D., Rodrigues, J. J., de Albuquerque, V. H. C. (2020). DCAVN: Cervical cancer
prediction and classification using deep convolutional and variational autoencoder network. Multimedia
Tools and Applications, 80, 1–17. DOI 10.1007/s11042-020-09607-w.
CMES, 2021, vol.128, no.2 519

101. Iizuka, O., Kanavati, F., Kato, K., Rambeau, M., Arihiro, K. et al. (2020). Deep learning models for
histopathological classification of gastric and colonic epithelial tumours. Scientific Reports, 10(1), 1–11.
DOI 10.1038/s41598-020-58467-9.
102. Khan, M. A., Sarfraz, M. S., Alhaisoni, M., Albesher, A. A., Wang, S. et al. (2020). Stomachnet: Optimal
deep learning features fusion for stomach abnormalities classification. IEEE Access, 8, 197969–197981. DOI
10.1109/ACCESS.2020.3034217.
103. Baptista, D., Ferreira, P. G., Rocha, M. (2021). Deep learning for drug response prediction in cancer.
Briefings in Bioinformatics, 22(1), 360–379. DOI 10.1093/bib/bbz171.
104. Mencattini, A., Di Giuseppe, D., Comes, M. C., Casti, P., Corsi, F. et al. (2020). Discovering the hidden
messages within cell trajectories using a deep learning approach for in vitro evaluation of cancer drug
treatments. Scientific Reports, 10, 7653. DOI 10.1038/s41598-020-64246-3.
105. Kuenzi, B. M., Park, J., Fong, S. H., Sanchez, K. S., Lee, J. et al. (2020). Predicting drug response
and synergy using a deep learning model of human cancer cells. Cancer Cell, 38(5), 672–684. DOI
10.1016/j.ccell.2020.09.014.
106. Coccia, M. (2020). Deep learning technology for improving cancer care in society: New direc-
tions in cancer imaging driven by artificial intelligence. Technology in Society, 60, 101198. DOI
10.1016/j.techsoc.2019.101198.
107. Swapna, G., Vinayakumar, R., Soman, K. P. (2018). Diabetes detection using deep learning algorithms. ICT
Express, 4(4), 243–246. DOI 10.1016/j.icte.2018.10.005.
108. Ayon, S. I., Islam, M. (2019). Diabetes prediction: A deep learning approach. International Journal of
Information Engineering & Electronic Business, 11(2), 21–27. DOI 10.5815/ijieeb.2019.02.03.
109. Padmapritha, T. (2019). Prediction of blood glucose level by using an LSTM based recurrent neural
networks. 2019 IEEE International Conference on Clean Energy and Energy Efficient Electronics Circuit for
Sustainable Development (INCCES), pp. 1–4. Krishnankoil, India.
110. Li, K., Daniels, J., Liu, C., Herrero, P., Georgiou, P. (2019). Convolutional recurrent neural networks
for glucose prediction. IEEE Journal of Biomedical and Health Informatics, 24(2), 603–613. DOI
10.1109/JBHI.2019.2908488.
111. Pavithra, M., Saruladha, K., Sathyabama, K. (2019). GRU based deep learning model for prognosis
prediction of disease progression. 2019 3rd International Conference on Computing Methodologies and
Communication (ICCMC), pp. 840–844. Erode, India, IEEE.
112. Perdomo, O., Rios, H., Rodríguez, F. J., Otálora, S., Meriaudeau, F. et al. (2019). Classification of diabetes-
related retinal diseases using a deep learning approach in optical coherence tomography. Computer Methods
and Programs in Biomedicine, 178, 181–189. DOI 10.1016/j.cmpb.2019.06.016.
113. Jyoti, O., Islam, N., Hasnain, F. M. S. (2020). Prediction of hepatitis disease using effective deep neural
network. 2020 IEEE International Conference for Innovation in Technology (INOCON), pp. 1–5. Bangluru,
India, IEEE.
114. Yu, Y., Wang, J., Ng, C. W., Ma, Y., Mo, S. et al. (2018). Deep learning enables automated scoring of liver
fibrosis stages. Scientific Reports, 8, 16016. DOI 10.1038/s41598-018-34300-2.
115. Xiao, W., Huang, X., Wang, J. H., Lin, D. R., Zhu, Y. et al. (2021). Screening and identifying hepatobiliary
diseases through deep learning using ocular images: A prospective, multicentre study. The Lancet Digital
Health, 3(2), e88–e97. DOI 10.1016/S2589-7500(20)30288-0.
116. Frid-Adar, M., Diamant, I., Klang, E., Amitai, M., Goldberger, J. et al. (2018). GAN-Based synthetic
medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing,
321, 321–331. DOI 10.1016/j.neucom.2018.09.013.
117. Lu, H., Tian, S., Yu, L., Lv, X., Chen, S. (2020). Diagnosis of hepatitis B based on Raman spectroscopy
combined with a multiscale convolutional neural network. Vibrational Spectroscopy, 107, 103038. DOI
10.1016/j.vibspec.2020.103038.
118. Wang, X., Tian, S., Yu, L., Lv, X., Zhang, Z. (2020). Rapid screening of hepatitis B using Raman spec-
troscopy and long short-term memory neural network. Lasers in Medical Science, 35(8), 1791–1799. DOI
10.1007/s10103-020-03003-4.
520 CMES, 2021, vol.128, no.2

119. Zhu, H., Guo, Q., Li, M., Wang, C., Fang, Z. et al. (2020). Host and infectivity prediction of wuhan 2019
novel coronavirus using deep learning algorithm. BioRxiv, DOI 10.1101/2020.01.21.914044.
120. Fu, M., Yi, S. L., Zeng, Y., Ye, F., Li, Y. et al. (2020). Deep learning-based recognizing covid-
19 and other common infectious diseases of the lung by chest CT scan images. MedRxiv. DOI
10.1101/2020.03.28.20046045.
121. Chae, S., Kwon, S., Lee, D. (2018). Predicting infectious disease using deep learning and big data. Interna-
tional Journal of Environmental Research and Public Health, 15(8), 1596. DOI 10.3390/ijerph15081596.
122. Wong, Z. S., Zhou, J., Zhang, Q. (2019). Artificial intelligence for infectious disease big data analytics.
Infection, Disease & Health, 24(1), 44–48. DOI 10.1016/j.idh.2018.10.002.
123. Alnazer, I., Bourdon, P., Urruty, T., Falou, O., Khalil, M. et al. (2021). Recent advances in medical
image processing for the evaluation of chronic kidney disease. Medical Image Analysis, 69, 101960. DOI
10.1016/j.media.2021.101960.
124. Wu, Y., Yi, Z. (2020). Automated detection of kidney abnormalities using multi-feature fusion convolutional
neural networks. Knowledge-Based Systems, 200, 105873. DOI 10.1016/j.knosys.2020.105873.
125. Navaneeth, B., Suchetha, M. (2020). A dynamic pooling based convolutional neural network approach
to detect chronic kidney disease. Biomedical Signal Processing and Control, 62, 102068. DOI
10.1016/j.bspc.2020.102068.
126. Bhaskar, N., Manikandan, S. (2019). A deep-learning-based system for automated sensing of chronic kidney
disease. IEEE Sensors Letters, 3(10), 1–4. DOI 10.1109/LSENS.2019.2942145.
127. Khamparia, A., Saini, G., Pandey, B., Tiwari, S., Gupta, D. et al. (2020). KDSAE: Chronic kidney disease
classification with multimedia data learning using deep stacked autoencoder network. Multimed. Tools
Appl., 79, 35425–35440. DOI 10.1007/s11042-019-07839-z.
128. Ma, F., Sun, T., Liu, L., Jing, H. (2020). Detection and diagnosis of chronic kidney disease using deep
learning-based heterogeneous modified artificial neural network. Future Generation Computer Systems, 111,
17–26. DOI 10.1016/j.future.2020.04.036.
129. Chen, G., Ding, C., Li, Y., Hu, X., Li, X. et al. (2020). Prediction of chronic kidney disease using adaptive
hybridized deep convolutional neural network on the internet of medical things platform. IEEE Access, 8,
100497–100508. DOI 10.1109/ACCESS.2020.2995310.
130. Collins, F. S., Varmus, H. (2015). A new initiative on precision medicine. New England Journal of Medicine,
372(9), 793–795. DOI 10.1056/NEJMp1500523.
131. Li, Y., Lv, Z., Zhao, J., Pan, Z. (2019). Improving performance of medical image fusion using histogram,
dictionary learning and sparse representation. Multimedia Tools and Applications, 78(24), 34459–34482.
DOI 10.1007/s11042-019-08027-9.
132. Lv, Z., Piccialli, F. (2020). The security of medical data on internet based on differential privacy technology.
ACM Transactions on Internet Technology (TOIT), 1–10. DOI 10.1145/3382769.
133. Kumar, R., Khan, A. A., Kumar, J., Zakria, A., Golilarz, N. A. et al. (2021). Blockchain-federated-learning
and deep learning models for COVID-19 detection using CT imaging. IEEE Sensors Journal, 1–1. DOI
10.1109/JSEN.2021.3076767.
134. Hesamian, M. H., Jia, W., He, X., Kennedy, P. (2019). Deep learning techniques for medical
image segmentation: Achievements and challenges. Journal of Digital Imaging, 32(4), 582–596. DOI
10.1007/s10278-019-00227-x.
135. Shen, D., Wu, G., Suk, H. I. (2017). Deep learning in medical image analysis. Annual Review of Biomedical
Engineering, 19, 221–248. DOI 10.1146/annurev-bioeng-071516-044442.
136. Yang, X., Zhang, T., Xu, C., Yan, S., Hossain, M. S. et al. (2016). Deep relative attributes. IEEE Transactions
on Multimedia, 18(9), 1832–1842. DOI 10.1109/TMM.2016.2582379.
137. Levine, A. B., Schlosser, C., Grewal, J., Coope, R., Jones, S. J. et al. (2019). Rise of the machines: Advances
in deep learning for cancer diagnosis. Trends in Cancer, 5(3), 157–169. DOI 10.1016/j.trecan.2019.02.002.
138. Afzal, M., Islam, S. R., Hussain, M., Lee, S. (2020). Precision medicine informatics: Principles, prospects,
and challenges. IEEE Access, 8, 13593–13612. DOI 10.1109/ACCESS.2020.2965955.
CMES, 2021, vol.128, no.2 521

139. Llovet, J. M., Montal, R., Sia, D., Finn, R. S. (2018). Molecular therapies and precision medicine for hepa-
tocellular carcinoma. Nature Reviews Clinical Oncology, 15(10), 599–616. DOI 10.1038/s41571-018-0073-4.
140. Le Tourneau, C., Borcoman, E., Kamal, M. (2019). Molecular profiling in precision medicine oncology.
Nature Medicine, 25, 711–712. DOI 10.1038/s41591-019-0442-2.
141. Wang, Y., Sun, S., Zhang, Z., Shi, D. (2018). Nanomaterials for cancer precision medicine. Advanced
Materials, 30(17), 1705660. DOI 10.1002/adma.201705660.
142. Fujiwara, N., Friedman, S. L., Goossens, N., Hoshida, Y. (2018). Risk factors and prevention of hep-
atocellular carcinoma in the era of precision medicine. Journal of Hepatology, 68(3), 526–549. DOI
10.1016/j.jhep.2017.09.016.
143. Zhang, S., Bamakan, S. M. H., Qu, Q., Li, S. (2018). Learning for personalized medicine: A comprehensive
review from a deep learning perspective. IEEE Reviews in Biomedical Engineering, 12, 194–208. DOI
10.1109/RBME.2018.2864254.
144. Panayides, A. S., Pattichis, M. S., Leandrou, S., Pitris, C., Constantinidou, A. et al. (2018). Radiogenomics
for precision medicine with a big data analytics perspective. IEEE Journal of Biomedical and Health
Informatics, 23(5), 2063–2079. DOI 10.1109/JBHI.2018.2879381.
145. Loomans-Kropp, H. A., Umar, A. (2019). Cancer prevention and screening: The next step in the era of
precision medicine. NPJ Precision Oncology, 3(1), 1–8. DOI 10.1038/s41698-018-0075-9.
146. Regel, I., Mayerle, J., Mahajan, U. M. (2020). Current strategies and future perspectives for precision
medicine in pancreatic cancer. Cancers, 12(4), 1024. DOI 10.3390/cancers12041024.
147. Steuer, C. E., Ramalingam, S. S. (2018). Tumor mutation burden: Leading immunotherapy to the era of
precision medicine. J. Clin. Oncol., 36(7), 631–632. DOI 10.1200/JCO.2017.76.8770.
148. Hamamoto, R., Suvarna, K., Yamada, M., Kobayashi, K., Shinkai, N. et al. (2020). Application of artificial
intelligence technology in oncology: Towards the establishment of precision medicine. Cancers, 12(12),
3532. DOI 10.3390/cancers12123532.
149. Ahmed, Z., Mohamed, K., Zeeshan, S., Dong, X. (2020). Artificial intelligence with multi-functional
machine learning platform development for better healthcare and precision medicine. Database, 2020,
baaa010. DOI 10.1093/database/baaa010.
150. Hulsen, T., Jamuar, S. S., Moody, A. R., Karnes, J. H., Varga, O. et al. (2019). From big data to precision
medicine. Frontiers in Medicine, 6, 34. DOI 10.3389/fmed.2019.00034.
151. Prosperi, M., Min, J. S., Bian, J., Modave, F. (2018). Big data hurdles in precision medicine and precision
public health. BMC Medical Informatics and Decision Making, 18, 139. DOI 10.1186/s12911-018-0719-2.
152. Hey, S. P., Gerlach, C. V., Dunlap, G., Prasad, V., Kesselheim, A. S. (2020). The evidence landscape in
precision medicine. Science Translational Medicine, 12(540). DOI 10.1126/scitranslmed.aaw7745.
153. Dienstmann, R., Vermeulen, L., Guinney, J., Kopetz, S., Tejpar, S. et al. (2017). Consensus molecular
subtypes and the evolution of precision medicine in colorectal cancer. Nature Reviews Cancer, 17(2), 79–92.
DOI 10.1038/nrc.2016.126.
154. Dayem Ullah, A. Z., Oscanoa, J., Wang, J., Nagano, A., Lemoine, N. R. et al. (2018). SNPnexus: Assessing
the functional relevance of genetic variation to facilitate the promise of precision medicine. Nucleic Acids
Research, 46(W1), W109–W113. DOI 10.1093/nar/gky399.
155. Tebeje, T. H., Klein, J. (2020). Applications of e-health to support person-centered health care at the time
of COVID-19 pandemic. Telemedicine and E-Health, 27, 150–158. DOI 10.1089/tmj.2020.0201.
156. Hossain, M. S. (2017). Cloud-supported cyber–Physical localization framework for patients monitoring.
IEEE Systems Journal, 11(1), 118–127. DOI 10.1109/JSYST.2015.2470644.
157. Bisio, I., Garibotto, C., Lavagetto, F., Sciarrone, A. (2019). When eHealth meets IoT: A smart wire-
less system for post-stroke home rehabilitation. IEEE Wireless Communications, 26(6), 24–29. DOI
10.1109/MWC.001.1900125.
158. Dubey, H., Monteiro, A., Constant, N., Abtahi, M., Borthakur, D. et al. (2017). Fog computing in medical
internet-of-things: Architecture, implementation, and applications. in Handbook of Large-Scale Distributed
Computing in Smart Healthcare, pp. 281–321. Cham: Springer.
522 CMES, 2021, vol.128, no.2

159. Moore, C. (2020). Medical internet of things-based healthcare systems: Wearable sensor-based devices,
patient-generated Big data, and real-time clinical monitoring. American Journal of Medical Research, 7(1),
41–47. DOI 10.22381/AJMR7120206.
160. Harvey, P., Toutsop, O., Kornegay, K., Alale, E., Reaves, D. (2020). Security and privacy of medical
internet of things devices for smart homes. 2020 7th International Conference on Internet of Things: Systems,
Management and Security (IOTSMS), pp. 1–6. Paris, France, IEEE.
161. Hossain, M. S., Muhammad, G. (2017). Emotion-aware connected healthcare big data towards 5G. IEEE
Internet of Things Journal, 5(4), 2399–2406. DOI 10.1109/JIOT.2017.2772959.
162. Zhou, Z., Yu, H., Shi, H. (2020). Human activity recognition based on improved Bayesian convolution
network to analyze health care data using wearable IoT device. IEEE Access, 8, 86411–86418. DOI
10.1109/ACCESS.2020.2992584.
163. Fouad, H., Hassanein, A. S., Soliman, A. M., Al-Feel, H. (2020). Analyzing patient health information
based on IoT sensor with AI for improving patient assistance in the future direction. Measurement, 159,
107757. DOI 10.1016/j.measurement.2020.107757.
164. Hossain, M. S., Muhammad, G., Alamri, A. (2019). Smart healthcare monitoring: A voice pathology
detection paradigm for smart cities. Multimedia Systems, 25, 565–575. DOI 10.1007/s00530-017-0561-x.
165. Al-Garadi, M. A., Mohamed, A., Al-Ali, A. K., Du, X., Ali, I. et al. (2020). A survey of machine and deep
learning methods for internet of things (IoT) security. IEEE Communications Surveys & Tutorials, 22(3),
1646–1685. DOI 10.1109/COMST.2020.2988293.
166. Alloghani, M., Baker, T., Al-Jumeily, D., Hussain, A., Mustafina, J. et al. (2020). Prospects of machine
and deep learning in analysis of vital signs for the improvement of healthcare services. Nature-Inspired
Computation in Data Mining and Machine Learning, pp. 113–136. Cham: Springer.
167. Wang, T., Lei, Y., Tian, Z., Dong, X., Liu, Y. et al. (2019). Deep learning-based image quality improvement
for low-dose computed tomography simulation in radiation therapy. Journal of Medical Imaging, 6(4),
043504. DOI 10.1117/1.JMI.6.4.043504.
168. Mohan, S., Thirumalai, C., Srivastava, G. (2019). Effective heart disease prediction using hybrid machine
learning techniques. IEEE Access, 7, 81542–81554. DOI 10.1109/ACCESS.2019.2923707.
169. Liu, Y., Kang, Y., Xing, C., Chen, T., Yang, Q. (2020). A secure federated transfer learning framework. IEEE
Intelligent Systems, 35(4), 70–82. DOI 10.1109/MIS.2020.2988525.
170. Chen, Y., Qin, X., Wang, J., Yu, C., Gao, W. (2020). Fedhealth: A federated transfer learning framework for
wearable healthcare. IEEE Intelligent Systems, 35(4), 83–93. DOI 10.1109/MIS.2020.2988604.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy