Brain Learning 2
Brain Learning 2
Brain Learning 2
https://doi.org/10.1007/s40120-019-00153-8
REVIEW
incomplete data sources, with the outcomes towards the development of novel diagnostic
being cumbersome to track. Identifying novel and prognostic tools. Deep learning techniques
clinical patterns amidst this complexity is defi- are particularly promising in the neurosciences
nitely not a trivial task [1–3]. where clinical diagnoses often rely on subtle
Modern representation learning methods symptoms and complicated neuroimaging
enable the automatic discovery of representa- modalities with granular and high-dimensional
tions needed to generate insights from raw data signals. In this article, we discuss the applications
[4]. Deep learning algorithms are an example of of deep learning in neurology and the ongoing
such representation learning approaches that challenges, with an emphasis on aspects relevant
hierarchically compose nonlinear functions to to the diagnosis of common neurologic disor-
transform raw input data into more sophisti- ders. However, our aim is not to provide com-
cated features that enable the identification of prehensive technical details of deep learning or
novel patterns [5]. Such approaches have its broader applications. We begin with a brief
proved to be essential in modern engineering overview of deep learning techniques followed
breakthroughs—from face recognition and self- by a review of applications in the clinical neuro-
driving cars to chat-bots and language transla- science field. We conclude the review with a
tion [6–12]. In medicine, the successful appli- short discussion on existing challenges and a
cation of deep learning algorithms to routine look to the future. This article is based on previ-
tasks has enabled a flood of academic and ously conducted studies and does not contain
commercial research, with publications on var- any studies with human participants or animals
ious applications growing from 125 published performed by any of the authors.
papers identified as machine learning publica-
tions in arXiv, the electronic scientific and
engineering paper archive, in 2000, to more FUNDAMENTALS OF DEEP
than 3600 by November of 2018 (see Fig. 1). LEARNING
The multidiscipline of clinical neurosciences
has similarly experienced the beginnings of an Machine learning is a subset of artificial intelli-
impact from deep learning, with movement gence that learns complex relationships among
Fig. 1 Machine learning publications in PubMed by year through 2018 showing the exponential growth of interest in the
field, as reported by the US National Library of Medicine of the National Institutes of Health [13]
Neurol Ther (2019) 8:351–365 353
variables in data [14]. The power of machine techniques often require manual feature engi-
learning comes from its ability to derive pre- neering and suffer from overhead that limits
dictive models from large amounts of data with their utility in scenarios that require near real-
minimal or, in some cases, entirely without the time decision-making.
need for prior knowledge of the data or any Deep learning differs from traditional
assumptions about the data. One of the most machine learning in how representations are
widely discussed modern machine learning automatically discovered from raw data. In
algorithms, the artificial neural network (ANN), contrast to ANNs, which are shallow feature
draws inspiration from biological neural net- learning techniques, deep learning algorithms
works that constitute mammalian brains. The employ multiple, deep layers of perceptrons
functional unit of the ANN is the perceptron, that capture both low- and high-level repre-
which partitions input data into separable cat- sentations of data, enabling them to learn richer
egories or classes [15]. When hierarchically abstractions of inputs [5]. This obviates the need
composed into a network, the perceptron for manual engineering of features and allows
becomes an essential building block for modern deep learning models to naturally uncover pre-
deep neural networks (DNNs), such as multi- viously unknown patterns and generalize better
layer perceptron classifiers. Similar examples of to novel data. Variants of these algorithms have
commonly used traditional machine learning been employed across numerous domains in
algorithms include linear regression (LR), logis- engineering and medicine.
tic regression, support vector machines (SVMs), Convolutional neural networks (CNNs) have
and the Naı̈ve Bayes classifier (Fig. 2). garnered particular attention within computer
These traditional machine learning methods vision and imaging-based medical research
have been important in furthering advance- [17, 18]. CNNs gather representations across
ments in medicine and genomics. As an exam- multiple layers, each of which learns specific
ple, LR has proven useful in the search for features of the image, much like the human
complex, multigene signatures that can be visual cortex is arranged into hierarchical layers,
indicative of disease onset and prognosis, tasks including the primary visual cortex (edge
which are otherwise too intricate and cumber- detection), secondary visual cortex (shape
some even for researchers with professional detection), and so forth [19]. CNNs consist of
training [16]. Although such tools have been convolutional layers in which data features are
very effective in parsing massive datasets and learned: pooling layers, which reduce the
identifying relationships between variables of number of features, and therefore computa-
interest, traditional machine learning tional demand, by aggregating similar or
Fig. 2 Breakdown of algorithm types in the machine learning family that are commonly used in medical subdomain
research and analyses
354 Neurol Ther (2019) 8:351–365
redundant features; dropout layers, which of data. As the two networks engage in this
selectively turn off perceptrons to avoid over- adversarial process, the fidelity of the generated
reliance on a single component of the network; data gradually improves. In some contexts, the
and a final output layer, which collates the resulting data have been utilized to augment
learned features into a score or class decision, existing datasets [29].
i.e., whether or not a given radiograph shows These strides in deep learning are largely due
signs of ischemia. These algorithms have to breakthroughs in computing capabilities and
achieved rapid profound success in image clas- the open-source nature of research in the field.
sification tasks and, in some cases, have mat- The application of graphics processing units to
ched board-certified human performance deep learning research has dramatically accel-
[20–24]. erated the size and complexity of algorithm
Recurrent neural networks and variants, such architectures and simultaneously reduced the
as long short-term memory (LSTM) and gated time to train such algorithms from months to
recurrent units, have revolutionized the analy- the order of days. The consequence has been
sis of time-series data that can be found in high-throughput research characterized by
videos, speech, and texts [25]. These algorithms rapid experimentation, ultimately enabling
sequentially analyze each element of input data more efficacious algorithms. In addition, the
and employ a gating mechanism to determine rise of open-source deep learning frameworks,
whether to maintain or discard information such as TensorFlow, Keras, PyTorch, Caffe, and
from prior elements when generating outputs. others, has increased accessibility to technical
In this manner, they efficiently capture long- advances and facilitated the sharing of ideas
term dependencies and have revolutionized and their rapid application across various
machine translation, speech processing, and domains [30, 31]. The truly collaborative nature
text analysis. of deep learning research has led to surprising
Autoencoders (AEs) are a class of unsuper- innovations and changed the landscape of
vised learning algorithms that discover mean- medical research and care.
ingful representations of data by learning a
lower-dimensional mapping from inputs to
outputs [26, 27]. They are composed of an LITERATURE REVIEW
encoder, which learns a latent representation of
the input, and a decoder, which reconstructs In this article, we review and summarize pub-
the input from the latent representation. By lished literature on the application of deep
constraining the latent representation to a learning to the clinical neurosciences. We used
lower dimensionality than the input, AEs are search engines and repositories such as Google
able to learn a compressed representation of Scholar, PubMed, ScienceDirect, and arXiv to
data that contains only the features necessary to identify and review existing literature and per-
reconstruct the input. Such algorithms are often formed keyword searches of these databases
employed to learn features that can be subse- using the following terms: ‘‘deep learning,’’
quently utilized in conjunction with the deep ‘‘machine learning,’’ ‘‘neurology,’’ ‘‘brain,’’ and
learning techniques previously discussed. ‘‘MRI.’’ Following a comprehensive review of
Generative adversarial networks are a newer the literature initially retrieved, we identified
class of algorithms aimed at generating novel 312 articles as containing one or more keywords
data that statistically mimic input data by associated with our queries. Of these articles,
approximating a latent distribution for the data 134 were subsequently identified as being rele-
[28]. Such algorithms are composed of two vant to the subject of this review. Following
competing (‘‘adversarial’’) networks: a genera- collation of the relevant articles, we grouped
tor, which produces synthetic data from noise articles first into broad modalities, namely
by sampling from an approximated distribu- image classification, image segmentation,
tion, and a discriminator, which aims to dif- functional connectivity and classification of
ferentiate between real and synthetic instances brain disorders, and risk prognostication.
Neurol Ther (2019) 8:351–365 355
Within these areas, we then grouped publica- Among the first such applications, Suk and Shen
tions into disease applications. We focused our utilized a stacked AE to learn multimodal brain
discussion on the clinical implications of the representations from structural MRI and posi-
developments in the field. tron emission tomography (PET), and incorpo-
rated those features with cerebrospinal fluid
biomarker data and clinical scores from the
DEEP LEARNING IN NEUROLOGY Mini-Mental State Examination (MMSE) and
Alzheimer’s Disease Assessment Scale-Cognitive
The deep learning techniques described above subscale (ADAS-Cog) to train an SVM classifier
are playing an increasingly crucial role in neu- that improved diagnostic accuracy [34]. Other
rological research, tackling problems within approaches pre-trained a stacked AE using nat-
several subdomains. First, radiological image ural images (everyday images) prior to training
classification and segmentation has been a tra- on brain MR images in order to learn more
ditional locus of deep learning development high-fidelity anatomical features, such as gray
efforts. Image classification and segmentation matter and structural deformities, for incorpo-
tasks are uniquely suited to deep learning due to ration into a CNN [35]. Variations on these
the high-dimensional nature of neuroimaging approaches have been used to incrementally
data which is unfavorable to manual analysis, improve diagnostic performance [36–42].
combined with the naturally digital nature of Whereas older approaches were limited to
most modern imaging. Secondly, deep learning two-dimensional (2D) slices of medical images
has been applied to functional brain mapping due to computational constraints, newer appli-
and correlational studies using functional cations have been able to incorporate the full
magnetic resonance imaging (fMRI) data for 3D volume of an imaging modality for AD
tasks such as prediction of postoperative sei- detection. Among the first such examples was
zure. Lastly, diagnostic prognostication with work by Payan and Montana in which they
deep learning using multiple data types, trained a sparse AE on 3D patches of MRI scans
including lab values, images, notes, among to learn a volumetric brain representation that
others, has been used to assign disease risk. In was used to pre-train a 3D CNN for AD diagnosis
the following sections, we discuss the successes [43]. More recently, Hosseini-Asl et al. used an
and challenges inherent in the deep learning adaptable training regime with a 3D CNN pre-
approaches adopted towards these tasks, as well trained by a convolutional AE to learn general-
as the limitations and difficulties that such izable AD biomarkers [44, 45]. This approach
methods face within the field of neurology and was notable because it allowed the transfer of
within medicine as a whole. learned features from the source CADDementia
dataset to the target ADNI dataset, resulting in
Medical Image Classification state-of-the-art AD diagnosis accuracy on an
external dataset. Analogous work with volu-
The first application of deep learning in medi- metric data has been conducted in the com-
cine involved the analysis of imaging modali- puted tomography (CT) domain to differentiate
ties, especially those for the detection of AD from brain lesions and the processes of
Alzheimer’s disease (AD) and other cognitive normal aging [46].
impairments. A variety of publicly available The most recent work has built on existing
databases, such as the Alzheimer’s Disease work in AD diagnosis and focused on predicting
Neuroimaging Initiative (ADNI) and Brain the onset of AD in at-risk patients in order to
Tumor Segmentation Benchmark (BraTS), have stem progression of the disease. Ding et al. used
become available to spur advancements in fluorine-18-fluorodeoxyglucose PET scans of the
neuroimaging analysis [32, 33]. brain derived from the ADNI database to train a
Early approaches used AEs in conjunction CNN to diagnose AD [47]. Unlike many inves-
with a classifier to distinguish AD, mild cogni- tigators before them, however, the authors
tive impairments (MCI) and healthy controls. evaluated the efficacy of their algorithm against
356 Neurol Ther (2019) 8:351–365
data from the long-term follow-up of patients segmentations by expert raters show consider-
who did not have AD at the time. Interestingly, able variation in images obscured by field arti-
they found that the algorithm predicted onset facts or where intensity gradients are minimal,
of AD on average 75.8 months prior to the final and rudimentary algorithms struggle to achieve
diagnosis on an independent dataset, which consistency in an anatomy that can vary con-
surpassed the diagnostic performance of three siderably from patient to patient [33]. In light of
expert radiologists. these factors, deep learning segmentation of
Deep learning-based image classification has neuroanatomy has become a prime target for
also been applied in the diagnosis of acute efforts in deep learning research.
neurologic events, such as intracranial hemor- Measurement of the performance of neu-
rhage (ICH) and cranial fractures, with the aim roanatomical segmentation algorithms has
of reducing time to diagnosis by optimizing been standardized by the BraTS, which was
neuroradiology workflows. Titano et al. trained established at the 2012 and 2013 Medical Image
a 3D CNN in a weakly supervised manner on Computing and Computer Assisted Interven-
37,236 CT scans to identify ICH for the pur- tions (MICCAI) conference [33]. Prior to the
poses of triaging patient cases [48]. They lever- establishment of this challenge, segmentation
aged a natural language processing algorithm algorithms were often evaluated on private
trained on 96,303 radiology reports to generate imaging collections only, with variations in the
silver-standard labels for each imaging study imaging modalities incorporated and the met-
and validated the efficacy of their CNN on a rics used to evaluate effectiveness. The estab-
subset of studies with gold standard labels gen- lishment of BraTS has been critical in
erated by manual chart review [49]. The inves- standardizing the evaluation of various models
tigators conducted a double-blind randomized for the determination of which to pursue in
control trial to compare whether the algorithm clinical practice. At the time of BraTS estab-
or expert radiologists could more effectively lishment, the models being evaluated were lar-
triage studies in a simulated clinical environ- gely simple machine learning models, including
ment and found that the CNN was 150-fold four random forest-based segmentation models
faster in evaluating a study and significantly [33]. Since then, there has been considerable
outperformed humans in prioritizing the most advancement in performance, largely based on
urgent cases. Subsequent studies have similarly the adoption of CNNs for anatomical
demonstrated the potential for deep learning to segmentation.
optimize radiology workflows in the diagnosis The traditional computational approach to
of ICH and detect as many as nine critical segmentation is to employ an atlas-based seg-
findings on head CT scans with sensitivity mentation, namely the FreeSurfer software,
comparable to that of expert radiologists which assigns one of 37 labels to each voxel in a
[50–52]. 3D MRI scan based on probabilistic estimates
[56]. In a recent comparative study, Wachinger
Medical Image Segmentation et al. designed and applied a deep CNN, called
DeepNAT, for the purposes of segmenting neu-
Segmentation of radiological brain images is roanatomy visualized in T1-weighted MRI scans
critical for the measurement of brain regions, into 25 different brain regions. The authors
including shape, thickness, and volume, that used the MICCAI Multi-Atlas Labeling chal-
are important for the quantification of struc- lenge, consisting of 30 T1-weighted images, in
tural changes within the brain that occur either addition to manually labeled segmentations
naturally or due to various disease processes [53, 57]. When the authors compared the cur-
[53]. Accurate structural classification is partic- rent clinical standard, FreeSurfer, which uses its
ularly important in patients with gliomas, the own anatomical atlas to assign anatomic labels,
most common brain tumor type, with less than to DeepNAT, they found that DeepNAT showed
a 2-year survival time [54, 55]. Manual statistically significant performance improve-
ments. Performance in segmentation was
Neurol Ther (2019) 8:351–365 357
LSTMs, achieved worse results, namely poorer potential to improve current standards for pre-
sensitivity and a higher hourly rate of false dicting clinical events.
positives [75, 76].
Genetic data has been another important
area of research and development for precision CHALLENGES
medicine. Predictive tasks in large-scale geno-
mic profiles face high-dimensional datasets that Despite the profound biomedical advances due
are often pared down by experts who hand-se- to deep learning algorithms, there remain sig-
lect a small number of features for training nificant challenges that must be addressed
predictive models [77]. In ASD, deep learning before such applications gain widespread use.
has played a particularly important role in We discuss some of the most critical hurdles in
determining the impact of de-novo mutations, the following sections.
including copy number variants and point
mutations, on ASD severity [78]. Using a deep Data Volume
CNN, Zhou et al. modeled the biochemical
impact of observed point mutations in 1790 Deep neural networks are computationally
whole-genome sequenced families with ASD, on intensive, multilayered algorithms with param-
both the RNA and DNA levels [78]. This eters on the order of millions. Convergence of
approach revealed that both transcriptional and such algorithms requires data commensurate
post-transcriptional mechanisms play a major with the number of parameters. Although there
role in ASD, suggesting biological convergence are no strict rules governing the amount of data
of genetic dysregulation in ASD. required to optimally train DNNs, empirical
Genomic data, either alone or in conjunc- studies suggest that tenfold more training data
tion with neuroimaging and histopathology, relative to the number of parameters is required
has provided cancer researchers a wealth of data to produce an effective model. It is no surprise
on which to perform cancer-related predictive then that domains, such as computer vision and
tasks [77, 79, 80]. Deep learning offers several natural language processing, have seen the most
advantages when working simultaneously with rapid progress due to deep learning given the
multiple data modalities, removing subjective wide availability of images, videos, and free-
interpretations of histological images, accu- form text on the Internet.
rately predicting time-to-event outcomes, and Biomedical data on the other hand is mostly
even surpassing gold standard clinical para- decentralized—stored locally within hospital
digms for glioma patient survival [80]. Using systems—and subject to privacy constraints that
high-powered histological slices and genetic make such data less readily accessible for
data, namely IDH mutation status and 1p/19q research. Furthermore, given the complexity of
codeletion, on 769 patients from The Cancer patient presentations and disease processes,
Genome Atlas (TCGA), Mobadersaney et al. reliable ground truth labels for biomedical
used a survival CNN (SCNN) to predict time-to- applications are extremely expensive to obtain,
event outcomes. The histological and genetic often requiring the efforts of multiple highly
model performed on par with manual histologic specialized domain experts. This paucity of
grading or molecular subtyping [80]. In a sec- labeled data remains an important bottleneck in
ond paper by this group, SCNNs were shown to the development of deep learning applications
outperform other machine learning algorithms, in medicine.
including random forest, in classification tasks
using genetic data from multiple tumor types, Data Quality
including kidney, breast, and pan-glioma can-
cers [77]. The ability of deep learning algo-
Healthcare data are fundamentally ill-suited for
rithms to reduce subjectivity in histologic
deep learning applications. Electronic medical
grading and disentangle complex relationships
records are highly heterogeneous, being
between noisy EEG or genetic data, has the
360 Neurol Ther (2019) 8:351–365
composed of clinical notes, a miscellany of made by algorithms and enable their wide-
various codes, and other patient details that spread adoption. As such, major efforts within
may often be missing or incomplete. Clinical the deep learning community to tackle prob-
notes consist of nuanced language and acro- lems of interpretability and explainability have
nyms that often vary by specialty and contain the potential to be particularly beneficial for
redundant information that provides an inac- facilitating the use of deep learning methods in
curate temporal representation of disease onset healthcare.
or progression. Diagnosis codes suffer from a
similar fate as they track billing for insurance Legal
purposes instead of health outcomes. This
inherent complexity makes it impossible for Medical malpractice rules govern standards of
deep learning algorithms to parse signal from clinical practice in order to ensure the appro-
noise. priate care of patients. However, to date, no
standards have been established to assign cul-
Generalizability pability in contexts where algorithms provide
poor predictions or substandard treatment rec-
Although existing deep learning applications ommendations. The establishment of such reg-
have garnered success in silico, their widespread ulations is a necessary prerequisite for the
adoption in real-world clinical settings remains widespread adoption of deep learning algo-
limited due to concerns over their efficacy rithms in clinical contexts.
across clinical contexts. Much of the concern is
based on the tendency of deep learning algo- Ethical
rithms to overfit to the statistical characteristics
of the training data, rendering them hyper- Incidental introduction of bias must be care-
specialized for a hospital or certain patient fully evaluated in the application of deep
demographic and less effective on the popula- learning in medicine. As discussed previously,
tion at-large [81, 82]. The siloed existence of deep learning algorithms are uniquely adept at
healthcare data in hospitals and the hetero- fitting to the characteristics of the data on
geneity of data across healthcare systems make which they are trained. Such algorithms have
the task of developing generalizable models the capability to perpetuate inequities against
even more difficult. And even when multi-in- populations underrepresented in medicine and,
stitutional data are acquired, the data are often by extension, in the very healthcare data used
retrospective in nature, which prevents practi- to train the algorithms. Furthermore, recent
cal assessment of algorithm performance. research evaluating algorithmic bias in a com-
mercial healthcare algorithm provides a cau-
Interpretability tionary tale on the importance of critically
evaluating the very outcomes algorithms are
The power of deep learning algorithms to map trained to predict [83].
complex, nonlinear functions can render them
difficult to interpret. This becomes an impor-
tant consideration in healthcare applications CONCLUSION
where the ability to identify drivers of outcomes
becomes just as important as the ability to Deep learning has the potential to fundamen-
accurately predict the outcome itself. In the tally alter the practice of medicine. The clinical
clinical setting, where clinical decision support neurosciences in particular are uniquely situ-
systems are designed to augment the decision- ated to benefit given the subtle presentation of
making capacity of healthcare professionals, symptoms typical of neurologic disease. Here,
interpretability is critical to convince healthcare we reviewed the various domains in which deep
professionals to rely on the recommendations learning algorithms have already provided
Neurol Ther (2019) 8:351–365 361
impetus for change—areas such as medical Disclosures. Aly Al-Amyn Valliani, Daniel
image analysis for improved diagnosis of AD Ranti and Eric Karl Oermann have nothing to
and the early detection of acute neurologic disclose.
events; medical image segmentation for quan-
titative evaluation of neuroanatomy and vas- Compliance with Ethics Guidelines. This
culature; connectome mapping for the article is based on previously conducted studies
diagnosis of AD, ASD, and ADHD; and mining and does not contain any studies with human
of microscopic EEG signals and granular genetic participants or animals performed by any of the
signatures. Amidst these advances, however, authors.
important challenges remain a barrier to inte-
gration of deep learning tools in the clinical Data availability. Data sharing is not
setting. While technical challenges surrounding applicable to this article as no datasets were
the generalizability and interpretability of generated or analyzed during the current study.
models are active areas of research and progress,
more difficult challenges surrounding data pri- Open Access. This article is distributed
vacy, accessibility, and ownership will necessi- under the terms of the Creative Commons
tate conversations in the healthcare Attribution-NonCommercial 4.0 International
environment and society in general to arrive at License (http://creativecommons.org/licenses/
solutions that benefit all relevant stakeholders. by-nc/4.0/), which permits any noncommer-
The challenge of data quality, in particular, may cial use, distribution, and reproduction in any
prove to be a uniquely suitable target for medium, provided you give appropriate credit
addressing using deep learning techniques that to the original author(s) and the source, provide
have already demonstrated efficacy in image a link to the Creative Commons license, and
analysis and natural language processing. indicate if changes were made.
Overcoming these hurdles will require the
efforts of interdisciplinary teams of physicians,
computer scientists, engineers, legal experts,
and ethicists working in concert. It is only in
REFERENCES
this manner that we will truly realize the
1. Jensen PB, Jensen LJ, Brunak S. Mining electronic
potential of deep learning in medicine to aug-
health records: towards better research applications
ment the capability of physicians and enhance and clinical care. Nat Rev Genet. 2012;13(6):395–405.
the delivery of care to patients.
2. Luo J, Wu M, Gopukumar D, Zhao Y. Big data
application in biomedical research and health care:
a literature review. Biomed Inform Insights.
ACKNOWLEDGEMENTS 2016;19(8):1–10.
7. Gilani SZ, Mian A. Learning from millions of 3D 20. Gulshan V, Peng L, Coram M, Stumpe MC, Wu D,
scans for large-scale 3D face recognition. 2017. Narayanaswamy A, et al. Development and valida-
http://arxiv.org/abs/1711.05942. tion of a deep learning algorithm for detection of
diabetic retinopathy in retinal fundus photographs.
8. Ramanishka V, Chen Y-T, Misu T, Saenko K. Toward JAMA. 2016;316(22):2402–10.
driving scene understanding: a dataset for learning
driver behavior and causal reasoning. In: Proceed- 21. Esteva A, Kuprel B, Novoa RA, et al. Dermatologist-
ings of IEEE conference on computer vision and level classification of skin cancer with deep neural
pattern recognition. Salt Lake City, UT. 2018. networks. Nature. 2017;542(7639):115–8.
pp. 7699–707.
22. Haenssle HA, Fink C, Schneiderbauer R, et al. Man
9. Maqueda AI, Loquercio A, Gallego G, Garcia N, against machine: diagnostic performance of a deep
Scaramuzza D. Event-based vision meets deep learning convolutional neural network for dermo-
learning on steering prediction for self-driving cars. scopic melanoma recognition in comparison to 58
2018. http://arxiv.org/abs/1804.01310. dermatologists. Ann Oncol. 2018;29(8):1836–42.
10. Mazaré P-E, Humeau S, Raison M, Bordes A. Train- 23. De Fauw J, Ledsam JR, Romera-Paredes B, et al.
ing millions of personalized dialogue agents. 2018. Clinically applicable deep learning for diagnosis
http://arxiv.org/abs/1809.01984. and referral in retinal disease. Nat Med.
2018;24(9):1342–50.
11. Zhang S, Dinan E, Urbanek J, Szlam A, Kiela D,
Weston J. Personalizing dialogue agents: I have a 24. Poplin R, Varadarajan AV, Blumer K, et al. Predic-
dog, do you have pets too? 2018. http://arxiv.org/ tion of cardiovascular risk factors from retinal fun-
abs/1801.07243. dus photographs via deep learning. Nat Biomed
Eng. 2018;2(3):158–64.
12. Wu Y, Schuster M, Chen Z, et al. Google’s neural
machine translation system: bridging the gap 25. Lipton ZC, Berkowitz J, Elkan C. A critical review of
between human and machine translation. 2016. recurrent neural networks for sequence learning.
http://arxiv.org/abs/1609.08144. 2015. http://arxiv.org/abs/1506.00019.
13. US National Library of Medicine National Institutes 26. Rumelhart DE, McClelland JL. Learning internal
of Health. PubMed. 2019. https://www.ncbi.nlm. representations by error propagation. In: Parallel
nih.gov/pubmed/?term=Machine?Learning. distributed processing: explorations in the
microstructure of cognition: foundations. Wacht-
14. Mitchell TM. The discipline of machine learning, endonk: MITP Verlags-GmbH & Co. KG; 1987.
vol. 9. Pittsburgh: School of Computer Science, pp. 318–62.
Carnegie Mellon University; 2006.
27. Hinton GE, Salakhutdinov RR. Reducing the
15. Rosenblatt F. The perceptron: A probabilistic model dimensionality of data with neural networks. Sci-
for information storage and organization in the ence. 2006;313(5786):504–7.
brain. Psychol Rev. 1958;65:386–408. http://dx.doi.
org/10.1037/h0042519. 28. Goodfellow IJ, Pouget-Abadie J, Mirza M, et al.
Generative adversarial networks. 2014. http://arxiv.
16. Ogutu JO, Schulz-Streeck T, Piepho H-P. Genomic org/abs/1406.2661.
selection using regularized linear regression models:
ridge regression, lasso, elastic net and their exten- 29. Shin H-C, Tenenholtz NA, Rogers JK, et al. Medical
sions. BMC Proc. 2012;6[Suppl 2]:S10. image synthesis for data augmentation and
anonymization using generative adversarial net-
17. Krizhevsky A, Sutskever I, Hinton GE. ImageNet works. In:Proc Third International Workshop,
classification with deep convolutional neural net- SASHIMI 2018, held in conjunction with MICCAI
works. In: Pereira F, Burges CJC, Bottou L, Wein- 2018, Granada, Spain, September 16, 2018. In:
berger KQ, editors. Advances in neural information Gooya A, Goksel O, Oguz I, Burgos N, editors.
processing systems, vol. 25. New York: Curran Simulation and synthesis in medical imaging.
Associates, Inc.; 2012; 1097–105. Cham: Springer International Publishing;
2018:1–11.
18. Szegedy C, Liu W, Jia Y, et al. Going deeper with
convolutions. 2014. http://arxiv.org/abs/1409.4842. 30. Shi S, Wang Q, Xu P, Chu X. Benchmarking state-
of-the-art deep learning software tools. 2016.
19. Saba L, Biswas M, Kuppili V, et al. The present and http://arxiv.org/abs/1608.07249.
future of deep learning in radiology. Eur J Radiol.
2019;114:14–24. 31. Liu J, Dutta J, Li N, Kurup U, Shah M. Usability
study of distributed deep learning frameworks for
Neurol Ther (2019) 8:351–365 363
33. Menze BH, Jakab A, Bauer S, et al. The Multimodal 45. Hosseini-Asl E, Ghazal M, Mahmoud A, et al. Alz-
Brain Tumor Image Segmentation Benchmark heimer’s disease diagnostics by a 3D deeply super-
(BRATS). IEEE Trans Med Imaging. vised adaptable convolutional network. Front
2015;34(10):1993–2024. Biosci. 2018;1(23):584–96.
34. Suk H-I, Shen D. Deep learning-based feature rep- 46. Gao XW, Hui R. A deep learning based approach to
resentation for AD/MCI classification. Med Image classification of CT brain images. In: 2016 SAI
Comput Comput Assist Interv. 2013;16(Pt computing conference (SAI). London, UK. 2016.
2):583–90. pp. 28–31. http://ieeexplore.ieee.org.
35. Gupta A, Ayhan M, Maida A. Natural image bases to 47. Ding Y, Sohn JH, Kawczynski MG, et al. A deep
represent neuroimaging data. In: Proceedings of learning model to predict a diagnosis of Alzheimer
30th international conference on machine learn- disease by using 18F-FDG PET of the brain. Radiol-
ing. vol. 28. Atlanta, GA. 2013. pp. 987–94. ogy. 2019;290(2):456–64.
36. Li F, Tran L, Thung K-H, Ji S, Shen D, Li J. Robust 48. Titano JJ, Badgeley M, Schefflein J, et al. Automated
deep learning for improved classification of AD/ deep-neural-network surveillance of cranial images
MCI Patients. Machine learning in medical imag- for acute neurologic events. Nat Med.
ing. New York: Springer International Publishing; 2018;24(9):1337–41.
2014:240–7.
49. Zech J, Pain M, Titano J, et al. Natural language-
37. Liu S, Liu S, Cai W, Pujol S, Kikinis R, Feng D. Early based machine learning models for the annotation
diagnosis of Alzheimer’s disease with deep learning. of clinical radiology reports. Radiology.
In: 2014 IEEE 11th international symposium on 2018;30:171093.
biomedical imaging (ISBI). Beijing, China. 2014.
pp. 1015–8. http://ieeexplore.ieee.org. 50. Arbabshirani MR, Fornwalt BK, Mongelluzzo
GJ,et al. Advanced machine learning in action:
38. Liu S, Liu S, Cai W, et al. Multimodal neuroimaging identification of intracranial hemorrhage on com-
feature learning for multiclass diagnosis of Alzhei- puted tomography scans of the head with clinical
mer’s disease. IEEE Trans Biomed Eng. workflow integration. NPJ Digit Med. 2018;1(1):9.
2015;62(4):1132–40.
51. Chilamkurthy S, Ghosh R, Tanamala S, et al. Deep
39. Suk H-I, Lee S-W, Shen D. Alzheimer’s disease learning algorithms for detection of critical findings
neuroimaging initiative. Latent feature representa- in head CT scans: a retrospective study. Lancet.
tion with stacked auto-encoder for AD/MCI diag- 2018;392(10162):2388–96.
nosis. Brain Struct Funct. 2015;220(2):841–59.
52. Lee H, Yune S, Mansouri M, et al. An explainable
40. Sarraf S, Tofighi G. Classification of Alzheimer’s deep-learning algorithm for the detection of acute
disease using fMRI data and deep learning convo- intracranial haemorrhage from small datasets. Nat
lutional neural networks. 2016. http://arxiv.org/ Biomed Eng. 2018;5:6. https://doi.org/10.1038/
abs/1603.08631. s41551-018-0324-9.
41. Suk H-I, Lee S-W, Shen D. Alzheimer’s disease 53. Wachinger C, Reuter M, Klein T. DeepNAT: deep
neuroimaging initiative. Deep sparse multi-task convolutional neural network for segmenting neu-
learning for feature selection in Alzheimer’s disease roanatomy. Neuroimage. 2018;15(170):434–45.
diagnosis. Brain Struct Funct. 2016;221(5):2569–87.
54. Ohgaki H, Kleihues P. Population-based studies on
42. Valliani A, Soni A. Deep residual nets for improved incidence, survival rates, and genetic alterations in
Alzheimer’s diagnosis. In: BCB. Boston, MA. 2017. astrocytic and oligodendroglial gliomas. J Neu-
p. 615. ropathol Exp Neurol. 2005;64(6):479–89.
43. Payan A, Montana G. Predicting Alzheimer’s dis- 55. Holland EC. Progenitor cells and glioma formation.
ease: a neuroimaging study with 3D convolutional Curr Opin Neurol. 2001;14(6):683–8.
364 Neurol Ther (2019) 8:351–365
56. Fischl B, Salat DH, Busa E, Albert M, Dieterich M, 68. Montufar GF, Pascanu R, Cho K, Bengio Y. On the
Haselgrove C, et al. Whole brain segmentation: number of linear regions of deep neural networks.
automated labeling of neuroanatomical structures In: Ghahramani Z, Welling M, Cortes C, Lawrence
in the human brain. Neuron. 2002;33(3):341–55. ND, Weinberger KQ, editors. Advances in neural
information processing systems, vol. 27. Red Hook:
57. Landman B, Warfield S. MICCAI 2012 workshop on Curran Associates, Inc.; 2014:2924–32.
multi-atlas labeling. In: Medical image computing
and computer assisted intervention conference. 69. Iidaka T. Resting state functional magnetic reso-
Nice, France. October 1–5, 2012. nance imaging and neural network classified autism
and control. Cortex. 2015;63:55–67.
58. Livne M, Rieger J, Aydin OU, et al. A U-Net deep
learning framework for high performance vessel 70. Chen H, Duan X, Liu F, et al. Multivariate classifi-
segmentation in patients with cerebrovascular dis- cation of autism spectrum disorder using fre-
ease. Front Neurosci. 2019;28(13):97. quency-specific resting-state functional
connectivity—a multi-center study. Prog Neu-
59. Loftis JM, Huckans M, Morasco BJ. Neuroimmune ropsychopharmacol Biol Psychiatry.
mechanisms of cytokine-induced depression: cur- 2016;4(64):1–9.
rent theories and novel treatment strategies. Neu-
robiol Dis. 2010;37(3):519–33. 71. Kuang D, Guo X, An X, Zhao Y, He L. Discrimina-
tion of ADHD based on fMRI data with deep belief
60. Menard C, Pfau ML, Hodes GE, et al. Social stress network. Intelligent computing in bioinformatics.
induces neurovascular pathology promoting New York: Springer International Publishing;
depression. Nat Neurosci. 2017;20(12):1752–60. 2014:225–32.
61. Lian C, Zhang J, Liu M, et al. Multi-channel multi- 72. Tjepkema-Cloostermans MC, de Carvalho RCV, van
scale fully convolutional network for 3D perivas- Putten MJAM. Deep learning for detection of focal
cular spaces segmentation in 7T MR images. Med epileptiform discharges from scalp EEG recordings.
Image Anal. 2018;46:106–17. Clin Neurophysiol. 2018;129(10):2191–6.
62. Jeong Y, Rachmadi MF, Valdés-Hernández MDC, 73. Tsiouris JL, Pezoulas VC, Zervakis M, Konitsiotis S,
Komura T. Dilated saliency U-Net for white matter Koutsouris DD, Fotiadis DI. A long short-term
hyperintensities segmentation using irregularity memory deep learning network for the prediction
age map. Front Aging Neurosci. 2019;27(11):150. of epileptic seizures using EEG signals. Comput Biol
Med. 2018;1(99):24–37.
63. Gootjes L, Teipel SJ, Zebuhr Y, et al. Regional dis-
tribution of white matter hyperintensities in vas- 74. Acharya UR, Oh SL, Hagiwara Y, Tan JH, Adeli H.
cular dementia, Alzheimer’s disease and healthy Deep convolutional neural network for the auto-
aging. Dement Geriatr Cogn Disord. mated detection and diagnosis of seizure using EEG
2004;18(2):180–8. signals. Comput Biol Med. 2018;1(100):270–8.
64. Karargyros A, Syeda-Mahmood T. Saliency U-Net: A 75. Truong ND, Nguyen AD, Kuhlmann L, Bonyadi MR,
regional saliency map-driven hybrid deep learning Yang J, Kavehei O. A generalised seizure prediction
network for anomaly segmentation. In: Medical with convolutional neural networks for intracranial
imaging 2018: computer-aided diagnosis. Interna- and scalp electroencephalogram data analysis.
tional Society for Optics and Photonics. Houston, 2017. http://arxiv.org/abs/1707.01976.
TX. 2018. 105751T.
76. Khan H, Marcuse L, Fields M, Swann K, Yener B.
65. Kuang D, He L. Classification on ADHD with deep Focal onset seizure prediction using convolutional
learning. In: 2014 international conference on networks. IEEE Trans Biomed Eng.
cloud computing and big data. Wuhan, China. 2018;65(9):2109–18.
2014. pp. 27–32. http://ieeexplore.ieee.org.
77. Yousefi S, Amrollahi F, Amgad M, et al. Predicting
66. Suk H-I, Wee C-Y, Lee S-W, Shen D. State-space clinical outcomes from large scale cancer genomic
model with deep learning for functional dynamics profiles with deep survival models. Sci Rep.
estimation in resting-state fMRI. Neuroimage. 2017;7(1):11707.
2016;1(129):292–307.
78. Zhou J, Park CY, Theesfeld CL, et al. Whole-genome
67. Meszlényi RJ, Buza K, Vidnyánszky Z. Resting state deep-learning analysis identifies contribution of
fMRI functional connectivity-based classification noncoding mutations to autism risk. Nat Genet.
using a convolutional neural network architecture. 2019;51(6):973–80.
Front Neuroinform. 2017;17(11):61.
Neurol Ther (2019) 8:351–365 365
79. Buda M, Saha A, Mazurowski MA. Association of chest radiographs: a cross-sectional study. PLoS
genomic subtypes of lower-grade gliomas with Med. 2018;15(11):e1002683.
shape features automatically extracted by a deep
learning algorithm. Comput Biol Med. 82. Zech JR, Badgeley MA, Liu M, Costa AB, Titano JJ,
2019;109:218–25. Oermann EK. Confounding variables can degrade
generalization performance of radiological deep
80. Mobadersany P, Yousefi S, Amgad M, et al. Pre- learning models. 2018. http://arxiv.org/abs/1807.
dicting cancer outcomes from histology and geno- 00431.
mics using convolutional networks. Proc Natl Acad
Sci USA. 2018;115(13):E2970–9. 83. Obermeyer Z, Mullainathan S. Dissecting racial bias
in an algorithm that guides health decisions for 70
81. Zech JR, Badgeley MA, Liu M, Costa AB, Titano JJ, million people. In: Proceedings of conference on
Oermann EK. Variable generalization performance fairness, accountability, and transparency. New
of a deep learning model to detect pneumonia in York: ACM; 2019. p. 89.