Deep learning-LSTM
Deep learning-LSTM
Multicellular LSTM-based deep learning model for aero-engine remaining useful life
prediction,
Reliability Engineering & System Safety,
Volume 216,
2021,
107927,
ISSN 0951-8320,
https://doi.org/10.1016/j.ress.2021.107927.
(https://www.sciencedirect.com/science/article/pii/S0951832021004439)
Abstract: The prediction of aero-engine remaining useful life (RUL) is helpful for
its operation and maintenance. Aiming at the challenge that most neural networks
(NNs), including long short-term memory (LSTM), cannot process the input data in
different update modes based on its importance degree, a novel variant of LSTM
named multicellular LSTM (MCLSTM) is constructed. The level division unit is
proposed to determine the importance degree of input data, and then multiple
cellular units are designed to update the cell states according to the data level.
Thus, MCLSTM can well mine different levels of degradation trends. Based on MCLSTM
and a deep NN (DNN), a deep learning model for RUL prediction is set up, where
MCLSTM and a branch of the DNN is used to extract health indicators (HIs) of aero-
engine from raw data, and the other part of the DNN is applied to generate the HIs
from human-made features and predict the RUL based on the concatenated HIs. The
proposed RUL prediction model is successfully applied to predict the RULs of aero-
engines via the Commercial Modular Aero Propulsion System Simulation datasets, and
the comparative results show that it has a better comprehensive prediction
performance than the commonly-used machine learning methods.
Keywords: RUL prediction; Degradation trend; Multi-resource data; Health feature;
Data level
Yucheng Liu, Wenyang Duan, Limin Huang, Shiliang Duan, Xuewen Ma,
The input vector space optimization for LSTM deep learning model in real-time
prediction of ship motions,
Ocean Engineering,
Volume 213,
2020,
107681,
ISSN 0029-8018,
https://doi.org/10.1016/j.oceaneng.2020.107681.
(https://www.sciencedirect.com/science/article/pii/S0029801820306739)
Abstract: Vessel motions due to the ocean waves contribute to maritime operational
safety and efficiency. Real-time prediction of deterministic ship motions in the
coming future seconds is essential in decision-making when performing motions
sensitive activities. The Long-Short Term Memory (LSTM) deep learning model
provides a potential way for nonlinear ship motions prediction due to its
capability in nonlinearity processing. Determination of a reasonable dimension of
the input vector is critical in training the LSTM model. Conventionally, the
optimal dimension for the input vector is selected by traversing an empirical
preset range. Hence, it suffers both high computational cost and poor adaptation in
determining the optimal input vector dimension. In the present work, an input
vector space optimization method is proposed based on the dependence hidden in ship
motion records of a sequence. Taking different correlation expressions into
consideration, both the Impulse Response Function (IRF) based and Auto-correlation
Function (ACF) based techniques are investigated for input vector space
optimization. Numerical simulations are carried out for vilification and comparison
purpose. The ACF technique is better in representing the auto-correlation hidden in
the stochastic ship motions. And the ACF-based LSTM model performs better in both
training efficiency and prediction accuracy.
Keywords: Ship motions real-time prediction; LSTM deep learning model; Input vector
space optimization; Impulse response function; Auto-correlation function
Iftikhar Rasheed, Fei Hu, Lin Zhang,
Deep reinforcement learning approach for autonomous vehicle systems for maintaining
security and safety using LSTM-GAN,
Vehicular Communications,
Volume 26,
2020,
100266,
ISSN 2214-2096,
https://doi.org/10.1016/j.vehcom.2020.100266.
(https://www.sciencedirect.com/science/article/pii/S2214209620300371)
Abstract: The success of autonomous vehicles (AVhs) depends upon the effectiveness
of sensors being used and the accuracy of communication links and technologies
being employed. But these sensors and communication links have great security and
safety concerns as they can be attacked by an adversary to take the control of an
autonomous vehicle by influencing their data. Especially during the state
estimation process for monitoring of autonomous vehicles' dynamics system, these
concerns require immediate and effective solution. In this paper we present a new
adversarial deep reinforcement learning algorithm (NDRL) that can be used to
maximize the robustness of autonomous vehicle dynamics in the presence of these
attacks. In this approach the adversary tries to insert defective data to the
autonomous vehicle's sensor readings so that it can disrupt the safe and optimal
distance between the autonomous vehicles traveling on the road. The attacker tries
to make sure that there is no more safe and optimal distance between the autonomous
vehicles, thus it may lead to the road accidents. Further attacker can also add
fake data in such a way that it leads to reduced traffic flow on the road. On the
other hand, autonomous vehicle will try to defend itself from these types of
attacks by maintaining the safe and optimal distance i.e. by minimizing the
deviation so that adversary does not succeed in its mission. This attacker-
autonomous vehicle action reaction can be studied through the game theory
formulation with incorporating the deep learning tools. Each autonomous vehicle
will use Long-Short-Term-Memory (LSTM)-Generative Adversarial Network (GAN) models
to find out the anticipated distance variation resulting from its actions and input
this to the new deep reinforcement learning algorithm (NDRL) which attempts to
reduce the variation in distance. Whereas attacker also chooses deep reinforcement
learning algorithm (NDRL) and wants to maximize the distance variation between the
autonomous vehicles.
Keywords: Autonomous vehicles; Deep reinforcement learning; Generative adversarial
network (GAN); Long-short-term-memory (LSTM); 5G; ITS-security
Ahmed A. Ewees, Mohammed A.A. Al-qaness, Laith Abualigah, Mohamed Abd Elaziz,
HBO-LSTM: Optimized long short term memory with heap-based optimizer for wind power
forecasting,
Energy Conversion and Management,
Volume 268,
2022,
116022,
ISSN 0196-8904,
https://doi.org/10.1016/j.enconman.2022.116022.
(https://www.sciencedirect.com/science/article/pii/S0196890422008147)
Abstract: The forecasting and estimation of wind power is a challenging problem in
renewable energy generation due to the high volatility of wind power resources,
inevitable intermittency, and complex fluctuation. In recent years, deep learning
techniques, especially recurrent neural networks (RNN), showed prominent
performance in time-series forecasting and prediction applications. One of the main
efficient RNNs is the long short term memory (LSTM), which we adopted in this study
to forecast the wind power from different wind turbines. We adopted the advances of
the metaheuristic optimization algorithms to train the LSTM and to boost its
performance by optimizing its parameters. The Heap-based optimizer (HBO) is a new
human-behavior-based metaheuristic algorithm that was inspired by corporate rank
hierarchy, and it was employed to solve complex optimization and engineering
problems. In this study, HBO is used to train the LSTM, and it showed significant
enhancement on the LSTM prediction performance. We used four datasets from the
well-known wind turbines in France, La Haute Borne wind turbines, to evaluate the
developed HBO-LSTM. We also considered several optimized LSTM models using several
optimization algorithms for comparisons, as well as several existing models. The
comparison outcome confirmed the capability of HBO to boost the prediction
performance of the basic LSTM model.
Keywords: Forecasting; Deep learning; Wind power; Heap-based optimizer; Long short
term memory
Xing-Ke Ma, Hong-Quan Huang, Bo-Rui Huang, Zhi-Wen Shen, Qing-Tai Wang, Yu-Yu Xiao,
Cheng-Lin Zhong, Hao Xin, Peng Sun, Kai-Ming Jiang, Lin Tang, Wei-Cheng Ding, Wei
Zhou, Jian-Bin Zhou,
X-ray spectra correction based on deep learning CNN-LSTM model,
Measurement,
Volume 199,
2022,
111510,
ISSN 0263-2241,
https://doi.org/10.1016/j.measurement.2022.111510.
(https://www.sciencedirect.com/science/article/pii/S0263224122007321)
Abstract: The two most important indicators in the measurement process for the X-
ray spectra are energy resolution and counting rate. However, in the actual
detection process, when the interval time between adjacent pulses is less than the
pulse shaping time, the pulses pile up. If the pile-up pulse is not separated and
recognized, then it greatly affects the X-ray spectrum’s accuracy. A method of X-
ray spectrum correction is proposed on the basis of a deep learning model, which
realizes the separation of the pile-up pulse by recognizing its parameters, and
then realizes the correction of the X-ray spectrum. Standard sources 55Fe and 238Pu
are used as the measurement objects, and the spectra correction method is used to
recognize the pile-up pulses. Measurement results show that the method can
effectively recognize the pile-up pulses, improve the spectrum’s counting rate, and
obtain more accurate X-ray spectra.
Keywords: X-ray spectra; Pile-up pulse; Deep learning; Pulse recognition; Radiation
measurement
Baiying Lei, Yuwen Zhang, Dongdong Liu, Yanwu Xu, Guanghui Yue, Jiuwen Cao, Huoyou
Hu, Shuangzhi Yu, Peng Yang, Tianfu Wang, Yali Qiu, Xiaohua Xiao, Shuqiang Wang,
Longitudinal study of early mild cognitive impairment via similarity-constrained
group learning and self-attention based SBi-LSTM,
Knowledge-Based Systems,
Volume 254,
2022,
109466,
ISSN 0950-7051,
https://doi.org/10.1016/j.knosys.2022.109466.
(https://www.sciencedirect.com/science/article/pii/S0950705122007353)
Abstract: Alzheimer’s disease (AD) is an incurable neurodegenerative disease. Mild
cognitive impairment (MCI) is often considered a critical time window for
predicting early conversion to Alzheimer’s disease (AD), with approximately 80% of
amnestic MCI patients developing AD within 6 years. MCI can be further categorized
into two stages (i.e., early MCI (EMCI) and late MCI (LMCI)). To identify EMCI
effectively and understand how it changes brain function, the brain functional
connectivity network (BFCN) has been widely used. However, the conventional methods
mainly focused on detection from a single time-point data, which could not discover
the changes during the disease progression without using multi-time points data.
Therefore, in this work, we carry out a longitudinal study based on multi-time
points data to detect EMCI and validate them on two public datasets. Specifically,
we first construct a similarity-constrained group network (SGN) from the resting
state functional magnetic resonance imaging (rs-fMRI) data at different time-
points, and then use a stacked bidirectional long short term memory (SBi-LSTM)
network to extract features for longitudinal analysis. Also, we use a self-
attention mechanism to leverage high-level features to further improve the
detection accuracy. Evaluated on the public Alzheimer’s Disease Neuroimaging
Initiative Phase II and III (ADNI-2 and ADNI-3) databases, the proposed method
outperforms several state-of-the-art methods.
Keywords: Mild cognitive impairment; Similarity-constrained group learning; SBi-
LSTM; Self-attention; Longitudinal study
Arsalan Mahmoodzadeh, Hamid Reza Nejati, Mokhtar Mohammadi, Hawkar Hashim Ibrahim,
Shima Rashidi, Tarik Ahmed Rashid,
Forecasting tunnel boring machine penetration rate using LSTM deep neural network
optimized by grey wolf optimization algorithm,
Expert Systems with Applications,
Volume 209,
2022,
118303,
ISSN 0957-4174,
https://doi.org/10.1016/j.eswa.2022.118303.
(https://www.sciencedirect.com/science/article/pii/S0957417422014361)
Abstract: Achieving an accurate and reliable estimation of tunnel boring machine
(TBM) performance can diminish the hazards related to extreme capital costs and
planning tunnel construction. Here, a hybrid long short-term memory (LSTM) model
enhanced by grey wolf optimization (GWO) is developed for predicting TBM-
penetration rate (TBM-PR). 1125 datasets were considered including six input
parameters. To vanish overfitting, the dropout technique was used. The effect of
input time series length on the model performance was studied. The TBM-PR results
of the LSTM-GWO model were compared to some other machine learning (ML) models such
as LSTM. The results were evaluated using root mean square error (RMSE), mean
absolute percentage error (MAPE), and correlation coefficient (R2). Finally, the
LSTM-GWO model produced the most accurate results (test: R2 = 0.9795; RMSE = 0.004;
MAPE = 0.009 %). The mutual information test revealed that input parameters of rock
fracture class and uniaxial compressive strength have the most and least impact on
the TBM-PR, respectively.
Keywords: Machine learning; Long short-term memory; Grey wolf optimization;
Metaheuristic optimization, Tunnel boring machine penetration rate
First Ali Agga, Second Ahmed Abbou, Yassine El Houm, Moussa Labbadi,
Short-Term Load Forecasting Based on CNN and LSTM Deep Neural Networks,
IFAC-PapersOnLine,
Volume 55, Issue 12,
2022,
Pages 777-781,
ISSN 2405-8963,
https://doi.org/10.1016/j.ifacol.2022.07.407.
(https://www.sciencedirect.com/science/article/pii/S2405896322008084)
Abstract: In the coming years, the world will witness a global transition towards
the adoption of photovoltaic technology for large-scale plants to produce
electricity at a grid scale, and more householders will also be encouraged to
produce their electricity. However, the reliance of the photovoltaic plants on
erratic weather conditions requires the development of solutions that could help in
preventing any electricity blackout or overproduction. Hence, comes the role of
forecasting models that help in overcoming that issue. In this work, two deep
learning models are developed and tested (LSTM, CNN). Both architectures will go
under several different configurations to witness the impact of changing the number
of hidden layers on the accuracy of the forecasts. The findings reveal that the
models behave differently when the number of layers changed over the different
configurations. In addition, two-time windows were considered (1-Day, 2-Days) for
even deeper insight..
Keywords: Deep Learning; CNN; LSTM; Load Forecast
Jun Zhang, Pengli Wu, Xia Xu, Ming Han, Bin Pan,
PCS-LSTM: A hybrid deep learning model for multi-stations joint temperature
prediction based on periodicity and closeness,
Neurocomputing,
Volume 501,
2022,
Pages 151-161,
ISSN 0925-2312,
https://doi.org/10.1016/j.neucom.2022.06.015.
(https://www.sciencedirect.com/science/article/pii/S0925231222007342)
Abstract: Temperature is one of the most important meteorological elements, which
affects the daily lives of people all over the world. Owing to the rapid
development of meteorological facilities, the number of meteorological observation
stations on earth is gradually increasing, which brings challenges to the spatial
association between stations. Many researchers focus on how to predict temperature
more accurately utilizing these associations. However, the existing deep learning
methods of temperature prediction have difficulty in capturing the interactions
between neighboring stations in the spatial dimension. In addition, in the time
dimension, the temperature in nature exhibits not only nearby variations but also
periodic characteristics, which further increases the difficulty of temperature
prediction. To solve the aforementioned two problems, we propose the periodicity
and closeness social long short-term memory (PCS-LSTM) model, which includes PS-
LSTM and CS-LSTM modules. Specifically, to model the relationships between multiple
meteorological observation stations, we utilized the social pooling in the PS-LSTM
and CS-LSTM modules to establish spatial associations. To further refine the
temperature variation, we combine PS-LSTM and CS-LSTM to model the periodicity and
closeness of the time series. Compared with the LSTM basic model, the experiments
show that the MAE of our model prediction results is reduced by 0.109°C in the next
24 h compared.
Keywords: Temperature prediction; LSTM; Periodicity; Closeness
Reuben Varghese Joseph, Anshuman Mohanty, Soumyae Tyagi, Shruti Mishra, Sandeep
Kumar Satapathy, Sachi Nandan Mohanty,
A hybrid deep learning framework with CNN and Bi-directional LSTM for store item
demand forecasting,
Computers and Electrical Engineering,
Volume 103,
2022,
108358,
ISSN 0045-7906,
https://doi.org/10.1016/j.compeleceng.2022.108358.
(https://www.sciencedirect.com/science/article/pii/S0045790622005754)
Abstract: In the era of ever-changing market landscape, enterprises tend to make
quick and informed decisions to survive and prosper in the competition. Decision
makers within an organization must be supplied with data in a way that could be
easily analyzed and comprehended to build strategies in order to achieve business
goals. Accurate demand forecasting of products is one of such decisions which is
crucial for retail operators to have a clear picture on the future demand of their
products and services. With a certainty in estimation, retailers might keep a check
on how many items to allocate, order and restock thus boosting their gross sales
and profits. Machine Learning approaches are widely used for demand forecasting of
different items. In this work, we have used the Store Item Demand Forecasting
Challenge dataset from Kaggle to implement our proposed framework. The main novelty
of this study was to build a coupled CNN-BiLSTM framework with Lazy Adam optimizer
to make an accurate forecast of product demand of store items. Various State-of-art
machine learning techniques like SGD (Stochastic Gradient Descent), Linear
Regression, K-Nearest Neighbour, Bagging, Random Forest, SVR, XgBoost (extreme
gradient boosting) and CNN-LSTM. for demand forecasting has been implemented and
the results were compared with the proposed model. On evaluation with metrics
including Mean Absolute Percentage Error (MAPE), R-Squared (R2) value and Mean
Absolute Error (MAE), it was observed that the proposed framework having more
accurecy as compare to the traditional approaches.
Keywords: Convolutional neural network; Bidirectional long short-term memory;
Product demand forecasting; Lazy Adam optimizer; Inventory prediction; Supply
chain; R-squared score; Mean absolute error; Mean absolute percentage error;
Machine learning; Time series analysis
Yaoran Chen, Yan Wang, Zhikun Dong, Jie Su, Zhaolong Han, Dai Zhou, Yongsheng Zhao,
Yan Bao,
2-D regional short-term wind speed forecast based on CNN-LSTM deep learning model,
Energy Conversion and Management,
Volume 244,
2021,
114451,
ISSN 0196-8904,
https://doi.org/10.1016/j.enconman.2021.114451.
(https://www.sciencedirect.com/science/article/pii/S0196890421006270)
Abstract: Short-term wind speed forecast is of great importance to wind farm
regulation and its early warning. Previous studies mainly focused on the prediction
at a single location but few extended the task to 2-D wind plane. In this study, a
novel deep learning model was proposed for a 2-D regional wind speed forecast,
using the combination of the auto-encoder of convolutional neural network (CNN) and
the long short-term memory unit (LSTM). The 12-hidden-layer deep CNN was adopted to
encode the high dimensional 2-D input into the embedding vector and inversely, to
decode such latent representation after it was predicted by the LSTM module based
on historical data. The model performance was compared with parallel models under
different criteria, including MAE, RMSE and R2, all showing stable and considerable
enhancements. For instance, the overall MAE value dropped to 0.35 m/s for the
current model, which is 32.7%, 28.8% and 18.9% away from the prediction results
using the persistence, basic ANN and LSTM model. Moreover, comprehensive
discussions were provided from both temporal and spatial views of analysis,
revealing that the current model can not only offer an accurate wind speed forecast
along timeline (R2 equals to 0.981), but also give a distinct estimation of the
spatial wind speed distribution in 2-D wind farm.
Keywords: Regional wind speed prediction; CNN; LSTM; Temporal series fitness;
Spatial distribution
Shaojiang Dong, Jiafeng Xiao, Xiaolin Hu, Nengwei Fang, Lanhui Liu, Jinbao Yao,
Deep transfer learning based on Bi-LSTM and attention for remaining useful life
prediction of rolling bearing,
Reliability Engineering & System Safety,
Volume 230,
2023,
108914,
ISSN 0951-8320,
https://doi.org/10.1016/j.ress.2022.108914.
(https://www.sciencedirect.com/science/article/pii/S0951832022005294)
Abstract: Many transfer learning methods focus on training models between domains
with large differences. However, the data feature distribution varies greatly in
different bearing degradation processes, which affects the prediction accuracy of
Remaining useful life (RUL). To solve this problem, a novel method for RUL
prediction with more refined transfer is proposed, which added failure behavior
judgment. Firstly, a failure behavior judgment method is proposed by using the
convolutional autoencoder (CAE) and Pearson correlation coefficient to determine
whether the bearing fails gradually or suddenly. Then, a multi-channel transfer
network is proposed for extracting multi-scale features of bearing degradation.
Each channel uses convolutional network and bidirectional long short-term memory
(Bi-LSTM) to extract global and temporal information. To reduce the difference in
feature distribution between the training and test bearing data, a domain adaptive
structure is added after feature fusion in each channel to enable the model to
learn domain invariant features. By applying this method to experimental data and
comparing it with other methods, the superiority and effectiveness of the proposed
method are verified.
Keywords: Rolling bearings; Remaining useful life prediction; Feature distribution;
Domain adaptive
Jayanthi E, Vallikannu R,
Enhancing the performance of asymmetric architectures and workload characterization
using LSTM learning algorithm,
Advances in Engineering Software,
Volume 173,
2022,
103266,
ISSN 0965-9978,
https://doi.org/10.1016/j.advengsoft.2022.103266.
(https://www.sciencedirect.com/science/article/pii/S0965997822001673)
Abstract: Asymmetric multiple core architectures are becoming increasingly popular
because of their QoS, energy efficiency, and high performance. Because of the
heterogeneous piping layout and execution procedure, the overall system performance
is highly influenced by workload characterization becomes a difficult issue.
Recently, it has become vital to create a performance evaluator for faster response
FPGA structures. With the embedded computerized visual hardware resources (e.g.
GPUs, FPGAs, and multi-core CPUs) and its associated software manages this
heterogeneous system design is tough for the researchers. A thorough analysis of
the performance of run-time and various embedded kernels' energy efficiencies to
assist developers in determining which embedded substrate is best for their
project. To solve these challenges, resource-based prediction models are developed
for FPGA-SoC relying on the characterization of intellectual workload, and its
performance is thoroughly examined. Xilinx FPGA frameworks are used to run a
variety of real-time applications that perform workloads depending on projected
processing cores. When compared to existing Ml classifiers, this LSTM predictor
achieves a mean of 96% of the experimental findings and also improves energy
efficiency.
Keywords: Performance Enhancement; Asymmetric architectures; LSTM; Workload
Characterization
Xiangling Li, Kang Xiao, Xianbing Li, Chunye Yu, Dongyan Fan, Zhixue Sun,
A well rate prediction method based on LSTM algorithm considering manual
operations,
Journal of Petroleum Science and Engineering,
Volume 210,
2022,
110047,
ISSN 0920-4105,
https://doi.org/10.1016/j.petrol.2021.110047.
(https://www.sciencedirect.com/science/article/pii/S0920410521016582)
Abstract: Manual operations such as changing the size of chokes as well as opening
and closing of the well have a great impact on oil and gas production from the
well. This scenario is not considered in most deep learning methods for predicting
productivity. Therefore, a deep learning method based on a long short-term memory
(LSTM) neural network model was established to predict well performance considering
the manual operations. The input dataset was composed of data related to choke
size, daily opening time series, and production; the first 90% of the dataset was
used as the training set and the remaining 10% was used as the test set. The deep
learning model was constructed using a LSTM module, regularization process, and
dropout network. The formulated LSTM model was proficient compared with a model
that did not consider the manual operation process, and showed better prediction
accuracy. Through multiple experiments, the production-related time step was
optimized at three, indicating that prediction for the subsequent step was most
relevant to the initial three step inputs. Overall, the operation of opening and
closing of wells, changing the size of chokes, and variations in daily production
time can be considered in our LSTM deep learning model, which provides more
reasonable results.
Keywords: Manual operations; Time series; LSTM; Deep learning method; Rate
prediction
Sachi Nandan Mohanty, E. Laxmi Lydia, Mohamed Elhoseny, Majid M. Gethami Al Otaibi,
K. Shankar,
Deep learning with LSTM based distributed data mining model for energy efficient
wireless sensor networks,
Physical Communication,
Volume 40,
2020,
101097,
ISSN 1874-4907,
https://doi.org/10.1016/j.phycom.2020.101097.
(https://www.sciencedirect.com/science/article/pii/S1874490720301737)
Abstract: Wireless sensor network (WSN) comprises a collection of sensor nodes
employed to monitor and record the status of the physical environment and organize
the gathered data at a central location. This paper presents a deep learning based
distributed data mining (DDM) model to achieve energy efficiency and optimal load
balancing at the fusion center of WSN. The presented DMM model includes a recurrent
neural network (RNN) based long short-term memory (LSTM) called RNN-LSTM, which
divides the network into various layers and place them into the sensor nodes. The
proposed model reduces the overhead at the fusion center along with a reduction in
the number of data transmission. The presented RNN-LSTM model is tested under a
wide set of experimentation with varying number of hidden layer nodes and signaling
intervals. At the same time, the amount of energy needed to transmit data by RNN-
LSTM model is considerably lower than energy needed to transmit actual data. The
simulation results indicated that the RNN-LSTM reduces the signaling overhead,
average delay and maximizes the overall throughput compared to other methods. It is
noted that under the signaling interval of 240 ms, it can be shown that the RNN-
LSTM achieves a minimum average delay of 190 ms whereas the OSPF and DNN models
shows average delay of 230 ms and 230 ms respectively.
Keywords: Deep learning; Energy efficiency; Recurrent neural network; WSN
Chenfei Ma, Chuang Lin, Oluwarotimi Williams Samuel, Lisheng Xu, Guanglin Li,
Continuous estimation of upper limb joint angle from sEMG signals based on SCA-LSTM
deep learning approach,
Biomedical Signal Processing and Control,
Volume 61,
2020,
102024,
ISSN 1746-8094,
https://doi.org/10.1016/j.bspc.2020.102024.
(https://www.sciencedirect.com/science/article/pii/S1746809420301804)
Abstract: Robotic arm control has drawn a lot of attention along with the
development of industrialization. The methods based on myoelectric pattern
recognition have been proposed with multiple degrees of freedom for years. While
these methods can support the actuation of several classes of discrete movements
sequentially, they do not allow simultaneous control of multiple movements in a
continuous manner like natural arms. In this study, we proposed a short connected
autoencoder long short-term memory (SCA-LSTM) based simultaneous and proportional
(SP) scheme that estimates continuous arm movements using kinematic information
extracted from surface electromyogram (sEMG) recordings. The sEMG signals
corresponding to seven classes of shoulder-elbow joint angle movements acquired
from eleven participants were preprocessed using max root mean square envelope.
Afterwards, the proposed SCA-LSTM model and two commonly applied models, namely,
multilayer perceptrons (MLPs) and convolutional neural network (CNN), were trained
and tested using the preprocessed data for continuous estimation of arm movements.
Our experimental results showed that the proposed SCA-LSTM model could achieve a
significantly higher estimation accuracy of approximately 95.7% that is
consistently stable across the subjects in comparison to the CNN (86.8%) and MLP
(83.4%) models. These results suggest that the proposed SCA-LSTM would be a
promising model for continuous estimation of upper limb movements from sEMG signals
for prosthetic control.
Keywords: Robotic arm control; Surface electromyogram; Simultaneous and
proportional control; Joint angle estimation; Deep learning
Yang Wang, Pandi Vijayakumar, Brij B. Gupta, Wadee Alhalabi, Audithan Sivaraman,
An improved entity recognition approach to cyber-social knowledge provision of
intellectual property using a CRF-LSTM model,
Pattern Recognition Letters,
Volume 163,
2022,
Pages 145-151,
ISSN 0167-8655,
https://doi.org/10.1016/j.patrec.2022.10.001.
(https://www.sciencedirect.com/science/article/pii/S016786552200294X)
Abstract: With the development of cutting-edge IT technologies, e.g. Big Data,
Knowledge Engineering, etc., traditional Intellectual Property (IP) services have
depicted high redundancy and low efficiency during management of such large-scale
of data. Recent advancement of Artificial Intelligence (AI) and Deep Learning (DL)
models has been accelerating relevant research activities being investigated on
Knowledge Graph (KG) schemes and applications in different domains, such as medical
services, social media, etc. However, when IP services and their cyber-social
provision are taken into account, relevant approaches suffer from unbalanced labels
against training results, and inappropriate evaluation metrics not well reflecting
the impact of the unbalance. In this paper, a deep learning model combining
Conditional Random Field and Bidirectional LSTM has been proposed, in order to
achieve named entity recognition with unbalanced labels. An adaptive metric, G-
Score was introduced to compare the fitting ability of models by evaluating the gap
between Precision and Recall. According to the results, the proposed model can
effectively recognize the potential named entities with outperformance over other
relevant models.
Keywords: Intellectual property; Named-entity recognition; LSTM; CRF; Knowledge
graph
Peipei Wang, Xinqi Zheng, Gang Ai, Dongya Liu, Bangren Zhu,
Time series prediction for the epidemic trends of COVID-19 using the improved LSTM
deep learning method: Case studies in Russia, Peru and Iran,
Chaos, Solitons & Fractals,
Volume 140,
2020,
110214,
ISSN 0960-0779,
https://doi.org/10.1016/j.chaos.2020.110214.
(https://www.sciencedirect.com/science/article/pii/S096007792030610X)
Abstract: The COVID-19 outbreak in late December 2019 is still spreading rapidly in
many countries and regions around the world. It is thus urgent to predict the
development and spread of the epidemic. In this paper, we have developed a
forecasting model of COVID-19 by using a deep learning method with rolling update
mechanism based on the epidemical data provided by Johns Hopkins University. First,
as traditional epidemical models use the accumulative confirmed cases for training,
it can only predict a rising trend of the epidemic and cannot predict when the
epidemic will decline or end, an improved model is built based on long short-term
memory (LSTM) with daily confirmed cases training set. Second, considering the
existing forecasting model based on LSTM can only predict the epidemic trend within
the next 30 days accurately, the rolling update mechanism is embedded with LSTM for
long-term projections. Third, by introducing Diffusion Index (DI), the
effectiveness of preventive measures like social isolation and lockdown on the
spread of COVID-19 is analyzed in our novel research. The trends of the epidemic in
150 days ahead are modeled for Russia, Peru and Iran, three countries on different
continents. Under our estimation, the current epidemic in Peru is predicted to
continue until November 2020. The number of positive cases per day in Iran is
expected to fall below 1000 by mid-November, with a gradual downward trend expected
after several smaller peaks from July to September, while there will still be more
than 2000 increase by early December in Russia. Moreover, our study highlights the
importance of preventive measures which have been taken by the government, which
shows that the strict controlment can significantly reduce the spread of COVID-19.
Keywords: Covid-19; LSTM; Rolling update mechanism; Modeling; Forecasting
Abdul Quadir Md, Sanjit Kapoor, Chris Junni A.V., Arun Kumar Sivaraman, Kong Fah
Tee, Sabireen H., Janakiraman N.,
Novel optimization approach for stock price forecasting using multi-layered
sequential LSTM,
Applied Soft Computing,
2022,
109830,
ISSN 1568-4946,
https://doi.org/10.1016/j.asoc.2022.109830.
(https://www.sciencedirect.com/science/article/pii/S1568494622008791)
Abstract: Stock markets can often be one of the most volatile places to invest.
Statistical analysis of past stock performance and external factors play a major
role in the decision to buy or sell stocks. These factors are all used to maximize
profits. Stock price index forecasting has been a subject of great research for
many years, and several machine learning and deep learning algorithms have been
proposed to simplify this complex task, but little success has been found so far.
In order to forecast stocks accurately, it is crucial to understand the context-
specific dependence of stock prices on their past values. The use of Long Short
Term Memory (LSTM), which is capable of understanding long-term data dependencies,
can help overcome this obstacle. In this context, this paper proposes a novel
optimization approach for stock price prediction that is based on a Multi-Layer
Sequential Long Short Term Memory (MLS LSTM) model which makes use of the adam
optimizer. Additionally, the MLS LSTM algorithm uses normalized time series data
divided into time steps to determine the relationship between past values and
future values in order to make accurate predictions. Furthermore, it eliminates the
vanishing gradient problem associated with simple recurrent neural networks. The
stock price index is forecasted by taking into account past performance information
along with past trends and patterns. The results illustrate that a 95.9% prediction
accuracy is achieved on the training data set and a 98.1% accuracy on the testing
data set with the MLS LSTM algorithm, which dramatically exceeds the performance of
other machine learning and deep learning algorithms. The mean absolute percentage
error was observed to be 1.79% on the training set and 2.18% on the testing set,
respectively. Moreover, the proposed model is able to estimate the stock price with
a normalized root mean squared error of 0.019, thus giving an accurate forecast and
making it a feasible real-world solution.
Keywords: Recurrent neural network; Forecasting; Stock market; Adam optimizer; Long
short-term memory
Ali Agga, Ahmed Abbou, Moussa Labbadi, Yassine El Houm, Imane Hammou Ou Ali,
CNN-LSTM: An efficient hybrid deep learning architecture for predicting short-term
photovoltaic power production,
Electric Power Systems Research,
Volume 208,
2022,
107908,
ISSN 0378-7796,
https://doi.org/10.1016/j.epsr.2022.107908.
(https://www.sciencedirect.com/science/article/pii/S0378779622001389)
Abstract: Climate change is pushing an increasing number of nations to use green
energy resources, particularly solar power as an applicable substitute to
traditional power sources. However, photovoltaic power generation is highly
weather-dependent, relying mostly on solar irradiation that is highly unstable, and
unpredictable which makes power generation challenging. Accurate photovoltaic power
predictions can substantially improve the operation of solar power systems. This is
vital for supplying prime electricity to customers and ensuring the resilience of
power plants’ operation. This research is motivated by the recent adoption and
advances in DL models and their successful use in the sector of energy. The
suggested model merges two deep learning architectures, the long short-term memory
(LSTM) and convolutional neural network (CNN). Using a real-world dataset from
Rabat, Morocco, as a case study to illustrate the effectiveness of the suggested
topology. According to error metrics, MAE, MAPE, and RMSE, the suggested
architecture CNN-LSTM performance exceeds that of standard machine learning and
single DL models in terms of prediction, precision, and stability.
Keywords: Solar energy; Short-term forecasting; Long short-term memory;
Convolutional Neural Network; CNN-LSTM; Photovoltaic Power; Forecasts; Hybrid
Models
Soufiane Belagoune, Noureddine Bali, Azzeddine Bakdi, Bousaadia Baadji, Karim Atif,
Deep learning through LSTM classification and regression for transmission line
fault detection, diagnosis and location in large-scale multi-machine power systems,
Measurement,
Volume 177,
2021,
109330,
ISSN 0263-2241,
https://doi.org/10.1016/j.measurement.2021.109330.
(https://www.sciencedirect.com/science/article/pii/S0263224121003286)
Abstract: Fault detection, diagnosis, identification and location are crucial to
improve the sensitivity and reliability of system protection. This maintains power
systems continuous proper operation; however, it is challenging in large-scale
multi-machine power systems. This paper introduces three novel Deep Learning (DL)
classification and regression models based on Deep Recurrent Neural Networks (DRNN)
for Fault Region Identification (FRI), Fault Type Classification (FTC), and Fault
Location Prediction (FLP). These novel models explore full transient data from pre-
and post-fault cycles to make reliable decisions; whereas current and voltage
signals are measured through Phasor Measurement Units (PMUs) at different terminals
and used as input features to the DRNN models. Sequential Deep Learning (SDL) is
employed herein through Long Short-Term Memory (LSTM) to model spatiotemporal
sequences of high-dimensional multivariate features to achieve accurate
classification and prediction results. The proposed algorithms were tested in a
Two-Area Four-Machine Power System. Training and testing data are collected during
transmission lines faults of different types introduced at various locations in
different regions. The presented algorithms achieved superior detection,
classification and location performance with high accuracy and robustness compared
to contemporary techniques.
Keywords: Multi-machine power system; Power transmission lines; Short-circuit
fault; Long short-term memory; Fault detection and isolation; Sequential deep
learning
Aditi Sakalle, Pradeep Tomar, Harshit Bhardwaj, Divya Acharya, Arpit Bhardwaj,
A LSTM based deep learning network for recognizing emotions using wireless
brainwave driven system,
Expert Systems with Applications,
Volume 173,
2021,
114516,
ISSN 0957-4174,
https://doi.org/10.1016/j.eswa.2020.114516.
(https://www.sciencedirect.com/science/article/pii/S095741742031160X)
Abstract: Positive and Negative emotions are experienced by the majority of
individuals in their day-to-day life. It is important to control access of negative
emotions because it may lead to several chronic health issues like depression and
anxiety. The purpose of this research work is to develop a portable brainwave
driven system for recognizing positive, negative, and neutral emotions. This
research considers the classification of four negative class of emotions using
genres sadness, disgust, angry, and surprise along with the classification of three
basic class of emotions i.e., positive, negative, and neutral. This paper
introduces a long short term memory deep learning (LSTM) network to recognize
emotions using EEG signals. The primary goal of this approach is to assess the
classification performance of the LSTM model. The secondary goal is to assess the
human behavior of different age groups and gender. We have compared the performance
of Multilayer Perceptron (MLP), K-nearest neighbors (KNN), Support Vector Machine
(SVM), LIB-Support Vector Machine (LIB-SVM), and LSTM based deep learning model for
classification. The analysis shows that, for four class of emotions LSTM based deep
learning model provides classification accuracy as 83.12%, 86.94%, 91.67%, and
94.12% for 50–50, 60–40, 70–30, and 10-fold cross-validations. For three class of
emotions LSTM based deep learning model provides classification accuracy as 81.33%,
85.41%, 89.44%, and 92.66% for 50–50, 60–40, 70–30, and 10-fold cross-validation.
The generalizability and reliability of this approach are evaluated by applying our
approach to publicly available EEG datasets DEAP and SEED. In compliance with the
self-reported feelings, brain signals of 18–25 years of age group provided the
highest emotional identification. The results show that among genders, females are
more emotionally active as compared to males. These results affirmed the potential
use of our method for recognizing positive, negative, and neutral emotions.
Keywords: Emotion recognition; Deep learning; EEG; Empirical mode decomposition;
LSTM
Ailing Tan, Yunxin Wang, Yong Zhao, Bolin Wang, Xiaohang Li, Alan X. Wang,
Near infrared spectroscopy quantification based on Bi-LSTM and transfer learning
for new scenarios,
Spectrochimica Acta Part A: Molecular and Biomolecular Spectroscopy,
Volume 283,
2022,
121759,
ISSN 1386-1425,
https://doi.org/10.1016/j.saa.2022.121759.
(https://www.sciencedirect.com/science/article/pii/S1386142522009088)
Abstract: This study proposed a deep transfer learning methodology based on an
improved Bi-directional Long Short-Term Memory (Bi-LSTM) network for the first time
to address the near infrared spectroscopy (NIR) model transfer issue between
samples. We tested its effectiveness on two datasets of manure and polyglutamic
acid (γ-PGA) solution, respectively. First, the optimal primary Bi-LSTM networks
for cattle manure and the first batch of γ-PGA were developed by ablation
experiments and both proved to outperform one-dimensional convolutional neural
network (1D-CNN), Partial Least Square (PLS) and Extreme Learning Machine (ELM)
models. Then, two types of transfer learning approaches were carried out to
determine model transferability to non-homologous samples. For poultry manure and
the second batch of γ-PGA, the obtained predicting results verified that the second
approach of fine-tuning Bi-LSTM layers and re-training FC layers transcended the
first approach of fixing Bi-LSTM layers and only re-training FC layers by reducing
the RMSEPtarget of 23.4275% and 50.7343%, respectively. Finally, comparisons with
fine-tuning 1D-CNN and other traditional model transfer methods further identified
the superiority of the proposed methodology with exceeding accuracy and smaller
variation, which decreased RMSEPtarget of poultry manure and the second batch of γ-
PGA of 7.2832% and 48.1256%, 67.1117% and 80.6924% when compared to that acquired
by fine-tuning 1D-CNN, Tradaboost-ELM and CCA-PLS which were the best of five
traditional methods, respectively. The study demonstrates the potential of the
Fine-tuning-Bi-LSTM enabled NIR technology to be used as a simple, cost effective
and reliable detection tool for a wide range of applications under various new
scenarios.
Keywords: Near infrared spectroscopy; Bi-directional Long Short-Term Memory;
Transfer learning; Fine-tuning; Manure; γ-PGA
Xinyu Dong, Rachel Wong, Weimin Lyu, Kayley Abell-Hart, Jianyuan Deng, Yinan Liu,
Janos G. Hajagos, Richard N. Rosenthal, Chao Chen, Fusheng Wang,
An integrated LSTM-HeteroRGNN model for interpretable opioid overdose risk
prediction,
Artificial Intelligence in Medicine,
2022,
102439,
ISSN 0933-3657,
https://doi.org/10.1016/j.artmed.2022.102439.
(https://www.sciencedirect.com/science/article/pii/S0933365722001919)
Abstract: Opioid overdose (OD) has become a leading cause of accidental death in
the United States, and overdose deaths reached a record high during the COVID-19
pandemic. Combating the opioid crisis requires targeting high-need populations by
identifying individuals at risk of OD. While deep learning emerges as a powerful
method for building predictive models using large scale electronic health records
(EHR), it is challenged by the complex intrinsic relationships among EHR data.
Further, its utility is limited by the lack of clinically meaningful
explainability, which is necessary for making informed clinical or policy decisions
using such models. In this paper, we present LIGHTED, an integrated deep learning
model combining long short term memory (LSTM) and graph neural networks (GNN) to
predict patients' OD risk. The LIGHTED model can incorporate the temporal effects
of disease progression and the knowledge learned from interactions among clinical
features. We evaluated the model using Cerner's HealthFacts database with over 5
million patients. Our experiments demonstrated that the model outperforms
traditional machine learning methods and other deep learning models. We also
proposed a novel interpretability method by exploiting embeddings provided by GNNs
and clustering patients and EHR features respectively, then conducted qualitative
feature cluster analysis for clinical interpretations. Our study shows that LIGHTED
can take advantage of longitudinal EHR data and the intrinsic graph structure of
EHRs among patients to provide effective and interpretable OD risk predictions that
may potentially improve clinical decision support.
Keywords: Opioid overdose; Opioid poisoning; Deep learning; Clinical decision
support; Electronic health records; Long short-term memory; Graph neural network
Weifang Mao, Huiming Zhu, Hao Wu, Yijie Lu, Haidong Wang,
Forecasting and trading credit default swap indices using a deep learning model
integrating Merton and LSTMs,
Expert Systems with Applications,
Volume 213, Part B,
2023,
119012,
ISSN 0957-4174,
https://doi.org/10.1016/j.eswa.2022.119012.
(https://www.sciencedirect.com/science/article/pii/S0957417422020309)
Abstract: Using macroeconomic and financial conditions to forecast credit default
swap (CDS) spreads is a challenging task. In this paper, we propose the Merton-LSTM
model, a modified LSTM model formed by integrating with the Merton determinants
model, to forecast the CDS indices. We provide the rigorous math behind the Merton-
LSTM model, which demonstrates that by leveraging the nonlinear learning ability of
LSTM with increased model capacity, the Merton-LSTM model is expected to learn the
inherent association between the Merton determinants and CDS spreads. Further, the
Merton-LSTM model is compared with the machine learning models LSTM, gated
recurrent unit (GRU), multilayer perceptron network (MLP), support vector machine
(SVM) and a typical stochastic series model in forecasting the two most liquid
five-year CDS indices, North America High Yield index (CDX.NA.HY) and North America
Investment Grade index (CDX.NA.IG) through the root mean squared error (RMSE) and
the Diebold-Mariano test. The comparison results show that the RMSEs of the Merton-
LSTM model are the lowest (6.2570–27.2000 for CDX.NA.HY and 1.3168–6.4772 for
CDX.NA.IG) compared to other competitive models. The superiority of the Merton-LSTM
model in forecasting performance is highlighted in long-term prediction even with a
forecasting horizon extended to 28 days. Simulated trading with different
thresholds and horizons is conducted in this study. We find that the Merton-LSTM
trading strategy yields the highest annualized Sharpe ratios and lowest maximum
losses at most thresholds and horizons, highlighting the economic significance of
the proposed model.
Keywords: Forecasting; Trading; Credit default swap; LSTM; Deep learning
Md. Kowsher, Anik Tahabilder, Md. Zahidul Islam Sanjid, Nusrat Jahan Prottasha, Md.
Shihab Uddin, Md Arman Hossain, Md. Abdul Kader Jilani,
LSTM-ANN & BiLSTM-ANN: Hybrid deep learning models for enhanced classification
accuracy,
Procedia Computer Science,
Volume 193,
2021,
Pages 131-140,
ISSN 1877-0509,
https://doi.org/10.1016/j.procs.2021.10.013.
(https://www.sciencedirect.com/science/article/pii/S1877050921020548)
Abstract: Machine learning is getting more and more advanced with the progression
of state-of-the-art technologies. Since existing algorithms do not provide a
palatable learning performance most often, it is necessary to carry on the trail of
upgrading the current algorithms incessantly. The hybridization of two or more
algorithms can potentially increase the performance of the blueprinted model.
Although LSTM and BiLSTM are two excellent far and widely used algorithms in
natural language processing, there still could be room for improvement in terms of
accuracy via the hybridization method. Thus, the advantages of both RNN and ANN
algorithms can be obtained simultaneously. This paper has illustrated the deep
integration of BiLSTM-ANN (Fully Connected Neural Network) and LSTM-ANN and
manifested how these integration methods are performing better than single BiLSTM,
LSTM and ANN models. Undertaking Bangla content classification is challenging
because of its equivocalness, intricacy, diversity, and shortage of relevant data,
therefore, we have executed the whole integrated models on the Bangla content
classification dataset from newspaper articles. The proposed hybrid BiLSTM-ANN
model beats all the implemented models with the most noteworthy accuracy score of
93% for both validation & testing. Moreover, we have analyzed and compared the
performance of the models based on the most relevant parameters.
Keywords: BiLSTM-ANN; LSTM-ANN; Supervised machine learning; Hybrid ML model;
Fusion of ML model; NLP
Haibo Huang, Xiaorong Huang, Weiping Ding, Siwen Zhang, Jian Pang,
Optimization of electric vehicle sound package based on LSTM with an adaptive
learning rate forest and multiple-level multiple-object method,
Mechanical Systems and Signal Processing,
Volume 187,
2023,
109932,
ISSN 0888-3270,
https://doi.org/10.1016/j.ymssp.2022.109932.
(https://www.sciencedirect.com/science/article/pii/S0888327022010007)
Abstract: The sound absorption and sound insulation performance of an acoustic
package (AP) system directly affect the noise, vibration and harshness performance
of a vehicle. Numerous studies have studied the optimization of vehicle sound
package, however, there are two deficiencies in the current research of sound
package: (1) The noise transmission path of acoustic package is complex and
hierarchical. Most of the related works focus on the data-driven part while
ignoring the knowledge attributes behind the acoustic package design problem, which
limits the further improvement of prediction and optimization of acoustic package
performance. (2) In using intelligent neural networks-based methods such as long
short-term memory (LSTM), reducing the learning rate during training gradually
narrows the search interval of a solution, and adjusting the learning rate in a
small range may tend to trap local optima. In this study, a knowledge- and data-
driven approach is proposed for the development of acoustic package systems. A
multiple-level multiple-object method is proposed as the knowledge model, and a
multilayer structure of the acoustic package system that contains the system,
subsystem and component layers is developed. In addition, an improved long short-
term memory model based on an adaptive learning rate forest, which can increase and
decrease the learning rate adaptively, is proposed as the data-driven model. The
knowledge- and data-driven method is applied to optimize the sound absorption and
insulation of the acoustic package system. In the experimental validation, the
effectiveness and robustness of the proposed method outperformed the traditional
direct mapping method and the conventional long short-term memory method.
Keywords: Noise, vibration and harshness; Acoustic package; Knowledge- and data-
driven; Long short-term memory; Multiple-level multiple-object
Yiyin Tang, Yalin Wang, Chenliang Liu, Xiaofeng Yuan, Kai Wang, Chunhua Yang,
Semi-supervised LSTM with historical feature fusion attention for temporal sequence
dynamic modeling in industrial processes,
Engineering Applications of Artificial Intelligence,
Volume 117, Part A,
2023,
105547,
ISSN 0952-1976,
https://doi.org/10.1016/j.engappai.2022.105547.
(https://www.sciencedirect.com/science/article/pii/S0952197622005371)
Abstract: In modern industrial processes, the data-driven soft sensor technology
has been widely used for the prediction of key quality variables. Due to the
important of dynamics and nonlinearity in industrial process data, deep learning
models like long short-term memory (LSTM) network are well suited for temporal
sequence dynamic modeling due to their excellent long-term memory function and
feature extraction capability. Furthermore, industrial processes generate a large
amount of process data with irregular sampling frequencies. However, traditional
LSTM cannot fully utilize the process data with irregular sampling frequency and
the guidance value of historical data samples for feature learning. To address
these issues, a novel semi-supervised LSTM with history feature fusion attention
(HFFA-SSLSTM) model is proposed in this paper. First, the semi-supervised learning
strategy is implemented in LSTM to fully utilize the unlabeled data and mine the
temporal sequence features of labeled samples and unlabeled samples with irregular
sampling frequencies. Then, a novel historical feature fusion attention (HFFA)
mechanism is developed, which utilizes historical hidden features to learn
attention scores for obtaining weighted historical information-related features.
Finally, the extracted features are combined to form the soft sensor model to
perform time series prediction tasks for key quality variables in industrial
processes. The experimental results on the actual industrial hydrocracking data set
demonstrate the effectiveness of the proposed HFFA-SSLSTM model and its possibility
of applicating in real industrial processes.
Keywords: Deep learning; Semi-supervised long short-term memory (SSLSTM);
Historical feature fusion attention (HFFA); Temporal sequence dynamic modeling;
Industrial processes
Chunlei Ji, Chu Zhang, Lei Hua, Huixin Ma, Muhammad Shahzad Nazir, Tian Peng,
A multi-scale evolutionary deep learning model based on CEEMDAN, improved whale
optimization algorithm, regularized extreme learning machine and LSTM for AQI
prediction,
Environmental Research,
Volume 215, Part 1,
2022,
114228,
ISSN 0013-9351,
https://doi.org/10.1016/j.envres.2022.114228.
(https://www.sciencedirect.com/science/article/pii/S0013935122015559)
Abstract: With the rapid development of economy, air pollution occurs frequently,
which has a huge negative impact on human health and urban ecosystem. Air quality
index (AQI) can directly reflect the degree of air pollution. Accurate AQI trend
prediction can provide reliable information for the prevention and control of air
pollution, but traditional forecasting methods have limited performance. To this
end, a dual-scale ensemble learning framework is proposed for the complex AQI time
series prediction. First, complete ensemble empirical mode decomposition adaptive
noise (CEEMDAN) and sample entropy (SE) are used to decompose and reconstruct AQI
series to reduce the difficulty of direct modeling. Then, according to the
characteristics of high and low frequencies, the high-frequency components are
predicted by the long short-term memory neural network (LSTM), and the low-
frequency items are predicted by the regularized extreme learning machine (RELM).
At the same time, the improved whale optimization algorithm (WOA) is used to
optimize the hyper-parameters of RELM and LSTM models. Finally, the hybrid
prediction model proposed in this paper predicts the AQI of four cities in China.
This work effectively improves the prediction accuracy of AQI, which is of great
significance to the sustainable development of the cities.
Keywords: Regularized extreme learning machine; Long short-term memory; CEEMDAN;
Whale optimization algorithm
Xingyu Tang, Peijie Zheng, Xueyong Li, Hongyan Wu, Dong-Qing Wei, Yuewu Liu, Guohua
Huang,
Deep6mAPred: A CNN and Bi-LSTM-based deep learning method for predicting DNA N6-
methyladenosine sites across plant species,
Methods,
Volume 204,
2022,
Pages 142-150,
ISSN 1046-2023,
https://doi.org/10.1016/j.ymeth.2022.04.011.
(https://www.sciencedirect.com/science/article/pii/S1046202322001001)
Abstract: DNA N6-methyladenine (6mA) is a key DNA modification, which plays
versatile roles in the cellular processes, including regulation of gene expression,
DNA repair, and DNA replication. DNA 6mA is closely associated with many diseases
in the mammals and with growth as well as development of plants. Precisely
detecting DNA 6mA sites is of great importance to exploration of 6mA functions.
Although many computational methods have been presented for DNA 6mA prediction,
there is still a wide gap in the practical application. We presented a convolution
neural network (CNN) and bi-directional long-short term memory (Bi-LSTM)-based deep
learning method (Deep6mAPred) for predicting DNA 6mA sites across plant species.
The Deep6mAPred stacked the CNNs and the Bi-LSTMs in a paralleling manner instead
of a series-connection manner. The Deep6mAPred also employed the attention
mechanism for improving the representations of sequences. The Deep6mAPred reached
an accuracy of 0.9556 over the independent rice dataset, far outperforming the
state-of-the-art methods. The tests across plant species showed that the
Deep6mAPred is of a remarkable advantage over the state of the art methods. We
developed a user-friendly web application for DNA 6mA prediction, which is freely
available at http://106.13.196.152:7001/ for all the scientific researchers. The
Deep6mAPred would enrich tools to predict DNA 6mA sites and speed up the
exploration of DNA modification.
Keywords: 6mA; DNA modification; Convolution neural network; Long-short term
memory; Feed-forward attention; Deep learning
Md. Ferdous Ahammed, A. Alim Molla, Rafiul Kadir, Mohammad Ismat Kadir,
Deep bidirectional LSTM for the signal detection of universal filtered multicarrier
systems,
Machine Learning with Applications,
Volume 10,
2022,
100425,
ISSN 2666-8270,
https://doi.org/10.1016/j.mlwa.2022.100425.
(https://www.sciencedirect.com/science/article/pii/S2666827022001001)
Abstract: Universal filtered multicarrier (UFMC) has emerged as a potential
waveform contender of orthogonal frequency division multiplexing (OFDM) for the
fifth generation (5G) and beyond wireless systems. In this paper, we propose a
bidirectional long short-term memory (Bi-LSTM)-based detector for the UFMC system.
The proposed detector directly detects the transmitted symbols using the deep
learning (DL)-based training data. The system is first trained with the aid of
training data and pilot symbols. The training tunes the DL-based network
parameters. During the testing phase, the signal is detected using the trained
network. The performance of the proposed scheme is compared with that of the DL-
aided OFDM system, and with the signal detection strategies using the conventional
channel estimation techniques. Our simulations show that the proposed Bi-LSTM-based
DL can flexibly and effectively detect UFMC signals.
Keywords: Universal filtered multicarrier; Deep learning; Long short-term memory
(LSTM); Bidirectional LSTM; Rayleigh fading channel
Ruoyu Yang, Shubhendu Kumar Singh, Mostafa Tavakkoli, Nikta Amiri, Yongchao Yang,
M. Amin Karami, Rahul Rai,
CNN-LSTM deep learning architecture for computer vision-based modal frequency
detection,
Mechanical Systems and Signal Processing,
Volume 144,
2020,
106885,
ISSN 0888-3270,
https://doi.org/10.1016/j.ymssp.2020.106885.
(https://www.sciencedirect.com/science/article/pii/S0888327020302715)
Abstract: The conventional modal analysis involves physically-attached wired or
wireless sensors for vibration measurement of structures. However, this method has
certain disadvantages, owing to the sensor’s weight and its low spatial resolution,
which limits the analysis precision or the high cost of optical vibration sensors.
Besides, the sensor installation and calibration in itself is a time consuming and
labor-intensive process. Non-contact computer vision-based vibration measurement
techniques can address the shortcomings mentioned above. In this paper, we
introduce CNN-LSTM (Convolutional Neural Network, Long Short-Term Memory) deep
learning based approach that can serve as a backbone for computer vision-based
vibration measurement techniques. The key idea is to use each pixel of an image
taken from an off the shelf camera, encapsulating the Spatio-temporal information,
like a sensor to capture the modal frequencies of a vibrating structure. Non-
contact “pixel-sensor” does not alter the system’s dynamics and is relatively low-
cost, agile, and provides measurements with very high spatial resolution. Our
computer vision-based deep learning model takes the video of a vibrating structure
as input and outputs the fundamental modal frequencies. We demonstrate, using
reliable empirical results, that “pixel-sensor” is more efficient, autonomous, and
accurate. Robustness of the deep learning model has been put to the test by using
specimens of a variety of materials, and varying dimensions and results have shown
high levels of sensing accuracy.
Keywords: CNN (convolutional neural network); LSTM (long short-term memory);
Computer vision; Modal analysis
Zengyi Lyu, Xiaowei Jia, Yao Yang, Keqi Hu, Feifei Zhang, Gaofeng Wang,
A comprehensive investigation of LSTM-CNN deep learning model for fast detection of
combustion instability,
Fuel,
Volume 303,
2021,
121300,
ISSN 0016-2361,
https://doi.org/10.1016/j.fuel.2021.121300.
(https://www.sciencedirect.com/science/article/pii/S0016236121011790)
Abstract: In this paper, we propose a deep learning model to detect combustion
instability using high-speed flame image sequences. The detection model combines
Convolutional Neural Network (CNN) and Long Short-Term Memory network (LSTM) to
learn both spatial features and temporal correlations from high-speed images, and
then outputs combustion instability detection results. We also visualize the
extracted spatial features and their temporal evolution to interpret the detection
process of model. In addition, we discuss the effect of different complexity of CNN
layers and different amounts of training data on model performance. The proposed
method achieves superior performance under various combustion conditions in swirl
chamber with high accuracy and a short processing time about 1.23 ms per frame.
Hence, we show that the proposed deep learning model is a promising detection tool
for combustion instability under various combustion conditions.
Keywords: Premixed swirling flame; Combustion instability; Deep learning;
Convolutional neural network; LSTM
Lin Lin, Xin Guan, Benran Hu, Jun Li, Ning Wang, Di Sun,
Deep reinforcement learning and LSTM for optimal renewable energy accommodation in
5G internet of energy with bad data tolerant,
Computer Communications,
Volume 156,
2020,
Pages 46-53,
ISSN 0140-3664,
https://doi.org/10.1016/j.comcom.2020.03.024.
(https://www.sciencedirect.com/science/article/pii/S0140366419321103)
Abstract: With the high penetration of large scale distributed renewable energy
generations, there is a serious curtailment of wind and solar energy in 5G internet
of energy. A reasonable assessment of large scale renewable energy grid-connected
capacities under random scenarios is critical to promote the efficient utilization
of renewable energy and improve the stability of power systems. To assure the
authenticity of the data collected by the terminals and describe data
characteristics precisely are crucial problems in assessing the accommodation
capability of renewable energy. To solve these problems, in this paper, we propose
an L-DRL algorithm based on deep reinforcement learning (DRL) to maximize renewable
energy accommodation in 5G internet of energy. LSTM as a bad data tolerant
mechanism provides real state value for the solution of accommodation strategy,
which ensures the accurate assessment of renewable energy accommodation capacity.
DDPG is used to obtain optimal renewable energy accommodation strategies in
different scenarios. In the numerical results, based on real meteorological data,
we validate the performance of the proposed algorithm. Results show considering the
energy storage system and demand response mechanism can improve the capacity of
renewable energy accommodation in 5G internet of energy.
Keywords: 5G internet of energy; Renewable energy accommodation; Deep reinforcement
learning; Demand response; LSTM
K.E. ArunKumar, Dinesh V. Kalaga, Ch. Mohan Sai Kumar, Masahiro Kawaji, Timothy M.
Brenza,
Comparative analysis of Gated Recurrent Units (GRU), long Short-Term memory (LSTM)
cells, autoregressive Integrated moving average (ARIMA), seasonal autoregressive
Integrated moving average (SARIMA) for forecasting COVID-19 trends,
Alexandria Engineering Journal,
Volume 61, Issue 10,
2022,
Pages 7585-7603,
ISSN 1110-0168,
https://doi.org/10.1016/j.aej.2022.01.011.
(https://www.sciencedirect.com/science/article/pii/S1110016822000138)
Abstract: Several machine learning and deep learning models were reported in the
literature to forecast COVID-19 but there is no comprehensive report on the
comparison between statistical models and deep learning models. The present work
reports a comparative time-series analysis of deep learning techniques (Recurrent
Neural Networks with GRU and LSTM cells) and statistical techniques (ARIMA and
SARIMA) to forecast the country-wise cumulative confirmed, recovered, and deaths.
The Gated Recurrent Units (GRU), Long Short-Term Memory (LSTM) cells based on
Recurrent Neural Networks (RNN), ARIMA and SARIMA models were trained, tested, and
optimized to forecast the trends of the COVID-19. We deployed python to optimize
the parameters of ARIMA which include (p, d, q) representing autoregressive and
moving average terms and parameters of SARIMA model include additional seasonal
terms which are denoted by (P, D, Q). Similarly, for LSTM and GRU based RNN models’
parameters (number of layers, hidden size, learning rate and number of epochs) are
optimized by deploying PyTorch machine learning framework. The best model was
chosen based on the lowest Mean Square Error (MSE) and Root Mean Squared Error
(RMSE) values. For most of the time-series data of the countries, deep learning-
based models LSTM and GRU outperformed statistical ARIMA and SARIMA models, with an
RMSE values that are 40 folds less than that of the ARIMA models. But for some
countries statistical (ARIMA, SARIMA) models outperformed deep learning models.
Further, we emphasize the importance of various factors such as age, preventive
measures and healthcare facilities etc. that play vital role on the rapid spread of
COVID-19 pandemic.
Keywords: COVID-19 pandemic; Gated Recurrent Units (GRUs); Long Short-Term Memory
(LSTM) cells; Recurrent Neural Networks (RNNs); Auto Regressive Integrated Moving
Average (ARIMA); Seasonal Auto Regressive Integrated Moving Average (SARIMA)
Zeng Chen, Huan Xu, Peng Jiang, Shanen Yu, Guang Lin, Igor Bychkov, Alexey Hmelnov,
Gennady Ruzhnikov, Ning Zhu, Zhen Liu,
A transfer Learning-Based LSTM strategy for imputing Large-Scale consecutive
missing data and its application in a water quality prediction system,
Journal of Hydrology,
Volume 602,
2021,
126573,
ISSN 0022-1694,
https://doi.org/10.1016/j.jhydrol.2021.126573.
(https://www.sciencedirect.com/science/article/pii/S002216942100620X)
Abstract: In recent years, water quality monitoring has been crucial to improve
water resource protection and management. Under the relevant laws and regulations,
environmental protection department agencies monitor lakes, streams, rivers, and
other types of water bodies to assess water quality conditions. The valid and high-
quality data generated from these monitoring activities help water resource
managers understand the existing pollution situations, energy consumption problems
and pollution control needs. However, there are inevitably many problems with water
quality data in the real world due to human mistakes or system failures. One of the
most frequently occurring issues is missing data. Although most existing studies
have explored classic statistical methods or emerging machine/deep learning methods
to fill gaps in data, these methods are not suitable for large-scale consecutive
missing data problems. To address this issue, this paper proposes a novel algorithm
called TrAdaBoost-LSTM, which integrates state-of-the-art deep learning theory
through long short-term memory (LSTM) and instance-based transfer learning through
TrAdaBoost. This model inherits the full advantages of the LSTM model and transfer
learning technique, namely the powerful ability to capture the long-term
dependencies among time series and the flexibility of leveraging the related
knowledge from complete datasets to fill in large-scale consecutive missing data. A
case study involving Dissolved Oxygen concentrations obtained from water quality
monitoring stations is conducted to validate the effectiveness and superiority of
the proposed method. The results show that the proposed TrAdaBoost-LSTM model not
only improves the imputation accuracy by 15%~25% compared with that of alternative
models based on the obtained performance indicators, but also provides potential
ideas for similar future research.
Keywords: Water quality; Transfer learning; LSTM; TrAdaBoost; Large-scale
consecutive missing data
Chao Peng, Yifan Tao, Zhipeng Chen, Yong Zhang, Xiaoyan Sun,
Multi-source transfer learning guided ensemble LSTM for building multi-load
forecasting,
Expert Systems with Applications,
Volume 202,
2022,
117194,
ISSN 0957-4174,
https://doi.org/10.1016/j.eswa.2022.117194.
(https://www.sciencedirect.com/science/article/pii/S0957417422005814)
Abstract: Generally, it is difficult to establish an accurate building load
forecasting model by using insufficient energy data. Although the transfer of
knowledge from similar buildings can effectively solve this problem, there is still
a lack of effective methods for both the selection of source domain buildings and
the use of transfer knowledge when many candidate buildings are available. In view
of this, this paper proposes a multi-source transfer learning guided ensemble LSTM
method for building multi-load forecasting (MTE-LSTM). Firstly, a two-stage source-
domain building matching method based on dominance comparison is developed to find
multiple source-domain buildings similar to the target building. Next, an LSTM
modeling strategy combining transfer learning and fine-tune technology is proposed,
which uses multiple source-domain data to generate multiple basic load forecasting
models for the target building. Following that, a model ensemble strategy based on
similarity degree is given to weight the output results of basic forecasting
models. Applications in many real buildings shows that the proposed building multi-
energy load forecasting method can obtain high-precision load forecasting results
when the target building data is relatively few.
Keywords: Building load forecasting; Transfer learning; Multi-source; LSTM
Haowen Hu, Xin Xia, Yuanlin Luo, Chu Zhang, Muhammad Shahzad Nazir, Tian Peng,
Development and application of an evolutionary deep learning framework of LSTM
based on improved grasshopper optimization algorithm for short-term load
forecasting,
Journal of Building Engineering,
Volume 57,
2022,
104975,
ISSN 2352-7102,
https://doi.org/10.1016/j.jobe.2022.104975.
(https://www.sciencedirect.com/science/article/pii/S2352710222009858)
Abstract: Accurate short-term load forecasting (STLF) plays an important role in
the daily operation of a smart grid. In order to forecast short-term load more
effectively, this article proposes an integrated evolutionary deep learning
approach based on complete ensemble empirical mode decomposition with adaptive
noise (CEEMDAN), improved grasshopper optimization algorithm (IGOA), and long
short-term memory (LSTM) network. First of all, CEEMDAN is used to decompose the
original data into a certain number of periodic intrinsic mode functions (IMFs) and
a residual. Secondly, the nonlinear strategy is used to improve the attenuation
coefficient of GOA, and the golden sine operator is introduced to update the
individual position of GOA. Then the improved GOA is used to optimize the
parameters of the LSTM model, which are the number of hidden neurons and learning
rate. The optimized LSTM is applied to the decomposed modal components. Finally,
the prediction results of each modal component are aggregated to get the real STLF
results. Through comparative experiments, the effectiveness of the CEEMDAN method,
the IGOA method, and the combined model is verified, respectively. The experimental
results show that the integrated evolutionary deep learning method proposed in this
article is an effective tool for STLF.
Keywords: Short-term load forecasting; Complete ensemble empirical mode
decomposition with adaptive noise; Improved grasshopper optimization algorithm;
Long short-term memory