computers-12-00170
computers-12-00170
Article
Evaluating User Satisfaction Using Deep-Learning-Based
Sentiment Analysis for Social Media Data in Saudi Arabia’s
Telecommunication Sector
Majed A. Alshamari
Department of Information Systems, College of Computer Sciences and Information Technology, King Faisal
University, Hofuf 31983, Saudi Arabia; smajed@kfu.edu.sa
Abstract: Social media has become common as a means to convey opinions and express the extent of
satisfaction and dissatisfaction with a service or product. In the Kingdom of Saudi Arabia specifically,
most social media users share positive and negative opinions about a service or product, especially
regarding communication services, which is one of the most important services for citizens who use
it to communicate with the world. This research aimed to analyse and measure user satisfaction
with the services provided by the Saudi Telecom Company (STC), Mobily, and Zain. This type
of sentiment analysis is an important measure and is used to make important business decisions
to succeed in increasing customer loyalty and satisfaction. In this study, the authors developed
advanced methods based on deep learning (DL) to analyse and reveal the percentage of customer
satisfaction using the publicly available dataset AraCust. Several DL models have been utilised in
this study, including long short-term memory (LSTM), gated recurrent unit (GRU), and BiLSTM,
on the AraCust dataset. The LSTM model achieved the highest performance in text classification,
demonstrating a 98.04% training accuracy and a 97.03% test score. The study addressed the biggest
challenge that telecommunications companies face: that the company’s services influence customers’
decisions due to their dissatisfaction with the provided services.
Citation: Alshamari, M.A.
Evaluating User Satisfaction Using
Keywords: user satisfaction; human–computer interaction; deep learning algorithms; sentiment
Deep-Learning-Based Sentiment analysis; social media; artificial intelligence; telecommunication
Analysis for Social Media Data in
Saudi Arabia’s Telecommunication
Sector. Computers 2023, 12, 170.
https://doi.org/10.3390/ 1. Introduction
computers12090170 Measuring users’ satisfaction is a critical part of assessing successful interaction be-
Academic Editors: Phivos Mylonas, tween humans and technologies. The telecommunications industry has emerged as a
Katia Lida Kermanidis and prominent sector in developed nations. The escalation of competition has been propelled
Manolis Maragoudakis by the proliferation of operators and advancements in technology [1]. Enterprises are
implementing diverse tactics to sustain themselves in this highly competitive marketplace.
Received: 26 June 2023
According to the extant literature [2], three principal strategies have been proposed to aug-
Revised: 15 August 2023
ment revenue generation: (1) procuring a new clientele, (2) upselling to the extant clientele,
Accepted: 23 August 2023
and (3) prolonging the retention duration of the clientele. Upon analysing these strategies
Published: 26 August 2023
while considering their respective return on investment (RoI), it has been determined that
the third strategy yields the greatest financial benefit [2]. This discovery corroborates the
idea that maintaining an existing customer is more cost effective than obtaining a new
Copyright: © 2023 by the author. one [3] and is also regarded as a less complicated tactic than the upselling technique [4]. To
Licensee MDPI, Basel, Switzerland. execute the third strategy, corporations must address the potential occurrence of customer
This article is an open access article churn, which describes the phenomenon of customers transitioning from one provider to
distributed under the terms and another [5].
conditions of the Creative Commons The pursuit of customer satisfaction is a key driver for telecommunication companies
Attribution (CC BY) license (https:// in the face of intense global competition. Numerous studies have established a positive cor-
creativecommons.org/licenses/by/ relation between customer satisfaction and both customer loyalty and customer churn [6–8].
4.0/).
2. Background of Study
Various methodologies have been used to forecast customer attrition in telecommuni-
cations firms. The majority of these methodologies employ machine learning (ML) and data
mining techniques. The predominant body of literature has centred on the implementation
of a singular data-mining technique for knowledge extraction, while alternative studies
have prioritised the evaluation of multiple approaches for the purpose of churn prediction.
In their study, Brandusoiu et al. [19] introduced a sophisticated data-mining approach
to predict churn among prepaid customers. This approach involved the use of a dataset con-
taining call details for 3333 customers, which included 21 distinct features. The dependent
churn parameter in this dataset was binary, with values of either ‘Yes’ or ‘No’. The features
encompass details pertaining to the quantity of incoming and outgoing messages as well as
voicemail for individual customers. The PCA algorithm was used by the author to perform
dimensionality reduction on the data. The research used three discrete ML algorithms—
specifically neural networks, a support vector machine (SVM), and Bayes networks—to
predict the churn factor. The author evaluated the algorithms’ performance using the
area under the curve (AUC) as a metric. The present study involved the computation of
the area under the receiver operating characteristic curve (AUC–ROC) for three distinct
ML models: Bayes networks, neural networks, and SVM. The AUC values acquired were
99.10%, 99.55%, and 99.70%, respectively. The current study used a restricted dataset that
was free from any instances of missing data. He et al. [20] proposed a model that employed
the neural network algorithm to tackle the problem of customer churn in a large telecom-
munications company in China that had a customer base of around 5.23 million. The metric
used to assess the precision of predictions was the general accuracy rate, which yielded a
score of 91.1%. Idris [21] addressed the issue of churn in the telecommunications industry
by presenting a methodology that employed genetic programming alongside AdaBoost.
The efficacy of the model was evaluated using two established datasets: one from Orange
Telecom and the other from cell2cell. The cell2cell dataset achieved an accuracy rate of 89%,
while the other dataset achieved a rate of 63%. Huang et al. [22] investigated customer
churn within the context of the big data platform. The aim of the researchers was to exhibit
noteworthy enhancement in churn prediction by leveraging big data, which is dependent
on the magnitude, diversity, and speed of the data. The handling of data derived from
the Operation Support and Business Support divisions of the largest telecommunications
corporation in China required the implementation of a big data platform to enable the
requisite manipulations. The use of the random forest algorithm was evaluated using the
AUC metric.
A rudimentary set theory-based churn prediction model was proposed by Makhtar
et al. [23] for the telecommunications sector. The rough set classification technique outper-
Computers 2023, 12, 170 4 of 24
formed the linear regression, decision tree, and voted perception neural network methods,
as indicated in the aforementioned research. The problem of skewed data sets in churn
prediction has been the subject of several studies. When the number of churned client
classes falls below the number of active customer classes, this phenomenon occurs. In
their research, Amin et al. [24] compared six alternative oversampling strategies in the
context of telecommunication churn prediction. The results showed that genetic algorithm-
based rules-generation oversampling algorithms outperformed the other oversampling
techniques evaluated.
Burez and Van den Poel [25] investigated the issue of imbalanced datasets in churn
prediction models. They conducted a comparative analysis of the efficacy of random
sampling, advanced undersampling, gradient-boosting models, and weighted random
forests. The model was evaluated using metrics such as AUC and Lift. The findings
indicate that the undersampling technique exhibited superior performance compared
to the other techniques tested. Individuals who use social media platforms, including
Twitter, Facebook, and Instagram, tend to provide commentary and evaluations regarding
a company’s offering because these platforms provide a means for users to express their
opinions and exchange ideas concerning products [9]. The process of sentiment analysis,
also referred to as feedback mining, involves the use of natural language processing (NLP),
statistical analysis, and ML to extract and classify feedback from textual inputs based
on criteria such as subjectivity and polarity recognition [6]. Individuals who use social
media platforms, including but not limited to Twitter, Facebook, and Instagram, have
been observed to provide commentary and evaluations regarding a company’s offerings
because these platforms provide an avenue for individuals to express their viewpoints
and exchange their perspectives on products [9]. The process of sentiment analysis, also
referred to as feedback mining, involves the use of NLP, statistical analysis, and ML to
extract and categorise feedback from textual inputs based on factors such as subjectivity
and polarity recognition [6]. Furthermore, Pavaloaia and colleagues provided a concise
definition of sentiment analysis as a social media tool that entails evaluating the presence
of positive and negative keywords in text messages linked to a social media post [10].
Recognition of the need for sentiment analysis is increasing [9,25]. This is attributed to
the growing demand for the estimation and organisation of unstructured data from social
media. The task of text mining is challenging because it involves the identification of topical
words across various subjects. To effectively categorise these words into either positive or
negative polarity, it is imperative to conduct sentiment analysis. Additionally, selecting
appropriate sentiment signals for real-time analysis is crucial in this process [26,27]. The
increasing prevalence of textual content sharing on social media has led to an increase in
the use of text-mining and sentiment analysis techniques [28–30].
The study conducted by [31] involved an analysis of consumer sentiment expressed in
a Jordanian dialect across the Facebook pages of multiple telecommunication businesses
in Jordan, as well as on Twitter. The four fundamental classifiers used for the manual
categorisation of all the gathered and processed attitudes are the SVM, K-nearest neighbour
(k-nn), naïve Bayesnaïve (), and decision tree (DT). The present study used its results to
exhibit the superiority of SVM over three other widely used sentiment classifiers. In [27], the
researchers aimed to ascertain the sentiment of user-generated content but were constrained
to classifying comments instead of the actual posts.
Furthermore, [32] employed Twitter as a medium for conducting sentiment analysis
by scrutinising tweets in the English language originating from diverse businesses in Saudi
Arabia. The researchers used K-nearest neighbour and naive Bayes algorithms to classify
attitudes into three categories: positive, negative, and neutral. These classifications were
based on their daily and monthly trend observations. Furthermore, the K-nearest neighbour
algorithm, an ML methodology, was employed to examine user sentiment in the present
investigation. Nonetheless, the exclusion of Arabic opinions may have resulted in a less
comprehensive dataset.
Computers 2023, 12, 170 5 of 24
The study used a sentiment analysis of Facebook posts as a means of assessing the
efficacy of social media posts in supporting effective self-marketing strategies on social
media platforms. A reference for this study is available. Furthermore, according to a
study conducted by [33], the implementation of sentiment analysis results in a rise in
negative sentiment among followers during phases of reduced user-generated activity. This
is noteworthy, as sentiment analysis consistently yields supplementary insights beyond
those derived from solely analysing comments, likes, and shares of articles. Research has
demonstrated that a single published article has the potential to generate a substantial
number of comments, which can be subjected to sentiment analysis using an ML-based
approach.
The researchers [34–36] used a range of deep learning approaches to establish the
correlation between several organizations and their clients, drawing from feedback, quality
assessments, comments, and surveys conducted across many domains. In the field of
natural language processing (NLP), these three approaches have garnered significant
interest because of their exceptional accuracy in text categorization analysis. These methods
have shown to be indispensable in many sectors, including commercial and consumer
interactions, as well as in predicting societal implications on future trends. The user has
provided a numerical sequence.
Figure 3. Positive and negative sentiments for (a) STC, (b) Mobily, and (c) Zain.
Figure
Figure 3. Positiveand
3. Positive andnegative
negative sentiments
sentiments for
for(a)(a)STC,
STC,(b)(b)
Mobily, andand
Mobily, (c) Zain.
(c) Zain.
Table 2 and
Table Figure
2 andFigure 44 permit
Figure4 a quick
permit aa quick comparison of sentiment frequencies between
Table 2 and permit quickcomparison
comparison of of
sentiment
sentimentfrequencies between
frequencies between
thethe
three telecommunication
three telecommunication providers, providing
providers, providing insights
insights intointo
the the overall
overall sentiments
sentiments as- as-
the three telecommunication providers, providing insights into the overall sentiments
sociated with
sociated witheach
eachcompany.
company.
associated with each company.
Figure 4. Comparison of the positive and negative sentiments for each company.
Figure
Figure Comparison
4.Arabic
4.
3.2.2. Comparison of the
of
Bigrams the positive
positive and
and negative
negative sentiments for each company.
Bigrams can be used for Arabic linguistic structures. They demonstrate Arabic text
3.2.2. Arabicand
structure Bigrams
meaning and capture the relationship between adjacent words [38]. Bigrams
are essentialcan
Bigrams to text
be analysis
used forand NLP. linguistic
Arabic Frequency,structures.
co-occurrence,
Theyanddemonstrate
context are ways to text
Arabic
structure and meaning and capture the relationship between adjacent words [38]. Bigrams
are essential to text analysis and NLP. Frequency, co-occurrence, and context are ways to
Computers 2023, 12, 170 8 of 24
Forget gate:
f t = σ W f x t + U f h { t −1} + b f (2)
Memory cell:
ct = f t ∗ c{t−1} + it ∗ tanh Wcxt + Uch{t−1} + bc (3)
Output gate:
o_t = σ Woxt + Uoh{t−1} + bo (4)
Hidden state:
ht = ot ∗ tanh(ct ), (5)
where xt is the current input, ht is the hidden state, ct is the memory cell state, it is the
input, ft is the forget state, ot is the output gate, and W, U and b are the network weights
and biases. Sigma and tanh represent the sigmoid and hyperbolic tangent activation
functions, respectively.
In this study, an LSTM model was developed for binary text classification tasks. The
model’s architecture consisted of the input, hidden, and output layers, as shown in Figure 5.
First, the model’s embedding layer was built. The embedding layer converted text into
numerical form for the neural network. The tokeniser’s index–word dictionary’s mappings
between word indices and words determined vocabulary size. The embedding layer’s
64 dimensions enabled dense vector representations of vocabulary words. Then, three
LSTM layers were added, the first layer with 512 and the second with 128 units, both
returned sequences = True, which returned all previous outputs from these LSTM layers.
The third layer was 64 units LSTM. Finally, a dense layer of two neurons completed the
model. In this dense layer, all neurons communicated with each other and with the layer
mputers 2023, 12, x FOR PEER REVIEW below. This layer used the sigmoid function for binary classification because it worked well. 12 of
Each class’s likelihood was compressed to 0 and 1, representing the negative and positive;
for more details of the model parameters, see Table 5.
FigureFigure
5. LSTM model.
5. LSTM model.
Parameter Value
Embedding layer Embedding(len(tokenizer.index_word) + 1, 64)
layer 1 LSTM (512, return_sequences = True)
Computers 2023, 12, 170 11 of 24
Parameter Value
Embedding layer Embedding(len(tokenizer.index_word) + 1, 64)
layer 1 LSTM (512, return_sequences = True)
layer 2 LSTM (128, return_sequences = True)
layer 3 LSTM (64)
Dense layer Dense (2, activation = ‘sigmoid’)
Optimiser Adam
Loss function Binary cross entropy
Batch size 256
Epochs 20
Callbacks Early stop with 5 patience
Figure 6. GRU
Figure model.
6. GRU model.
Parameter Value
Embedding layer Embedding(len(tokenizer.index_word) + 1, 64)
Computers 2023, 12, 170 12 of 24
Parameter Value
Embedding layer Embedding(len(tokenizer.index_word) + 1, 64)
GRU layer 1 GRU (512, return_sequences = True)
GRU layer 2 GRU (128, return_sequences = True)
GRU layer 3 GRU (64)
Dense layer Dense (2, activation = ‘sigmoid’)
Optimiser Adam
Loss function Binary cross entropy
Batch size 256
Epochs 20
Callbacks Early stop with 5 patience
Parameter Value
Embedding layer Embedding(len(tokenizer.index_word) + 1, 64)
Bidirectional layer 1 LSTM (128, return_sequences = True)
Bidirectional layer 3 LSTM (32)
Dense layer Dense (2, activation = ‘sigmoid’)
Optimiser RMSprop
Loss function Binary crossentropy
Batch size 256
Epochs 20
Callbacks Early stop with 5 patience
Like the forward LSTM, the internal memory state is updated based on the input
and memory states at each time step. Bidirectional LSTM networks combine forward and
backward LSTM layers to obtain a more complete picture of the input sequence at a given
time. Routing the output layer through fully connected layers generates the final output.
Despite their higher training and inference costs, bidirectional LSTM networks are often
a good choice for sequence modelling due to their improved performance and ability to
capture bidirectional dependencies.
The BiLSTM model has several key components for text analysis, as shown in Figure 7.
An embedding layer converts each word or token into a dense vector in continuous space.
Len(tokeniser.index_word) + 1 vocabulary and 64 embedding dimensions were used in
this model’s embedding layer. Each word in the input text becomes a 64-dimensional
semantic vector. Next is a 128-unit LSTM bidirectional layer. LSTM and RNN are op-
timised for sequential data processing by long-term dependencies and a memory state.
Bidirectional layers analyse input sequences. Return sequences = True ensures that the
layer returns the hidden state output for each time step in the sequence without disrupting
Computers 2023, 12, 170 13 of 24
the data’s natural order. After that, a 32-unit LSTM bidirectional layer uses a second. This
layer gathers contextual information
layer gathers contextual from the input
information fromsequence’s forward and
the input sequence’s backward
forward and backward
motions. Fully connected,
motions. Fullytheconnected,
final densethelayer has two
final dense units.
layer Since
has two the Since
units. model thepredicts
model predicts
two classes—satisfactory and unsatisfactory—this
two classes—satisfactory layer completes
and unsatisfactory—this the classification.
layer completes Each Each
the classification.
class’s
class’s probability probability
estimates estimates areto
are compressed compressed to a single
a single value value0between
between and 1 by0 and 1 by the dense
the dense
layer’s ‘sigmoid’layer’s ‘sigmoid’
activation activation function.
function.
Figure 7. BiLSTMFigure
model.7. BiLSTM model.
Table Neural
3.5.5. Convolutional 7. BiLSTM model parameters.
Networks
CNNs, a type of DL model, process and analyse grid-like
Parameter Value data, such as images and
sequences
Embedding layer[47]. They were developed for computer vision but are now used
Embedding(len(tokenizer.index_word) + 1,in64)
NLP [13].
CNNs learn
Bidirectional hierarchical
layer 1 representations ofLSTM input(128,
datareturn_sequences
automatically. Convolution,
= True) pool-
ing, and fully
Bidirectional layer 3connected neurons achieve this. CNNs LSTM have revolutionised
(32) many fields,
including
Dense layercomputer vision, by performing well in image
Dense classification,
(2, activation object detection,
= ‘sigmoid’)
and segmentation.
Optimiser RMSprop
Loss CNNs
functioncan process text and audio signals, makingBinary them crossentropy
useful for NLP. One-dimensional
CNNs use
Batch size sequence data such as word embeddings or character-level
256 encodings. Parsed
sentences
Epochs and documents are converted into neural network-readable
20 numerical represen-
tations. One-dimensional filters or kernels scanEarly
Callbacks inputstop
datawithin 5the convolutional layer to
patience
apply convolutional operations across sequential dimensions. The pooling layer reduces
the convolutional 3.5.5. Convolutional
layer’s feature maps,NeuralandNetworks
the fully connected layers continue processing
and learning relevant feature
CNNs, combinations.
a type of DL model, process and analyse grid-like data, such as images and
One-dimensional
sequencesCNNs
[47]. are
They useful for text classification,
were developed for computersentiment analysis,
vision but are now used named-
in NLP [13].
entity recognition, CNNsandlearn hierarchical
speech representations
recognition. To improve of NLP
inputperformance,
data automatically.
1D CNNsConvolution,
have pool-
ing, and fully connected
been modified architecturally. neuronsimprove
These changes achieve this. CNNs haveofrevolutionised
the modelling one-dimensional many fields,
including computer vision,
sequential data and feature extraction for NLP. by performing well in image classification, object detection,
and segmentation.
CNNs can process text and audio signals, making them useful for NLP. One-dimen-
3.5.6. CNN-LSTM Network
sional CNNs use sequence data such as word embeddings or character-level encodings.
The CNN-LSTM model uses the LSTM’s temporal dependencies and the CNN’s spatial
Parsed sentences and documents are converted into neural network-readable numerical
features becauserepresentations.
the CNN feedsOne-dimensional
the LSTM [14]. CNN-LSTM models
filters or kernels scancombine spatial
input data in thefeature
convolutional
extraction powerlayer
andtosequential
apply convolutional operations across sequential dimensions.sequential
modelling precision. CNNs and LSTMs process The pooling layer
data to extract features. It can
reduces the be used forlayer’s
convolutional videofeature
analysis, sentiment,
maps, and text
and the fully classification,
connected layers continue
requiring spatialprocessing
and temporal data. The
and learning CNN
relevant extracts
feature spatial features from input data,
combinations.
while the LSTM handles sequential or
One-dimensional temporal
CNNs dependencies.
are useful for text classification, sentiment analysis, named-
entity recognition,
Each component of the model and speech recognition.
architecture To improve
has several crucialNLP performance,
parameters. The 1D CNNs have
model
can process up tobeen
themodified architecturally.
tokeniser’s index–word These changes length
dictionary improveplusthe modelling of one-dimensional
one distinct word or
token. Each word sequential
or tokendata and
in the feature
input hasextraction for NLP.
a dense vector representation with 64 dimensions
that captures its semantic meaning. The Conv1D layer’s 128 filters detect patterns in the
input data and serve as feature detectors. The Conv1D layer extracts features from 5-word
or token blocks with a kernel size of 5. Conv1D uses ReLU, a nonlinear activation function,
to better capture intricate data patterns. The LSTM layer’s 64 units, which determine
Computers 2023, 12, 170 14 of 24
the dimensionality of the hidden state and the number of memory cells, are essential
for capturing complex temporal dependencies. Two binary classification units in the
dense layer generate output probabilities. The model’s architecture is made up of these
parameters, which affect data processing and learning. Table 8 shows the parameters of the
CNN-LSTM model. The architecture of the CNN-LSTM model for analysis sentiment of
customer satisfaction from social media is presented in Figure 8.
Parameter Value
Vocabulary size len(tokenizer.index_word) + 1
Embedding dimension 64
Conv1D filters 128
Conv1D kernel size 5
Conv1D activation ReLU
LSTM units 64
Dense units 2
Dense activation Sigmoid
Optimiser Adam
Loss function Binary Cross Entropy
Metrics Accuracy
Batch size 256
Number of epochs 20
Computers 2023, 12, x FOR PEER REVIEW 16 of 25
Callbacks Early stop with 5 patience
Figure 8.
Figure 8.CNN-LSTM
CNN-LSTMModel.
Model.
4. Experimental Results
In this section, the experimental setup and results are presented, in which different
models were assessed based on standard evaluation metrics.
Computers 2023, 12, 170 15 of 24
4. Experimental Results
In this section, the experimental setup and results are presented, in which different
models were assessed based on standard evaluation metrics.
TP + TN
Accuracy = × 100 (6)
FP + FN + TP + TN
Confusion Matrix: The confusion matrix analyses model predictions. True positive
(TP), true negative (TN), false positive (FP), and false negative (FN) are shown above. In
addition to the F1 score, the confusion matrix can measure precision, recall, sensitivity,
and specificity.
Precision: Precision, or positive predictive value, measures how well a model predicts
the future. Precision was calculated by dividing the correct results by positive and negative
findings. This metric shows the percentage of confirmed positive instances predicted
correctly. Precision can be calculated by Equation (7).
TP
Precision = × 100 (7)
TP + FP
Sensitivity: Sensitivity (recall) or TP rate (TPR) measures a model’s positive event
detection accuracy. Sensitivity was calculated by dividing the number of correct diag-
noses by the number of incorrect diagnoses. This metric shows the percentage of positive
achievements. Sensitivity can be calculated using Equation (8).
TPs
Sensitivity = (8)
TPs + FPs
Specificity: Specificity is a model’s ability to identify outliers. The actual negative rate
can be calculated using Equation (9). This represented the percentage of negative cases that
were accurately detected.
TNs
Speci f icity = (9)
TNs + FNs
F1 score: The F1 score balances precision and recall to assess a model’s performance.
The F1 score was calculated by dividing the harmonic mean of precision and recall accuracy
Computers 2023, 12, 170 16 of 24
ratings by their sum, as in Equation (10). When recall and precision are equal or datasets
are unbalanced, the F1 score is useful.
precision × Sensitivity
F1 − score = 2 ∗ × 100 (10)
precision + Sensitivity
5. Results
This section presents the results of various DL models, namely BiLSTM, CNN-LSTM,
GRU, and LSTM, for sentiment analysis of Arabic customer satisfaction. Several evaluation
metrics, such as accuracy, precision, and the F1 score, were used to assess the quality of these
models. Table 9 shows the results of the DL models. The training accuracy for BiLSTM was
97.84%, while the test accuracy was 96.40%. With a sensitivity of 91.67% and a specificity
of 98.58 percent, it showed a healthy middle ground. The overall classification ability
was measured by an AUC score of 96.44% and an F1 score of 94.14%, which considered
both precision and recall. CNN-LSTM scored 96.82% on the accuracy test, which was
slightly higher than BiLSTM’s score of 96.80%. Its specificity remained high, at 98.58%,
while its sensitivity increased to 93.1%. In spite of a slight drop in AUC (96.17%), the F1
score improved to 94.86%. The test results showed that GRU, similar to CNN-LSTM, had a
sensitivity of 93.02% and a specificity of 98.58%. However, it improved upon the previous
version’s AUC score of 96.57% and F1 score of 94.86%.
Training Test
Model Sensitivity Specificity F1 Score AUC
Accuracy Accuracy
Name (%) (%) (%) Score (%)
(%) (%)
BiLSTM 97.84 96.40 91.67 98.58 94.14 96.44
CNN-
97.82 96.82 93.02 98.58 94.86 96.17
LSTM
GRU 98.07 96.82 93.02 98.58 94.86 96.57
LSTM 98.04 97.03 93.34 98.72 95.19 96.35
When compared to other models, LSTM achieved the best results. Its test accuracy
was 97.03%, which was nearly as high as its 98.04% training accuracy. LSTM also had the
highest sensitivity (93.34%) and specificity (98.72%) of all the models, indicating that it was
the best at making the right positive and negative identifications. It performed admirably
across the board, with an F1 score of 95.19% and an AUC of 96.35%. Figure 9 shows a
comparison of the performance of the models.
The models’ performance on the task was very high. However, LSTM excelled above
all other models in terms of accuracy, sensitivity, specificity, F1 score, and AUC.
The LSTM model trained for 20 epochs and early stopping at 8 epochs. The perfor-
mance of the model in training accuracy was 98.04%, and the testing accuracy was 97.03%,
as shown in Figure 10a,b. The model achieved a sensitivity of 93.34%, a specificity of
98.72%, and an F1 score of 95.19%. Additionally, the model achieved an AUC of 96.35%.
was 97.03%, which was nearly as high as its 98.04% training accuracy. LSTM also had the
highest sensitivity (93.34%) and specificity (98.72%) of all the models, indicating that it
was the best at making the right positive and negative identifications. It performed admi-
rably across the board, with an F1 score of 95.19% and an AUC of 96.35%. Figure 9 shows
a comparison of the performance of the models.
Computers 2023, 12, 170 17 of 24
The models’ performance on the task was very high. However, LSTM excelled above
all other models in terms of accuracy, sensitivity, specificity, F1 score, and AUC.
F1 Score (%)
Specificity (%)
Sensitivity (%)
88 90 92 94 96 98 100
Figure 9.
Figure 9. Comparison
Comparison of
of performance
performance models.
models.
The LSTM model trained for 20 epochs and early stopping at 8 epochs. The perfor-
mance of the model in training accuracy was 98.04%, and the testing accuracy was 97.03%,
as shown in Figure 10a,b. The model achieved a sensitivity of 93.34%, a specificity of
98.72%, and an F1 score of 95.19%. Additionally, the model achieved an AUC of 96.35%.
Figure
Figure 10.Training
10. Trainingplots
plotsand
andtesting
testing accuracy
accuracy and
and loss
loss for
forLSTM
LSTMmodels:
models:(a)(a)accuracy (b)(b)
accuracy loss.
loss.
TheGRU
The GRU model
model trained
trained forfor
2020 epochs
epochs and
and early
early stopping
stopping at 7atepochs.
7 epochs.
TheThe perfor-
performance
mance of the model in training accuracy was 98.07%, and the testing accuracy
of the model in training accuracy was 98.07%, and the testing accuracy was 96.82%, as was 96.82%,
as shown
shown in Figure
in Figure 11a,b.11a,b. The model
The model achieved
achieved a sensitivity
a sensitivity of 93.2%,
of 93.2%, a specificity
a specificity of
of 98.58%,
98.58%, and an F1 score of 94.86%. Additionally, the model achieved
and an F1 score of 94.86%. Additionally, the model achieved an AUC of 96.57%. an AUC of 96.57%.
The BiLSTM model was trained for 20 epochs and early stopping at 12 epochs. The
performance of the model in training accuracy was 97.84%, and the testing accuracy was
96.40%, as shown in Figure 12a,b. The model achieved a sensitivity of 91.67%, a specificity
of 98.58%, and an F1 score of 94.14%. Additionally, the model achieved an AUC of 96.44%.
The BiLSTM model was trained for 20 epochs and early stopping at 12 epochs. The
performance of the model in training accuracy was 97.82%, and the testing accuracy was
96.82%, as shown in Figure 12a,b. The model achieved a sensitivity of 93.02%, a specificity
of 98.58%, and an F1 score of 94.86%. Additionally, the model achieved an AUC of 96.17%.
This study’s customer satisfaction level findings help improve services and retain
regular clients. This research detailed the models’ sensitivity, specificity, and positive and
negative predictive values, as described in Figure 13. With only 35 FPs and 84 FNs, LSTM
achieved 2704 true positives and a 1177 TN accuracy ratio. Among the 2700 positive results
and 1173 negative results that GRU found, there were only 39 FPs and 88 FNs. Exactly the
Figure 11. Training plots and testing accuracy and loss for GRU models: (a) accuracy (b) loss.
Computers 2023, 12, 170 Figure 10. Training plots and testing accuracy and loss for LSTM models: (a) accuracy (b) loss. 18 of 24
The GRU model trained for 20 epochs and early stopping at 7 epochs. The perfor-
mance of the model in training accuracy was 98.07%, and the testing accuracy was 96.82%,
same
as numbers
shown in of true 11a,b.
Figure positives
The (2700), FPs (39), aFNs
model achieved (88), and
sensitivity of true negatives
93.2%, (1177)
a specificity of were
generated
98.58%, by
andboth
an F1BiLSTM and CNN-LSTM.
score of 94.86%. Additionally, the model achieved an AUC of 96.57%.
Figure 12. Training plots and testing accuracy and loss for BiLSTM models: (a) accuracy (b) loss.
This study’s customer satisfaction level findings help improve services and retain
regular clients. This research detailed the models’ sensitivity, specificity, and positive and
negative predictive values, as described in Figure 13. With only 35 FPs and 84 FNs, LSTM
achieved 2704 true positives and a 1177 TN accuracy ratio. Among the 2700 positive re-
sults and 1173 negative results that GRU found, there were only 39 FPs and 88 FNs. Ex-
actly the same numbers of true positives (2700), FPs (39), FNs (88), and true negatives
Figure 12. Training plots and testing accuracy and loss for BiLSTM models: (a) accuracy (b) loss.
Figure 12. Training
(1177) plots and
were generated testing
by both accuracy
BiLSTM and and loss for BiLSTM models: (a) accuracy (b) loss.
CNN-LSTM.
This study’s customer satisfaction level findings help improve services and retain
regular clients. This research detailed the models’ sensitivity, specificity, and positive and
negative predictive values, as described in Figure 13. With only 35 FPs and 84 FNs, LSTM
achieved 2704 true positives and a 1177 TN accuracy ratio. Among the 2700 positive re-
sults and 1173 negative results that GRU found, there were only 39 FPs and 88 FNs. Ex-
actly the same numbers of true positives (2700), FPs (39), FNs (88), and true negatives
(1177) were generated by both BiLSTM and CNN-LSTM.
Figure
Figure 13. Training
13. Training plots
plots andand testingaccuracy
testing accuracy and
andloss
lossforfor
CNN-LSTM models:
CNN-LSTM (a) accuracy
models: (b) (b) loss.
(a) accuracy
loss.
Although there were some differences between the models in terms of the propor-
tions of correct predictions, incorrect predictions and FNs, all of them performed a re-
spectable job. LSTM had the highest proportion of correct positive and negative identifi-
cations, demonstrating its superior ability to detect customer satisfaction. The confusion
Computers2023,
Computers 2023,12,
12,170
x FOR PEER REVIEW 21
19 of 25
of 24
Although there were some differences between the models in terms of the proportions
of correct predictions, incorrect predictions and FNs, all of them performed a respectable
job. LSTM had the highest proportion of correct positive and negative identifications,
demonstrating its superior ability to detect customer satisfaction. The confusion metrics of
Computers 2023, 12, x FOR PEER REVIEW 21 of 25
the deep learning models is presented in Figure 14. Figure 15 shows a comparison of the
confusion metrics of DL models.
FigureFigure
14. Confusion matrix
14. Confusion matrixplots ofthe
plots of the study
study models:
models: (a) (b)
(a) LSTM LSTMGRU(b) GRU (c)
(c) BiLSTM (d)BiLSTM
CNN- (d) CNN-
Figure 14. Confusion matrix plots of the study models: (a) LSTM (b) GRU (c) BiLSTM (d) CNN-LSTM.
LSTM.LSTM.
TN
TN
FN
FN
Fp
Fp TP
Figure 15.
Figure 15. Comparison
Comparison of
of model
model performance
performance based
based on
on positive
positive and
and negative
negativefactors.
factors.
Computers 2023, 12, x FOR PEER REVIEW 22 of 25
Computers 2023, 12, 170 20 of 24
6. Discussion
6. Discussion
The phenomenon of customer churn represents a significant challenge and a top pri-
orityThe phenomenon
for major of customer
corporations. Owing to churn representsimpact
its significant a significant challenge
on corporate and apar-
revenues, top
priority for major corporations. Owing to its significant impact on corporate revenues,
ticularly within the telecommunications industry, companies are actively pursuing strat-
particularly within the telecommunications industry, companies are actively pursuing
egies to forecast potential customer churn. Hence, identifying the determinants that con-
strategies to forecast potential customer churn. Hence, identifying the determinants that
tribute to customer attrition is crucial to implementing appropriate measures aimed at
contribute to customer attrition is crucial to implementing appropriate measures aimed
mitigating this phenomenon. Our study’s primary objective was to create a churn predic-
at mitigating this phenomenon. Our study’s primary objective was to create a churn
tion model that can aid telecommunication operators in identifying customers who are at
prediction model that can aid telecommunication operators in identifying customers who
a higher risk of churning.
are at a higher risk of churning.
This paper used Arabic tweets from Saudi telecommunications companies. The new
This paper used Arabic tweets from Saudi telecommunications companies. The new
restrictions on Twitter prevent data collection from tweets using the Python scripter. The
restrictions on Twitter prevent data collection from tweets using the Python scripter. The
restrictions, which were put in place in January 2023, limit the number of tweets a single
restrictions, which were put in place in January 2023, limit the number of tweets a single
user or application can collect in a given period. This makes it more difficult to collect
user or application can collect in a given period. This makes it more difficult to collect large
large datasets
datasets of tweets,
of tweets, whichwhich is often
is often necessary
necessary for mining
for data data mining and other
and other research
research pur-
purposes.
poses. This study compared four models for predicting customer satisfaction.
This study compared four models for predicting customer satisfaction. Models such as Models
such asGRU,
LSTM, LSTM, GRU, BiLSTM,
BiLSTM, and CNN-LSTM
and CNN-LSTM were tested. were
Thetested.
researchThe researchthe
confirmed confirmed the
significance
significance
of customers’ ofuse
customers’
of socialuse of social
media media
to share theirtoexperiences,
share their experiences,
both good and both good
bad, withanda
bad, with a company’s services or products. Figure 16 shows the ROC of
company’s services or products. Figure 16 shows the ROC of the deep learning models. the deep learning
models.
The The problem
problem wasbysolved
was solved by and
creating creating and DL
training training
methodsDL methods on the open-source
on the open-source AraCust
AraCust dataset. The LSTM model stood out because it had
dataset. The LSTM model stood out because it had the highest training andthe highest training and test
test accuracy
accuracy for text classification: 98.04% and 97.03%,
for text classification: 98.04% and 97.03%, respectively. respectively.
(a) (b)
(c) (d)
Figure 16.
Figure 16. Training plotsand
Trainingplots andtesting accuracyand
testing accuracy and loss
loss of the
of the study
study models:(a)
models: (a)LSTM
LSTM (b)
(b) GRU
GRU
(c) BiLSTM (d) CNN-LSTM.
(c) BiLSTM (d) CNN-LSTM.
Computers 2023, 12, 170 21 of 24
The comparison results of proposed deep learning models and existing models for
sentiment analysis for Arabic customer satisfaction on the racist dataset are presented
in Table 10. This is related to the telecommunication sectors of Saudi Arabia. Almuqren
et al. [49] roposed two models: Bi-GRU and LSTM. The BiG RU model achieved an accuracy
of 95.16%, while the LSTM model achieved 94.66% accuracy. Aftan and Shah [50] proposed
three other models: RNN, CNN, and AraBERT. The AraBERT model achieved 94.33%
accuracy, the RNN model achieved an accuracy of 91.35%, and the CNN model achieved
88.34% accuracy. Almuqren et al. [46] proposed a SentiChurn model and obtained an
accuracy of 95.8%. In this study, we proposed several DL models; the best accuracy result
achieved by an LSTM model was 97.03%, and it also achieved the highest accuracy among
the existing studies.
Table 10. Comparison of various proposed DL models and existing research papers.
7. Conclusions
The significance of conducting research in the telecommunications industry lies in
its potential to enhance the interaction between users and technologies and, therefore, to
improve companies’ profitability. It is widely acknowledged that the ability to forecast
customer churn is a critical revenue stream for telecommunications enterprises. Therefore,
the objective of this study was to construct a predictive system for customer churn in Saudi
Arabian telecommunications companies. This study used DL and sentiment analysis to
make important decisions about increasing customer loyalty and satisfaction. This research
can help the telecommunications industry better serve its customers and address their
concerns as social media continues to shape public opinion. This study used sentiment
analysis to assess customer satisfaction with STC, Mobily, and Zain services, and to inform
business decisions. The study confirmed social media’s value as a platform for consumers
to share their positive and negative experiences with a company’s products or services.
Communication is vital to Saudi life and, so, online discussions are inevitable. In this study,
sophisticated DL models were trained on the available online dataset AraCust, which
was collected from Arabic tweets. The proposed models in this study were LSTM, GRU,
BiLSTM, and CNN-LSTM. The LSTM model had the highest training (98.04%) and test
accuracy (97.03%) in text classification. The model’s superior sensitivity to identifying
customer satisfaction showed its potential to help telecommunications providers reduce
customer churn caused by dissatisfaction with their offerings. The researcher aimed to
enhance their existing research model by incorporating sophisticated DL techniques, such
as transform models and time series models, to enhance its precision.
This research paper provided a substantial contribution to the domain of customer
satisfaction analysis in the Arabic language. It is a crucial area of investigation, given
the population of Arabic speakers in the world. The study effectively showed the ability
of different deep learning models to accurately predict customer satisfaction through
analysing Arabic tweets. This study highlighted the importance of social media platforms
as valuable mediums through which customers can share their experiences, which helps
business owners improve service quality and maintain customer loyalty.
Computers 2023, 12, 170 22 of 24
Funding: This research was funded by the Deputyship for Research and Innovation, Ministry of
Education in Saudi Arabia, grant number 523.
Data Availability Statement: The datasets are available at the following link: DOI: https://doi.org/
10.7717/peerj-cs.510/supp-2, accessed on 15 May 2023.
Conflicts of Interest: The author declares no conflict of interest.
References
1. Gerpott, T.J.; Rams, W.; Schindler, A. Customer retention, loyalty, and satisfaction in the German mobile cellular telecommunica-
tions market. Telecommun. Policy 2001, 25, 249–269. [CrossRef]
2. Wei, C.P.; Chiu, I.T. Turning telecommunications call details to churn prediction: A data mining approach. Expert Syst. Appl. 2002,
23, 103–112. [CrossRef]
3. Qureshii, S.A.; Rehman, A.S.; Qamar, A.M.; Kamal, A.; Rehman, A. Telecommunication subscribers’ churn prediction model
using machine learning. In Proceedings of the Eighth International Conference on Digital Information Management, Islamabad,
Pakistan, 10–12 September 2013; pp. 131–136.
4. Ascarza, E.; Iyengar, R.; Schleicher, M. The perils of proactive churn prevention using plan recommendations: Evidence from a
field experiment. J. Mark. Res. 2016, 53, 46–60. [CrossRef]
5. Adwan, O.; Faris, H.; Jaradat, K.; Harfoushi, O.; Ghatasheh, N. Predicting customer churn in telecom industry using multilayer
preceptron neural networks: Modelling and analysis. Life Sci. J. 2014, 11, 75–81.
6. Afful-Dadzie, E.; Nabareseh, S.; Oplatková, Z.K.; Klímek, P. Enterprise competitive analysis and consumer sentiments on social
media: Insights from telecommunication companies. In Proceedings of the 3rd International Conference on Data Management
Technologies and Applications (DATA 2014), Vienna, Austria, 29–31 August 2014; pp. 22–32.
7. Aghakhani, N.; Asllani, A. A Text-mining approach to evaluate the importance of information systems research themes. Commun.
IIMA 2020, 18, 3. Available online: https://scholarworks.lib.csusb.edu/ciima/vol18/iss1/3/?utm_source=scholarworks.lib.
csusb.edu%2Fciima%2Fvol18%2Fiss1%2F3&utm_medium=PDF&utm_campaign=PDFCoverPages (accessed on 2 June 2023).
[CrossRef]
8. Alalwan, A.A.; Rana, N.P.; Dwivedi, Y.K.; Algharabat, R. Social media in marketing: A review and analysis of the existing
literature. Telemat. Inform. 2017, 34, 1177–1190. [CrossRef]
9. El Rahman, S.A.; Alotaibi, F.A.; Alshehri, W.A. Sentiment analysis of Twitter data. In Proceedings of the 2019 International
Conference on Computer and Information Sciences (ICCIS 2019), Aljouf, Saudi Arabia, 3–4 April 2019.
10. Pavaloaia, V.D.; Teodor, E.M.; Fotache, D.; Danileţ, M. Opinion mining on social media data: Sentiment analysis of user preferences.
Sustainability 2019, 11, 4459. [CrossRef]
11. Aldhyani, T.H.H.; Alsubari, S.N.; Alshebami, A.S.; Alkahtani, H.; Ahmed, Z.A.T. Detecting and Analyzing Suicidal Ideation on
Social Media Using Deep Learning and Machine Learning Models. Int. J. Environ. Res. Public Health 2022, 19, 12635. [CrossRef]
12. Susanti, C.E. The effect of product quality and service quality towards customer satisfaction and customer loyalty in traditional
restaurants in East Java. In Proceedings of the International Conference on Managing the Asian Century, Singapore, 11–13 July
2013; Springer: Singapore, 2013; pp. 383–393.
13. Abiodun, R. Development of mathematical models for predicting customers satisfaction in the banking system with a queuing
model using regression method. Am. J. Oper. Manag. Inf. Syst. 2017, 2, 86–91.
14. Mugion, R.G.; Musella, F. Customer satisfaction and statistical techniques for the implementation of benchmarking in the public
sector. Total Qual. Manag. Bus. Excell. 2013, 24, 619–640. [CrossRef]
15. Al-Ghamdi, S.M.; Sohail, M.S.; Al-Khaldi, A. Measuring consumer satisfaction with consumer protection agencies: Some insights
from Saudi Arabia. J. Consum. Mark. 2007, 24, 71–79. [CrossRef]
16. The Communication and Information Technology Commission. Annual Report of (CITC). Available online: https://www.cst.gov.
sa/en/mediacenter/reports/Documents/PR_REP_013Eng.pdf (accessed on 2 June 2023).
17. Hassounah, M.; Raheel, H.; Alhefzi, M. Digital response during the COVID-19 pandemic in Saudi Arabia. J. Med. Internet Res.
2020, 22, e19338. [CrossRef]
18. Digital 2019 Saudi Arabia. Available online: https://www.slideshare.net/DataReportal/digital-2019-saudi-arabia-january-2019
-v01 (accessed on 2 June 2023).
19. Brandusoiu, I.; Toderean, G.; Ha, B. Methods for churn prediction in the prepaid mobile telecommunications industry. In
Proceedings of the International Conference on Communications, Kuala Lumpur, Malaysia, 22–27 May 2016; pp. 97–100.
20. He, Y.; He, Z.; Zhang, D. A study on prediction of customer churn in fixed communication network based on data mining. In
Proceedings of the Sixth International Conference on Fuzzy Systems and Knowledge Discovery, Tianjin, China, 14–16 August
2009; Volume 1, pp. 92–94.
Computers 2023, 12, 170 23 of 24
21. Idris, A.; Khan, A.; Lee, Y.S. Genetic programming and AdaBoosting based churn prediction for telecom. In Proceedings of
the IEEE International Conference on Systems, Man, and Cybernetics (SMC), Seoul, Republic of Korea, 14–17 October 2012;
pp. 1328–1332.
22. Huang, F.; Zhu, M.; Yuan, K.; Deng, E.O. Telco churn prediction with big data. In Proceedings of the ACM SIGMOD International
Conference on Management of Data, Melbourne, Australia, 31 May–4 June 2015; pp. 607–618.
23. Makhtar, M.; Nafis, S.; Mohamed, M.; Awang, M.; Rahman, M.; Deris, M. Churn classification model for local telecommunication
company based on rough set theory. J. Fundam. Appl. Sci. 2017, 9, 854–868. [CrossRef]
24. Amin, A.; Anwar, S.; Adnan, A.; Nawaz, M.; Howard, N.; Qadir, J.; Hawalah, A.; Hussain, A. Comparing oversampling techniques
to handle the class imbalance problem: A customer churn prediction case study. IEEE Access 2016, 4, 7940–7957. [CrossRef]
25. Tul, Q.; Ali, M.; Riaz, A.; Noureen, A.; Kamranz, M.; Hayat, B.; Rehman, A. Sentiment analysis using deep learning techniques: A
review. Int. J. Adv. Comput. Sci. Appl. 2017, 8, 424–433. [CrossRef]
26. Pang, B.; Lee, L. Opinion mining and sentiment analysis. In Foundations and Trends in Information Retrieval; Alet Heezemans:
Rotterdam, The Nederland, 2008; Volume 2, pp. 1–135.
27. Vieira, S.T.; Rosa, R.L.; Rodríguez, D.Z.; Ramírez, M.A.; Saadi, M.; Wuttisittikulkij, L. Q-meter: Quality monitoring system for
telecommunication services based on sentiment analysis using deep learning. Sensors 2021, 21, 1880. [CrossRef]
28. Chiu, S.T.; Susanto, H.; Leu, F.Y. Detection and defense of DDoS attack and flash events by using Shannon entropy. In Innovative
Mobile and Internet Services in Ubiquitous Computing, Proceedings of the IMIS 2022, Kitakyushu, Japan, 29 June–1 July 2022; Lecture
Notes in Networks and Systems; Barolli, L., Ed.; Springer: Cham, Switzerland, 2022; Volume 496.
29. Heru, S.; Leu, F.Y.; Alifya, K.S.S. Genetics algorithm approaches of cheminformatics reengineering process. J. Biomed. Sci. 2022, 4,
1523–1530.
30. Setiana, D.; Norsarah, S.; Besar, N.; Anna, T.; Nasution, M.; Susanto, H. Technology disruption in the time of the digital ecosystem
society’s adoption: Cyber bullying phenomenon—The truth or hoax? In Handbook of Research on Big Data, Green Growth, and
Technology Disruption in Asian Companies and Societies; IGI Global: Hershey, PA, USA, 2022; pp. 238–255.
31. Alamsyah, A.; Paryasto, M.; Putra, F.J.; Himmawan, R. Network text analysis to summarize online conversations for marketing
intelligence efforts in telecommunication industry. In Proceedings of the 2016 4th International Conference on Information and
Communication Technology (ICoICT 2016), Bandung, Indonesia, 25–27 May 2016.
32. Najadat, H.; Al-Abdi, A.; Sayaheen, Y. Model-based sentiment analysis of customer satisfaction for the Jordanian telecommunica-
tion companies. In Proceedings of the 2018 9th International Conference on Information and Communication Systems (ICICS
2018), Irbid, Jordan, 3–5 April 2018; pp. 233–237.
33. Qamar, A.M.; Ahmed, S.S. Sentiment classification of Twitter data belonging to Saudi Arabian telecommunication companies. Int.
J. Adv. Comput. Sci. Appl. 2017, 8, 395–401.
34. Hang, A.; Li, B.; Wang, W.; Wan, S.; Chen, W. MII: A Novel Text Classification Model Combining Deep Active Learning with
BERT. Comput. Mater. Contin. 2020, 63, 1499–1514.
35. Gabhane, M.D.; Suriya, D.S.B.A. Churn Prediction in Telecommunication Business using CNN and ANN. J. Posit. Sch. Psychol.
2022, 6, 4672–4680.
36. DiPietro, R.; Hager, G.D. Deep learning: RNNs and LSTM. In Handbook of Medical Image Computing and Computer Assisted
Intervention; Zhou, S.K., Rueckert, D., Fichtinger, C.A.I., Eds.; The Elsevier and MICCAI Society Book Series; Academic Press:
Cambridge, MA, USA, 2020; pp. 503–519.
37. Almuqren, L.; Cristea, A. AraCust: A Saudi Telecom Tweets corpus for sentiment analysis. PeerJ Comput. Sci. 2020, 7, e510.
[CrossRef] [PubMed]
38. Hathlian, N.F.B.; Hafez, A.M. Subjective text mining for Arabic social media. In Cognitive Analytics: Concepts, Methodologies, Tools,
and Applications; IGI Global: Hershey, PA, USA, 2020; pp. 1483–1495.
39. Sun, W.; Cai, Z.; Li, Y.; Liu, F.; Fang, S.; Wang, G. Data processing and text mining technologies on electronic medical records: A
review. J. Healthc. Eng. 2018, 2018, 4302425. [CrossRef]
40. Webster, J.J.; Kit, C. Tokenization as the initial phase in NLP. In Proceedings of the 14th International Conference on Computational
Linguistics (COLING 1992), Nantes, France, 23–28 August 1992; Volume 4.
41. Barabas, P.; Kovacs, L. Efficient encoding of inflection rules in NLP systems. Acta Marisiensis Ser. Technol. 2012, 9, 11.
42. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [CrossRef]
43. Cambria, E.; White, B. Jumping NLP curves: A review of natural language processing research. IEEE Comput. Intell. Mag. 2014, 9,
48–57. [CrossRef]
44. Medsker, L.R.; Jain, L.C. Recurrent neural networks. Des. Appl. 2001, 5, 64–67.
45. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [CrossRef] [PubMed]
46. Dey, R.; Salem, F.M. Gate-variants of gated recurrent unit (GRU) neural networks. In Proceedings of the 2017 IEEE 60th
International Midwest Symposium on Circuits and Systems (MWSCAS), Boston, MA, USA, 6–9 August 2017; pp. 1597–1600.
47. Huang, Z.; Xu, W.; Yu, K. Bidirectional LSTM-CRF models for sequence tagging. arXiv 2015, arXiv:1508.01991.
48. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in
convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [CrossRef]
Computers 2023, 12, 170 24 of 24
49. Almuqren, L.; Alrayes, F.S.; Cristea, A.I. An empirical study on customer churn behaviours prediction using Arabic twitter
mining approach. Future Internet 2021, 13, 175. [CrossRef]
50. Aftan, S.; Shah, H. Using the AraBERT model for customer satisfaction classification of telecom sectors in Saudi Arabia. Brain Sci.
2023, 13, 147. [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.