published article
published article
Abstract—Sign language is a communication way that Various works were introduced under the umbrella of
enables deaf and dump people to communicate with others. sign language translation. In the work [3], the authors
Different communication systems are developed to cover presented gestures translator system composed of 4 modules.
various languages including Arabic language. The goal of this The modules were sensing unit, processing unit, voice
study was to design a low cost smart glove system that storage unit, and wireless communication unit. Through a set
recognizes hands gestures in Arabic sign language. Through a of flex sensors, the gestures were processed by PIC16F877A.
set of flex sensors and a tilting sensing module for right and left Based on a vision-based technique, continuous Arabic Sign
hands, gestures were converted into a vocalized speech. Language (ArSL) recognition was introduced by Assaleh et
Further, by developing Android application “Smart Gloves” a
al [1]. The system was developed using spatio-temporal
text speech was appeared. It facilitates user interface with the
feature extraction and Hidden Markov Models (HMM). The
proposed system. Indeed, Bluetooth module was used for
wireless communication between the system and a mobile average recognition rate of words was 94%. Another
phone. The glove system has been designed for words level and approach was adopted in [5] to recognize continuous ArSL
sentences level as well. Results point to 90% of recognition using two DG5-V hand data gloves. A digital camera was
rate. The system has proven its ability to translate hands utilized to synchronize hand movement with words. Further,
gestures into Arabic Sign Language in spoken and written words classification has been carried out by a Modified K-
ways. Nearest Neighbor (MKNN) approach. This approach
achieved sentences recognition rate of 98.9%. In a relevant
Keywords—Arabic sign language; glove; Hands Gestures; study, isolated Arabic words were translated into sign
Communication; Bluetooth. language [7]. The vision-based technique was utilized by two
modes. The first one has been used to translate from deaf
I. INTRODUCTION and dump people to the vocal people and the second one was
vice versa.
One of the most significant challenges that face deaf and
dump people is to communicate with others. Different The aim of this work was to develop a low-cost smart
techniques are developed to facilitate communication with glove system to translate hands gestures into a vocalized
those people using sign language. In general, the utilized speech. Moreover, a smart phone android application was
techniques are classified to be either vision-based or glove- developed to translate gesture as a text speech. It facilitates
based [1, 2]. Through vision-based methodology, gestures user interface with the glove system. Thus, a low cost
are captured by a camera and then images features are solution is presented for deaf and dumb individuals to enable
extracted to interpret gestures meaning [3, 4]. In glove-based them for communication with normal people in a consistent
technique, the user wears a glove and through a set of way.
sensors, gestures are translated into audio or text speech [5,
The article is formed as follows. Materials and methods
6].
are described and discussed in section II. Testing the system
Although image-based techniques are more realistic, they and results are presented in section III. Finally, section IV
require complex processes such as features extraction and concludes this work and the future work is covered.
images processing. Moreover, the background and the
position of the camera are crucial for this technique [2]. II. MATERIALS AND METHODS
Unlikely, glove-based technique is characterized by a less
computational complexity, fast response, and portability [2]. The system is designed to translate hands gestures into
In general, both techniques are utilized to interpret the sign ArSL. A glove-based technique was chosen for this purpose.
language in forms of isolated words or continuous words [1, In implementation, a hardware module and a software
7, 8]. module were integrated to form the proposed system. The
hardware module consists of a pair of tailored gloves,
microcontroller (Arduino MEGA), Bluetooth module, MP3
Vo = Vi ( ) (1)
Fig.1. Block diagram of the proposed system TABLE I. STANDARD MEASUREMNTS OF FINGERS BENDING AT 3 DIFFERENT
ANGLES
Finger F1 F2 F3 F4 F5
1) The glove : Tailored pair of gloves were designed for 0 degree 295 317 390 360 204
right and left hands using 5 flex sensors and a tilting sensing
45 degree 210 235 275 210 160
module each. They were embedded togther on a nylon
material-based glove as shown in Fig.2. Each sensor was 90 degree 155 165 200 145 110
fixed on thumb, index, middle, ring, and pinky fingers
respectively. In addition a gyroscope was mounted on the b) Gyroscope: Arabic words have their own sign
glove to indicate orientation of the hand palm. language. In fact, some words have similar finger bending
but they are different in hand motion. Thus, these words
require differention in fingers bending orientation and hand
palm motion. Therefore, a trible-axis micro electro
mechanical system (MEMS) module MPU-6050 is used. It
contains a gyroscope to distinguish hand orientation in X ,Y,
and Z directions and accelerometer to detect hand palm
motion. It provides a wide range of features including 16-bit
ADCs.
2) The microcontroller : Arduino Mega 2560 has been
selected to be a microcontroller. It features with a low
power consumption as well as low cost[10]. It has 16
analog input pins, 54 digital input/output pins, 4 UARTs,
and USB connection.
3) HC-05 Bluetooth: A wireless communication was
carried out by using HC-05 Bluetooth module. It allows
Fig.2. Designed glove model
sending data from the microcontroller to a mobile phone or a
304
computer. In our case, the data was sent from Arduino board shown in Fig. 5, while the simulation design of one glove
to a mobile phone asynchronously. (the master) is illustrated in Fig. 6.
4) Arduino MP3 shield: Arduino MP3 shield is an
onboard 3w/channel stereo audio amplifier [11]. It is
employed to derive a loudspeaker to come out with the sound
of translated gestures. In this application, a 3W/4ȍ
loudspeaker was used.
B. Software Module
System development as shown in Fig.4 starts with hands
gestures that are converted by flex sensors to analog signals.
The analog signals are converted to digital signals and
processed by the microcontroller. Each word/sentence has
its gesture; thereby each one has its own serial measurements
[12]. By recording these measurements and comparing them
with coded template of serial values, words are recognized
and vocalized; otherwise the sequence is repeated. The code
is written by Arduino C language as it supported by Arduino. Fig.5. Overall proposed system containing the gloves, the
Real challenge of this study was writing the code to read microcontroller, and the power supply units
Arabic words. In general, Arduino software is written only
for English; therefore a new code was developed to
recognize and vocalize Arabic words.
305
TABLE II. SAMPLES OF TRANSLATED GESTURES IN FORMS OF WORDS AND
EXPRESSIONS
No Word No Expression
1 ϲΑ (My father) 1 ϢϜϋΎϤΘγ ϦδΤϟ ήϜη (Thanks for
your attention)
2 ϰϣ (My mother) 2 ϯΩήϔϤΑ ΔγέΪϤϟ ϰϟ· ΐϫΫ (Go to the
school alone)
3 ήμϣ (Egypt) 3 ϰΑ ΐΣ Ύϧ (I love my father)
4 ήϜη ( Thank you) 4 ϰϣ ΐΣ Ύϧ (I love my mother)
5 ΔγέΪϣ (School) 5 ˯ΎϤδϟ ϰϓ ΔόρΎγ βϤθϟ (The sun is
shining in the sky)
6 ϥΎϋϮΟ (Hungry) 6 ϡΎότϟ Ϊϳέϭ ϥΎϋϮΟ Ύϧ (I am hungry
and I need food)
7 ϥΎθτϋ (Thirsty) 7 ΏΎΘϜϟ ήϗ (I read the book)
8 ϒΎΧ (Afraid) 8 ϡΪϘϟ Γήϛ ΐόϟ (I play football)
9 ϞϤόϟ (Work) 9 ϰϨϤϴϟ ϯΪϴΑ ΐΘϛ (I write with my
right hand)
10 ΐϫΫ (Go) 10 ΐόΘϟΎΑ ήόη (I am tired)
11 ϢγϹ (Name) 11 ϡϼψϟ Ϧϣ ϒΎΧ Ύϧ ( I afraid of the
dark)
12 ΐΣ (Love) 12 ϞϤόϟ ϥΎϘΗ· ΐΠϳ (You must master
the work)
Fig.7 A signal of vocalized word “Egypt” in Arabic language 13 ϡΎότϟ (Food) 13 ωέΎθϟ ήΒϋ (Cross the street)
14 ˯ΎϤδϟ (Sky) 14 ϢϠϘϟ Ϊϳέ (I need the pen)
15 βϤθϟ (Sun) 15 ΔγέΪϟ ϲϓ ϰϧΪϋΎδϳ ϰΧ (My brother
helps me to study)
16 ωέΎθϟ (Street)
17 ΐΘϛ (Write)
18 ήϗ (Read)
19 Ϊϴϟ (Hand)
20 ϢϠϘϟ (Pen)
306
ACKNOWLEDGMENT
Sincere thanks and appreciation go to Eng. Rabie Gouda,
Eng. Abdullah Taha, and Eng. Ahmed Bahaa for their
assistance in writing the code and testing of the system.
REFERENCES
[1] K. Assaleh, T. Shanableh, M. Fanaswala, F. Amin, and H. Bajaj,
“Continuous Arabic sign language recognition in user dependent
mode,” Journal of Intelligent Learning Systems and Applications, vol.
2, pp. 19-27, 2010.
[2] P. Lokhande, R. Prajapati, and S. Pansare, “Data gloves for sign
language recognition system,” in Proc. of National Conference on
Emerging Trends in Advanced Communication Technologies, June
2015, pp. 1-4.
[3] K. Gunasekaran and R. Manikandan, “Sign language to speech
translation system using PIC microcontroller,” International Journal
of Engineering and Technology. vol. 5, no. 4, pp. 1024-1028, 2013.
[4] K.V. Fale, A. Phalke, P. Chaudhari, and P. Jadhav, “Smart glove:
gesture vocalizer for deaf and dump people,” International Journal of
Innovative Research in Computer and Communication Engineering,
vol.4 no.4, pp. 6800-6806, 2016.
[5] N. Tubaiz, T. Shanableh, and K. Assaleh, “Glove-based continuous
Arabic sign language recognition in user dependent mode,” IEEE
Transactions on Human Machine Ssystems, vol. 45, no. 4, pp. 526-
533, 2015.
[6] D. J. Sturman and D. Zeltzer, “A survey of glove-based input,” IEEE
Fig.9. Screenshot of Smart Gloves application
Computer Graphics and Applications, vol. 14, no. 1, pp. 30-39, 1994.
[7] N.B. Ibrahim, M.M. Selim, and H.H. Zayed, “An automatic Arabic
sign language recognition system,” Journal of King Saud Uninversity-
Computer and Information Sciences, vol. 30, no. 4, pp.470-477, 2018.
IV. CONCLUSION [8] Suharjito, R. Anderson, F. Wiryana, M. Chandra, and G. P. Kusuma,
A smart glove-based system is designed to reduce the “Sign language recognition application systems for deaf-mute people:
a review based on input-process-output,” in Proc. of 2nd International
communication gap between the deaf/dump people and the Conference on Computer Science and Computational Intelligence,
normal people. It translates hands gestures into ArSL in the Bali, Indonesia, October 2017, pp. 441-448.
forms of words and expressions as well. The outcome is a [9] M. I. Sadek, M. N. Michael, and H.A. Mansour, “Glove-based
vocalized speech. Moreover, Android application “Smart continuous Arabic sign language recognition in user dependent
Gloves” is developed to translate gestures into a text speech. mode,” in Proc. of 34th National Radio Science Conference, Port Said,
In fact, the system is characterized by ease of interface, Egypt, March 2017, pp. 380-388.
portability, no prior training, and low costs. [10] V. Oza and P. Mehta, “Arduino robotic hand: survey paper,” in Proc.
of International Conference on Smart City and Emerging Technology,
The accuracy of the system is relatively high. It depends Mumbai, India, January 2018, pp. 1-5.
on the physical characteristics of the user. Once, the system [11] A. Nayyar and V. Puri, “A review of Arduino board's, Lilypad's &
is well tightened on the user, it acts in a good way. The Arduino shields,” in Proc. of 3rd International Conference on
Computing for Sustainable Global Development, New Delhi, India,
system can be generalized for other sign languages such as March 2016, pp. 1485-1492.
American Sign Language. In addition, the system can be
[12] A. Sengupta, T. Mallick, and A. Das, “A cost effective designand
programmed to translate more gestures for both words and implementation of Arduino based sign language interpreter ,” in Proc.
expressions levels. One limitation of this system is to adapt of Devcies for Integrated Circuits, Kalyani, India, March 2019, pp.
for more than one user in the same time. It requires 12-15.
modifying size of the board pod. In general, the proposed
system presents an efficient way for communication with
deaf/dump people by interpreting their hands gestures in
spoken and written forms.
307