Human Centered Computing PDF
Human Centered Computing PDF
Human Centered Computing PDF
Bo Hu (Eds.)
LNCS 10745
Human Centered
Computing
Third International Conference, HCC 2017
Kazan, Russia, August 7–9, 2017
Revised Selected Papers
123
www.allitebooks.com
Lecture Notes in Computer Science 10745
Commenced Publication in 1973
Founding and Former Series Editors:
Gerhard Goos, Juris Hartmanis, and Jan van Leeuwen
Editorial Board
David Hutchison
Lancaster University, Lancaster, UK
Takeo Kanade
Carnegie Mellon University, Pittsburgh, PA, USA
Josef Kittler
University of Surrey, Guildford, UK
Jon M. Kleinberg
Cornell University, Ithaca, NY, USA
Friedemann Mattern
ETH Zurich, Zurich, Switzerland
John C. Mitchell
Stanford University, Stanford, CA, USA
Moni Naor
Weizmann Institute of Science, Rehovot, Israel
C. Pandu Rangan
Indian Institute of Technology, Madras, India
Bernhard Steffen
TU Dortmund University, Dortmund, Germany
Demetri Terzopoulos
University of California, Los Angeles, CA, USA
Doug Tygar
University of California, Berkeley, CA, USA
Gerhard Weikum
Max Planck Institute for Informatics, Saarbrücken, Germany
www.allitebooks.com
More information about this series at http://www.springer.com/series/7409
www.allitebooks.com
Qiaohong Zu Bo Hu (Eds.)
•
Human Centered
Computing
Third International Conference, HCC 2017
Kazan, Russia, August 7–9, 2017
Revised Selected Papers
123
www.allitebooks.com
Editors
Qiaohong Zu Bo Hu
Wuhan University of Technology Fujitsu Laboratories of Europe Ltd.
Wuhan Hayes
China UK
LNCS Sublibrary: SL3 – Information Systems and Applications, incl. Internet/Web, and HCI
www.allitebooks.com
Preface
In the past decade, a gradual yet steady paradigm shift was the refocusing of tech-
nologies from machine-oriented concepts, algorithms, and automats toward the max-
imization of human potentials and the fulfilment of human needs and human
well-being. Consequently, the development of computational power and computational
efficiency has been set in the context of effortless technology utilization by both
individuals and societies. With the recent advances in human-machine interfaces,
wireless and mobile network technologies, and data analytics, we finally set out on the
long journey toward making computer services truly human-centric, instead of limiting
human capacities to suit computer requirements. Instead of playing a prominent role in
the behavior of individuals and societies, machines and their very existence have
become ever more subtle: smart devices have been weaved into the everyday fabric;
analytic power has penetrated into all aspects of our daily activities from family to
work (and everything in between); and automation has begun to displace segments of
our routines. For an area as active as human-centered computing, it is not possible to
cover the entire thematic spectrum. Instead, the Human-Centered Computing
(HCC) conference aims to present a selection of examples of new approaches, methods,
and achievements that can underpin the aforementioned paradigm shift.
HCC 2017 was the third in the series, following successful events in Phnom Penh,
Cambodia (2014), and Columbo, Sri Lanka (2016). The HCC 2017 papers present a
balance between conceptual and empirical work, between design and evaluation
studies, and between theoretical and applied research.
All HCC 2017 submissions went through rigorous paper evaluation and selection
process. Each paper was peer-reviewed by the Program Committee and selected
reviewers and meta-reviewed by senior Program committee members. Based on these
recommendations, the program co-chairs made acceptance decisions and classified the
accepted papers into the following categories: full regular papers, short papers, and
position papers.
It has been another year of hard-working and selfless contribution. As the confer-
ence Organizing Committee, we are grateful to all members of the Technical Program
Committee — it was their hard work that enabled us to identify a set of high-quality
submissions reflecting the trends and interests of the paradigm shift. We would like to
extend our gratitude to the international Advisory Committee for its invaluable advice
and guidance. Finally, our special thanks also goes to the additional reviewers, student
volunteers and members of the local organization team, who are the key elements that
made HCC2017 a successful event.
www.allitebooks.com
Organization
Conference Co-chairs
Max Talanov ITIS, Kazan Federal University, Russia
Yong Tang South China Normal University, China
Organizing Committee
Chengzhou Fu South China Normal University, China
Vlada Kugurakova Kazan Federal University, Russia
Qiaohong Zu Wuhan University of Technology, China
Publication Committee
Bo Hu Fujitsu Labs of Europe, UK
Philip Moore Birmingham City University, UK
Secretariat
Jizheng Wan Coventry University, UK
www.allitebooks.com
Abstracts of Invited Keynotes
www.allitebooks.com
IoT+AI: Opportunities and Challenges
Huadong Ma
The Internet of Things (IoT) can enable the interconnection and integration of the
physical word and the cyber space, and has been widely considered as the kernel
technology for sensing the urban environments and providing smart services further. At
the same time, the rapid development of Artificial Intelligence (AI) brings many
opportunities to IoT. In this talk, we first introduce the challenges of urban sensing
networks. Combing AI theory, we discuss some researches on sensing, networking and
computing, and service in the IoT environment. Finally, we outline the prospects of IoT
development.
Biograph
Dr. Huadong Ma is a Chang Jiang Scholar Professor and Executive Dean, School of
Computer Science, Beijing University of Posts and Telecommunications (BUPT),
China. He is also Director of Beijing Key Lab of Intelligent Telecommunications
Software and Multimedia, BUPT. He is Chief Scientist of the project “Basic Research
on the Architecture of Internet of Things” supported by the National 973 Program of
China from 2010 to 2013. He received his PhD degree in Computer Science from the
Institute of Computing Technology, Chinese Academy of Science in 1995. From 1999
to 2000, he held a visiting position in the Department of Electrical Engineering and
Computer Science, The University of Michigan, Ann Arbor, USA. His current research
focuses on sensor networks and Internet of things, multimedia computing, and he has
published over 200 papers in journals (such as ACM/IEEE Transactions) or Confer-
ences (such as IEEE INFOCOM, ACM MM) and 4 books on these fields. As a
co-author, he got the best paper award in IEEE ICPADS2010 and the best student
paper award in IEEE ICME2016. He was awarded National Funds for Distinguished
Young Scientists in 2009. He serves for Chair of ACM SIGMOBILE China.
www.allitebooks.com
Machine Learning for Real-World:
Is Deep Learning Enough?
Adil Khan
The real world is unpredictable; it is full of noise and filled with novel scenarios.
Therefore, it is almost impossible to provide the machine learning models with a
complete representation of such a world at the time of training, forcing us to work with
an insufficient picture of our world. That is why one of the biggest problems that the
field of machine learning still faces today is the inability of the learned models to
generalize well to scenarios that are different from the ones seen at the time of training.
In this talk, we will explore this problem in four different areas: Human Activity
Recognition, Emotion Recognition in Text, User Authentication, and Medical Image
Analysis. We will see the results of applying deep learning models to these problems,
and discuss ways that may be used in addition to the use of such models to achieve
optimum performance.
Biograph
www.allitebooks.com
Contents
Quantified Self: Big Data Analysis Platform for Human Factor Efficacy
Evaluation of Complex System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 648
Chaoqiang Li, Wenjun Hou, Xiaoling Chen, and Hao Li
Abstract. MST clustering algorithm can detect data clusters with irregular
boundaries. For a weighted complete graph the feasible solutions to the MST
problem is non-unique. Membrane computing known for its characteristics of
distribution and maximal parallelism can properly reduce the complexity of
processing a MST of a graph. This paper combines MST clustering algorithm
and membrane computing by designing a specific P system. The designed P
system realizes the process of an improved MST clustering by collecting all
feasible solutions to the MST problem together preserving proper edges and
deleting redundant heavy edges. The improved MST clustering method effi-
ciently enhances the quality of clustering and proved to be feasible through an
instance.
1 Introduction
Clustering is the process of partitioning a set of data objects into subsets, making each
subset a cluster, such that objects in identical clusters are similar to each other, yet
dissimilar to objects in other clusters. Clustering has been widely used in many
applications such as business intelligence, image pattern recognition, Web search,
biology, and security [1].
There are many kinds of algorithms in the literature to solve clustering problems,
mainly including partitioning methods, hierarchical methods, density-based methods
and grid-based methods and so on. Among various kinds of clustering methods, the
minimum spanning tree (MST) clustering algorithm is known to be capable of
detecting clusters with irregular boundaries, because it does not assume a spherical
shaped clustering structure of the underlying data. MST constructing problems has
been investigated by researchers since 1926 which also makes it a relatively mature
algorithm. Many researches on this algorithm have promised close to linear time
complexity of construction cost [2, 3]. Standard MST clustering algorithms basically
include sorting the edges in constructed graph and removing the edges with heaviest
weight. Applying MST algorithm to solve clustering problems has been investigated,
but in practical applications, clustering results of MST algorithms are easy to be
affected by outliers [3].
2 Preliminary
2.1 P System
Membrane computing, introduced by PAUN in 1998, takes the living cell as multi-
hierarchical structural regions, which are surrounded by the so-called membranes.
There are three main investigated variants of P systems, cell-like, tissue-like, and
neural-like. Cell-like P System imitates the function and structure of the cells, and it
includes the membrane structure, rules and objects as basic elements. Cell-like
arrangements of membranes correspond to trees. And some P systems have been
proposed to solve computer science related problems, like NP problems, arithmetic
operations, matrix vector computation and image processing. P systems have also been
proved to be effective when being studied with clustering problems [6].
Membranes divide the whole system into different regions. The skin membrane is
the outermost membrane. A membrane is a basic membrane if there are no membranes
in it and a membrane is a non-elemental membrane otherwise. Rules and objects exist
in regions. Usually the objects are indicated by strings or characters. Rules are used to
process objects or membranes in corresponding region. The rules are executed
uncertainly and maximum concurrently. Cell-like P System can be further divided into
three types from according to kinds of rules: transition P system, P system with
communication rules and P system with active membranes [7]. The basic membrane
structure is shown in Fig. 1.
In general, a P system of degree m is a construct:
Y
¼ ðV; T; C; H; l; w1 ; w2 ; . . .; wm ; ðR1 ; q1 Þ; ðR2 ; q2 Þ; . . .; ðRm ; qm ÞÞ
Where:
1. V is an alphabet. Elements in it are called objects;
2. T V is the output objects;
An Improved MST Clustering Algorithm Based on Membrane Computing 3
Skin
Elementary membrane
2 5
Membranes
Regions
7 8
6
3
4
Environment
1
3. C V − T is the catalyst. These catalysts neither change their numbers nor their
kinds in rules. But rules cannot be executed without these catalysts;
4. H = {1, 2, m} is the set of membrane labels.
5. µ is a membrane structure; each membrane has its label;
6. wi ði ¼ 1; 2; mÞ is the objects in membrane i;
7. The basic rule is in the form of (u ! v), u is a string composed of objects in V and
v is a string in the form of v = v′ or v = v′d. v′ is a string over {here, out, in 2 V,
1 j m}.
8. Ri is the set of the rules in region i. qi is the precedence relation which defines the
partial order relation over Ri . High priority rule is executed prior [8].
The rules are used in maximum parallel and uncertainly in each membrane when
calculating. So space of exponential growth can be generated in linear operation steps.
This is very helpful to solve the computationally hard problems within feasible time.
The P system will halt after some steps if no more rules can be executed and these
objects in output membrane is the final result. The P system will not halt if rules
are always executed, then this calculation is invalid, and there is no result being
outputted [9].
MST related theories have been widely used for data classification in the field of
pattern recognition and image processing for about forty years. As classical algorithms
rely on either the idea of grouping data around some ‘centers’ or the idea of separating
data points using some regular geometric curve like a hyper-plane, they generally do
not work well when the boundaries of the clusters are very complex. An MST is quite
invariant to detailed geometric changes in the boundaries of clusters. As long as the
relative distances between clusters do not change significantly, the shape complexity of
a cluster has very little effect on the performance of our MST-based clustering algo-
rithms. And the process is quite simple. Remove the k−1 largest weighted edges from
the constructed minimum spanning tree, and k sub-trees obtained can just be regarded
as k clusters [10]. Two key advantages in representing a set of multi-dimensional data
as an MST are: 1. The simple structure of a tree facilitates efficient implementations of
rigorous clustering algorithms, which otherwise are highly computationally challeng-
ing; and 2. As an MST-based clustering does not depend on detailed geometric shape
of a cluster, it can overcome many of the problems faced by classical clustering
algorithms.
Representing a set of multi-dimensional data points as a simple tree structure will
lose some of the inter-data relationship. But during the process of simplifying the data
set into a MST, essential intra-data information is remained in MST tree. Each cluster
corresponds to one sub-tree, which does not overlap the representing sub-tree of any
other clusters. And through MST representation, we can convert a multi-dimensional
clustering problem to a tree partitioning problem, just like finding a particular set of tree
edges and cutting them. And finding a globally optimal solution for a combinatorial
optimization problem is often possible.
www.allitebooks.com
An Improved MST Clustering Algorithm Based on Membrane Computing 5
Where w0ij is the distance between point ai and aj . To facilitate calculation, Wnn
0
is
totally converted into corresponding integer matrix Wnn as follows:
2 3
w11 w12 . . . w1n
6 w21 w22 . . . w2n 7
Wnn ¼6
4 ...
7
... ... ... 5
wn1 wn2 . . . wnn
Accordingly, every wij here means the weight of edge eij connecting vertex ai and aj .
The first part of MST clustering process is to construct minimum spanning trees:
We set V as vertex set and E as edge set, and they both are empty sets initially; rank all
edges in ascending order; then add the shortest edge eij (with the smallest weight) into
the set E, and accordingly add the associated two vertex ai and aj into the set V; other
edges are added into set E gradually, which are supposed to be connected to one of the
existing vertices and with smallest weight in the remaining edges; after an edge is
added into E, we add the other vertex connected to it into V; while once there are two
identical vertices in V, it means that there is a circle in tree. And this edge will be
abandoned. Above steps are repeated until |V| = n and |E| = n−1(|V| is the number of
vertices in V; |E| is the number of edges in E).
For a complete graph, especially a graph with a mass of vertex, there are edges with
identical weights. Therefore, the minimum spanning tree of a given complete graph is
not unique and only if the shape of the minimum spanning tree is not unique. To find
all feasible solutions, we are supposed to choose every different edge with the same
weights every time when constructing a MST. But obviously, such enumerating pro-
cess will take too much time.
One of the most advantages of membrane computing is its parallelism. Membrane
structure is hierarchical, and computing rules in regions evolve react and communicate
synchronously. For this reason, we apply this parallelism to constructing all minimum
spanning trees of a complete graph at a time by duplicating and generating new reaction
membranes with rules. When there are more than one edges can be chosen, we can
construct new computing regions, where we add every feasible new edge into the tree.
When the process of adding edges and vertices to corresponding sets ends, the next
step is to partition the produced MST. An edge with a large weight means that the two
vertices connected by it are distinct from each other. For this reason, we choose to
delete k−1 maximum value edges from the existed set E to form k clusters. Since there
are more than one MST feasible solutions collected, for these always appearing short
edges, we can conclude that they are correct elements of sub-tree which should be
preserved in partitions. We rearrange edges which used to belong to set E but now
appear in different membranes as parts of MST solutions with their frequency of
occurrences in descending order. Then we delete redundant edges one by one. For
edges with the same value, we preferentially preserve ones with higher frequency; for
edges with same frequency, we preferentially delete ones with heavier weights. To
make sure that every point is kept in set V, when an edge is to be deleted, the existence
of two adjacent vertices must be checked. Repeat this process until |E| = N−K.
6 P. Gong and X. Liu
θ1 c11a1a2 ...an
1
λ
i0
MST feasible solutions and partition the dataset. The designed P system is shown as
Fig. 2.
Y
¼ ðO; l; M0 ; M1 ; Mi0 ; R0 ; R1 ; Ri0 ; qÞ
Where:
O ¼ fc11 a1 a2 . . .an h1 kg specifies the initial collection of objects in the P system;
l ¼ ½0 ½1 1 ½i0 i0 0 specifies the initial membrane structure of the P system;
M0 ¼ fc11 a1 a2 . . .an g specifies the initial objects in membrane 0;
M1 ¼ fh1 g specifies the initial object in membrane 1;
Mi0 ¼ fkg specifies that there is nothing in the output membrane i0 initially;
Rule in R0 :
w
r1 ¼ fcij ai aj ! ciðj þ 1Þ ai aj ½Uij ij in1 j1 i; j ng [
fciðn þ 1Þ ! cði þ 1Þði þ 2Þ j1 i; j ng [ fcnðn þ 1Þ ! kg
Rules in R1 :
w w 1
r2 ¼ fh1 Uij ij ! h1 Uij ij j1 i; j ng
Rules in Ri0 :
t1 m1 1 m2 1 t 1
r15 ¼ fgt Pm
i1 j1 Pi2 j2 Pit jt ! g
1 m2 mt
Pi1 j1 Pi2 j2 Pm
it jt j
t 2 N þ ; mi 2 N þ ; m1 m2 mt ; 1 i; j ng
ðm Þ0 ðm Þ0 ðm Þ0 wi wi wi
r16 ¼ fgtmn Pi1 j11 Pi2 j22 Pin1n1
jn1 ! Pi1 j1 Pi2 j2 Pin1 jn1 j
11 22 n1 n1 j j j
1 t; m 2 N þ ; m1 m2 mn1 ; 1 is ; js ng
wi wi wi wi 1 wi2 j2 1 wi jn1 1
r17 ¼ fPi1 j11 1 Pi2 j22 2 . . .Pin1n1jn1 ! Pi1 j11 1 Pi2 j2 . . .Pin1n1jn1
j j n1 j j
j1 is ; js ng
Þ0 1 ðwis js Þ0 1
ðwis ðwis jsnk Þ0 1
. . .. . .Pin1 jn1 g
js
g2 Pis js Pi1 j1 1 1 Pi j 2 2
2 2
By counting the amount of objects to find the shortest edges, the sorting procedure
is combined with the tree construction. Meanwhile the parallelism of computation
reduces the time complexity of finding all feasible solutions to OðnÞ.
Rule r4 continually identifies edges which are connected to the existed points and
calculates which the minimum one is. When utis ¼ 0 appears, it means the next edge
with minimum weight is find out. Adding a minimum weight edge, two identical
objects vi may appear in same membrane. That means there is a circle in the minimum
spanning tree. Rule r5 are activated to dissolve the according uis to avoid forming a
circle. Then rule r6 generates relative objects to represent the edge and point.
If the number of current edges with minimum weight is greater than 1, rule r7
activates the current membrane to divide into two, with one containing all elements
including the new added edge but another with all elements except the new edge.
The number of object f is set as n−1, and as every edge is added, it reduces by 1.
When there is no f in membrane1, it means that n−1 edges have been added into the
graph. Object h2 generates object d 0 without the constraint of f which catalyzes other
objects than Pij to dissolve themselves. It also promote Pij conveyed into membrane i0 .
While if there are objects f but no other edges to be added, rule r12 r13 r14 dissolve the
current membrane and other objects.
The second computing stage of this structure happens in membrane i0 . Rule r8
r11 sends the chosen edges for a minimum spanning tree into membrane i0 . In every
membrane1, executing rule r8 r11 , there is a feasible solution to the minimum
spanning tree. As a consequence, there will be more than n–1 edges existing in
membrane sent to i0 , including identical edges. Rule r15 r19 is executed in membrane
i0 . The number of object g indicates the number of edges in membrane i0 . We choose
the edge with maximum weights and highest frequency. Rule r18 executes until the
cardinal of g2 reducing to n−k (n−1−(k−1) = n−k), which means there are only n−k
different edges left. At this time, Rule r19 finds out the final n−k edges and preserves
them in the result.
4 Instance Analysis
In order to verify whether the designed P system has better clustering effect, we
introduce a simple example in this section. In this test, 10 points need to be divided into
3 clusters, and data-points in same clusters are similar to each other but dissimilar to
data-points in other clusters. The example points are shown as Fig. 3.
The modified similarity matrix of this set is as below.
2 3
0 1 1 8 10 13 10 10 17 25
6 1 0 2 5 5 10 9 13 20 26 7
6 7
6 1 2 0 5 9 8 5 5 10 16 7
6 7
6 8 5 5 0 2 1 2 10 13 13 7
6 7
6 10 5 9 2 0 5 8 20 25 25 7
W1;10 ¼6
6 13
7
6 10 8 1 5 0 1 9 10 8 7
7
6 10 9 5 2 8 1 0 4 5 5 7
6 7
6 10 13 5 10 20 9 4 0 1 5 7
6 7
4 17 20 10 13 25 10 5 1 0 2 5
25 26 16 13 25 8 5 5 2 0
10 P. Gong and X. Liu
0
0 1 2 3 4 5 6 7
Fig. 3. Instance data set (a1 = (1,2), a2 = (1,3), a3 = (2,2), a4 = (3,4), a5 = (2,5), a6 = (4,4),
a7 = (4,3), a8 = (4,1), a9 = (5,1), a10 = (6,2)).
The designed P system realizes the minimum spanning tree clustering algorithm
with its maximal parallelism and distributed manner. Now we give simple analysis of
how the P system executives in this example.
w
At first, rule r1 in the environment sends objects uij ij ð1 i; j 10Þ into membrane
w
1 to represent distances between data-points. Among all these uij ij ð1 i; j 10Þ, rules
r2 r3 pick u12 u13 u46 u67 to start the construction, because they have the minimum
weight 1. Therefore there are four membrane1 in parallel to continue next constructing
steps, and each of them starts from different edges. For the membrane1 starting from
u12 , the edge with next minimum weight is e13 . Then rule r5 estimate whether adding
e13 will form a circle or not. If a circle is to form, this edge will be abandoned.
Otherwise, the edge is added. After the judgment, u13 evolves into q3 v3 x3 P13
according to rule r6 . Then the next object u23 is to be judged. Since there are v1 v2 v3
here, rule r5 dissolves u23 , and the next edges with minimum weight 5–edges
e24 e25 e34 e27 e38 continue this process. As there are more than one objects to be added,
these previous objects in the membrane is copied 5 times and membrane 1 split into 5
identical membranes according to rule r7 . For one of them, for example, the one
containing e25 , the next minimum edge to be added is e54 ; and the one containing e34 ,
e46 with weight 1 is next to be added. These procedures will repeat until no f or uij is
left. Finally there are 26 minimum spanning trees constructed as shown in Fig. 4. And
according to rule r10 , edges in these trees are sent into membrane i0 . The membrane
structure containing all feasible solutions is shown as Fig. 4.
The final partition step is executed in membrane i0 with rules r15 r19 . In this
example, we can see edges in all minimum spanning tree feasible solutions including
e12 e13 e46 e78 e89 e9;10 e67 e45 e24 e25 e34 e38 e56 e37 . They are sorted by rules in
descending order by their cardinal and retained the first 9(n−1) edges only by rule r15
and r16 . Rule r17 r18 and r19 pick the longest 2(k−1) edges – e78 e24 and delete them by
their weights. Finally edges e12 e13 e46 e89 e9;10 e67 e45 are reserved in the graph. The
clustering result is fa1 a2 a3 g,fa4 a5 a6 a7 g,fa8 a9 a10 g.
An Improved MST Clustering Algorithm Based on Membrane Computing 11
1 p12 p13 p24 p45 p46 p67 p78 p89 p910 d ' i0
1 p12 p13 p25 p24 p46 p67 p78 p89 p910 d '
λ 01
1 p12 p13 p25 p56 p64 p67 p78 p89 p910 d '
……
1 p12 p13 p25 p24 p46 p67 p78 p89 p910 d '
In this paper, we construct a specific P system where all feasible solutions to the
minimum spanning tree problem of a complete graph can be calculated. Then by
integrating these feasible solutions and deleting some large weight edges the final
partition result is formed in the output membrane. Through the example test, we can
see that the designed P system can effectively find quality clustering result in a dis-
tributed and parallel manner. The future work is mainly focused on optimizing the
process of finding solutions with the purpose of reducing identical solutions and
computation complexity and making our system run more efficiently.
References
1. Han, J., Pei, J., Kamber, M.: Data Mining: Concepts and Techniques. Elsevier (2011)
2. Fredman, M., Willard, D.: Trans-dichotomous algorithms for minimum spanning trees and
shortest paths. In: Proceedings of the 31st Annual IEEE Symposium on Foundations of
Computer Science, pp. 719–725 (1990)
3. Zhong, C., Malinen, M., Miao, D., et al.: A fast minimum spanning tree algorithm based on
k-means. Inf. Sci. 295, 1–17 (2015)
4. Zhong, C., Miao, D., Wang, R.: A graph-theoretical clustering method based on two rounds
of minimum spanning trees. Pattern Recogn. 43(3), 752–766 (2010)
5. Păun, G.: Computing with membranes. J. Comput. Syst. Sci. 61(1), 108–143 (2000)
12 P. Gong and X. Liu
6. Xue, J., Liu, X.: Lattice based communication P systems with applications in cluster
analysis. Soft. Comput. 18(7), 1425–1440 (2014)
7. Paun, G., Rozenberg, G., Salomaa, A.: The Oxford Handbook of Membrane Computing.
Oxford University Press Inc., New York (2010)
8. Sosík, P., Rodríguez-Patón, A.: Membrane computing and complexity theory: A charac-
terization of PSPACE. J. Comput. Syst. Sci. 73(1), 137–152 (2007)
9. Zhang, G.X., Pan, L.Q.: A survey of membrane computing as a new branch of natural
computing. Chin. J. Comput. 33, 208–214 (2010)
10. Grygorash, O., Zhou, Y., Jorgensen, Z.: Minimum spanning tree based clustering
algorithms. In: 2006 18th IEEE International Conference on Tools with Artificial
Intelligence (ICTAI 2006), pp. 73–81. IEEE (2006)
A Method for Correcting the Leaning of AR
Two-Dimensional Codes Based on LSM
1 Introduction
Augmented reality (AR) is with the development of virtual reality technology and
produce a new kind of computer application and the technology of human-computer
interaction [1]. It combines the computer-generated virtual environment with the real
scene of the user by means of photoelectric display technology, interactive technology,
multi-sensor technology, computer graphics and multimedia technology, so that the user
can confirm from the sensory effect that the virtual environment is an integral part of
the surrounding real scene. Unlike immersive virtual reality, augmented reality tech‐
nology is mainly based on the existing real world, to provide users with a new sensory
composite visual effects, to expand the human cognitive and perceived ability of the
world. Augment reality technology not only has the strong sense of reality, modeling
on small workload, and more secure and reliable.
Mobile augmented reality applications require mobile device software and hardware,
using the device camera to capture real-world images, calculate the relevant information,
integration of virtual scenes, and finally output to the screen, projectors and other display
devices [4]. When the two-dimensional barcode image is acquired and decoded by the
camera of the mobile phone, the acquired two-dimensional code is inevitably inclined.
When the tilt angle exceeds a certain range, the decoded barcode can not be decoded
correctly. Therefore, it is necessary to tilt the two-dimensional code.
2 Overview of 2D Code
of the QR code. In the symbol, the dark module represents the binary “1”, the light
module represents the binary “0” [9].
According to the structure of the QR code, firstly locate the position mark and record
the space coordinates of its left boundary, construct the data set {X, Y}, where X, Y is
the input vector, used to store the spatial coordinates of the boundary point of the
contour [2].
After rotation, the left edge of the 2D barcode must be perpendicular to the X axis,
then:
∑n ( ( ) ( ))2
min 𝜔 xi , yi − 𝜔 x, y (1)
i=1
According to the data set {X, Y} constructed, it can be linear regression according
to f(x) = ωx + b, we can see that the regression error of sample points is:
ei = f (x) − Y = 𝜔x = b − Y (2)
∑
n
( ( ) ( ))2
min 𝜔 xi , yi − 𝜔 x, y
i=1
n ( ( ))2
∑ (3)
= min ei − b − Y − e − b − Y
i=1
∑n ( )2
= min ei
i=1
16 C. Miao et al.
Therefore, only the data set {X, Y} regression. When the inaccuracy ei variance is
minimum, the parameter ω is the tilt vector. In this way, the optimization problem of
the image tilt angle is transformed into the process of regression of the data set {X, Y}
to identify the parameter ω.
Suppose there exists a univariate regression model f(x) = ax + b, and the regression
of the random variable Y on the independent variable is f(x). The univariate regression
model is:
( )
y = ax + b + ε ε ∼ N 0, 𝜎 2
By Eq. 3 we can see that when the deviation is the minimum angle is the tilt, the
value is:
∑n−1 ( )2
min 𝛿 2 = min yi − axi − b (4)
i=0
The value of a obtained here is the slope of the straight line after the left edge of the
fitting, and the inclination angle of the two-dimensional code in the horizontal direction
can be obtained.
Fig. 3. Flowchart
There are two kinds of formulas for converting the color image into grayscale image:
Here, we use the formula Y = (R + G)/2 and ignore the B component of the two-
dimensional code image gray processing. The adaptive threshold segmentation algo‐
rithm is adopted in the binarization threshold processing. The basic idea of the algorithm
is that the average gray value of S pixels is calculated when traversing the whole image.
When the pixel value of a pixel is lower than this value, set to black, otherwise set to
white.
18 C. Miao et al.
In practice, the camera in take of QR code image will have different degree of distortion.
Any geometric distortion can be defined by an equation that transforms the non-distorted
( ′ ′)
coordinate system (X, Y) into the distortion coordinate system X , Y , which is gener‐
ally of the form:
{
x′ = h1 (x, y)
y′ = h2 (x, y) (9)
Let f(x, y) be the original image without distortion and g(x, y) be the result of f(x, y)
distortion. The distortion process is known, and is defined by the functions h1(x, y) and
h2(x, y)
( )
g x′ , y′ = f (x, y) (10)
Equation (10) that should appear in the image in pixels (x, y) on the grey value due
to the distortion, and appear in the (x, y), the distortion problem might be solved by
mapping transformation. In the case of known g(x′ , y′ ), h1(x, y) and h2(x, y), the resto‐
ration process is as follows:
1. Find the corresponding positions in g(x′ , y′ ) for each point (x0, y0) in f(x, y): (M,
N) = [h1(x0, y0), h2(x0, y0)]. M and N are the coordinate values of the spatial points,
respectively. (M, N) does not coincide with any point in g(x′ , y′ ), since M and N are
not necessarily integers.
( ) ( )
2. Find the point x1′ , y1′ nearest to (M, N) in g(x′ , y′ ), let f(x0, y0) = g x1′ , y1′ ,
( )
that is, give the grayscale values of g x1′ , y1′ to f(x0, y0), according to this way
point by point until the entire image is finished, and the geometry is corrected.
After the distortion correction, we record the left border of the bar code, using (5,6) to
calculate the tilt angle 𝛼 = 𝚊𝚛𝚌 𝚝𝚊𝚗 a; After the bar code is obtained, the rotation formula,
that is
[ ] [ ][ ]
xnew cos a sin a xold
= (11)
ynew − sin a cos a yold
A Method for Correcting the Leaning of AR Two-Dimensional Codes 19
If the Eq. (11) is used directly, for a wide (W) * high (H) image, the rotation requires
4WH multiplication and 2WH addition, and the algorithm is highly computationally. In
fact, for a height of H for the two-dimensional code image, the vertical direction of the
projection maximum deviation is:
Ymax = H tan a
The deviations for the jth column and the ith row in the image are, respectively (Fig. 4).
( )
j
Δyj = ×W
Ymax + 1
( )
i
Δxi = ×H
Xmax + 1
Some points that are not in the integer position may be generated during the image
rotation transformation. This requires an algorithm for gray-scale interpolation to
produce a smooth mapping that maintains continuity and connectivity. Interpolation
methods include the nearest neighbor interpolation, bilinear interpolation and high
order interpolation. The nearest neighbor interpolation is a simple interpolation
algorithm, but the nearest neighbor interpolation algorithm will produce a clear
image of the zigzag boundary. For the QR code with only black and white, bilinear
interpolation can be used to produce a satisfactory image restoration effect [8]. The
mathematical model is shown in Fig. 5:
20 C. Miao et al.
After this series of operations, the original two-dimensional code image into a more
standard image, as shown (Fig. 6):
This paper mainly focuses on the tilt correction method of 2-D codes in mobile AR. The
LSM effectively improves the decoding time and decoding success rate. Start from the
server to download the three-dimensional virtual model loaded into memory. Then, the
model rendering output to the phone screen, and ultimately achieve the target QR code
to enhance the reality show. However, this method only supports the identification of a
single QR code. The next step is to consider a method that supports multiple QR code
recognition and combine with the current excellent local feature point algorithm to
realize fast and effective recognition of multiple QR codes.
References
1. Belimpasakis, P., Selonen, P., You, Y.: Bringing user-generated content from internet services
to mobile augmented reality clients. In: 2010 Cloud-Mobile Convergence for Virtual Reality
Workshop (CMCVR), pp. 14–17. IEEE (2010)
2. Trier, O.D., Jain, A.K.: Goal-directed evaluation of binarization methods. IEEE Trans PAMI
17(12), 1191–1201 (1995)
3. Chen, Y.-Y., Shi, P.-F.: Study and application of two-dimensional bar code. Tech. Meas.
Control 23(12), 17–19 (2006)
4. Schmalstieg, D., Wagner, D.: Experiences with handheld augmented reality. In: Proceedings
of Mixed and Augmented Reality, 6th IEEE and ACM International Symposium, pp. 3–18
(2007)
22 C. Miao et al.
5. Liu, Y., Liu, M.-Y., Shang, Z.-H.: Multi-level thresholding method for fast response matrix
codes. Appl. Res. Comput. 8, 177–179 (2006)
6. Liang, Y.-H., Wang, Z.-Y.: Two-dimensional bar code tilt detection based on projection under
high noise conditions. Microcomput. Inf. 22, 232–234 (2006)
7. Azuma, R., Bailot, Y., Behringer, R., et al.: Recent advances in augmented reality. Comput.
Graph. Appl. 21(6), 34–47 (2001)
8. White, S., Feiner, S., Kopylec, J.: Virtual vouchers: prototyping a mobile augmented Reality
user interface for botanical species identification. In: IEEE Symposium on 3DUserInterfaces
2006, 3DUI2006, pp. 119–126. IEEE (2006)
9. Pavlidis, T.: A new paper/computer interface: Two-dimensional symbologies. IEEE Comput.
Mag. 2, 145–151 (2000)
The Associated Algorithm of Target Batch Number Based
on Gaussian Mixture Clustering
Abstract. This article provides some questions about the large quantities of
target track information, unclear sensor batch number, etc. referring to target
fusion, and provide the clustering algorithm of Gaussian Mixture Distribution to
realize the division of all the batch numbers of target track and associated oper‐
ation. According to the processing to the Gaussian distribution of information by
maximum likelihood estimation and through continuous iterative and refinement
of the clusters divided, the tracking information detected by all radar sensors of
each target batch number can be obtained precisely in the end. This algorithm
provides rapid and precise working conditions for target fusion.
1 Introduction
When the formation of ship combat synergistically, naval and radars in the fleet will
report all the detected objectives about the main naval, which will create influences on
the warship equipment attacked precisely according to the information.
Not only this, the objectives detected by radars can only feedback its physical infor‐
mation and can not precisely judge the target batch number defined by human. In the
current literature, radars’ objective fusion rarely have research, which aims at the two
sides. In order to simplify the target quantity, avoid unnecessary redundant target infor‐
mation. The associated algorithm of target batch number in this article will connect the
prototype cluster in the ensemble learning to solve the related problems between target
information and target batch number.
The command system of formation of ship receives periodically signals from different
radars in formation. Here it is set as M pieces of sensors, with each sensor as m. It can
periodically make up instructions and send N pieces of tracking information processed
by itself and under the unified coordinate system. Here the ensemble of communication
of multiple batches of tracking information sent by No.m sensor is set as Dm and among
it, some batch of tracking information is dmn(PS: m = 1, 2,…… as the tracking infor‐
mation detected by NO.m sensor. n = 1, 2…… as the NO.n target detected).
Target batch correlation is the basis of target fusion. Exactly differentiating DS is the
precondition of fusion algorithm accuracy. In the war of formation of ship, because it
is very difficult to recognize the target without exactly differentiating a large amount of
tracking information detected, the accuracy of the algorithm is very important.
Firstly define some basic knowledge. As it is mentioned earlier, target tracking
information includes a great amount of data, such as a sensor’ serial number, target
nationality, target type, longitude and latitude, height and a series of data. Here the
number of attributes is set as k and set dmn = {xmn1, xmn2,…, xmnk}, the NO.n target
tracking information detected by NO.m radar as a k-dimensional vector. xmnk is the value
of NO.k of a some batch tracking information (such as, the above target type’s value on
xijk is “the air”), while k is the dimension of the tracking information vector.
The result of batch correlation is a list, with a batch number on each line. Batch
numbers correspond to the tracking information collection under the target detected by
the related sensor. The collections can not be controlled strictly or intersected and the
parallel operation of these collections is D, which is all tracking information detected
by all radars. Here we call every subset as a “cluster”. Through this divide, each cluster
stands for tracking information collection of the same target detected by the radar sensor.
After the target batch correlates with tracking information, the list formed by all clusters
report to fusion center to perform blend operation.
It is formally said that D can be divided into h pieces of disjoint clusters {Ci|i = 1,
2…, h}, with Ci ∩ Ci’ ≠ ∅. Accordingly, each cluster’s marker is set as the target batch
number. We use λ = {λ1, λ2,…, λh} to represent tracking information clusters’ marker,
including the value of λh as the target batch number.
Above we introduce is some basic knowledge about association algorithm. The
association algorithm the article studies will be researched by connecting clusters under
unsupervised learning. We just need to divide the D collection formed by all tracking
information into multiple joint clusters. But we need to solve the problem of distance
calculation.
We know that Gaussian distribution is fixed by two parameters among it. The two
parameters are mean vector μ and covariance matrix G. We define Gaussian mixture
distribution first:
∑
h
( )
PM(x) = αi p x|μi , G (3.1.1)
i=1
The Associated Algorithm of Target Batch Number 25
Among it, the latter part p(x|μi, G) is the probability density of Gaussian distribution.
Suppose that all tracking information is produced by Gaussian mixture distribution, the
specific process is to fix different Gaussian mixture distribution probability density
according to α1, α2,…, αh. After fix the probability density function, put different tracking
information in different Gaussian distributions. In here we can see that Gaussian mixture
distribution is the mixed composition of multiple Gaussian distribution. Each Gaussian
distribution’s probability is controlled by a parameter αh. The probability is how much
the the size is about some tracking information has been specifically related to one
Gaussian distribution.
It can be seen that three parameters αi, μi, G among it should be fixed if defining
Gaussian mixture distribution. Suppose that all tracking information has been
completed the correlation, how to fix the target batch number of some tracking
information, which means how to fix cluster marker of the tracking information
cluster some tracking information is in. Suppose that the tracking information cluster
of some tracking information dij is in is z. In Gaussian mixture distribution, zmn is the
Gaussian distribution some sample is in. The value range is 1~h. But we have no way
to know its specific value. According to Bayes’ theorem, posteriori distribution of
zij is
( ) ( )
( ) p zmn = h ⋅ PM dmn |zmn = h
PM zmn = h|dmn = ( )
PM dmn
(3.1.2)
α ⋅ p(dmn |μi , G)
= ∑h i
α ⋅ p(dmn |μi , G)
l=1 l
Which means we have got the posteriori probability of each tracking information
cluster dmn is in, marked as γmnz. When we want to know the batch number of some target
tracking information, it is only to find the max one the dmn corresponds to.
1 ∑
αi = γmnz (3.2.4)
mn
After get every parameter’s calculation method and use iterative process of Gaussian
mixture clustering, which means getting posterior probability of each Gaussian distri‐
bution through current parameters and current tracking information. Then through the
formula, update three parameters till reach the maximum value of iterative or LL(D)
with a little improvement or no improvement, stop updating the parameters and finally
make sure the tracking information included among every tracking information. And
the corresponding marker of every tracking information is target batch. In this way, the
overview that flow the target tracking information relates is shown in following figure.
The Associated Algorithm of Target Batch Number 27
We define a measure that evaluates the clarity of the tracking information to prove the
advantages of this algorithm. Davies-Bouldin index:
( ( ) ( ))
1∑ avg Ci + avg Cj
h
DB = max ( )
h i=1 j≠i dcen μi , μj
Among them, dcen(…) is to calculate the distance between the center of two clusters,
avg(C) is the average distance of each trace information in the cluster C. So, we can see,
the denominator in the DB index is the distance between the two tracking information
clusters, the larger the denominator, the lower the similarity of the two clusters, and the
molecules represent the tracking information aggregation in two clusters. So, the smaller
the value of the DB index, the better the result of the algorithm. In a tracking information
data set of 11 ships, the results are obtained 0.32~0.33, it’s better than the general infor‐
mation.
The algorithm promoted in this article controls all the tracking information within
the k-dimension space, proceeds loop iteration and fix the collection of tracking infor‐
mation detected by multi sensors each target corresponds to, which realizes the connec‐
tive work of precise target batch number and complete the work within the quadratic
time.
References
1. Wan, S.: Clustering of multisensor data fusion method. J. Syst. Eng. Theory Pract. (05) (2008)
2. Xinrui, Z., he, J.: Multi-sensor information fusion technology. Comput. Meas. Control (09)
(2007)
3. Hongfeng ailin, W., Single Showers: Research progress of foreign military information fusion
theory and application. Lightning Control (04) (2007)
4. Ma, J., Jixiang, S.: Information fusion model of neural network to describe. J. Avionics
Technol. (01) (1998)
5. Xin, W., Qidan, Z., Shuli, S.: Unfettered global optimal weighted measurement fusion
estimation. Comput. Eng. Appl. (24) (2010)
6. Hu, W., Shenyuan, Y.: The multi-sensor track correlation. Appl. Sci. Technol. (12) (2005)
Density Peaks Clustering Based on Improved
RNA Genetic Algorithm
1 Introduction
2 Related Works
The proposed DPC-RNAGA is based on DPC and RNA-GA. This section provides
brief reviews of DPC and RNA-GA.
(
1 if x\0
vð x Þ ¼ ð1Þ
0 if x 0
Where dc is cutoff distance, qi is defined as the number of points that are adjacent to
the point vi , dc is an adjustable parameter and the only variable in Formula (1). Here we
30 L. Ren and W. Zang
use the other way to calculate the local density called exponential kernel function, it is
defined as follows:
X 2 !
dij
qi ¼ exp ð2Þ
j6¼i
dc
The computation of di is quite simple. The minimum distance between the point vi
and any other points with higher densities, denoted by di , is defined as:
(
minj:qj [ qi d vi ; vj if 9js:t:qj [ qi
di ¼ ð3Þ
maxj d vi ; vj if otherwise
Only those points with relative high qi and high di are considered as clustering
centers. The points with high qi and high di are also called as peaks. After clustering
centers have been found, DPC assigns each remaining points to the same cluster as its
nearest neighbor with higher density.
Genetic operators and penalty factor. The tournament method is adopted as the
selection method in DPC-RNAGA [5]. However, in the special circumstance, It was
found that the search field is biased after several experiments. So we propose the
penalty factor d to get rid of inferior individuals whose fitness value stays the last d
percentage of all. If the initial population is Q1 ¼ fq1 ; q2 ; ; qm g, the remaining ratio
of the parent generation is r, so the number of individuals remaining from the parents is
r m.
Selective crossover operator is designed to enhance the ability in exploring the
searching space in this paper. According to the crossover probability pc , the two parents
exchange the bases between the two crossover points to produce two offspring
individuals.
In this paper, we adopted the ordinary mutation operators but with adaptive
probability. We divide the nucleotide bases into two parts: left and right parts. Cor-
respondingly, the adaptive probabilities pml (left) and pmh (right) are described as
follows:
8
< pml ¼ a1 þ 1 þ exp½baa1 ðgg Þ
ð4Þ
0
: pmh ¼ a1 þ b1
1 þ exp½aaðgg0 Þ
Density Peaks Clustering Based on Improved RNA Genetic Algorithm 31
Where a1 denotes the initial mutation probability of pml, b1 denotes the range of
transmutability. The parameter g is evolutionary generation, and g0 denotes the gen-
eration where the maximum of mutation probability is reached, and aa is the speed of
change.
Fitness function. In this paper, Rand index was adopted as the fitness function of
RNA-GA. The Rand index is defined as the ratio of the same data pairs from the two
clustering results and all the data pairs [6]. The value of rand index ranges from 0
through 1, and the value close to 1 shows that the clustering is better.
From Table 1, we can see that the time complexity of DPC-RNAGA is of the same
order as DPC. Although it is higher than K-means, it is still acceptable. And it doesn’t
affect the application of the algorithm.
32 L. Ren and W. Zang
4 Experiments
5 Conclusion
A clustering algorithm based on RNA genetic algorithm and finding density peaks was
proposed in this paper. This method computes the local density of data points with the
exponential kernel methods, which reduced the dependence on the parameter dc .
RNA-GA is used to search the optimal density and distance from other data point with
higher density, which makes the clustering algorithms more effective. Penalty factor is
added to the RNA-GA, which avoids the search region getting wrong way and
accelerates convergence. The power of DPC-RNAGA was tested on several synthetic
datasets and real-world datasets. The results demonstrate that DPC-RNAGA is pow-
erful in finding clustering centers and recognizing clusters regardless of their shape and
size. The comparison analysis of several clustering algorithms shows that our proposed
method can be fit to many different datasets and perform better than the existing DPC,
K-means and Max_Min_SD methods.
References
1. Mohebi, A., Aghabozorgi, S.R., Teh, Y.W., Herawan, T., Yahyapour, R.: Iterative big data
clustering algorithms: a review. Softw. Pract. Exper. 46(1), 107–129 (2016). https://doi.org/
10.1002/spe.2341
2. Gorunescu, F.: Data Mining –Concepts, Models and Techniques. ISRL, vol. 12. Springer,
Heidelberg (2011). https://doi.org/10.1007/978-3-642-19721-5
3. Rodriguez, A., Laio, A.: Clustering by fast search and find of density peaks. Science 344
(6191), 1492–1496 (2014)
4. Sun, K., Geng, X., Ji, L.: Exemplar component analysis: a fast band selection method for
hyperspectral imagery. IEEE Geosci. Remote Sensing Lett. 12(5), 998–1002 (2015). https://
doi.org/10.1109/LGRS.2014.2372071
5. Zhu, Q., Ning, W., Li, Z., Zhu, Q., Ning, W., Li, Z.: Circular genetic operators based RNA
genetic algorithm for modeling proton exchange membrane fuel cells. Int. J. Hydrog. Energy
39(31), 17779–17790 (2014)
6. Dubes, R.C.: Cluster analysis and related issues. In: Handbook of Pattern Recognition &
Computer Vision (2015). 996 p
7. Chang, H., Yeung, D.: Robust path-based spectral clustering. Pattern Recogn. 41(1), 191–
203 (2008). https://doi.org/10.1016/j.patcog.2007.04.010
8. KBache, M.L.: Uci machine learning repository (2013). http://archive.ics.uci.edu/ml
9. Dubey, A.K., Gupta, U., Jain, S.: Analysis of k-means clustering approach on the breast
cancer wisconsin dataset. Int. J. Comput. Assist. Radiol. Surg. 11(11), 2033–2047 (2016).
https://doi.org/10.1007/s11548-016-1437-9
10. Bian, W., Tao, D.: Max-min distance analysis by using sequential SDP relaxation for
dimension reduction. IEEE Trans. Pattern Anal. Mach. Intell. 33(5), 1037–1050 (2011).
https://doi.org/10.1109/TPAMI.2010.189
Local Homogeneous Weighted Spiking Neural P Systems
1 Introduction
Membrane computing [1] is one of the recent branches of natural computing, whose aim
is to construct powerful computing models and intelligent algorithms by abstracting
ideas from a single living cell and from complexes of cells. The obtained models are
distributed and parallel computing devices, usually called P systems. There are three
main classes of P systems investigated: cell-like P systems [1], tissue-like P systems [2]
and neural-like P systems [3]. Spiking Neural P systems (SN P systems, for short) are
a class of neural-like P systems, which are inspired by the way of biological neuron in
human brain processing information and communicating with each other by means of
electrical impulses (spikes) [3].
In the research of SN P system, the results mainly focus on three aspects: theory,
application and implement. Theoretical level is divided into three aspects: generating
sets of numbers [4–10], generating languages [11, 12], computing any Turing comput‐
able functions [13, 14]. In recent years, we have studied a variety of SNP variants and
analyzed their ability to generate sets of numbers [4, 6–8, 15, 16]. Particularly, Zeng
proposed the homogeneous SNP systems in 2009 [4], and proved that homogeneous
SNP systems have Turing universality in generating and accepting modes. In 2015, Song
proposed homogenous SNP Systems with Inhibitory Synapses (through inhibitory
www.allitebooks.com
Local Homogeneous Weighted Spiking Neural P Systems 35
synapses, spikes is converted to the anti-spikes) [16], and proved the systems have
Turing universality in the case of using only the spiking rules.
Inspired by various of SNPS, this article proposed a variant of SNP system based on
a new biological fact and study the ability of generating sets of numbers in two modes
of the new system, besides that, we analyze the time need to execute the system. The
biological fact is that although neurons are “designed” very similar, but there are still
some differences between different neurons. According to functional division, neurons
can be divided into three categories: sensory neurons, connecting neurons, motor
neurons. In the nerve center, neuronal cell bodies having the same function gather
together to regulate a physiological activity of body (the definition of nerve center). That
is to say, there are some differences in the rules of the neuron, but the neurons of the
same rules will be gathered together to cooperate with each other to complete a certain
physiological function. We introduce this biological phenomenon into SN P systems
and the SNP systems have local homogeneity.
In LHWSN P systems, the rules used by neurons in the same function module are
identical, and the weights on the synapse can be positive integer and negative integer.
When the spike passes through the synapse with negative weights, the spike is changed
into the anti-spike and the number of the spike is doubled according to the absolute value
of the weight.
In this paper, we proved that the LHWSN P systems have Turing universality. First,
we conclude that the delay of rules can be removed but the systems still have Turing
universality. Second, we use negative weight to express inhibitory synapse so the
forgetting rule in neurons is removed. Third, the proposing of local homogeneous
reduces the total time needed to execute the system.
In the next section, we introduce prerequisites and the definition of LHWSN P
systems. In Sect. 3, we investigate the computation power of the system. Conclusions
and remarks are drawn in Sect. 4.
2.1 Background
In the study of the computing power of SN P system, most of them are proved by
simulating calculation process of the register machine. A register machine is a construct
( )
M = m, H, l0 , lh , R , where m is the number of registers, H is the set of instruction labels,
l0 is the start label, lh is the halt label, and R is the set of instructions. The instructions
are of the following forms:
( )
• li : ADD(r), lj , lk (add 1 to the register r and then go to one of the instructions with
label lj and lk, non-deterministically chosen),
( )
• li : SUB(r), lj , lk (if register r is non-empty, then subtract 1 from it and go to the
instruction with label lj, otherwise go to the instruction with label lk),
• lh: HALT (halt the calculation of the register machine and regard the number in register
1 as the result of the register machine).
Register machine has two kinds of working modes: generating mode and accepting
mode. In generating mode, the register machine computes number n as follows: starting
M with all registers empty, the initial instruction with label l0 is applied, and the instruc‐
tions are applied as indicated by labels. When the register machine M proceeds to the
halt instruction, number n stored at that time in the first register is said to be computed
by M. In accepting mode (using deterministic ADD instructions), a random number is
stored in the first register (other registers are empty). If the computation starting from
the initial configuration eventually halts, the number is said to be accepted by M.
In this section, we define the LHWSN P system. The definition is complete, but famil‐
iarity with the basic elements of homogeneous SN P systems, anti-spikes and SNP with
weights (e.g. from [4, 7, 19]) is helpful.
An LHWSN P system of degree m ≥ 1 is a construct of the form:
∏ ( )
= O, σ1 , σ2 , … σm , syn, in, out
Where
• o = {a, ā } is the alphabet, where a is spike and ā is anti-spike;
( )
• σ1 , σ2 , … σm are neurons of the form σi = ni , R with 1 ≤ i ≤ m where
(a) ni is a natural number representing the initial number of spikes in σi;
(b) R is set of rules in each neuron of the following forms:
1. E∕ac → a is the spiking rule, where E is the regular expression over {a}.
(c is integer and c ≥ 1);
2. as → λ is the forgetting rule, with the restriction that for any s ≥ 1 and any
spiking rule E∕ac → a;d, as ∉ L(E), where L(E) is set of regular languages
associated with regular expression E and λ is the empty string;
3. aa → λ.
Local Homogeneous Weighted Spiking Neural P Systems 37
• syn ⊆ {1, 2, ..., m} × {1, 2, ..., m} × Z (Z is integer) is set of synapses between neurons,
where i ≠ j, z ≠ 0 for each (i, j, z) ∈ syn, and for each (i, j) ∈ {1, 2, ..., m} × {1, 2, ..., m}
there is at most one synapse (i, j, z) in syn.
• In, out ∈ {1, 2, ..., m} indicate the input and output neurons respectively.
In LHWSN P systems, the rules used by neurons in the same function module are
identical, but the rules of different functional modules are different. A neuron may
contain a number of spikes or anti-spikes, but only one of them. In LHWSN P systems,
spiking rules (E∕ac → a) can be applied in any neuron as follows: if neuron σi contains
k spikes a with ak ∈ L(E) and k ≥ c, the spiking rule E∕ac → a is enabled to be applied.
By using the rule, c spikes a are consumed, thus k − c spikes a remain in the neuron σi,
and one spike a is sent to all neurons σj such that (i, j, z) ∈ syn. Rules of the form
as → λ, s ≥ 1 are forgetting rules with the restriction as ∉ L(E) (that is to say, a neuron
cannot apply the spiking rules and forgetting rules at the same moment), where L(E) is
set of regular languages associated with regular expression E and λ is the empty string.
If neuron σi contains exactly s spikes, the forgetting rule as → λ can be applied, by which
s spikes can be removed from the neuron.
The set of synapses is denoted by (i, j, z). Specifically, neuron σi spikes at certain step
sending a spike to neuron σj along the synapse (i, j, z), if z > 0, neuron σj will receive z
spikes; if z < 0, neuron σj will receive −z anti-spikes. The result of a computation of the
∏
system is defined as the time internal of first two spikes being emitted to the envi‐
ronment by the output neuron at steps t1 , t2 in the form of number t2 − t1; we say that this
∏ ∏
number is generated by . The set of all numbers computed in this way by is denoted
(∏)
by N2 (the subscript 2 indicates that we only consider the distance between the first
two spikes of any computation.)
∏
An LHWSN P system can also work in the accepting mode. A number n is intro‐
duced in a specified neuron in the form of f(n) spikes by reading spike train from the
environment through the input neuron. If the computation eventually halts, then number
∏ ∏ (∏)
n is said to be accepted by . The set of numbers accepted by is denoted by Nacc
(the subscript acc indicates the system works in the accepting mode).
We denote by Nα LHWk SNPm , α ∈ {2, acc} all sets of numbers generated or accepted
by LHWSN P systems of degree m. If the parameter m is not bounded, then it is replaced
with ∗. The absolute value of the weight is not more than K.
In this section, we proved that the LHWSN P systems have a universal computation in
the generating mode and the accepting mode. In the following proofs, a directed graph
is used to represent the structure of the system: the neurons are placed in the nodes of
the graph connecting with each other by the synapses; the output neuron has an outgoing
synapse, emitting spikes to the environment.
38 M. Liu and F. Qi
li
-1
li1 r
-1
5
li2 li3
lk lj
Assume at step t, the system starts to simulate a SUB instruction li of M. Having one
spike inside, neuron σli fires at step t emitting a spike to neurons σl1i and σr. The spike
arriving at neuron σr will be changed into an anti-spike due to the negative weight, and
there are the following two cases in neuron σr.
Local Homogeneous Weighted Spiking Neural P Systems 39
– Before receiving the anti-spike, if the number stored in register r is not zero, then
neuron σr contains at least 5n spikes. After the annihilation, the number of spikes in
neuron σr is decremented to 5(n − 1) + 4, which simulates decreasing the number in
register r by one. With 5(n − 1) + 4 spikes inside, neuron σr fires at step t + 1 by using
( )+
the rule a4 a5 ∕a4 → a and sends one spike to neuron σl2i , σl3i. The spike arriving at
neuron σl2i will be changed into an anti-spike due to the negative weight. Meanwhile,
neuron σl2i receives one spike from neurons σl1i . The anti-spike in neuron σl2i will be
annihilated by the spike from neuron σl1i . So neuron σl2i ends with no spike or anti-
spike inside and remains inactive. By receiving the spike, neuron σl3i fires at step
t + 2, and the neuron σlj will be fired at step t + 3 to start to simulate instruction lj of
M. It is worth to point out that 5 spikes are sent back to neuron σr at step t + 3 from
neuron σl3i, which represents that the number in register r is still n.
– Before receiving the anti-spike, if the number in register r is zero, then neuron σr has
no spikes. After receiving the anti-spike, neuron σr has one anti-spike. At step t + 1,
neuron σl1i fires sending one spike to neuron σl2i . At step t + 2, neuron σl2i sends a spike
to neuron σlk and neuron σr. In neuron σr, the anti-spike is annihilated, and it keeps
inactive. At step t + 3, neuron σlk receives a spike from neuron σl2i and becomes active,
which starts to simulate instruction lk of M.
The simulation of SUB instruction is correct: system P starts from neuron σli and
ends in neuron σlj(if the number stored in register r is great than 0 and decreased by one),
or in neuron σlk(if the number stored in register r is 0).
( )
Module 𝐀𝐃𝐃 shown in Fig. 2— simulating the ADD instruction li : ADD(r), lj , lk .
The neurons of ADD module are identical, and the rules contained in the neurons are
a → a , a2 → a, a3 → a, and a3 ∕a2 → a.
r
li
l 1i
3
l 2
i
l 3
i l i4
-1 -1 2
5
l i l 6
i
lj lk
lh
2 2
h1
1
-1
2 -2 4
h3 out
h2
2 2
Assume now that the computation in M halts (that is, the halting instruction is
∏
reached), which means that neuron σlh in has two spikes and fires. The evolution of
the numbers of spikes in the neurons during the computation of the FIN module is shown
in Table 1.
Table 1. The numbers of spikes in neurons of FIN module during the process outputting the
computational result.
Neuron Step
t t+1 t+2 t+3 ··· t+n t + n+1 t + n+2 t + n+3
𝜎lh 1 0 0 0 ··· 0 0 0 0
𝜎1 5n 5n + 2 5(n−1) + 2 5(n−2) + 1 ··· 5+1 1 1 0
𝜎h1 0 2 0 0 ··· 0 0 0 0
𝜎h2 0 0 2 2 ··· 2 2 0 0
𝜎h3 0 0 2 0 ··· 0 0 2 0
𝜎out 0 0 0 2 ··· 0 0 0 2
According to the function of the above three modules, it is clear that the register
∏
machine M can be correctly simulated by LHWSN P system . Therefore, LHWSN P
systems can characterize NRE. It is also worth to mention that there are only standard
spiking rules, and no forgetting rule is used. Moreover, we can see that the maximum
number of rules in neurons is 4(in ADD module), so we greatly reduce the number of
rules in the neurons.
42 M. Liu and F. Qi
in d4
-2
7
d1 d2 d3
-2
-2
3 2
l0 1
Table 2. The numbers of spikes in the neurons of INPUT module during the process of reading
spike train 10n−11
Neuron Step
t t+1 t+2 t+3 ··· t+n t + n+1 t + n+2 t + n+3
𝜎in 14 10 8 8 ··· 9 5 3 3
𝜎d1 0 7 7 7 ··· 7 14 10 10
𝜎d2 0 1 1 1 ··· 1 2 0 0
𝜎d3 0 1 1 1 ··· 1 2 0 0
𝜎d4 0 1 0 0 ··· 0 1 0 0
𝜎l0 0 0 0 0 ··· 0 0 1 0
𝜎1 0 0 5 5×2 ··· 5(n−1) 5n 5n 5n
Local Homogeneous Weighted Spiking Neural P Systems 43
During the process of reading spike train 10n−11, the evolution of the numbers of
spikes in the neurons of INPUT module is shown in Table 2.
In accepting module, ADD module is deterministic and its form is
( ( ))
li : ADD(r), lj , which is indicted in Fig. 6. The neurons of this system contain rule
a → a.
li
r lj
Module SUB remains unchanged (as shown in Fig. 1), module FIN is removed, with
neuron σlh remaining in the system, but without outgoing synapses. When neuron σlh
receives two spikes, it means that the computation of register machine M reaches
instruction σlh and stops. Having two spikes inside, neuron σlh spikes (a2 → a is enabled)
∏′
without emitting any spike out and the system halts.
As explained above, it is obtained that any set of natural numbers accepted by deter‐
∏′
ministic register machines can be accepted by LHWSN P system .
Literature [20] proposed that spiking rules in neurons see the regular expressions as
criteria of spiking and determining whether a regular expression satisfied may be a NP
hard problem. This paper reduce the number of rules in neurons, greatly reducing the
time of determining whether a regular expression is satisfied and then reducing the time
of the whole module running.
We assume that the time to judge whether the spiking rules are satisfied is T1, and
whether the forgetting rules are satisfied is T2. Each step of the system needs to deter‐
mine whether the spiking and forgetting rules can be executed. In the following, We
will compare the time required for LHWSNPS and HWSNPS [4] in each module.
In the ADD module of INHWSPS, each neuron has 4 spiking rules, and the ADD
module takes 5 steps from start to finish, So the total time required for the ADD module
is 20T1; In ADD module of HWSNPS, each neuron has five spiking rules and two
forgetting rules, and four steps are needed to finish ADD module, so the total time is
20T1 + 8T2. Similarly, we can calculate the time consumed by two systems in the SUB
module, FIN module and INPUT module respectively. Comparison of execution time
of two systems can be seen in Table 3.
44 M. Liu and F. Qi
4 Final Remarks
Acknowledgment. This work was supported by the Natural Science Foundation of China (No.
61502283). Natural Science Foundation of China (No. 61472231). Natural Science Foundation
of China (No. 61640201).
References
1. Păun, G.: Computing with membranes. J. Comput. Syst. Sci. 61(1), 108–143 (2000)
2. Freund, R., Păun, G., Pérez-Jiménez, M.J.: Tissue-like P systems with channel-states. Theor.
Comput. Sci. 330, 101–116 (2005)
3. Ionescu, M., Păun, G., Yokomori, T.: Spiking neural P systems. Fund. Inf. 71(2–3), 279–308
(2006)
4. Zeng, X., Zhang, X., Pan, L.: Homogeneous spiking neural P systems. Fundamenta
Informaticae 97(1–2), 275–294 (2009)
5. Ibarra, O.H., Păun, A., Rodríguez-Patón, A.: Sequential SNP systems based on min/max spike
number. Theor. Comput. Sci. 410(s30–s32), 2982–2991 (2009)
6. Cavaliere, M., Ibarra, O.H., Păun, G., et al.: Asynchronous spiking neural P systems. Theoret.
Comput. Sci. 410(24), 2352–2364 (2009)
7. Pan, L., Păun, G.: Spiking neural P systems with anti-spikes. Int. J. Comput. Commun. Control
4(3), 273–282 (2009)
8. Song, T., Pan, L., Păun, G.: Asynchronous spiking neural P systems with local
synchronization. Inf. Sci. 219, 197–207 (2013)
Local Homogeneous Weighted Spiking Neural P Systems 45
9. Zhang, X., Zeng, X., Pan, L.: Weighted spiking neural p systems with rules on synapses.
Fundamenta Informaticae 134(1–2), 201–218 (2014)
10. Song, T., Zou, Q., Liu, X., et al.: Asynchronous spiking neural P systems with rules on
synapses. Neurocomputing 151, 1439–1445 (2015)
11. Chen, H., Freund, R., Ionescu, M., et al.: On string languages generated by spiking neural P
systems. Fundamenta Informaticae 75(75), 141–162 (2007)
12. Zhang, X., Zeng, X., Pan, L.: On string languages generated by spiking neural P systems with
exhaustive use of rules. Nat. Comput. 7(4), 535–549 (2008)
13. Păun, A., Păun, G.: Small universal spiking neural P systems. BioSystems 90(1), 48–60
(2007)
14. Pan, L., Zeng, X.: Small universal spiking neural P systems working in exhaustive mode.
IEEE Trans. NanoBiosci. 10(2), 99–105 (2011)
15. Song, T., Shi, X., Jinbang, X.U.: Reversible spiking neural P systems. Front. Comput. Sci.
Sel. Publ. Chin. Univ. 7(3), 350–358 (2013)
16. Song, T., Wang, X.: Homogenous spiking neural P systems with inhibitory synapses. Neural
Process. Lett. 42(1), 199–214 (2014)
17. Rozenberg, G., Salomaa, A.: Handbook of Formal Languages: Beyond Words. Springer,
Heidelberg (1997). https://doi.org/10.1007/978-3-642-59126-6
18. Hopcroft, J.E., Motwani, R., Ullman, J.D.: Introduction to automata theory, languages, and
computation, 2nd edn. ACM Sigact News 5(8), 60–65 (2001)
19. Wang, J., Hoogeboom, H.J., Pan, L., et al.: spiking neural P systems with weights. Neural
Comput. 22(10), 2615–2646 (2010)
20. Leporati, A., Zandron, C., Ferretti, C., et al.: On the computational power of spiking neural
P systems. Int. J. Unconventional Comput. 5(5), 459–473 (2009)
Weight-Improved K-Means-Based Consensus Clustering
School of Management Science and Engineering, Shandong Normal University, Jinan, China
15554130027@163.com, xls3366@163.com, sdxyliu@163.com
Abstract. Many consensus clustering methods ensemble all the basic partition‐
ings (BPs) with the same weight and without considering their contribution to
consensus result. We use the Normalized Mutual Information (NMI) theory to
design weight for BPs that participate in the integration, which highlights the
contribution of the most diverse BPs. Then an efficient approach K-means is used
for consensus clustering, which effectively improves the efficiency of combina‐
torics learning. Experiment on UCI dataset iris demonstrates the effective of the
proposed algorithm in terms of clustering quality.
1 Introduction
There is no single clustering algorithm can performs best for all data sets [1], and can
discover all types of cluster shapes and structures [2]. Consensus clustering approached
are able to integrate multiple clustering solutions obtained from different data sources
into a unified solution, and provide a more robust, stable and accurate final result [3].
However, the previous research still has some limitations.
Firstly, high quality BPs are beneficial to the performance of consensus clustering
yet the partitions with poor quality will lead to worse consensus result. But most studies
tend to integrate all BPs, and they do not filter poor BPs. Secondly, the diversity between
BPs also have great impact on consensus clustering, diverse BPs which means the BP
that has different mutual information with other BPs will have different contribution to
the consensus result. However, there are few references explore impact of the number
of BPs to consensus clustering neither did they take into account the diversity of BPs
into the integration process.
We proposed weight-improved K-means-based consensus clustering (WIKCC).
Firstly, we design weight for each BP participating in the integration. Specifically,
we generate multiple BPs and measure the quality of each BP using normalized Rand
index Rn [6], and sort the BPs in the increasing order of Rn, then we explore the
influence of the number of BPs on consensus clustering, based on the above explo‐
ration we can choose an appropriate number of better BPs for consensus clustering,
which can minimize the number of BPs in quality assurance. After that we construct
the co-occurrence matrix with the selected BPs, and calculate the similarity of two
BPs with Normalized Mutual Information (NMI) method [4] according to the co-
occurrence matrix. Then weight of each BP is designed according to NMI values
which reflect a single BP to overall BPs’ similarity. K-means-based method [2] make
special attention for its simplicity and high efficiency. So we transform consensus
clustering to K-means clustering.
Mutual information is used to measure the shared information of the two distributions.
We compute the NMI between two BPs, and the greater the value of NMI means the
lower difference, which will result in lower wi. { }
Given two BPs results πi with Ki clusters, πi = C1(i) , C(i) , … , C (i)
and πj with Kj
{ } 2 Ki
(j) (j) (j)
clusters, πj = C1 , C2 , … CK the mutual information between two results is defined
j
as follows:
( )
( ) 2I1 πi , πj
NMI πi , πj = ( ) ( ) (1)
I2 πi + I2 πj
| (i) (j) | | |
( ) ∑ ∑ ||Ch ∩ Cl || n|Ch(i) ∩ Cl(j) |
I1 πi , πj = log | |
| (i) | (j) (2)
h l n |Ch | |Cl |
| |
( ) ∑ |Ch(i) | |C(i) |
I2 πi = − log h (3)
h n n
( ) ∑ |Cl(j) | |Cl(j) |
I2 πj = − log (4)
l n n
For a single BP the average mutual information can be defined as:
( ) 1 ∑r ( )
H 𝜋i = NMI 𝜋i , 𝜋k , (i = 1, 2, … r) (5)
r−1 k=1,k≠i
{ } { }
Where h ∈ 1, 2, … , ki , l ∈ 1, 2, … , kj is one of the cluster result label of πi and
| | | (j) |
πj, |C(i) |, |C | respectively represent the number of the data set belong to cluster C(i) in πi
| h || l | h
| |
and C(j) in πj, |C(i) ∩ Cl(j) | is the number of the dataset belong both to C(i) and C(j), r is the
l | h | h l
number of the BPs.
( )
The greater H 𝜋i indicate that cluster member 𝜋i share more information with other
cluster members. The weight is defined as:
1
w,m = ( ) (6)
H 𝜋i
48 Y. Wang et al.
The weight is bigger as the greater diversity between two base clustering.
In this section, we first introduce the co-occurrence matrix which is used for records the
situation of sharing dataset between two BPs. Table 1 shows an example.
(i)
In Table 1, BPs: 𝜋 ∗ and 𝜋i contain k and ki clusters respectively, nKK represents the
i
(i)
number of the objects that belongs to both CK and CK , then let
∑Ki (i) i
∑r | (b) |
Where xl(b) is an n × Ki binary data set matrix with |xl,i | = 1.
i=1 | |
Weight-Improved K-Means-Based Consensus Clustering 49
We use the K-means algorithm to integrate the BPs, suppose r BPs are integrated to
a result 𝜋 ∗, mk represent the centroid of the Ck in 𝜋 ∗ as follows:
⟨ ⟩
mk = mk,1 , … , mk,i , … , mk,r , with (11)
⟨ ⟩
mk,i = mk,i1 , … , mk,ij , … , mk,iKi , (12)
The centroids can be computed by the Co- occurrence matrix, and mk is a vector of
∑r
k dimension. The element in the vector is computed by the number of shared data
i=1 i
set between current cluster and all of the clusters of BPs.
By using the co-occurrence matrix and the binary data set the consensus clustering
are transformed to the K-means clustering, that is:
∑r ∑k ∑ ( )
max wi U(𝜋, 𝜋 ∗ ) = min f xl(b) , mk (14)
i=1 k=1 xl∈Ck
get the weighted binary dataset X (b)′, by running K-means on weighted binary dataset
X (b)′, we can get final consensus result 𝜋 ∗.
4 Experimental Results
We present experiment on UCI dataset iris. The normalized Rand index (Rn) [6] is
adopted. Its value usually range between [0,1]. The higher value, indicate that the higher
quality of clustering. We demonstrate the cluster validity of WIKCC by comparing it
with two well-known consensus clustering algorithms the K-means-based algorithm
(KCC) [2], the hierarchical algorithm (HCC) [5].
We run K-means algorithm 100 times with the initialized number of clusters randomized
√
within [K, n] to generate 100 basic partitionings (BPs) for consensus clustering; K is
the true class of data set, n is the number of the instances, the squared Euclidean distance
is used for the distance function, the quality of each BPs is measured by Rn, the distri‐
bution of quality of BPs is shown as Fig. 2.
30
25
20
BPs
15
10
0 Rn
0 0.2 0.4 0.6 0.8
iris
As can be seen in Fig. 2, the distribution of the clustering quality of the BPs show
that there is a large proportion of BPs with poor quality, but only quite a small proportion
of BPs with relatively high quality. This shows that the incorrect pre-specified number
of classes will lead to weak clustering result.
In order to In order to determine a suitable number of BPs for WIKCC, we explore the
influence of the number of BPs on consensus clustering. In the above experiment, r BPs
have generated, and r = 100. We randomly select a part of BPs to obtain the subset
∏r
, with r = 10, 20, …, 90. For each r we do KCC [2] algorithm 100 times to get 100
Weight-Improved K-Means-Based Consensus Clustering 51
consensus clustering result. The distribution of the quality of consensus clustering result
for different subset is shown as Fig. 3.
0.9
0.8
0.7
n
R
0.6
0.5
0.4
10 20 30 40 50 60 70 80 90
The number of BPs
iris
As shown in Fig. 3, when r < 50, the quality of the consensus result present increasing
trend with the increase of r, but when r > 50 the result fluctuate in a mall range and
nearly tend to be stable, it imply that 50 may be the appropriate number of BPs for
WIKCC. Based on above exploration we chose the BPs with the quality of the top 50
BPs for WIKCC.
We compare the WIKCC with KCC and HCC, we choose top 50 better BPs for each
method and run on the iris dataset for 10 times.
0.85
0.8
0.75
n
R
0.7
0.65
WIKCC
0.6 KCC
HCC
0.55
0 2 4 6 8 10
We can see in Fig. 4. The WIKCC shows significantly higher than KCC, and outper‐
forms better than the HCC in term of the quality of consensus clustering. In addition,
comprising the Figs. 2 and 4, we can see that consensus clustering is much better than
almost all the basic clustering result obtained by K-means, this indicates that, the
consensus clustering method can find the real cluster structure more accurately than a
single traditional clustering algorithm by integrating the commonality of many basic
52 Y. Wang et al.
clustering results, so it can obtain a more stable and accurate clustering result by
ensemble multiple weak BPS.
5 Concluding Remarks
We explore the influence of the number of BPs on the consensus clustering and chose
appropriate number better BPs for WIKCC. The weight is designed by the NMI method
between two BPs based on co-occurrence matrix. Finally, the experiment on iris demon‐
strates that WIKCC outperforms the state-of-the-art well-known KCC and HCC algo‐
rithms in terms of clustering quality. In the future, we will explore the more other factors
that have influence on the performance of KCC, and we will consider more other factors
when designing the weights.
References
1. Strehl, A., Ghosh, J.: Cluster ensembles—a knowledge reuse framework for combining
multiple partitions. J. Mach. Learn. Res. 3(12), 583–617 (2002)
2. Wu, J., Liu, H., Xiong, H., et al.: K-means-based consensus clustering: a unified view. IEEE
Trans. Knowl. Data Eng. 27(1), 155–169 (2015)
3. Yu, Z., Luo, P., You, J., et al.: Incremental semi-supervised clustering ensemble for high
dimensional data clustering. IEEE Trans. Knowl. Data Eng. 28(3), 701–714 (2016)
4. Vinh, N.X., Epps, J., Bailey, J.: Information theoretic measures for clusterings comparison: is
a correction for chance necessary? In: Proceedings of the 26th Annual International Conference
on Machine Learning, pp. 1073–1080. ACM (2009)
5. Fred, A.L.N., Jain, A.K.: Combining multiple clusterings using evidence accumulation. IEEE
Trans. Pattern Anal. Mach. Intell. 27(6), 835–850 (2005)
6. Tan, P.-N., Steinbach, M., Kumar, V.: Introduction to Data Mining. Addison-Wesley, Reading
(2005)
A Novel Trace Clustering Technique Based on Constrained
Trace Alignment
1
Nanjing University of Aeronautics and Astronautics, Nanjing, People’s Republic of China
pwang@nuaa.edu.cn
2
Shanghai Polytechnic University, Shanghai, People’s Republic of China
{watan,aqtang}@sspu.edu.cn
3
Beihang University, Beijing, People’s Republic of China
hukai@buaa.edu.cn
1 Introduction
Process discovery is one of the most crucial process mining tasks that entails the
construction of process models from event logs of information systems [1]. The most
arduous challenge for process discovery is tackling the problem that discovery algo‐
rithms are unable to generate accurate and comprehensible process models out of event
logs stemming from highly flexible environments.
Trace Clustering is an efficient solution, which clusters the traces such that each
of the resulting clusters corresponds to coherent sets of cases that can each be
adequately represented by a process model [3]. Figure 1 shows the basic procedure for
trace clustering.
Nevertheless, most currently available trace clustering techniques are not precise
enough due to the indiscriminate treatment on the activities captured in traces. As a
result, the impacts of some important activities are reduced and some typical process
information may be distorted or even lost during comparison.
To address the drawback, this paper presents a novel similarity measurement based
on constrained traces alignment. First, some typical causal sequences that reflect the
“backbone” of process are identified. Then, these sequences are exploited as constraints
to guarantee the priority of important activities in traces. Subsequently, we suggest two
clustering strategies that agree with the process mining perspective. The agglomerative
hierarchical clustering (AHC) was selected for its embedded flexibility on abstraction
level to provide us an overall insight into the complex process. And the spectral clus‐
tering has a good recommendation about the number of clusters corresponding to the
generic abstraction level.
In brief, this work contributes by proposing a novel constrained trace similarity
measurement to guarantee the priority of important process episodes and subsequently
adapting two appropriate clustering techniques into process mining perspective. In
addition, experiments on real-life logs prove the improvements achieved by our method
relative to six existing methods.
The rest of the paper is organized as follows: Sect. 2 provides a brief overview of
related works. Next, Sect. 3 introduces our novel constrained trace similarity measure‐
ment and the process-adaptive clustering strategies we selected. And Sect. 4 discusses
the experiment results. Finally, Sect. 5 draws conclusions and spells out directions for
future work.
A Novel Trace Clustering Technique Based on Constrained Trace Alignment 55
2 Related Work
The main distinction between trace clustering techniques is the clustering bias (distance/
similarity measures) they proposed. Existing approaches in literature can be classified
into two major categories.
3 Approach Design
⎧ |a > b| − |b > a|
⎪ | L | | L | if a ≠ b
| a > b | + | b > a| + 1
|a ⇒L b| = ⎪ | | | |
| ⎨
L L
| |a > a| (2)
⎪ | L | if a = b
⎪ | a > a| + 1
⎩ | L |
length(CLCS(σi , σj , Ci,j ))
Sim(σi , σj ) = (3)
max(length(σi ), length (σj ))
4 Experiments
We used the ProM1 framework which has been developed to support process mining
algorithms to perform the experiments. The data is from Dutch Financial Institute2. And
we adopted the HeuristicsMiner to derive the process model as it has the best capability
to deal with real-life event logs. The approaches to compare with are presented as
follows: DWS Mining [5], Trace Clustering [13], GuideTree Miner [2, 3], Sequence
Clustering [6] and ActiTraC [15].
1
http://www.processmining.org.
2
http://www.win.tue.nl/bpi/doku.php?id=2012:challenge.
60 P. Wang et al.
We evaluate the results with respect to their model complexity, as they are measured
by a comprehensive list of metrics reported in [12]:
We made comparisons with different number of clusters for three different abstraction
levels. Here, level 1 represents the original trace set, i.e. there is only one cluster. Level
2 stands for 2/3/4 clusters while level 3 contains 5/6/7 clusters. We calculated |A| by
taking their corresponding nodes weighted average and the same as |N|.
The aggregated results are presented in Table 1. All the data has been depicted to
the radarplots in Fig. 5. We can see that all cluster techniques lead to models with lower
complexity than the original log file. However, the DWS, the ActiTraC and the Trace
Clustering approaches lead to clusters whose models have higher density values than
the unclustered one though they perform well in the other metrics. The smaller area and
more balanced capabilities shown in the radarplots from two abstraction levels proved
the effectiveness of our constraints.
Moreover, Fig. 6 depicts Constrained Trace Clustering at different levels. With the
increasing number of clusters, there is only a little improvement in all aspects. Consid‐
ering the extra elaboration on more clusters, it is inefficient and meaningless to set the
number of clusters to a higher value. This is consistent with the spectral clustering. Just
as the eigenvalues scatterplot shown in Fig. 7, only the first three clusters are well sepa‐
rated. Therefore, the spectral clustering can guide us to correctly capture the right level
of abstraction by providing a good recommendation about the number of clusters instead
of the iterations on different hierarchies.
Acknowledgment. This work is supported in part by the National Natural Science Foundation
of China under Grant No. 61672022, Key Disciplines of Computer Science and Technology of
Shanghai Polytechnic University under Grant No. XXKZD1604, the Fundamental Research
Funds for the Central Universities and Foundation of Graduate Innovation of Shanghai
Polytechnic University, and Foundation of Graduate Innovation Center in NUAA under Grant
No. kj20161601.
References
6. Ferreira, D.R.: Applied sequence clustering techniques for process mining. In: Handbook of
Research on Business Process Modeling, pp. 492–513 (2009)
7. García BañUelos, L., Dumas, M., La Rosa, M., De Weerdt, J., Ekanayake, C.C.: Controlled
automated discovery of collections of business process models. Inf. Syst. 46, 85–101 (2014)
8. Greco, G., Guzzo, A., Pontieri, L., Sacca, D.: Discovering expressive process models by
clustering log traces. IEEE Trans. Knowl. Data Eng. 18, 1010–1027 (2006)
9. Günther, C.W., van der Aalst, W.M.P.: Fuzzy mining – adaptive process simplification based
on multi-perspective metrics. In: Alonso, G., Dadam, P., Rosemann, M. (eds.) BPM 2007.
LNCS, vol. 4714, pp. 328–343. Springer, Heidelberg (2007). https://doi.org/
10.1007/978-3-540-75183-0_24
10. Hastie, T., Friedman, J., Tibshirani, R.: The Elements of Statistical Learning, vol. 167.
Springer, New York (2009). https://doi.org/10.1007/978-0-387-21606-5
11. von Luxbur, U.: A tutorial on spectral clustering. Stat. Comput. 17, 395–416 (2007)
12. Reijers, H.A., Mendling, J.: A study into the factors that influence the understandability of
business process models. Trans. Sys. Man Cyber. Part A 41, 449–462 (2011)
13. Song, M., Günther, C.W., van der Aalst, W.M.P.: Trace clustering in process mining. In:
Ardagna, D., Mecella, M., Yang, J. (eds.) BPM 2008. LNBIP, vol. 17, pp. 109–120. Springer,
Heidelberg (2009). https://doi.org/10.1007/978-3-642-00328-8_11
14. Tsai, Y.T.: The constrained common subsequence problem. Inf. Process. Lett. 88, 173–176
(2003)
15. Weerdt, J.D., Broucke, S.V., Vanthienen, J., Baesens, B.: Active trace clustering for improved
process discovery. IEEE Trans. Knowl. Data Eng. 25, 2708–2720 (2013)
A Fast Local Image Descriptor Based on Patch
Quantization
Tian Tian1(B) , Fan Yang1 , Kun Zheng2 , Hong Yao1(B) , and Qian Gao3
1
College of Computer Science, China University of Geosciences,
No. 388 Lumo Road, Wuhan 430074, China
{tiantian,yaohong}@cug.edu.cn
2
Faculty of Information Engineering, China University of Geosciences,
No. 388 Lumo Road, Wuhan 430074, China
3
School of Mathematics and Statistics,
Huazhong University of Science and Technology,
No. 1037 Luoyu Road, Wuhan 430074, China
1 Introduction
As one of the vital problems in image processing and understanding, image
representation is a core research of many computer vision applications, such
as image registration, image retrieval, and object classification. This task can
be generally divided into two main steps: the extraction of feature points and
the description of these keypoints [9]. During the past decades, Scale Invariant
Feature Transform (SIFT) [11] proposed by Lowe was once a benchmark for
its excellent performance. And Speeded Up Robust Features (SURF) [2] which
exhibits much higher efficiency than SIFT has then become a de facto standard.
SURF addresses the issue of speed, but, since the SURF feature is still a
vector of floating point values, the storage and computation of features is still
expensive. Thus, a number of techniques are proposed from different aspects to
reduce memory consumption and speed up matching of features. The first direct
c Springer International Publishing AG 2018
Q. Zu and B. Hu (Eds.): HCC 2017, LNCS 10745, pp. 64–75, 2018.
https://doi.org/10.1007/978-3-319-74521-3_8
www.allitebooks.com
A Fast Local Image Descriptor Based on Patch Quantization 65
2 Methods
2.1 Image Patch Quantization
Quantization methods have been thoroughly studied in the field of signal pro-
cessing, including the coding and compression of different signals. Image quan-
tization represents the image with finite gray levels, whose idea is exactly the
same as we desire for image descriptors – represent the most primary informa-
tion with less data. Inspired by this, we attempt to use vector quantization on
image patches for descriptor construction. Quantizing an image patch refers to
the quantization of the gray scales. A scalar quantization can be understood as
a mapping of an input into a finite number of output. A straight forward way is
the uniform quantization, which divides the gray scales equally. It is the simplest
and fastest way of implementation, and the intensity mapping function for pixels
within a patch is as following:
(p(x, y) − minv) Nq
f (p) = (1)
maxv − minv + 1
where
where minv is defined in Eq. (2). CDF counts the number of pixels whose inten-
sities are below each gray level, which provides us an expression for quantity
partition. For a CDF function, it is obvious that CDF (maxv) = Npatch (pixel
number of the whole patch), and CDF (minv) = N (minv) (number of pixels
with the smallest intensity). So the intensity mapping function can be written
in two forms:
A Fast Local Image Descriptor Based on Patch Quantization 67
(CDF (i) − CDF (minv)) Nq
f (p) = (5)
CDF (maxv) − CDF (minv) + 1
(CDF (i) − N (minv)) Nq
= (6)
Npatch − N (minv) + 1
50 50
400
350
40 40
cumulative pixel numbers
300
pixel numbers
pixel numbers
30 250 30
200
20 20
150
100
10 10
50
0 0 0
50 100 150 200 50 100 150 200 50 100 150 200
pixel intensity pixel intensity pixel intensity
Once the intensity mapping function is given, we can obtain a matrix by mapping
all pixels into new values within the patch. Then by vectorizing it, we get the
feature vector as the patch descriptor. Our descriptors take the form of integers,
within the range of 0 to Nq , where Nq equals the number of quantization levels.
A diagram shown as Fig. 2 depicts the descriptor construction process.
The main parameters our descriptors may concern should be the quantization
levels and patch size. Too few levels will result in a loss of feature distinctiveness,
but an excess of levels will also overlook the intention of feature extraction.
As for the size of an image patch, a proper size is also demanded. Oversize
patches will no longer contribute to description effectiveness but cause a waste on
computation power. Usually, parameters are always trade-offs between precision
and recall, accuracy and efficiency. We provide some experiments on parameter
68 T. Tian et al.
Quantization
selection in Sect. 3.2, from which, some empirical parameters can be concluded
for practical uses.
We discover the descriptor quantized via CDF is actually a order-
permutation-based method, which should be invariant to monotonic illumination
changes [17]. As we know in combinatorics, permutation is a bijective mapping
of a finite set onto itself. It can be understood as a sequence involving every
element in the finite set which appears only once. We use X = (x1 , x2 , ..., xn )
to denote the vector of an image patch, and vectorization is arbitrary (e.g, row-
major) but fixed for all patches. Let Z = (0, ..., 0, 1, ...1, ..., Nq − 1, ..., Nq − 1) be
a vector associated with X, where Z contains all the quantized values of each
element (pixel intensity) in X. Z is a fixed vector and we define a bijective map-
ping or action π to denote the corresponding relationship between X and Z via
the mapping function in Eq. (6). From the CDF-based quantization process, we
know that this action basically concerns the orders of the elements rather than
the values. Hence, the descriptor D is a permutation on Z, where π −1 is applied
on Z. As we imagine, the descriptor based on patch quantization via CDF is
promising to be robust to monotonic changes.
3 Experiments
In this section, we compare our method with several competing descriptors, and
a real-world dataset that is publicly available is employed in the experiments.
Parameters of the quantization-based descriptors are discussed, and performance
evaluation including recognition rates and computation time is studied.
2.93 GHz i7 CPU and Windows 7 Professional 64-bit OS, since OpenCV provides
a convenient implementation of most state-of-the-art features.
Although our descriptor can be computed on arbitrary kinds of keypoint
detector, we first carry out the comparison on SURF points. And in the sit-
uation that considers scaling and rotation correction, we would rather employ
the AGAST [12] detector that BRISK feature uses for its fast speed. After key-
points are extracted from each images, 1000 points are selected from them based
on their rankings of Harris corner measure [7]. We evaluate the performance of
descriptors using the recognition rate as BRIEF does. As for the strategy to
determine what is a correct match, we choose the Nearest Neighbor Distance
Ratio (NNDR) [13]. This strategy is more strict than the Nearest Neighbor one,
and produce less but more accurate correspondences. The distance ratio is set
as 0.9 in all the experiments.
We first test the impact of parameters in our methods in order to find a group of
well-performed parameters. Figure 4 plots the variations of patch size and num-
ber of quantization levels when the other parameter is fixed. As the patch size
increases, recognition rate of image pair matching is improving explicitly. The
curves increase as an exponential function and approach the upper limit when
the patch size is above 16. Hereafter, larger patch size will not contribute to an
obvious improvement on performance but bring in more computation require-
ment. Results of Quantization levels have the similar findings as the former
parameter. Recognition rate increases rapidly when the quantization level grows
under 8, but starts to level off with slight fluctuations after that. To sum up,
we consider a patch of 16 × 16 which is quantized to 8 levels is a good choice of
empirical parameters. Therefore, we use this group of parameters in the latter
tests against other descriptors.
70 T. Tian et al.
100 96
90 94
92
80
90
70
recognition rates(%)
recognition rates(%)
88
60
86
50
84
40
82
30
80
Q1 Q1
20 78
Q2 Q2
10 76
0 10 20 30 40 50 0 5 10 15 20 25 30 35
patch size number of quantization levels
(a) (b)
recognition rates(%)
recognition rates(%)
40 80
30 60
20 40
10 20
0 0
1|2 1|3 1|4 1|5 1|6 1|2 1|3 1|4 1|5 1|6
image pairs image pairs
Leuven−light Ubc−compression
100 100
recognition rates(%)
recognition rates(%)
80 80
60 60
40 40
20 20
0 0
1|2 1|3 1|4 1|5 1|6 1|2 1|3 1|4 1|5 1|6
image pairs image pairs
Bikes−blur Tree−blur
80 80
recognition rates(%)
60 recognition rates(%) 60
40 40
20 20
0 0
1|2 1|3 1|4 1|5 1|6 1|2 1|3 1|4 1|5 1|6
image pairs image pairs
Graf−viewpoint Wall−viewpoint
80 100
recognition rates(%)
recognition rates(%)
80
60
60
40
40
20
20
0 0
1|2 1|3 1|4 1|5 1|6 1|2 1|3 1|4 1|5 1|6
image pairs image pairs
Leuven−light Ubc−compression
100 100
recognition rates(%)
recognition rates(%)
80 80
60 60
40 40
20 20
0 0
1|2 1|3 1|4 1|5 1|6 1|2 1|3 1|4 1|5 1|6
image pairs image pairs
Bikes−blur Boat−zoomrotation
80 100
recognition rates(%)
recognition rates(%)
80
60
60
40
40
20
20
0 0
1|2 1|3 1|4 1|5 1|6 1|2 1|3 1|4 1|5 1|6
image pairs image pairs
0.9
0.8
0.7
0.6
NCC
0.5
0.4
0.3
0.2
SURF
0.1 BRISK
Q2
0
0 50 100 150 200
Patches#
good performance under slight noise and begins to degrade gradually when noise
keeps increasing.
Finally we provide the time efficiency estimation of the proposed method.
Time complexity and CPU running time are included as the evaluation of time
analysis. The uniform quantization can be computed in linear time with no
doubt, and no extra space is need either. And the CDF-based quantization
requires an auxiliary array to store the CDF function, which is also O(n) in
time complexity. Merely, it operates two linear-time computations, one for the
CDF and the other for the mapping, hence it should be more time-consuming
than the uniform one. Table 1 lists the description and matching time of different
descriptors on image pair Wall 1-2. For each image, top 1000 points are selected
according to Harris measure from SURF keypoints, and time shown in the table
is the average computation time per point. For a fair assessment, only descrip-
tion time is compared here and all the descriptors except SURF do not take
scale and rotation into account in the description process. It is seen from the
table that the Q1 quantization method shows a definite superiority on descrip-
tion time consumption. Though Q2 requires more time than the former, it shares
a similar rank with BRIEF. In general, our methods perform competitively well
as state-of-the-art features and require very little computation time.
100
90
80
70
recognition rates(%)
60
50
40
30
20 SURF
BRISK
10 Q2
0
0.001 0.002 0.003 0.004 0.005 0.01 0.02
sigma
4 Conclusion
In this paper, we present a simple quantization-based description method for
local image patch. More detailed, we propose two different ways to carry out
quantization based on uniform thresholds on either gray scales or pixel quan-
tities. Generally, descriptors using both quantization methods have very good
time efficiency, and perform well on texture scene, illumination changes, com-
pression distortions and image blurring. Specifically, the CDF-based descriptor
requires more computation time yet yields better results. We think it is a fea-
sible attempt to introduce quantization into fast descriptor building besides bit
sampling and comparison of BRIEF-like descriptors and intensity permutation
of LUCID-like descriptors. The quantization-based methods have similar advan-
tages as permutation-based descriptors and consume as little computation power
as binary ones. Our method can be applied as an alternative of SURF-like and
BRIEF-like descriptor under the circumstance of real-time requirement, lim-
ited computation power and minor image deformation. Future work may aim
at developing the robustness to undergo greater deformation and warping, and
binarized construction methods utilizing quantization is worthy to be discovered
for better storage efficiency.
References
1. Alahi, A., Ortiz, R., Vandergheynst, P.: FREAK: fast retina keypoint. In: 2012
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 510–
517. IEEE (2012)
2. Bay, H., Tuytelaars, T., Van Gool, L.: SURF: speeded up robust features. In:
Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006, Part I. LNCS, vol. 3951,
pp. 404–417. Springer, Heidelberg (2006). https://doi.org/10.1007/11744023 32
3. Bjornson, E., Ottersten, B.: Post-user-selection quantization and estimation of cor-
related frobenius and spectral channel norms. In: IEEE 19th International Sympo-
sium on Personal, Indoor and Mobile Radio Communications (PIMRC 2008), pp.
1–6. IEEE (2008)
4. Calonder, M., Lepetit, V., Strecha, C., Fua, P.: BRIEF: binary robust independent
elementary features. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010,
Part IV. LNCS, vol. 6314, pp. 778–792. Springer, Heidelberg (2010). https://doi.
org/10.1007/978-3-642-15561-1 56
5. Chandrasekhar, V., Takacs, G., Chen, D.M., Tsai, S.S., Reznik, Y., Grzeszczuk,
R., Girod, B.: Compressed histogram of gradients: a low-bitrate descriptor. Int. J.
Comput. Vis. 96(3), 384–399 (2012)
A Fast Local Image Descriptor Based on Patch Quantization 75
6. Gionis, A., Indyk, P., Motwani, R., et al.: Similarity search in high dimensions via
hashing. In: VLDB, vol. 99, pp. 518–529 (1999)
7. Harris, C., Stephens, M.: A combined corner and edge detector. In: Alvey Vision
Conference, Manchester, UK, vol. 15, p. 50 (1988)
8. Hua, G., Brown, M., Winder, S.: Discriminant embedding for local image descrip-
tors. In: IEEE 11th International Conference on Computer Vision (ICCV 2007),
pp. 1–8. IEEE (2007)
9. Huang, Z., Kang, W., Wu, Q., Chen, X.: A new descriptor resistant to affine
transformation and monotonic intensity change. Comput. Vis. Image Underst. 120,
117–125 (2014)
10. Leutenegger, S., Chli, M., Siegwart, R.Y.: BRISK: binary robust invariant scalable
keypoints. In: 2011 IEEE International Conference on Computer Vision (ICCV),
pp. 2548–2555. IEEE (2011)
11. Lowe, D.G.: Object recognition from local scale-invariant features. In: The pro-
ceedings of the seventh IEEE International Conference on Computer Vision, vol.
2, pp. 1150–1157. IEEE (1999)
12. Mair, E., Hager, G.D., Burschka, D., Suppa, M., Hirzinger, G.: Adaptive and
generic corner detection based on the accelerated segment test. In: Daniilidis, K.,
Maragos, P., Paragios, N. (eds.) ECCV 2010, Part II. LNCS, vol. 6312, pp. 183–196.
Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15552-9 14
13. Mikolajczyk, K., Schmid, C.: A performance evaluation of local descriptors. IEEE
Trans. Pattern Anal. Mach. Intell. 27(10), 1615–1630 (2005)
14. Ni, Z.S.: B-SIFT: a binary sift based local image feature descriptor. In: 2012 Fourth
International Conference on Digital Home, pp. 117–121. IEEE (2012)
15. Oliva, A., Torralba, A.: Modeling the shape of the scene: a holistic representation
of the spatial envelope. Int. J. Comput. Vis. 42(3), 145–175 (2001)
16. Rublee, E., Rabaud, V., Konolige, K., Bradski, G.: ORB: an efficient alternative
to SIFT or SURF. In: 2011 IEEE International Conference on Computer Vision
(ICCV), pp. 2564–2571. IEEE (2011)
17. Singh, M., Parameswaran, V., Ramesh, V.: Order consistent change detection via
fast statistical significance testing. In: IEEE Conference on Computer Vision and
Pattern Recognition (CVPR 2008), pp. 1–8. IEEE (2008)
18. Tian, T., Sethi, I., Ming, D., Zhang, Y., Ma, J.: LIPID: local image permutation
interval descriptor. In: 2013 12th International Conference on Machine Learning
and Applications (ICMLA), vol. 2, pp. 513–518. IEEE (2013)
19. Torralba, A., Fergus, R., Weiss, Y.: Small codes and large image databases for
recognition. In: IEEE Conference on Computer Vision and Pattern Recognition
(CVPR 2008), pp. 1–8. IEEE (2008)
20. Tuytelaars, T., Schmid, C.: Vector quantizing feature space with a regular lattice.
In: IEEE 11th International Conference on Computer Vision (ICCV 2007), pp.
1–8. IEEE (2007)
21. Wang, C., Duan, L.Y., Wang, Y., Gao, W.: PQ-WGLOH: a bit-rate scalable local
feature descriptor. In: 2012 IEEE International Conference on Acoustics, Speech
and Signal Processing (ICASSP), pp. 941–944. IEEE (2012)
22. Ziegler, A., Christiansen, E., Kriegman, D., Belongie, S.J.: Locally uniform com-
parison image descriptor. In: Advances in Neural Information Processing Systems,
pp. 1–9 (2012)
A Method of Large - Scale Log Pattern Mining
1
School of Electronic Engineering, Beijing University of Posts and Telecommunications,
Beijing 100876, China
SallyLi0863@163.com
2
Institute of Network Technology, Beijing University of Posts and Telecommunications,
Beijing 100876, China
1 Introduction
At present, there are several methods to deal with log, and methods based on the
pattern are widely used in log analysis. In this method, one raw log message can be
divided into two parts: the constant part and the variable part. For telecommunication
network equipment log, the variable part contains a lot of valid information, such as the
location of the module that issued the log, the actions performed by the operator, and
the time of the log. However, when the volume of log increases to a certain extent, due
to the huge variable data, despite using this method, the results will take up a lot of
storage space.
Therefore, we investigated the various log preprocess methods and for the charac‐
teristics of the network device logs, we have taken certain ways based on the natural
language model, making complex log becomes more suitable for storing and mining. In
order to validate our method, we used it and several other typical log parser to compare
the result in the test set of different kinds of logs and the real network device data set,
which proves the accuracy and efficiency of our method.
2 Log Parser
Parser
For example, in [5], the author proposes a STE method to judge whether it is a log
template word or a variable word based on the location and length of the word in the
log text, which can determine the log pattern.
This is the usual method for log analysis, and its speed is faster and performance is
better, but because the device log parameters varies and the data format is very irregular,
leading to the poor quality and large redundancy for the pattern set, which has a great
impact on the effect of the data mining work.
In order to verify the performance of our parser, we chose three typical parsers to
compare with ours. And these parser’s source code can be find in the [1].
2.2.1 LKE
This method is proposed to parse free-form text log for anomaly detection in distributed
systems, and it is made up by the following steps: (1) remove the parameters according
A Method of Large - Scale Log Pattern Mining 79
to the established rules (2) measure raw log similarity by the weighted edit distance (3)
cluster similar raw log keys together (4) Log template generation [2].
2.2.2 IPLoM
This method is proposed for automatic event log analysis, which includes three step
hierarchical partitioning process: (1) Partition by log length. The first step is to use the
token count heuristic to partition the log messages, because the log messages that have
the same line format are likely to have the same token length (2) Partition by token
position. By counting the words in the same position, the method will sort the log by
the count. (3) Partition by search for mapping. By searching for mapping relationships
between the set of unique tokens in two token positions, the log is divided into smaller
partition. (4) Log template generation [3].
2.2.3 LogSig
To understand and optimize system behaviors, LogSig is proposed to generate system
events from textual log messages. LogSig works in three steps: (1) Word pair generation.
Raw log data are converted to a set of word pairs to record both the word and its position.
(2) Log Clustering. Based on the word pairs, a potential value is calculated for each log
message to decide which cluster the log message potentially belongs to. (3) Log template
generation. In each cluster, the log messages are leveraged to generate a log template
[1, 4].
From the results we can see that for the small data set of log data, IPLoM algorithm
show a great advantage in speed, our algorithm is running faster than the other two
parsers. In terms of accuracy, our parser has an advantage in the analysis of certain logs
(Figs. 3 and 4).
myparser
IPLoM
LKE
LogSig
2500
2000
1500
1000
500
0
SOSP BGL HPC proxifier zookeeper
We listed the running time of each parser, where LKE was unable to perform log
parsing due to memory overflow, and we could see that IPLoM still had a great advantage
in speed, but when there was no standard regular expressions, its results contain a lot of
redundancy. Our parser will give less redundant and more accurate results within a
tolerable time range.
2.4.3 Summary
Comparing our performance with the other three classic parsers, we found that although
our parser was not as fast as IPLoM, it showed good adaptability and speed when dealing
with a large number of telecommunications device logs data. After that, we have carried
out further data mining to understand the data and find the information in the log data.
3 Log Analysis
3.1 Background
3.1.1 Word Vector
To apply the machine learning method to the natural language processing field, the most
basic problem is the representation of the language symbol. So far, the most commonly
used method for natural language processing is One-hot Representation, which means
that n words are n-dimensional vectors, each vector is 1 in a dimension and the other
dimension is zero. However, this method will cause the lexical gap problem, there is no
connection between words and words.
Therefore, the Distributed Representation method is proposed, that is, using low-
dimensional real vector to represent vocabulary. The biggest advantage of Distributed
Representation is that it can make meaning-related words relatively similar in distance.
At the same time, the word vector will show many special properties, as shown below
(Fig. 5).
The third layer of the network is represented by the node Yi, using the softmax
function to normalize the output value into probability, and the final Y is calculated as:
Y = b + Wx + U * tanh(d + Hx)
U is the parameter from the hidden layer to the output layer, the majority of the model
compute operation is centered on the matrix multiplication of the U and hidden layers.
Finally, we use the stochastic gradient descent method to optimize the model, then we
can get word vector [8].
A Method of Large - Scale Log Pattern Mining 83
3.1.3 Word2vec
Word2vec is a tool launched by Google to calculate words vector, which has been gained
comprehensive attention because of its efficiency and convenience. It is based on the
neural network and the natural language model. By the relationship between words and
sentences, it can calculate a word vector for each word, and we can compare the word
similarity by the distance between the word vectors. Based on the principle of word2vec,
we have present a log mining method that can get literally similar logs or logs containing
a log of important links.
In the results, we can find that some similar log patterns produced by word2vec have
literal similarity, which means that word2vec can help us to optimize the log parsing
and find else log patterns need to be classified except which has the different digital
parameters or whose edit distance is less than a certain value. At the same time, word2vec
84 L. Li et al.
can help us discover log patterns that are literally unrelated but have similar meanings
or links, which is of great importance to subsequent log analysis.
4 Conclusion
Our algorithm uses the neural network language model for the first time to analyze the
telecommunication equipment log, and obtains the similarity pattern. At the same time,
we design the log analysis method to adapt to the log of the telecommunication equip‐
ment, and verify the effectiveness of the method by experiment.
Through the experiment and the comparison of the results, we can find that our parser
obtains a better analytical effect for the telecommunications server log data in the toler‐
able time. Subsequent analysis, whether using word2vec for similar patterns discovery,
or the use of other data mining methods to explore, such as abnormal point recognition
and correlation analysis, can be based on our processing results for analysis.
However, our analytical methods are also deficient, for example, we can find patterns
that have similar characteristics in the order of occurrence, but how these patterns are
applied specifically to some telecommunications systems problem, such as system error
prediction, auto log analysis system without expert, we also need to continue exploring
and researching.
References
1. He, P., et al.: An evaluation study on log parsing and its use in log mining. In: IEEE/IFIP
International Conference on Dependable Systems and Networks. IEEE Computer Society, pp.
654–661 (2016)
2. Fu, Q., Lou, J., Wang, Y., Li, J.: Execution anomaly detection in distributed systems through
unstructured log analysis. In: Proceedings of International Conference on Data Mining, ICDM
2009 (2009)
3. Makanju, A., Zincir-Heywood, A., Milios, E.: Clustering event logs using iterative
partitioning. In: Proceedings of International Conference on Knowledge Discovery and Data
Mining, KDD 2009 (2009)
4. Tang, L., Li, T., Perng, C.: LogSig: generating system events from raw textual logs. In:
Proceedings of ACM International Conference on Information and Knowledge Management,
CIKM 2011, pp. 785–794 (2011)
5. Kimura, T., lshibashi, K., Mori, T., Shiomoto, K.: Spatio-temporal factorization of log data
for understanding network events. In: INFOCOM 2014 Proceedings. IEEE (2014)
6. Juneja, P., Kundra, D., Sureka, A.: Anvaya: an algorithm and case-study on improving the
goodness of software process models generated by mining event-log data in issue tracking
systems. Support. Care Cancer 6(6), 539–541 (2015)
7. Hamooni, H., Debnath, B., Xu, J., et al.: LogMine: fast pattern recognition for log analytics.
In: CIKM (2016)
8. Bengio, Y., et al.: A neural probabilistic language model. J. Mach. Learn. Res. 3(6), 113 (2003)
More Efficient Filtration Method for Big Data
Order-Preserving Matching
Wenchao Jiang ✉ , Dexi Lin, Sui Lin, Chuanjie Li, and Aobing Sun
( )
Abstract. Data matching and retrieval aims at finding out similar substrings with
the pattern P in the given data set T. This problem has wide applications in big
data analysis. A liberalized verification rule is proposed first, and then a similarity
computing based order preserving matching method is presented. Theory analysis
indicates our method runs in linear. Furthermore, the experimental results show
that our method can improve effectively the precision ratio and the recall ratio.
More qualified matching results can be detected compared with the state of the
art of this problem.
1 Introduction
Fast and accurate data matching and retrieval is one of the key problems in big data
applications such as video retrieval, stock analysis and prediction. Based on the context
and application scenarios, the data objections can be abstracted into a series of vectors
with different properties. Furthermore, the different properties can be integrated into a
number through reduction or conversion. Consequently, data matching problem will be
transformed into string or number matching problem which is one kind of well known
problems in pattern recognition. Given a set of numbers T of length n and a pattern P of
∑
length m, both being numbers or strings over a finite alphabet , the task of string
matching is to find all the substrings u in T which have the same relative order as P, and
|u| = |P|. For example, let P = (10, 22, 15, 30, 20, 18, 27) and T = (22, 85, 79, 24, 42,
27, 62, 40, 32, 47, 69, 55, 25), then the relative order of P matches the substring u = (24,
42, 27, 62, 40, 32, 47) of T [1].
Several online [3–7] and one offline solution [2] have been proposed for the string
matching problem. Kim et al. [3] and Kubica et al. [4] presented solutions based on the
Knuth-Morris-Pratt algorithm (KMP) [8], the KMP algorithm is mutated such that is
determines if the text contains substring with the same relative order as that of the pattern
using the order-borders table. Kim et al. [3] utilized the prefix representation method to
find the rank of each number in the prefix, and this method was further optimized using
the nearest neighbor representation to overcome the overhead involved in computing
the rank function. Later, Cho et al. [5, 6] gave a sublinear solution based on the bad
character heuristic of the Boyer-Moore algorithm [9]. Almost at the same time,
Belazzougui et al. [7] derived an optimal sublinear solution. The state of the art for this
problem is the filtration method [1] which is presented by Chhabra et al. All the verifi‐
cation rules in previous researching demanded the values and the positions of the
numbers in both T and P must be coherent strictly. So, some actual matching results
may be discarded. Moreover, almost all the earlier researching were focused on the time
complexity analysis while ignored the precision ratio analysis and recall ratio analysis.
They didn’t research whether some actual matching results were missing despite the
algorithm became more and more fast.
In view of the above problem, a similarity computing based string or number order
preserving matching method is presented. Based on the preprocessing, i.e. binary trans‐
forming and filtration, our method proposes a novel verification rule which can guarantee
more candidate results can be found out. Then, a similarity computing based sorting
method is presented which can ensure all the matching results are listed according to
the similarities with the pattern P. Theory analysis indicates our method is sublinear.
The experimental results show that our method can improve effectively the precision
ratio and the recall ratio compared with the newest method at present.
This paper is organized as follows. Section 2 presents our solution. Section 2.4
analyses our approach. The experimental results are given and discussed in Sect. 3.
Section 4 concludes this paper.
2 Our Solution
The state of the art for order-preserving matching is the filtration method [1] which was
presented by Chhabra et al. We call the filtration method as MT.C in this paper. The
MT.C transform the original data T and the pattern P into binary string T’ and P’
according to formulation (1). Then searching for the substring with the same relative
order with P in T can be transformed into searching P’ in the analogously T’. In the
above example, P’ = 101001 and T’ = 100101001100. Each occurrence is a match
candidate which is verified following the numerical order of the positions in the original
pattern P.
{ {
1, ti−1 < ti 1, pi−1 < pi
ti = , ti ∈ T ; pi = , pi ∈ P (1)
0, ti ≤ ti+1 0, pi ≤ pi+1
The MT.C method made the matching process simpler and more efficient than the
earlier solutions. However, some deficiencies could be found because of the too harsh
filtration and verification rules. The MT.C method required the order of the numbers is
strictly coherent according to the values and the corresponding positions in P and T. For
example, as shown in Fig. 1(a), the MT.C method can’t find the differences between T
and P if the max-number increases or the mini-number decreases immensely in T.
Moreover, as shown in Fig. 1(b), the MT.C method can’t find the differences either if
More Efficient Filtration Method for Big Data 87
certain subsection data in T jumps suddenly while the variation trend maintains as a
whole. Furthermore, to find out the matching substrings from the candidate strings which
were produced through filtration algorithm, the numbers of P must be sorted, and the
verification processing demands the positions and the size relations between T and P
must be coherent strictly. This rule would result in the looseness of some actual matching
substrings. As shown in Fig. 1(c), according to the MT.C method, the gray node x in T
can only change between dashed line a and dashed line b which represent the right
neighbor y and the left neighbor z respectively in the sorting of T. Once the gray node
x changes above the dashed line a or down the dashed line b, the MT.C method will
regard that T is not matching with P. Then the corresponding substring will be discarded
consequently. Actually, we can find that this is not the case especially when |y-z| is small
enough. For example, when x changes to x’ which is slightly larger than y or changes
to x” which is slightly smaller than z, the MT.C method will discard the substring. But
we think the substring is still similar with P in most actual applications such as data
retrieval.
ui ≤ uj ⇔ vi ≤ vj for 1 ≤ i, j ≤ m. (2)
This rule demands all numbers in u and v are strictly coherent with both the sizes
and their positions. So, some actual matching results would be discarded as shown in
Fig. 1(c). To overcome this deficiency, a different rule is presented as following. Two
∑
strings u = u1u2…um and v = v1v2…vm of the same length over . The numbers of
v = v1v2…vm are sorted firstly and the result is a sequential table r as formulation (3).
|u − ur[i]−1 ||
|u − u | | r[i]+1
| r[i] r[i]−1 | ≤ or
2
|u |
− ur[i]−1 | (4)
|ur[i]+1 − ur[i] | ≤ | r[i]+1 for vr[i]−1 ≤ vr[i] ≤ vr[i]+1 , 1 ≤ i ≤ m.
| | 2
Two strings u and v are called order-isomorphic, written u ≈ v, if formulation (4)
holds. Apparently, multiple matching results would be found out according to formu‐
lation (4) once the pattern is given. But, the similar extent of the multiple matching
results are different. To distinguish the similar level of multiple candidate results, a
similarity computing method should be given based on the optimized verification
method. Considering the possible concussion range of some numbers in T which have
the same positions with relevant numbers in P. The similarity function f, as shown in
Fig. 2, between the candidate substring u and the pattern P is defined as formulation (5).
∑
f (u, P) = |g(ui ) − g(pi )|, 1 ≤ i ≤ m. (5)
Where, ui and pi are the relevant numbers which have the same position order in the
candidate substring u and the pattern P. m is the length of u and P. g is the reduction
function, shown in formulation (6), which can reduce all the numbers in u and p into the
same zone [0,1]. Min(u) and Min (P) are separately the Min-values in u and P. Max(u)
and Max(P) are separately the Max-values in u and P.
As shown in Fig. 2, the similarity between the candidate substring u and the pattern
P can be computed easily as the sum of all the differences, represented by Diff.i
(1 ≤ i ≤ n), between the relevant nodes of u and P according to formulation (5).
As described above, our solution includes four steps: binary transformation, filtration,
verification and similarity computing. The binary transformation can be processed
according to [1], and the filtration can be conducted using any exact string matching
algorithm. Supposing the binary transformation and filtration have been finished, the
More Efficient Filtration Method for Big Data 89
pseudo-code of our solution based on the optimization and the similarity computing can
be described as Algorithm 1 and Algorithm 2.
1) Xt=AnticipationBinary (Tr);
2) Xp=AnticipationBinary (P);
4) Sum+=Abs(Xt[i] - Xp[i]);
5) Return Sum.
1) intput A;
2) Max=A.Max;
3) Min=A.Min;
5) If (Max==Min);
6) B[i]=0;
7) else
8) B[i]=(A[i]-Min)/(Max-Min);
9) Return B.
90 W. Jiang et al.
Compared with [1], the main differences of our solution are the verification and
similarity computing. Furthermore, the main modification in verification is a liberalized
verification rule is implemented and more matching results can be detected. So, the time
complexity of verification doesn’t be changed and only the O(SC) need be analyzed.
Supposing the numbers in P and T are integers and they are statistically independent
of each other and the distribution of numbers is discrete uniform. Supposing LT = n,
LP = m. In the worst case, the similarity computing process requires O(m(n-m)) simi‐
larity computing operations. In most cases, LT is usually very large while LP is usually
a constant small enough. So, O(m(n-m)) ≈ O(nm) on average which is equal with MT.C
method. So, we can conclude O(MS) = O(MT.C) which are all sublinear.
3 Experimental Results
Our experiments used linear string matching algorithm KMP [8] as the filtration method.
The tests were run on single node of Tianhe-2 super computer with configuration CPU
E5-2692 v2 12*2 2.20 GHz. All the algorithms were implemented in C# in the 64-bit
mode.
3.1 Effectiveness
To explain the superiority of our method, two special data sets are generated based on
one basic data set. The basic data set was given as T = (10, 14, 12, 15, 13, 28, 36, 32,
24, 38, 26) and P = (10, 14, 12, 15, 13, 19, 21, 20, 17, 22, 18). The first special data set
shown in Fig. 3(a) was generated by multiplying n (1 < n < 100) to the last six numbers
in T. The second special data set shown in Fig. 3(b) was generated by multiplying n
(1 < n < 100) to only the 10th number in T.
More Efficient Filtration Method for Big Data 91
From Fig. 3(a) and (b), we can find that the data T and the pattern P become more
and more dissimilar with the increasing of n. But, the MT.C method couldn’t find the
differences among them. Through computing the similarities between T and P according
to the technologies presented in Sect. 2, the variation trend of the similarity between T
and P with the increasing of n was shown in Fig. 4. According to the definition of
similarity shown as formulation (5), the more bigger the value of the similarity is, the
more dissimilar the substring is compared with P. We can find that the similarity
increases with the increasing of n and becomes converging with the infinite increasing
of n. The point of inflection means that the data set T couldn’t be considered similar
with P according to our method. At the same time, we can find that the points of inflection
are different with different data sets. So there is no stable point of inflection for different
data sets in our method. The appearance of the point of inflection depend on the data set
itself.
Imitating the data generation method in [1], we generated two special data set and
one random data set. The random data set contains 100000000 random integers between
0 and 230. The lengths of patterns (LP) were picked as 10, 15 and 30. The experimental
results showed that our method can find more matching substrings than the MT.C
method. For example, when the pattern length was picked as 10, the MT.C method can
only find 15 results. However, our method can find 135 results which were sorted
according to the similarities. The first 10 most similar results were shown in Figs. 5, 6
and 7 separately when the length of pattern were picked as 10, 15 and 30. The red line
represents the pattern, and the blue lines represent the missing results using MT.C
method while are detected using our solution. The black lines represent the matching
results which can be detected using both MT.C method and our solution.
Furthermore, to compare the time consuming between our method and the MT.C
method. We enlarged the length of T from 10000000 to 49000000 gradually with an
increment of 1000000. Ten experiments were conducted and the average time
consuming was computed with the different lengths of patterns such as 5, 30 and 50.
The statistical results were shown in Fig. 8. Because our method could find more results
than the MT.C method, and more similarity computing operations were needed. So the
overall executing time appear a small amount of growth on equal conditions. But, the
executing time growth decreased quickly with the increasing of the length of the pattern.
Because the number of matching results decreased quickly with the increasing of the
length of the pattern.
4 Conclusion
Fast and accurate data matching and retrieval is a key problem in big data applications
such as video retrieval, stock analysis and bioinformatics etc. A similarity computing
based string or number order preserving matching method is presented. Theory analysis
indicates our method is sublinear. Furthermore, the experimental results show that our
method can improve effectively the precision ratio and the recall ratio compared with
the newest method at present under the same conditions. Compared with former research
works for order-preserving matching problem, our solution liberalized the verification
rules on certain extent and so more qualified matching results can be detected and found
out.
References
1. Chhabra, T., Tarhio, J.: A filtration method for order-preserving matching. Inf. Process. Lett.
116(2), 71–74 (2016)
2. Chhabra, T., Külekci, M.O., Tarhio, J.: Alternative algorithms for order-preserving matching.
In: Prague Stringology Conference (2015)
3. Chhabra, T., Giaquinta, E., Tarhio, J.: Filtration algorithms for approximate order-preserving
matching. In: Iliopoulos, C., Puglisi, S., Yilmaz, E. (eds.) SPIRE 2015. LNCS, vol. 9309, pp.
177–187. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-23826-5_18
4. Cantone, D., Faro, S., Külekci, M.O.: An efficient skip-search approach to the order-
preserving pattern matching problem. In: Prague Stringology Conference (2015)
5. Knuth, D.E., Morris, J.H., Pratt, V.R.: Fast pattern matching in strings. SIAM J. Comput.
6(2), 323–350 (1977)
6. Crochemore, M., Iliopoulos, C.S., Kociumaka, T., Kubica, M., Langiu, A., Pissis, Solon P.,
Radoszewski, J., Rytter, W., Waleń, T.: Order-preserving incomplete suffix trees and order-
preserving indexes. In: Kurland, O., Lewenstein, M., Porat, E. (eds.) SPIRE 2013. LNCS,
vol. 8214, pp. 84–95. Springer, Cham (2013). https://doi.org/10.1007/978-3-319-02432-5_13
7. Kim, J., Eades, P., Fleischer, R., et al.: Order-preserving matching. Theoret. Comput. Sci.
525(4), 68–79 (2013)
8. Boyer, R.S.: A fast string searching algorithm. Commun. ACM 20(10), 762–772 (1977)
9. Cho, S., Na, J.C., Park, K., Sim, J.S.: Fast order-preserving pattern matching. In: Widmayer,
P., Xu, Y., Zhu, B. (eds.) COCOA 2013. LNCS, vol. 8287, pp. 295–305. Springer, Cham
(2013). https://doi.org/10.1007/978-3-319-03780-6_26
10. Cho, S., Na, J.C., Park, K., et al.: A fast algorithm for order-preserving pattern matching. Inf.
Process. Lett. 115(2), 397–402 (2015)
11. Belazzougui, D., Pierrot, A., Raffinot, M., Vialette, S.: Single and multiple consecutive
permutation motif search. In: Cai, L., Cheng, S.-W., Lam, T.-W. (eds.) ISAAC 2013. LNCS,
vol. 8283, pp. 66–77. Springer, Heidelberg (2013). https://doi.org/
10.1007/978-3-642-45030-3_7
Blind Image Quality Assessment via Analysis of GLCM
1 Introduction
At present, digital images, as the carrier of massive information, have greatly enriched
people’s life as well as drastically facilitated the communication among people [1, 2].
Yet image distortion remains a stubborn problem in image transmission system. There‐
fore, it is indispensable to establish efficient methods for image quality assessment
(IQA).
Generally, IQA method can be split into two major categories: subjective and objec‐
tive assessment methods. Currently, objective IQA algorithm has been widely studied
because it is easy to implement and portability. Given the available information of the
pristine image, objective assessment method can be further classified into full-reference
IQA (FR-IQA), reduced-reference IQA (RR-IQA) and no-reference IQA (NR-IQA).
Since both FR-IQA and RR-IQA methods use information of the original reference
image, so they are limited to special situations. In this paper, we mainly focus on the
NR-IQA method.
At present, NR-IQA method can be broadly divided into two classes, i.e., training-
based opinion-aware metric and opinion-unaware metric. The former one requires a
training process to create a regression model for predicting image quality. For example,
Moorthy and Bovik [3] proposed a two-step framework that called BIQI. Specifically,
each distortion type was trained with a regression model. In such case, the distortion
type of image can be obtained through these models. Subsequently, the image statistical
properties are gradually applied into IQA and have been proved effectively. For instance,
Saad et al. [4] provided a NR-IQA algorithm under the hypothesis that the statistics
features of discrete cosine transform (DCT) coefficients change regularly along with
image quality. Although these methods have achieved meaningful performance, they
require training procedure. To tackle the problem, metrics, which don’t require human
opinion scores and any regression model, have been proposed. Xue et al. [5] used a set
of cluster centroid with quality label as a codebook to predict image quality, called QAC.
Natural image quality evaluator (NIQE) [6] established a completely blind BIQA metric
by fitting the quality-aware features to a multivariate Gaussian (MVG) model. Although
the training process is not required, their performances need to be further improved. In
this paper, we propose a new blind image quality assessment method based on training.
It should be mentioned that the above methods mainly rely on mathematical statistics
method but without full consideration of the HVS characteristics. GLCM can effectively
describe image feature by measuring statistical characteristic of image in multi-direction
and multi-scale. By considering characteristics of HVS and the variety of computing
method for GLCM, this paper presents a simple yet effective BIQA metric. Figure 1
shows the pipeline of our method. It can be divided into the following three parts: calcu‐
lation of GLCM, feature extraction and image quality prediction.
Tx = [−1 0 1] (1)
Blind Image Quality Assessment via Analysis of GLCM 97
Ty = Tx′ (2)
Gx = Tx * I (3)
Gy = Ty * I (4)
where ‘*’ denotes convolution. Finally, the gradient map G is calculated as:
|Gx | + |Gy |
G= (5)
2
Wavelet transform decomposes image into multi-scale and multi-direction. The
image is usually transformed along horizontal, vertical and diagonal directions. And
then the decomposition sub-graphs in those three directions are usually denoted by HL,
LH and HH, respectively [15]. In this paper, the wavelet decomposition scale is set to
1, which gets good results.
As image distortion always induces the structural degradation, we desire to evaluate
image quality by utilizing image structure information. Image gradient and decompo‐
sition sub-graphs are complementary to each other in representing rich image structure.
On the one hand, image gradient describes the global image structure while misses
orientation information. On the other hand, wavelet decomposition reflects image
features in different orientation, while ignores global structure. Hence, their combination
ensures integrity of the image structure information.
(2) GLCM matrix calculation
Usually, image distortion brings about a significant change of image statistic char‐
acteristics. GLCM can provide image statistic characteristics in different directions and
at different scales in spatial domain, so it can describe image characteristics from various
aspects. Based on this, in this paper, the GLCM matrixes of the above image structure
maps are calculated.
The GLCM is composed of the joint probability density between image gray tones.
There are three important parameters in GLCM: angle (θ), quantized gray tones (L) and
distance (d). Firstly, the image is quantized to L gray tones. Then, the probability of
occurrences of the pair of gray tones i and j in original image is expressed in P(i, j, d,
θ) (i = 1, 2, …, L, j = 1, 2, …, L). Each entry (i, j) is depart at a distance d in angle θ.
Finally, the GLCM can be denoted as [P(i, j, d, θ)]L×L, where P(i, j, d, θ) is the element
of [P(i, j, d, θ)]L×L in the i-th row and j-th column.
In [7], fourteen features were extracted from GLCM to represent image properties from
multiple perspectives. Currently, researchers usually used part of them in view of the
98 G. Yue et al.
redundancy among them [8]. In this paper, we employ four commonly used features,
namely contrast, energy, correlation, and homogeneity, to extract quality sensitive
features for IQA. Those four selected features involve local and global image charac‐
teristics. Among them, contrast and energy describe the overall characteristics of the
image. Specifically, contrast describes image definition. Energy reflects the image
distribution as well as roughness. On the contrary, correlation and homogeneity are local
image descriptors. Concretely, correlation illustrates the local correlation of image
grayscale. Homogeneity measures local change of image grayscale. Overall, the selected
features can reflect both local and global features of image, to a certain extent. Therefore,
they can be applied into IQA problem.
Although we have demonstrated the feasibility of GLCM in IQA problem, how to
choose the parameters, i.e., θ, L, d, is still a thorny problem. Research shows that HVS
is more sensitive to the horizontal and vertical image information than the oblique
direction [9]. Moreover, different viewing distances produce various perception for
HVS. HVS focuses on outline of image at large viewing distance, while at small distance,
it will pay attention to image details [10]. And for GLCM, small scale in GLCM can
describe characteristics of fine image structure, while large scale obtains characteristics
of rough image structure. Inspired by these, we extract GLCM in multi-direction and
multi-scale. Specifically, θ is set as 0° and 90° to highlight the sensitive direction of HVS,
the distance d is set as 1, 2, 4, and 8 for simulating the variation of viewing distance,
and L is set as 8. Since distortion also corrupts the brightness information, to avoid its
loss, we also extract the above features on distorted image. Overall, the GLCM for
gradient image, decomposed high-frequency sub-images (HL, LH and HH after one
scale wavelet decomposition) and distorted image is calculated in two directions (0° and
90°) at 4 scales, resulting in eight GLCM matrices for each calculated image. A total of
40 GLCM matrices are attained, followed by four features extraction for each GLCM
matrices.
The proposed method is tested on three public databases: LIVE II [12], TID2013 [13]
and CSIQ [14] database. In LIVE II database, we test the proposed algorithm on all of
the five distortion types, i.e., JPEG2000 compression (JP2K), JPEG compression
(JPEG), white noise (WN), Gaussian blur (Gblur) and transmission errors in the JP2K
using Fast-fading Rayleigh channel model (FF). In TID2013 and CSIQ databases, four
distortion types are tested, namely JP2K, JPEG, WN and Gblur. Three general IQA
criteria, i.e., Spearman rank order correlation coefficient (SROCC), Pearson linear
correlation coefficient (PLCC) and root-mean-squared error (RMSE), are employed for
performance evaluation. A better performance means a value close to 1 for PLCC and
SROCC while a value close to 0 for RMSE.
In order to verify the effectiveness of the proposed method, we select two public FR
algorithms (SSIM [15] and PSNR) and several mainstream NR methods (QAC [5], BIQI
[3], ILNIQE [14], GM-LOG [16] and YCLTYCbCr [17]) for comparison. For ROI-
BRISQUE, because the source code is not obtained, we directly use the experiment
results on LIVE II database provided in the original paper for comparison. Since the
proposed method is based on training, we divide the image set into two non-overlapping
image sets: training set and testing set. The training set contains 80% of the reference
images and corresponding distortion versions of them, and the testing set is comprised
by the residual images. After the random train-test split is repeated 1000 times, the
median performance is taken as the final results.
Table 1 shows the performance tested on the entire database. For better observation, the
top three performed algorithms are highlighted in bold. As we can see, the performance
of the proposed method always lies in top three. In fact, compared with those IQA
methods in Table 1, our method achieves the best performance in all three databases.
Table 1. SROCC, PLCC and RMSE (median value across 1000 train-test trials) of SSIM, PSNR,
QAC, BIQI, NIQE, ILNIQE, GM-LOG and YCLT-YCbCr on the overall database of LIVE II,
TID2013 and CSIQ respectively.
Database Metric SSIM PSNR QAC BIQI IL-NIQE GM- YCLT- Pro.
LOG YCbCr
LIVE II PLCC 0.9397 0.9122 0.8755 0.8909 0. 9011 0.9539 0.9354 0.9581
SROCC 0.9244 0.8000 0.8803 0.8899 0. 8996 0.9503 0.9348 0.9524
RMSE 7.9001 14.0604 11.1677 10.5236 12.1186 8.1723 5.9386 6.6445
CSIQ PLCC 0.9294 0. 9281 0.8645 0.9101 0.8991 0.9408 0.8980 0.9482
SROCC 0.9274 0. 8550 0.8338 0.8925 0.8854 0.9228 0.8869 0.9432
RMSE 0.1040 0.1483 0.1411 0.1198 0.1491 0.0950 0.1295 0.0890
TID2013 PLCC 0.8591 0.9149 0.8273 0.8048 0.9001 0.9439 0.8789 0.9512
SROCC 0.8291 0.9058 0.8188 0.7846 0.8714 0.9282 0.8690 0.9377
RMSE 0.7232 0.5908 0.7670 0.8254 0.6728 0.4629 0.9017 0.4293
100 G. Yue et al.
4 Conclusion
In this paper, we propose a simple yet efficient blind IQA metric based on GLCM statistic
model of image structure. To verify the performance of our proposed method, we
conducted a set of experiments on LIVE II, CSIQ and TID2013 databases. We apply it
on the entire database, and the experimental results demonstrate that our predicted scores
is more accuracy than two public FR-IQA algorithms and the mainstream NR-IQA
methods. In summary, we can draw the conclusion that the proposed method obtains
excellent performance in BIQA.
Acknowledgement. This work was supported by the National Natural Science Foundation of
China under Grant 61471262, the International (Regional) Cooperation and Exchange under
Grant 61520106002.
References
1. Gu, K., Lin, W.S., Zhai, G.T., Yang, X.K.: No reference quality metric of contrast-distorted
images based on information maximization. IEEE Trans. Cybern. 47, 4559–4565 (2016)
2. Gu, K., Zhai, G.T., Yang, X.K., Zhang, W.J.: Using free energy principle for blind image
quality assessment. IEEE Trans. Multimed. 17(1), 50–63 (2015)
3. Moorthy, A.K., Bovik, A.C.: A two-step framework for constructing blind image quality
indices. IEEE Signal Process. Lett. 17(5), 513–516 (2010)
4. Saad, M.A., Bovik, A.C., Charrier, C.: Blind image quality assessment: a natural scene
statistics approach in the DCT domain. IEEE Trans. Image Process. 21(8), 3339–3352 (2012)
5. Xue, W.F., Zhang, L., Mou, X.Q.: Learning without human scores for blind image quality
assessment. In: Computer Vision and Pattern Recognition, pp. 995–1002 (2013)
6. Mittal, A., Soundararajan, R., Bovik, A.C.: Making a completely blind image quality
analyzer. IEEE Signal Process. Lett. 20(3), 209–212 (2013)
7. Haralick, R.M., Shanmugam, K., Dinstein, I.: Textural features for image classification. IEEE
Trans. Syst. Man Cybern. smc-3(6), 610–621 (1973)
8. Gadelmawla, E.S.: A vision system for surface roughness characterization using the gray
level co-occurrence matrix. NDT E Int. 37(7), 577–588 (2004)
9. Lambrecht, C.J.V.D.B., Verscheure, O.: Perceptual quality measure using a spatiotemporal
model of the human visual system. In: Electronic Imaging: Science & Technology, pp. 450–
461. International Society for Optics and Photonics (1996)
10. Gu, K., Wang, S.Q., Yang, H., Lin, W.S.: Saliency guided quality assessment of screen
content images. IEEE Trans. Multimed. 18(6), 1 (2016)
11. Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines. ACM Trans. Intell.
Syst. Technol. 2(3), 389–396 (2007). Article 27
12. Sheikh, H.R., Sabir, M.F., Bovik, A.C.: A statistical evaluation of recent full reference image
quality assessment algorithms. IEEE Trans. Image Process. 15(11), 3440–3451 (2006)
13. Ponomarenko, N., Lukin, V., Zelensky, A., Egiazarian, K., Carli, M., Battisti, F.: TID2008-
a database for evaluation of full-reference visual quality assessment metrics. Adv. Mod.
Radioelectron. 10(4), 30–45 (2009)
14. Larson, E.C., Chandler, D.M.: Categorical image quality (CSIQ) database (2010). http://
vision.okstate.edu/csiq
Blind Image Quality Assessment via Analysis of GLCM 101
15. Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error
visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004)
16. Xue, W., Mou, X., Zhang, L., Bovik, A.C., Feng, X.: Blind image quality assessment using
joint statistics of gradient magnitude and Laplacian features. IEEE Trans. Image Process.
23(11), 4850–4862 (2014)
17. Wu, Q.B., Li, H.L., Meng, F.M., Ngan, K.N., Luo, B., Huang, C., Zeng, B.: Blind image
quality assessment based on multichannel feature fusion and label transfer. IEEE Trans.
Circuits Syst. Video Technol. 26(3), 425–440 (2016)
An Improved Ranked K-medoids Clustering
Algorithm Based on a P System
1 Introduction
The rest of the paper is organized as follows. Section 2 presents a variant of the
ranked k-medoids algorithm. Section 3 design a specific P system to realize the
algorithm we proposed. Finally, Sect. 4 makes conclusions.
P ¼ ðO; M 0 ; M 1 ; M 2 ; . . .M k ; ch; c0 ; qÞ
104 B. Zhang et al.
Where:
O ¼ a11 ; c1;0 ; h1 ; n1 ; a1 ; a2 ; . . .an ; #0
h i
ch ¼ ½ 1 ½ 2 ; . . .½ k ½ c0
r0 ¼ w0;0 ; R0
R0 :
n o
x
r1 ¼ aij ai aj ! aiðj þ 1Þ ai aj Uij ij j1 i; j n
[ aiðn þ 1Þ ! aði þ 1Þ;1 j1 i n [ anðn þ 1Þ ! k
r2 ¼ U11
w11
! Dw1111 U11 w11
n o n o
w w 1Þ h þ w w w w 1Þ
r3 ¼ bhij Uij ij Uiðjiðjþþ1Þ ! biðj þ 1Þij iðj þ 1Þ Uij ij Uiðjiðjþþ1Þ [ bhiðn þ 1Þ ! biði þ 1Þ
j0 \ i \ j n
n 0 0 o
cp þ 1 h \ 0; 0 \ i \ j \ n
w w 1Þ w w 1Þ
r4 ¼ bhij Dij ij Uiðjiðjþþ1Þ cp ! bhiðj þ 1Þ Uij ij Diðjiðjþþ1Þ
n o
w
[ cp Dij ij ai aj ! ðAi Þ1 Aj 2 jp ¼ ðn 1Þ!
q w11 w12
g U11 U12 U1n U21 Unn
w1n w21 wnn
!
r5 ¼
gqwij U11 U12 U1n
w11 w12
U21 Unn
w1n w21 wnn
n o
[ gqwij aj ! Aj t j2\t k
An Improved Ranked K-medoids Clustering Algorithm 105
8
9
< U xij c ! c qþ1
1qij þ 1 . . . 1qij þ 1 1qij þ 1 =
ij i;q i;q þ 1 1ij
r6 ¼ in1 in2 ink
: ;
xij ¼ 0; 1 i; j n; 0 q \ n
n o n o
[ ci;ðn þ 1Þ ! cði þ 1Þ;0 j1 i n [ cnðn þ 1Þ ! e
r7 ¼ ci;q Ui1xi1 Ui2xi2 . . .Uinxin ! ci;q Ui1xi1 1 Ui2xi2 1 . . .Uinxin 1 j1 i n; 0 q n
n o
r8 ¼ vt Ati aj 1pij np ! vt Ati Gpj int np þ 1 j1 i; j n; 1 t k; 1 p m
n o
r9 ¼ vk þ 1 nm ! n1 ðg1 Þin1 ðg1 Þin2 . . .ðg1 Þink
8 xij xij xij ! 9
>
> d k ai A1j1 A2j2 . . .Akjk Uij1 1 Uij2 2 . . .Uijk k ! >
>
< =
r10 ¼ d k
ð a Þ
i inp A 1j1 in1 A 2j2 in2 . . . A kjk ink
>
> ai >
>
: ;
xijp ¼ 0; 1 j1 ; j2 ; . . .jk n
( xij xij xij xij 1 xij 1 xij 1
)
Uij1 1 Uij2 2 . . .Uijk k ! Uij1 1 Uij2 2 . . .Uijk k
r11 ¼
xij ¼ 0; 1 j1 ; j2 ; . . .jk n
n o
h
ri ¼ wi;0 ; Ri ; 1 i k Where:
wt;0 ¼ #0
Rt ð1 t kÞ :
n o
r10 ¼ ðbu ! buai Þai 1 i n; u Otj [ ap j1 p n
n o
r20 ¼ bu ! ðbuÞinc u Otj [ ap j1 p n [ f1 ! kg
( )
0
Q
0 Atp G1j1 G2j2 . . .Gmjm 1qpj11 1qpj22 . . .1qpjmm ! Atp G1j1 G2j2 . . .Gmjm 1qpj11 1qpj22 . . .1qpjmm 1p p
r3 ¼
j1 j1 ; j2 ; . . .jm ; p n; 1 t k
8 9
> gi Giji G1j1 G2j2 . . .Gði1Þjði1Þ Gði þ 1Þjði þ 1Þ . . . >
>
> >
>
>
> >
. . .1qjimjm 1jipp ! >
q Þ qði þ 1Þ q
>
> G A 1q1 1q2 . . .1jiðji1 1 >
>
< mjm tp ji j1 ji j2 ði1Þ ji jði þ 1Þ =
r40 ¼ gi Giji G1j1 G2j2 . . .Gði1Þjði1Þ Gði þ 1Þjði þ 1Þ . . .
>
> >
>
>
> >
. . .1qjimjm 1ji ji >
q Þ qði þ 1Þ Q
>
> Gmjm Atp 1qji1j1 1qji2j2 . . .1jiðji1 1 >
>
>
:
ði1Þ ji jði þ 1Þ
>
;
j1 j1 ; j2 ; . . .jm ; p n; 1 i m
( )
Q Q Q Q
0 #h Atp Giji 1p p 1ji ji ! #h þ Qji Qp Atp Giji 1p p 1ji ji
r5 ¼
j1 t k; 1 ji ; p n; 1 i m
r60 ¼ #h0 Atp Giji ! #0 Gip Atji ljh0 [ 0; 1 p; ji n; 1 i m; 1 t k
[ f#h0 ! #0 mjh0 0g [ gi ! gi þ 1 [ gm þ 1 ! k
n o
Q ji
r70 ¼ 1Qp 1ji ! kj1 ji ; p n
p
n o
r80 ¼ li m j G1j1 G2j2 . . .Gmjm Atp 1 ! baj1 aj2 . . .ajm Atp v1 in0 j1 j1 ; j2 . . .jm ; p n; 1 t k
n o
q ¼ ri [ rj j1 i \ j 11 [ ri0 [ rj0 j1 i \ j 8
106 B. Zhang et al.
The P system is supposed to distribute the points into 3 clusters with the given
parameter m of the value 3. Diagram 2 depicts the original state of the 13 points.
First, calculate the distance between every two points, then find points (1, 1), (5, 7),
(8, 2) that satisfy maximum distance method. After that, the rules in three membranes
are executed at the same time and find the final clustering medoids by hostility values.
Finally, sent the objects to the nearest clusters.
Eventually, the clustering result was attained that the 13 points were classified into
3 clusters. And the clustering effect sketch is shown in Fig. 3.
4 Conclusions
Acknowledgment. This work is supported by the Natural Science Foundation of China (Nos.
61170038, 61472231, 61402187, 61502535, and 61572523).
References
1. Han, J., Kambr, M.: Data mining concepts and techniques, Elsevier Inc., USA (2012). ch. 8
2. Razavi Zadegan, S.M., Mirzaie, M., Sadoughi, F.: Ranked k-medoids: a fast and accurate
rank-based partitioning algorithm for clustering large datasets. Knowl.-Based Syst. 39(2),
133–143 (2013)
3. Zhang, T.: BIRCH: an efficient data clustering method for very large databases.
ACM SIGMOD Record 25(2), 103–114 (1996)
4. Kaufman, L., Rousseeuw, P.J.: Finding Groups in Data: An Introduction to Cluster Analysis
(1991)
5. Sander, J., Ester, M., Kriegel, H.P., et al.: Density-based clustering in spatial databases: the
algorithm GDBSCAN and its applications. Data Min. Knowl. Disc. 2(2), 169–194 (1998)
6. Fraley, C., Raftery, A.E.: Model-based clustering, discriminant analysis, and density
estimation. J. Am. Stat. Assoc. 97(458), 611–631 (2002)
7. Wang, W., Yang, J., Muntz, R.R.: STING: a statistical information grid approach to spatial
data mining. In: International Conference on Very Large Data Bases, pp. 186–195. Morgan
Kaufmann Publishers Inc. (1997)
8. Peng, H., Wang, J., Pérez-Jiménez, M.J., et al.: An unsupervised learning algorithm for
membrane computing. Inf. Sci. 304(C), 80–91 (2015)
9. Niu, Y., Xiao, J., Jiang, Y.: Time-free solution to 3-coloring problem using tissue P systems.
Chin. J. Electron. 25(3), 407–412 (2016)
A Novel Inverse-Operation Based Group Undo/Redo
Algorithm for Feature-Based 3D Collaborative CAD
Systems
1
School of Information Management, Wuhan University, Wuhan, China
2
School of Computer Science, Wuhan University, Wuhan, China
fzhe@whu.edu.cn
1 Introduction
Feature-based modeling approaches have played a relevant role for qualitative knowl‐
edge specification and integration in collaborative designing since the 70s [1–3], and
feature-based CAD systems are currently considered the state-of-art technologies for
product modeling [4–7]. Undo/Redo is a standard and core function for nearly all human-
computer interaction applications. It can help to remove some erroneous operations and
enable a user to explore a new system by try-and-failure. Most importantly, undo/redo
is an essential tool to grant that a document is the result of designers’ intentions by
removing the effects of undesired operations. An Undo/Redo mechanism should under‐
stand which operation is to be undone when a user issues an undo command by different
interaction methods. Besides, the undo effect demands that the result of a certain oper‐
ation is eliminated as if it has never been executed. This leaves us a question that how
the effect of an undo target be removed from the 3D model. Our previous study inves‐
tigate both full-rerun and full checkpoint strategy[8, 9], although the correctness of our
group Undo/Redo algorithms can be promised, the performance is still not satisfying.
2 Related Works
Several undo models are proposed for multi-user Undo/Redo. Each model sets a frame‐
work for developing specific undo and redo algorithms. The script model [11] is one of
the earliest group Undo/Redo solutions. Regional undo was also applied in spreadsheets
[12] by allowing users to select a part of the spreadsheet and perform undo on the cells.
Treating an undo as a concurrent inverse operation is another popular solution for group
undo. For the first time, Berlage introduced the selective undo model that adds the
reverse operation of the selected command to the current context [13]. In collaborative
textual editors, delete and insert can be the inverse operation for each other. DistEdit
[14] is the first OT based selective undo solution. The undo target is transposed to the
end of the history buffer and the inverse command of the transposed operation is
executed. The adopted [15, 16] also considered an inverse as the concurrent operation
of operations generated after O. Neither DisEdit nor adOPTed supports the full featured
selective undo. Sun has proposed the framework of AnyUndo to support “Anytime,
Anywhere undo”. The proposed algorithms GOTO and COT both integrate do and
selective undo [17, 18]. Azurite [19, 20], provides general selective undo in a code editor.
Azurite uses the inverse model, by adding the inverse text editing operation to the end
of the history, and provides a variety of user interfaces to resolve conflicts. Logoot-Undo
[21] is the group undo algorithm supporting AnyUndo under the CRDT framework.
110 Y. Cheng et al.
The document is manipulated at the line-level. The effects of that an inserted line is
undone is actually controlled by changing the visibility of the line. Conflicts between
two concurrent undos are therefore avoided because if an undo cannot find a line given
its unique identifier, it can therefore be discarded.
Applying the inverse model to graphical editors seems easier than to painting
systems. On one hand, in an object-oriented system, the inverse operation can be easily
generated by retrieving the designing history, and then executing some system defined
operation by using the properties of the graphical object before the latest modification
[13, 22]. In pixel-based drawing systems, such as Photoshop, it is the historical infor‐
mation of every pixel that needs to be recorded which is resource consuming. In Wang’s
solution for undo in bitmap systems [23], only pixels of the drawing area and connections
between operations are recorded. Aquamarine [24, 25] is a system developed on the
basis of Photoshop for selective undo using the script model. Another undo scheme
developed under the script model is the group Undo/Redo method for 3D collaborative
CAD systems developed by our previous research [25].
We build an ontology, called Part Feature Ontology (PFO), to describe the semantics of
a part from the perspectives of individual settings and parameter values. We also intend
to use this ontology to categorize features in a 3D collaborative CAD system. The classes
defined in PFO are: modeling feature, topology info, B-rep Operation, Reference Info
and Sketch Info. Six relations are defined, which are: is-a, has-topology-attribute, has-
reference-attribute, has-boolean-operation, as-dependency-attribute, and has-sketch.
We categorize features according to their effects on the volume. The subclasses of
modeling feature include convex feature, concave feature, edge transition feature and
primitive feature.
The inverse operation of a convex feature creating operation is to remove any topological
entities that extrude outward the datum plane without considering the shape of the
feature. The construction process of the inverse operation is presented in Algorithm 1.
A Novel Inverse-Operation Based Group Undo/Redo Algorithm 111
Function Inverse(O)
1: Obtain the topological entities that belong to the feature created by O. Typically, these
topological entities are protrusive above the original part model. This is accomplished by
parsing the semantic descriptor OWL file FDO with OWL API parser so as to obtain the
attribute TopologicalFaceSet;
2: For every topological face f from the TopologicalFaceSet, execute the delete(f) so as to
remove the revealed faces from the part model;
3: Detect the broken area on the datum plane, create a face in the same shape ,and merge
the face with the plane.
4: return ;
Figure 1 gives us an example. Three feature instances are added to the same base
block feature by different modeling methods. O1 adds the convex feature by sweeping
the face fp along a curve. O2 adds the convex feature by rotating. O3 adds the feature by
extruding. The features are in different forms. It is worth noting that, in some situations,
when deleting some revealed faces from the boundary, there leaves a broken area on the
datum plane. In this case, the plane should be repaired. As it is illustrated in Fig. 2, when
the revealed faces from the round protrusion are deleted, there leaves a circular broken
area on the datum plane where the protrusion located.
Fig. 1. Inverse operations for different convex Fig. 2. An example of the broken area
feature instances
112 Y. Cheng et al.
Fig. 3. Inverse operations for different concave Fig. 4. Inverse operations for different edge
feature instances transition feature instances
A Novel Inverse-Operation Based Group Undo/Redo Algorithm 113
Function Inverse(O)
1: Parse the FDO and extract the information, namely, the category the feature instance
belongs to, the dimension parameters and the reference information.
2: The MWS for the concave feature is constructed. It is then transformed to the proper
position by using the same reference information extracted in step 1;
3: Unite the MWS with the current part model so as to refill the hollow space that O
created;
4: Delete faces from the MWS that are extruding outward the part and repair the broken
area;
5: return ;
Algorithm 3 The inverse operation for edge transition feature creating operations
Function Inverse(O)
1: Parse the FDO and extract the information, namely, the category the feature instance
belongs to, the dimension parameters.
2: The MWS for the transition edge is constructed. It is then transformed to the proper
position;
3: If O selects a convex edge for edge transition, the MWS is united with the part. if O
selects a concave edge for edge transition, the MWS is subtracted from the part;
4: return ;
114 Y. Cheng et al.
4 Experiments
In this example, three geographically dispersed sites build and modify models with both
do and undo operations. How the boundary of the part evolves with the continuous
execution of both do and undo at each site is given in Fig. 5 and the expected effect of
every operation is described in details in Table 1. Whenever an operation is performed,
the related hybrid feature descriptor is generated. Observed from Fig. 5, the collaboration
activity involves three stages. The operations issued in the first stage are pure modeling
operations. The second stage contains concurrent do and undo. The third stage is a typical
scenario of concurrent undo. The emphasis will be laid on the group undo process.
When undo is issued by site1, its intention is to delete the round hole feature created
by O5 from site1. It is carried out immediately at site1. The undo command is sent to
other sites in the form of UNDO(⟨1, 1⟩, ⟨1, 3⟩). Being an instance of a concave feature,
the MWS for the round hole is created given the length of the hole and the radius of the
circular end-face. The MWS is then transformed given the locating information
containing in the reference information descriptor of FDO5. By uniting it with the part,
O5 is successfully undone. When undo1 is received by a remote site, O5 is then identified
with the assistance of the tuple ⟨1, 3⟩. The undo process is carried in the same process
with how the undo is performed at the local site site1.
Undo2 and undo3 are concurrently generated at site0 and site2. However, they are
not in a conflictive relation. Undo2’s intention is to delete the round protrusion created
by O3 from site1. It is carried out immediately at site0. The undo command is sent to
other sites in the form of UNDO(⟨0, 1⟩, ⟨1, 1⟩). The revealed faces from the round
protrusion are deleted. Then, the broken area left on the base feature is repaired. The
topology information are required are from FDO3. When undo2 is received by a remote
site, O3 is then picked out with the assistance of the tuple ⟨1,1⟩.
A Novel Inverse-Operation Based Group Undo/Redo Algorithm 115
The intention of undo3 is to delete the small round hole within the base cylinder
created by O4 from site1. It is carried out immediately at site2 and sent to other sites in
the form of UNDO(⟨2, 1⟩,⟨2, 1⟩). The MWS for this round hole is created given its
length and the radius. It is then transformed given the same reference information
acquired from FDO4. By uniting it with the part, O4 is successfully undone. O7 is undone
as well because of its dependency relation with O4.
5 Conclusions
Acknowledgement. The work is supported by the National Natural Science Foundation of China
(NSFC Grant Nos. 61472289 and 61502353) and Youth Innovation Corps Fund of Humanities
and Social Sciences, Wuhan University. The authors would like to thank all the reviewers for their
constructive comments and suggestions.
116 Y. Cheng et al.
References
1. He, F., Han, S.: A method and tool for human-human interaction and instant collaboration in
CSCW-based CAD. Comput. Ind. 57(8), 740–751 (2006)
2. Jing, S., He, F., Han, S., et al.: A method for topological entity correspondence in a replicated
collaborative CAD system. Comput. Ind. 60(7), 467–475 (2009)
3. Li, X., He, F., Cai, X., et al.: CAD data exchange based on the recovery of feature modelling
procedure. Int. J. Comput. Integr. Manuf. 25(10), 874–887 (2012)
4. Li, X., He, F., Cai, X., et al.: A method for topological entity matching in the integration of
heterogeneous CAD systems. Integr. Comput. Aided Eng. 20(1), 15–30 (2013)
5. Zhang, D., He, F., Han, S., et al.: Quantitative optimization of interoperability during feature-
based data exchange. Integr. Comput. Aided Eng. 23(1), 31–50 (2016)
6. Wu, Y., He, F., Zhang, D., et al.: Service-oriented feature-based data exchange for cloud-
based design and manufacturing. IEEE Trans. Serv. Comput. PP(99), 1 (2015)
7. Wu, Y., He, F., Han, S.: Collaborative CAD synchronization based on a symmetric and
consistent modeling procedure. Symmetry 9(4), 59 (2017)
8. Cheng, Y., He, F., Cai, X., et al.: A group Undo/Redo method in 3D collaborative modeling
systems with performance evaluation. J. Netw. Comput. Appl. 36(6), 1512–1522 (2013)
9. Cheng, Y., He, F., Wu, Y., et al.: Meta-operation conflict resolution for human-human
interaction in collaborative feature-based CAD systems. Cluster Comput. 19(1), 237–253
(2016)
10. Ressel, M., Nitsche-Ruhland, D., Gunzenhauser, R.: An integrating, transformation-oriented
approach to concurrency control and undo in group editors. In: Proceeding of Conference on
Computer Supported Cooperative Work (CSCW), pp. 288–297 (1996)
11. Archer Jr., J.E., Conway, R., Schneider, F.B.: User recovery and reversal in interactive
systems. ACM Trans. Program. Lang. Syst. 6(1), 1–19 (1984)
12. Kawasaki, Y., Igarashi, T.: Regional undo for spreadsheets (Demo). In: 2004 Adjunct
Proceedings UIST, 2 pages (2004)
13. Berlage, T., Genau, A.A.: Framework for shared applications with a replicated architecture.
In: 1993 ACM UIST, Atlanta, GA, pp. 249–257 (1993)
14. Prakash, A., Knister, M.J.: A framework for undoing actions in collaborative systems. ACM
Trans. Comp.-Hum. Inter. 1(4), 295–330 (1994)
15. Ressel, M., Gunzenhauser, R.: Reducing the problems of group undo. In: ACM Conference
on Supporting Group Work (1999)
16. Ressel, M., Nitsche-Ruhland, D., Gunzenhauser, R.: An integrating, transformation oriented
approach to concurrency control and undo in group editors. In: Proceeding of ACM
conference on computer supported co-operative work, November 1996, pp. 288–297 (1996)
17. Sun, C.Z.: Undo as concurrent inverse in group editors. ACM Trans. Comput. Hum. Interact.
9(4), 309–361 (2002)
18. Sun, D., Sun, C.: Context-based operational transformation in distributed collaborative
editing systems. IEEE Trans. Parallel Distrib. Syst. 20(10), 1454–1470 (2009)
19. Yoon, Y., Koo, S., Myers, B.A.: Visualization of fine-grained code change history. In: IEEE
VL/HCC 2013, pp. 119–126 (2013)
20. Yoon, Y., Myers, B.A.: Supporting selective undo in a code editor. In: ICSE 2015, Florence,
Italy (2015)
21. Weiss, S., Urso, P., Molli, P., et al.: Logoot-undo: distributed collaborative editing system
on P2P networks. IEEE Trans. Parallel Distrib. Syst. 21(8), 1162–1174 (2010)
22. Myers, B.A.: Scripting graphical applications by demonstration. In: SIGCHI 1998, pp. 534–
541 (1998)
A Novel Inverse-Operation Based Group Undo/Redo Algorithm 117
23. Wang, X.Y., Bu, J.J., Chen, C.: Achieving undo in bitmap-based collaborative graphics
editing systems. In: The 2002 ACM Conference on Computer Supported Cooperative Work
(2002)
24. Myers, BA., Lai, A., Le, T.M., et al.: Selective undo support for painting applications. In:
Human Factors in Computing Systems, pp. 4227–4236 (2015)
25. www.cs.cmu.edu/NatProg/aquamarine.html
Two Steps Method Using Polarization to Resist
Phase Noise for Self-interference Cancelation
in Full-Duplex
Abstract. Phase noise caused by the local oscillator affects both the
cancelation of self-interference and the reception of the desired signal. In
this paper, a novel two steps method exploiting polarization is proposed
to resist the effect of phase noise on the self-interference cancelation in
full-duplex. This method can convert the multiplicative noise to additive
noise with polarization signal processing in two steps, which can cancel
the effect of phase noise on both the self-interference and desired signal.
It should be noted that this method needs no prior information of the
phase noise. Theory analysis and numerical simulations show that the
effect of phase noise is reduced and the amount of cancelation is also
evaluated in condition of different phase noise values.
1 Introduction
Full-duplex is a modality of communication that allows a node to transmit and
receive signal simultaneously at the same frequency band. It has higher spec-
tral efficiency, higher throughput, and lower transmission delay than the tradi-
tional half duplex system, which satisfies the requirements of 5G very well [1–5].
However, massive self-interference (SI) caused by the local transmitter coupling
into the local receiver is an inevitable problem to resolve. Self-interference trans-
mitted through a short path is 15–100 dB stronger than the desired signal [6].
Therefore the cancelation of self-interference is one of the key factors influencing
the full-duplex communication. Moreover, it has been demonstrated that the
phase noise is one of the bottleneck of the self-interference cancelation [7].
Currently, relevant scholars put forward many self-interference cancelation
methods at the receiver, which can be classified into three categories, antenna
This work is supported by Chinese National Nature Science Foundations (61501050)
and (61271177).
c Springer International Publishing AG 2018
Q. Zu and B. Hu (Eds.): HCC 2017, LNCS 10745, pp. 118–129, 2018.
https://doi.org/10.1007/978-3-319-74521-3_14
Two Steps Method Using Polarization to Resist Phase Noise 119
cancelation [2,5,9], radio frequency cancelation (RFC) [1,3,6,8], and digital can-
celation [10–13]. Antenna cancelation and radio frequency cancelation is on the
premise of ignoring the phase noise, then the self-interference can be suppressed
to 10–15 dB stronger than the desired signal. Digital cancelation utilizes one pilot
siganl to track the phase noise and uses compensation structure to improve the
amount of self-interference cancelation. The effect of phase noise caused by sep-
arate local oscillator or the same local oscillator on self-interference cancelation
is analyzed [10]. Inter carrier interference (ICI) and common phase error (CPE)
caused by the phase noise is handled through estimating and compensating the
phase noise, the amount of cancelation can be improved 7–10 dB in [11,12]. The
closed-form solution between the amount of cancelation and 3 dB bandwidth of
phase noise is derived [13]. However, the existing methods dealing with phase
noise only focus on self-interference signal but ignore the effect of phase noise
on the desired signal. In this paper, a novel polarization self-interference can-
celation method is proposed to cancel the effect of phase noise both on the
self-interference signal and the desired signal without any prior information of
the phase noise.
The polarization state (PS) of the signal is a substantive character which can
be used to carry information such as time, frequency, space, and code character
[14]. Polarization modulation technology have been proposed, and the oblique
projection filtering is used to distinguish the desired signal from the noise [15].
The PS of the signal is decided by the amplitude ratio and phase difference.
Taking advantage of the PS definition, the basic idea of this paper is that, phase
noise does not affect the phase difference between the two branches but only
affect the absolute phase, thus the phase noise has no effect on the PS of the
signal. We have proved the theoretical feasibility in the previous work [16], and
we will further consider the effect of phase noise on both self-interference and
desired signal in this paper.
In this paper, the PS of the signal can be used to cancel the effect of the
phase noise on the self-interference and desired signals, since it is not affected
by the phase noise. The algorithm can be divided into two steps by converting
the multiplicative noise to additive noise with polarization signal processing.
The first step is converting the effect of phase noise on the self-interference
signal to the desired signal and white noise, and using the reconstructed signal
feedback from transmitter to cancel the self-interference signal. The second step
is converting the effect of phase noise on the desired signal to white noise, then
recovering the desired signal. Both of the two steps does not need the prior
information of the phase noise. This method uses the polarization matching to
receive the desired signal and improve the amount of self-interference cancelation.
Theory analysis and numerical simulation show the effect of phase noise on self-
interference cancelation. The amount of cancelation can be improved by 0–10 dB
in condition of the different phase noise values. This paper mainly analyzes the
effect of phase noise on the self-interference cancelation, other parameters like
the nonlinearity of power amplifier of transmitter and receiver, I/Q (in phase and
120 F. Liu et al.
where the NhL (t) and NvL (t) express the H and V branch of NL (t), and the two
branches satisfies Gaussian distribution of random variable, NhL (t) and NvL (t)
are independent and identical with mean 0 and variance is σ 2 2.
base band, which is affected by phase noise. The traditional algorithm is esti-
mation and compensation [11–13], thus the estimation error of phase noise will
affect the cancelation of self-interference. The proposed elimination algorithm
in domain of polarization is shown in Fig. 2. This method can be achieved by
converting the multiplicative noise to additive noise with polarization signal pro-
cessing in two steps. The first step is converting the effect of phase noise on the
self-interference to the desired signal and the white noise using stokes vector pro-
cessing in polarization domain, then using the reconstructed signal to cancel the
self-interference signal. The second step is recovering the desired signal using the
same principle. Then the mixed signal enters into the Match Receiving module
(MR), at last it enters into the Demodulation Module (DM).
After the mixed signal being down-converted, the mixed signal is expressed as
T
Eq. (9), in which the NhL (t) NvL (t) can be written as Eq. (10).
NhL (t) nhc + jnhs
= (10)
NvL (t) nvc + jnvs
The nhc and nhs express the in-phase and quadrature components of NhL (t),
the nvc and nvs has the similar expression. In order to analyze conveniently, the
signal transformed through stokes, as expression (11).
2 2
S1 = |yLH (t)| − |yLV (t)|
S2 = 2 |yLH (t)| |yLV (t)| cos(ϕyLH (t) − ϕyLV (t)) (11)
S3 = 2 |yLH (t)| |yLV (t)| sin(ϕyLH (t) − ϕyLV (t)),
where S1 , S2 , S3 express the three components of stokes. yLH (t) and yLV (t)
express the H and V branch of yL (t). Through (9) and (11), yL (t) expressed as
Two Steps Method Using Polarization to Resist Phase Noise 123
S1 = (Ehi2
− Evi
2 2
) + (Ehs − Evs
2
) + (n2hc + n2hs ) − (n2vc + n2vs )
+ 2(Ehi Ehs cos(φst (t) − φit (t)) − Evi Evs cos(φst (t) − φit (t) + (δs − δi )))
T
E cos(φst (t) − φit (t)) + Ehi
+2 hs
Ehs sin(φst (t) − φit (t))
cos(φit (t) + φr (t)) sin(φit (t) + φr (t)) nhc
· (12)
− sin(φit (t) + φr (t)) cos(φit (t) + φr (t)) nhs
T
E cos(δs + φst (t) − φit (t)) + Evi cos(δi )
−2 vs
Evs sin(δs + φst (t) − φit (t)) + Evi sin(δi )
cos(φit (t) + φr (t)) sin(φit (t) + φr (t)) nvc
· ,
− sin(φit (t) + φr (t)) cos(φit (t) + φr (t)) nvs
T T T
[14], nhc1 nhs1 and nvc1 nvs1 have the same distribution with nhc nhs
T
and nvc nvs , then the yL (t) can be calculated as
Ehi Ehs j(φst (t)−φit (t)) NhL1 (t)
yL (t) = + e + , (17)
Evi ejδi Evs ejδs NvL1 (t)
the NhL1 (t) = nhc1 + jnhs1 , NvL1 (t) = nvc1 + jnvs1 . Comparing Eqs. (9) and
(17), the phase noise of self-interference is converted to the desired signal and
white noise, and the distribution of white noise has not changed. In the first
rotation, it is not simply multiply the e−j(φit (t)+φr (t)) both sides of the Eq. (9).
because the φit (t), φst (t), φr (t) is unknown. It just uses the character of phase
noise in frequency domain. Therefore it is unnecessary to estimate phase noise
in this paper, which is the strong aspect compared with the method in time and
frequency domain. Using the canceling signal ycl (t) from transmitter through
wired feedback, the residual signal is Eq. (18).
Suppose:
nhc2 cos(φst (t) − φit (t)) sin(φst (t) − φit (t)) nhc1
= · , (19)
nhs2 − sin(φst (t) − φit (t)) cos(φst (t) − φit (t)) nhs1
nvc2 cos(φst (t) − φit (t)) sin(φst (t) − φit (t)) nvc1
= · , (20)
nvs2 − sin(φst (t) − φit (t)) cos(φst (t) − φit (t)) nvs1
T T
nhc2 nhs2 and nvc2 nvs2 express the H and V branch of white noise
after the second rotation. Using the same method as the first step, the Eq. (18)
can be written as Eq. (21).
Ehs NhL2 (t)
yLR (t) = + , (21)
Evs ejδs NvL2 (t)
In Eq. (21), the NhL2 (t) = nhc2 + jnhs2 , NvL2 (t) = nvc2 + jnvs2 . From the
Eq. (21), the effect of phase noise on the desired signal is converted to white noise,
Two Steps Method Using Polarization to Resist Phase Noise 125
and the distribution of white noise has not changed. Then using the minimum
variance criterion, we can get the polarization state of the desired signal. Then
the polarization of the desired signal can be written as Eq. (22).
∧
Ps = [ cos(εs ) sin(εs )ejδs ]T , (22)
∧
Ps is the polarization of the desired signal estimated at the receiver. After the
(MR), the mixed signal changed as Eq. (23).
∧
yLR1 (t) = Ps1 ·yLR (t)
T
cos(εs ) Ehs NhL2 (t)
= · jδs +
sin(εs )e−jδs E
vs e NvL2 (t)
−jδs E hs −jδs NhL2 (t)
= [ cos(εs ) sin(εs )e ] + [ cos(εs ) sin(εs )e ]
Evs ejδs NvL2 (t)
= s(t) + no (t) ,
(23)
∧
In Eq. (23), the yLR1 (t) is the mixed signal after MR, Ps1 is the Hermitian
∧
Transpose of Ps , no (t) is the white noise after MR.
NhL2 (t)
no (t) = [ cos(εs ) sin(εs )e−jδs ] (24)
NvL2 (t)
After the MR, the desired signal s(t) has been recovered. From the Eq. (23), the
phase noise caused by the local oscillator of transmitter and receiver is canceled.
The white noise is full polarization after the MR, thus the power of white noise
keeps only in half. In the meanwhile, the amount of self-interference cancelation
is calculated as Eq. (25).
SINRout
ηSINR (dB) = 10log , (25)
SINRin
ηSINR (dB) expresses the amount of self-interference cancelation. SINRin and
SINRout express the ratio between the desired signal and self-interference add
white noise before and after signal processing at the receiver. Then the Eq. (25)
can be written as Eq. (26).
b2 (t)
ηSINR (dB) = 10log 2(1 + ) . (26)
σ2
Comparing the Eq. (26) and the amount of self-interference cancelation in
[7,9,10,12], the phase noise caused by the local oscillator of transmitter and the
receiver is canceled, and the amount of self-interference cancelation is improved.
paper. This part shows the effect of phase noise on the signal and the process of
how to cancel the effect of the phase noise intuitionally.
In this section, the ratio between H and V branch is set as 1, the phase
difference is set as 45 in degree. The variance of phase noise is set as 0.018 in
radian. The effect of phase noise on the self-interference is shown in Figs. 3 and 4.
The spectrum of the self-interference without phase noise is shown in Fig. 3,
then Fig. 4 shows the spectrum of self-interference effected by phase noise. By
comparing Figs. 3 and 4, we can obtain that the spectrum of self-interference
has been spreaded due to the phase noise. The center power value of the self-
interference decreased. The total power of the signal unchanged at the same time.
The effect of phase noise on the desired signal is the same as the self-interference.
The waveform of the mixed signal at the receiver in time domain is shown in
Fig. 5(a). The self-interference is 90 dB higher than the desired signal. Because of
the effect of the self-interference, phase noise and white noise, the desired signal
has been distorted seriously. After the first step cancelation, the self-interference
is canceled, whose phase noise is converted to the desired signal and white noise
as shown in Fig. 5(b). After cancelation by the second step, the phase noise of
the desired signal is converted to the white noise as shown in Fig. 5(c). After the
two steps, the effect of phase noise on self-interference cancelation is canceled,
and the desired signal is recovered.
Fig. 5. Waveform of the mixed signal before cancelation (a), after the first cancelation
(b), after the second cancelation (c)
140
Cancelation in polarization domain in theory
Cancelaiton in polarization domain in ADS
120 Cancelation in traditioanl domain in theory L=30
The cancelation of self−interference in dB
80
60
40
20
0
0 0.005 0.01 0.015 0.02
The standard deviation of Phase Noise in radian
decreased as the power of phase noise increased. The green (rectangle) and pink
(pentagram) lines express the amount of cancelation affected by phase noise after
traditional self-interference elimination algorithm [11,12]. Although the amount
of cancelation increased under the condition without suppression, it decreased
when the phase noise increases. The red (triangle) line expresses the amount of
self-interference cancelation using the method proposed in this paper. The black
(roundness) line expresses the amount of self-interference cancelation simulated
in ADS. These two lines suggest that by utilizing the cancelation algorithm pro-
posed in this paper, the amount of cancelation remained unchanged as the phase
noise increased. We can conclude that the effect of phase noise on the cancelation
of self-interference has been canceled by using the two steps method algorithm
proposed in this paper.
5 Conclusion
Based on the unitary matrix rotation characteristics in polarization domain, the
method in this paper has canceled the effect of phase noise on the self-interference
cancelation and recovering of the desired signal. This method can be achieved
by converting the multiplicative noise to additive noise with polarization signal
processing in two steps. The first step is converting the effect of phase noise on
the self-interference to the desired signal and the white noise using stokes vector
processing in polarization domain, then using the reconstructed signal to cancel
the self-interference signal. The second step is recovering the desired signal using
the same principle. Theory analysis and numerical simulation show that effect
of phase noise on self-interference cancelation is canceled, and the amount of
cancelation is improved on the premise of recovering the desired signal.
References
1. Zhang, Z., Long, K., Vasilakos, A.V., Hanzo, L.: Full-duplex wireless communica-
tions: challenges, solutions, and future research directions. Proc. IEEE 104, 1369–
1409 (2016)
2. Everett, E., Sahai, A., Sabharwal, A.: Passive self-interference suppression for full-
duplex infrastructure nodes. IEEE Trans. Wireless Commun. (2013, in press)
3. Debaillie, B., van den Broek, D.-J., Lavn, C., van Liempd, B., Klumperink, E.A.M.,
Palacios, C., Craninckx, J., Nauta, B.: Analog/RF solutions enabling compact full-
duplex radios. IEEE J. Sel. Areas Commun. 32(9), 1662–1673 (2014)
4. Liu, G., Yu, F.R., Ji, H., Leung, V.C.M., Li, X.: In-band full-duplex relaying for
5G cellular networks with wireless virtualization. IEEE Netw. 29, 54–61 (2015)
5. Foroozanfard, E., Franek, O., Tatomirescu, A., Tsakalaki, E., de Carvalho, E.,
Pedersen, G.F.: Full-duplex MIMO system based on antenna cancellation tech-
nique. Electron. Lett. 50(16), 1116–1117 (2014)
6. Sahai, A., Patel, G., Sabwarwal, A.: Pushing the limits of full duplex: design and
real-time implementation. Rice University, Technical report REE1104 (2011)
7. Sahai, A., Patel, G., Dick, C., Sabharwal, A.: On the impact of phase noise on
active cancelation in wireless full-duplex. IEEE Trans. Veh. Technol. 62(9), 4494–
4510 (2013)
Two Steps Method Using Polarization to Resist Phase Noise 129
8. Wang, J., Zhao, H., Tang, Y.: A RF adaptive least mean square algorithm for
self-interference cancellation in co-frequency co-time full duplex systems. In: IEEE
Symposium Transactions on Communications (2014)
9. Duarte, M., Dick, C., Sabharwal, A.: Experiment-driven characterization of full-
duplex wireless systems. IEEE Trans. Wireless Commun. 11(12), 4296–4307 (2012)
10. Syrjala, V., Valkama, M., Anttila, L., Riihonen, T., Korpi, D.: Analysis of oscilla-
tor phase-noise effects on self-interference cancellation in full-duplex OFDM radio
transceivers. IEEE Trans. Wireless Commun. 13(6), 2977–2990 (2014)
11. Syrjala, V., Yamamoto, K.: Self-interference cancellation in full-duplex radio
transceivers with oscillator phase noise. In: European Wireless (2014)
12. Ahmed, E., Eltawil, A.M., Sabharwal, A.: Self-interference cancellation with phase
noise induced ICI suppression for full-duplex systems. In: Signal Processing for
Communications Symposium, GLOBECOM (2013)
13. Shao, S., Quan, X., Shen, Y., Tang, Y.: Effect on phase noise of digital self-
interference cancelation in wireless full duplex. In: IEEE International Conference
on Acoustic, Speech and Signal Processing (ICASSP) (2014)
14. Zhuang, Z., Xiao, S., Wang, X.: Radar Polarization Information Processing and
Application. National Defense Industry Press, Beijing (1999)
15. Cao, B., Zhang, Q.Y., Jin, L.: Polarization division multiple access with polariza-
tion modulation for LOS wireless communications. EURASIP J. Wireless Com-
mun. Network. (2011)
16. Liu, F., Jia, S., Guo, C., Feng, C.: Exploiting polarization to resist phase noise for
digital self-interference cancellation in full-duplex. In: IEEE International Confer-
ence on Communications (ICC) (2016)
17. Benedetto, S., Poggiolin, P.: Theory of polarization shift keying modulation. IEEE
Trans. Commun. 40(4), 708–721 (1992)
Sparse Linear Method Based Top-N Course
Recommendation System with Expert Knowledge
and L0 Regularization
1
School of Control and Engineering, Shandong University, Jinan, China
2
Shandong University of Finance and Economics, Jinan, China
3
Shandong University of Science and Technology, Qingdao, China
lianjianlianjian@163.com
1 Introduction
The emergence and rapid development of Internet have greatly affected the traditional
viewpoint on choosing courses by providing detailed course information. As the number
of courses conforming to the students’ has tremendously increased, the above-mentioned
problem has become how to determine the courses mostly suitable for the students
accurately and efficiently. A plethora of methods and algorithms [2, 3, 11, 15] for course
recommendation have been proposed to deal with this problem. Most of the methods
designed for recommendation system can be grouped into three categories, including
collaborative [1, 8], content-based [7, 14], and knowledge-based [5, 8, 17], which have
been applied in different fields such as [4] proposed a collaborative filtering embedded
with an artificial immune system to the course recommendation for college students.
The rating from professor was exploited as ground truth to examine the results.
Inspired by the idea form [4] and the optimization framework in [9], we propose a
sparse linear based method for top-N course recommendation with expert knowledge
as the ground truth. This method extracts the coefficient matrix for the courses in the
experimental results of state-of-the-art methods and our method are evaluated with the
course recommendations presented by seven experts with voting strategy.
The rest of the paper is organized as follows. In Sect. 2, we describe the details of
our proposed method. In Sect. 3 the dataset that we used in our experiments and the
experimental results are presented. In Sect. 4 the discussion and conclusion are given.
2 Our Method
In the following content, tj and si are introduced to denote each course and each student
in course recommender system, respectively. The whole student-course taken will be
represented by a matrix A of size m × n, in which the entry is 1 or 0 (1 denotes that the
student has taken the course, 0 vice versa).
In this paper, we introduce a Sparse Linear Method (SLIM) to implement top-N
course recommendation. In this approach, the score of course recommendation on each
un-taken student/course item tj of a student si is computed as a sparse aggregation of
items that have been taken by si, which is shown in Eq. (1).
where ā is the initial course selection of a specific student and wj is the sparse vector of
aggregation coefficients. The model of SLIM with matrix is represented as:
Ā = AW (2)
Where overlineA is the initial value of student/course matrix, A denotes the latent
binary student-course item matrix, W denotes the n × n sparse matrix of aggregation
coefficients, in which j − th column corresponds to wj as in Eq. (1), and each row of
C(ci ) is the course recommendation scores on all courses for student si. The final course
recommendation result of each student is completed through sorting the non-taken
courses in decreasing order, and the top-N courses in the sequences are recommended.
In our method, the initial student/course matrix is extracted from the learning
management system of a specific University in China. With the extracted student/course
matrix of size m × n, the sparse matrix W size of n × n in Eq. (2) is iteratively optimized
by alternate minimization method. Different from the objective function previously
proposed in [9] shown in Eq. (3), our proposed method is shown in Eq. (4).
1 𝛽1 | ‖2
min ‖ A − AW ‖ ‖ ‖
‖2 + 2 |W ‖F + 𝜆1 ‖A‖1
2
(3)
W≥0 2 ‖
1 𝛽2 | ‖ 2
min ‖ A − AW ‖ ‖ ‖
‖2 + 2 |W ‖F + 𝜆2 ‖A‖0 + 𝜇|∇A|0
2
(4)
W≥0 2 ‖
Where ‖.‖F denotes the Frobenius norm for matrix, ‖W‖1 is the item-wise L1 norm,
‖W‖0 denotes the entry-wise L0 norm that stands for the number of entries with zero
Sparse Linear Method Based Top-N Course Recommendation System 133
value. The data term ‖A − AW‖ is exploited to measure the difference between the
calculated model and the training dataset. The LF − norm, L1 − norm, and L0 − norm
are exploited to regularize the entries of the coefficient matrix W, A, and ∇A, respectively.
The parameters 𝛽1 , 𝛽2 , 𝜆2, and 𝜇 are used to constrain the weights of regularization terms
in the objective functions.
In our proposed final objective function, the LF norm is introduced to transfer the
optimization problem into elastic net problem [19], which prevents the potential over
fitting. Moreover, the L1 norm in Eq. (3) is changed to L0 norm in our proposed objective
function. This novel norm L0 [12, 13, 16] is introduced to constrain the sparseness of
the A and ∇A.
Due to the independency of the columns in matrix W, the final objective function in
Eq. (4) is decoupled into a set of objective functions as follows:
1‖ ‖ 2 𝛽2 | ‖ ‖ | |
min ‖aj − aj wj ‖ + |wj ‖2
+ 𝜆2 ‖aj ‖ + 𝜇|∇aj |
wj ≥0,wj ,j=0 2 ‖ ‖2 2 | ‖F ‖ ‖0 | |0 (5)
where aj is the j-th column of matrix A, wj denotes j-th column of matrix W. As there
are two unknown variables in each Eq. (5), which is a typical ill-posed problem. Thus,
this problem need to be solved by alternate minimization method. In each iteration, one
of the two variables is fixed and the other variable is optimized.
Subproblem1: computing wj
The wj computation sub-problem is represented by the minimization of Eq. (6):
1‖ ‖2 𝛽2 ‖ ‖2
‖a − aj wj ‖ + ‖wj ‖ (6)
2‖ j ‖2 2 ‖ ‖F
Through eliminating the L0 terms in Eq. (5), the function Eq. (6) has a global
minimum, which can be computed by gradient descent. The analytical solution to
Eq. (6) is shown in Eq. (7):
⎛ ( ) ⎞
−1 ⎜
F aj ⎟
wj = F ⎜
( ) 𝛽2 ( ( )∗ ( ) ( )∗ ( )) ⎟ (7)
⎜F a + F 𝜕x ⋅ F 𝜕x + F 𝜕y ⋅ F 𝜕y ⎟⎠
⎝ j
2
where F(.) and F −1 (⋅) denotes the Fast Fourier Transform (FFT) and reverse FFT,
respectively .F()∗ is the complex conjugate of F(⋅).
1‖ ‖2 ‖ ‖ ‖ ‖
‖aj − aj wj ‖ + 𝜆2 ‖aj ‖ + 𝜇‖∇aj ‖ (8)
2‖ ‖2 ‖ ‖0 ‖ ‖0
1‖ ‖2 ‖ ‖2 ‖ ‖2 ( )
‖aj − aj wj ‖ + 𝜆2 ‖aj − h‖ + 𝜇‖∇aj − v‖ + 𝜆 ‖h‖0 + ‖v‖0 (9)
2‖ ‖2 ‖ ‖2 ‖ ‖2
3 Experimental Results
3.1 Datasets
In order to testify the performance of our proposed method and implement the method
in practical scenarios, we gather the data from five classes of information management
specialty for the learning management system of our University. The data records of the
courses and students were extracted from the Department of Management Information
System, Shandong University of Finance and Economics and the Department of Elec‐
tronic Engineering Information Technology at Shandong University of Sci&Tech. The
most important information of the courses and students is mainly about the grades
corresponding to the courses. All of the students from the information management
specialty are freshmen in our University. Most of them have taken the courses of the
first year in their curriculum except three students have failed to go up to the next grade.
Thus, firstly we eliminate the records of the three students. Meanwhile, we collect the
knowledge including the programming skill that they have mastered through a ques‐
tionnaire. The courses that they have taken and the content that have grasped are
combined in the final datset. A part of the dataset is shown in Table 1, where 1 denotes
that the si student has mastered the tj course, and 0 denotes the opposite.
After gathering the data of the students from the five classes, comparing experiments
between state-of-the-art methods and our method are conducted. We choose several
state-of-the-art methods including collaborative filtering methods itermkNN, userkNN,
and the matrix factorization methods PureSVD.
3.2 Measurement
The knowledge from several experts on the courses in information management
specialty are adopted as ground truth in the experimental process. To measure the
performance of the comparing methods, we introduce the Hit Rate (HR) and the Average
Reciprocal Hit-Rank (ARHR) in the experiments, which are defined as shown in
Eqs. (11) and (12).
#hits
HR = (11)
#students
where #hits denotes the number of students whose course in the testing set is recom‐
mended by the expert, too. #students denotes the number of all students in the dataset.
1 ∑ 1 #hits
ARHR = (12)
#students i=1 pi
In this section, the experimental results calculated from the practical dataset. Table 2
shows the experimental results of the comparing methods in top-N course recommen‐
dation.
136 J. Lin et al.
Where HRi, ARHRi denotes the performance for classi, respectively. The experi‐
mental results shown in Table 2 demonstrate that our proposed method outperforms
state-of-the-art methods in most of course recommendations both in the HR and ARHR.
It shows that the sparse regularization term based on the prior knowledge from the
observation in our method are suitable for solving the problem of course recommenda‐
tion.
In order to illustrate the performance of our proposed method according to the
number of courses and topics included in the experimental testing. It shows in Fig. 2
that a higher accuracy is obtained when the number of courses increases. Meanwhile,
the courses included in our experiments are divided into 32 different topics, Fig. 3 shows
that the accuracy is also higher when there are more relative courses.
Fig. 2. Accuracy of our proposed recommendation system method due to the number of courses
Sparse Linear Method Based Top-N Course Recommendation System 137
Fig. 3. Accuracy of our proposed recommendation system method due to the number of topics
4 Conclusion
Acknowledgments. This work was financially supported by the Teaching Reform Research
Project of Undergraduate Colleges and Universities of Shandong Province (2015M111,
2015M110, Z2016Z036) and the Teaching Reform Research Project of Shandong University of
Finance and Economics (2891470), Teaching Reform Research Project of Undergraduate
Colleges and Universities of Shandong Province (2015M136). SDUST Young Teachers Teaching
Talent Training Plan (BJRC20160509); SDUST Excellent Teaching Team Construction Plan;
Teaching research project of Shandong University of Science and Technology (JG201509 and
qx2013286); Shandong Province Science and Technology Major Project (No. 2015ZDX
X0801A02).
138 J. Lin et al.
Disclosures. The authors declare no conflict of interest. The founding sponsors had no role in the
design of the study; in the collection, analyses, or interpretation of data; in the writing of the
manuscript, and in the decision to publish the results.
References
1. Ahn, H.: Utilizing popularity characteristics for product recommendation. Int. J. Electron.
Commer. 11(2), 59–80 (2006)
2. Albadarenah, A., Alsakran, J.: An automated recommender system for course selection. Int.
J. Adv. Comput. Sci. Appl. 7(3) (2016)
3. Aher, S.B., Lobo, L.M.R.J.: A comparative study of association rule algorithms for course
recommender system in e-learning. Int. J. Comput. Appl. 39(1), 48–52 (2012)
4. Chang, P., Lin, C., Chen, M.: A hybrid course recommendation system by integrating
collaborative filtering and artificial immune systems. Algorithms 9(3), 47 (2016)
5. Chen, Y., Cheng, L., Chuang, C.: A group recommendation system with consideration of
interactions among group members. Expert Syst. Appl. 34(3), 2082–2090 (2008)
6. Christakopoulou, E., Karypis, G.: HOSLIM: higher-order sparse linear method for top-N
recommender systems. In: Tseng, V.S., Ho, T.B., Zhou, Z.-H., Chen, A.L.P., Kao, H.-Y.
(eds.) PAKDD 2014. LNCS (LNAI), vol. 8444, pp. 38–49. Springer, Cham (2014). https://
doi.org/10.1007/978-3-319-06605-9_4
7. Kim, E., Kim, M., Ryu, J.: Collaborative filtering based on neural networks using similarity.
In: Wang, J., Liao, X.-F., Yi, Z. (eds.) ISNN 2005. LNCS, vol. 3498, pp. 355–360. Springer,
Heidelberg (2005). https://doi.org/10.1007/11427469_57
8. Mclaughlin, M., Herlocker, J.L.: A collaborative filtering algorithm and evaluation metric
that accurately model the user experience, pp. 329–336 (2004)
9. Ning, X., Karypis, G.: SLIM: sparse linear methods for top-n recommender systems, pp. 497–
506 (2011)
10. Ning, X., Karypis, G.: Sparse linear methods with side information for top-n recommendations,
pp. 155–162 (2012)
11. Omahony, M.P., Smyth, B.: A recommender system for on-line course enrolment: an initial
study, pp. 133–136 (2007)
12. Pan, J., Hu, Z., Su, Z., Yang, M.: Deblurring text images via L0-regularized intensity and
gradient prior, pp. 2901–2908 (2014)
13. Pan, J., Lim, J., Su, Z., Yang, M.H.: L0-regularized object representation for visual tracking.
In: Proceedings of the British Machine Vision Conference. BMVA Press (2014)
14. Philip, S., Shola, P.B., John, A.O.: Application of content-based approach in research paper
recommendation system for a digital library. Int. J. Adv. Comput. Sci. Appl. 5(10) (2014)
15. Sunil, L., Saini, D.K.: Design of a recommender system for web based learning, pp. 363–368
(2013)
16. Xu, L., Lu, C., Xu, Y., Jia, J.: Image smoothing via L0 gradient minimization. In: International
Conference on Computer Graphics and Interactive Techniques, vol. 30, No. 6 (2011). 174
17. Yap, G., Tan, A., Pang, H.: Dynamically-optimized context in recommender systems,
pp. 265–272 (2005)
18. Zheng, Y., Mobasher, B., Burke, R.: CSLIM: contextual slim recommendation algorithms,
pp. 301–304 (2014)
19. Zou, H., Hastie, T.: Regularization and variable selection via the elastic net. J. Roy. Stat. Soc.
Ser. B-Stat. Methodol. 67(2), 301–332 (2005)
The Research on the Container Truck Scheduling Based
on Fuzzy Control and Ant Colony Algorithm
1 Introduction
The loading/unloading equipment at most container terminals in China are mainly Quay
Cranes (QCs), Yard Cranes (YCs) and CTs. The scheduling system of CTs is well known
to possess complex logistics system characteristics for there are many factors effecting
scheduling, such as time, space, resources, and uncertain factors. Optimize the sched‐
uling system is meaningful, which can improve the efficiency of loading and unloading.
The physical entity resource network and the control decision-making information
network were integrated into the modeling and optimization architecture using Harvard
architecture and Agent based computing [1, 2]. In [3], Relationship between transport
tasks and service of CTS has been taken as a contract net using the fuzzy set theory and
method. The dispatching model based on Contract Network Protocol (CNP) using bidir‐
ectional negotiation is provided and fuzzy reasoning process of dispatching decisions is
suggested. But the distance from the current position of CT to the loading/unloading
position is calculated by the established coordinate system. In the case of road conges‐
tion, CT should choose a smooth route, so the distance should be the length of the route.
Recently, the path planning problem is a hot research area [4, 5], the main algorithms
used are genetic algorithm, ant colony algorithm and so on.
This paper proposes Multi-Agent contract net protocols based on two-way negotia‐
tion mechanism. The fuzzy control system is designed to couple multiple factors on the
tasks to assess to their dispatching emergency degree. The steps of ant colony algorithm
considering obstruction status were designed to find the optimal route. After fuzzifica‐
tion and defuzzification, the decision query table of dispatching plan is obtained.
Container terminal consists of containers, ships, handling equipment (QC, YC, and CT)
communication equipment, berths, container yards and human resource, etc. Superior
and subordinate subsystems have command and obedience relationships, and parallel
subsystems have collaboration, consultation, competitive relationships. Figure 1 shows
the hierarchical structure of the whole system.
As we can see, there are three types of individual agents, namely Control Agent
(fixed), Execution Agent (moving with wireless mobile terminal) and Operation
(Mobile) Agent (dynamically generated).
The Research on the Container Truck Scheduling Based on Fuzzy Control 141
Operation Agents are created dynamically and freely by the Central Processing
Agent throughout the whole system just like computer software, and they are assigned
by the Control Agent, executed by the Execution Agent.
In the traditional static scheduling mode, the truck is unloaded at half of the distance. A
dynamic dispatching strategy in which multiple tasks are matched with multiple CTs is
adapted in this study, which can guarantee the truck in the idle state can be quickly put
into the operation point which needs trucks. So how to arrange a truck for the task should
be considered. In this paper, considering the load of communication and consultation
efficiency in system, Multi-Agent contract net protocols based on two-way negotiation
mechanism was adopted to achieve the best arrangement. Figure 2 shows the contract
net.
The structure of fuzzy reasoning controller is shown in Fig. 3. Here are the two inputs,
namely, predicted travel distance, priority of handling operation. The output is
dispatching emergency degree. Dispatching decision knowledge set consists of n subset,
R1, R2, … Rn. Each subset contains some knowledge of dispatching decision, which is
described in a matrix form according to certain rules. The input information is given
according to bidding document and database of Agents. The meaning or interpretation
of the function module is based on the input linguistic variables to determine its value
corresponding fuzzy sets A1, A2; Effect of fuzzy reasoning is to use and expertise subset
R1, R2,… Rn, and the information of synthesis module A to obtain comprehensive
information on the fuzzy relation composition operations in order to generate output
result. The result is also indicated by fuzzy sets. Linguistic matching function with the
fuzzy decision module is translate the result into output language, or the result of this
reasoning is converted to exact amount of that task dispatching emergency degree, to
which Agents can refer to bid.
Heavy weights and the closed space on the dock make the truck more likely to be
involved in traffic crash which leads to road congestion. So the travel time and distance
from the current position of CT to the loading/unloading position cannot just calculated
according to the shortest route. Through judging obstruction status, the steps of ant
colony algorithm were designed to find the optimal route. Then the predicted travel
distance and time can be calculated.
4.1 The Ant Colony Algorithm Design for the Optimal Route of Container Truck
Definition and Control of Congestion
Definition and Control of Congestion. Traffic flow model is a kind of math equation
to express the correlation of traffic parameters like speed, density and flux etc. Obstruc‐
tion status can be judged according to related data gathered by GPRS. Take the spur
track i → j for instance.
The Research on the Container Truck Scheduling Based on Fuzzy Control 143
( )
uij = uijf 1 − kij ∕kij0
Suppose the average minimum distance headway at the port of congestion density
as hijs, and average minimum time headway as hijt, then
kij0 = 1000∕hijs
The judging procedure about the congestion for the spur track is as follows:
Step 1: If qij > qijm, then switch to step 3, otherwise, switch to step 2.
Step 2: If uij < uijf , then switch to step 3.
Step 3: The spur track is under congestion, and the jam knots are recorded in the
congestion table Ck.
Confirming Principle of Allowed Knot Set allowedk. Considering the reality of dock,
the container is abstracted into some grids. The dynamic control over the change of ants
position comes to be true through modifying Relation Matrices by utilizing coordinate
to confirm the direct pre-knot or sub-knot.
The allowed knot set allowedk equals n minus Ck and Pk at knot i for ant number k,
in which n is the total number of knots and Pk is the knots passed by ant k. There is no
repetition at the same knot for the consequent knots considering the traffic rules of
container terminals.
The Steps of Ant Colony Algorithm of Optimal Route for Container Trucks.
Step 1: Initialization. To set up a starting point and ending point and set the maximum
number of cycles Nmax. To initialize the control parameters α, β, Q (user-
defined). At the initial, m ants are randomly put on n nodes and the amount of
information on each path is equal. The data gathered by GPRS for every spur
track are: uij, hijs, hijt, qij.
144 M. Yu et al.
Step 2: Put the initial knot of ant k into Pk to search till find target knot. If does, end
the loop, and turn to step 3. If doesn’t find the target knot, determine allowedk
and work out the transfer probability pkij, then adjust knot and keep searching.
The transfer probability from knot i to j of ant k can be defined as:
⎧ 𝜏ij𝛼 𝜂ij𝛽
⎪ ∑ , s ∈ allowedk
pij = ⎨
k
𝜏is𝛼 𝜂is𝛽
⎪ s∈allowedk
⎩ 0, else
Step 3: End a loop iteration when the last ant find the target knot and record the length
Lk.
Step 4: Update the pheromone intensity
If the arrival time is between t and t + 1, the pheromone intensity is to be updated
according to the following formula:
∑
m
Δ𝜏ij = Δ𝜏 k
ij
k=1
{
Q
if e(i, j) ∈ Tk (t)
Δ𝜏ijk = Lk
0 else
Q is a constant denoting the pheromone intensity. Lk is the gross route length covered
in repetition by the ant number k. ρ symbolizes the left-over elements of pheromone,
and ρ ought to be defined as a number less than 1 in order to avoid the endless accumu‐
lating of pheromone.
Step 5: To compare all the routes passed by the ants and the minimum route length is
the optimal route. To output the result, the algorithm is to be ceased.
Case Analysis. The optimal route and the shortest route are shown in Figs. 4 and 5,
respectively.
The speed of the spur tracks which are under the congestion is 10 km/h, and the speed
of other spur tracks are 25 km/h. The length of each grid in the map is 50 m, and the
oblique line is simplified into two straight segments. The travel time and distance
between the current position of CT to the loading/unloading position are calculated. The
distance and time of the optimal route is respectively 1.25 km, 0.05 h, and the shortest
route respectively 0.95 km, 0.06 h. The application of ant colony algorithm can help to
search out the optimal route while the route is under the congestion. The study of optimal
route is helpful in instructing the running of container trucks so as to avoid or relieve
traffic congestion.
The Research on the Container Truck Scheduling Based on Fuzzy Control 145
4.2 Fuzzification
L=P−N
The fuzzy linguistic value is defined as unimportant, a little important, important,
very important and extremely important, and the corresponding fuzzy set is P1, P2, P3,
P4, VP.
Fuzzy Inference and Defuzzification. Method of centroid is used in this system. Clear
outputs corresponding to each of the input values are calculated through MATLAB tool.
Table 3 shows the decision query table.
In practical work, the terminal management system can directly query the table for
the dispatching emergency degree.
The Research on the Container Truck Scheduling Based on Fuzzy Control 147
The two axis represents the input of the system (L, R), and the vertical axis represents
the output (S). It can be seen from this figure that S increases with the decrease of L and
R. Space surface is smooth, illustrating the design of membership functions and fuzzy
rules are basically correct.
5 Case Analysis
There are four operations requested, shown in Table 4, and three idle CT, shown in
Table 5. After CT dispatching Agent announce the four tasks, three free CT Agents gain
task dispatching emergency degree by fuzzy reasoning controller, shown in Tables 6, 7
and 8.
148 M. Yu et al.
evaluated dispatching emergency degree of CT301 is the highest (shown in Table 9),
CT dispatching Agent rewards Agent of CT301 and meanwhile refuses to application
of CT311 and CT312.
Then CT311 and CT312 select tasks to bid, whose dispatching assessed value of
emergency is the highest among the remaining, namely QC111 and YC208. CT
dispatching Agent respectively issue orders to CT311 and CT312 after evaluating.
6 Conclusion
Acknowledgements. This paper has been funded by the National Natural Science Foundation
of China (No. 71672137 and No. 61503291).
References
1. Li, B., Li, W.: Container terminal logistics systems collaborative scheduling based on multi-
agent systems. Comput. Integr. Manuf. Syst. 17, 2502–2513 (2011)
2. Li, B., Li, W., Zhang, Y.: Agent-based modeling and simulation for vehicle dispatching at
container terminals. J. Syst. Simul. 20(19), 5158–5161 (2008)
3. Meng, Y., Shaomei, W.: Study on scheduling system based on multi-agent of container
terminal. In: Proceedings 2006 10th International Conference on Computer Supported
Cooperative Work in Design, Nanjing, China, pp. 579–584 (2006)
4. Li, G.R., Yang, D.B., Ren, D.W.: Path optimization algorithm of dynamic scheduling for
container truck. Jiaotong Yunshu Gongcheng Xuebao 12(3), 86–91 (2012)
5. Cao, Q., Zhao, F.: Port trucks route optimization based on GA-ACO. Syst. Eng. Theor. Pract.
33(7), 1820–1828 (2013)
Researches on the Analysis Framework of Application
Layer Communication Protocol Based on SQLite
1
China Shipbuilding Industry Systems Engineering Research Institute, Beijing, China
xwy0987@sina.com
2
School of Control and Computer Engineering, North China Electric Power University,
Hebei, China
weifengxu@163.com
Abstract. Since the concept of mobile Internet, big data and cloud computing
has been proposed, the format of the information undergoes a tremendous change,
information gradually appear problems in large quantities, the complexity and the
data form is not fixed. This is not a small impact and challenge to the traditional
fixed form protocol analysis. Dynamic processing and analysis of data is very
important for the data processing flexibility, easy scalability and better fault
tolerance. Dynamic is mainly reflected in the dynamics of dynamic definition,
dynamic analysis and dynamic processing. In the dynamic analysis of the data,
you don’t need to set the specific format of the data in the program, only the
framework of dynamic analysis need to be constructed in the sending and
receiving programs, the programs can read the format of data automatically, and
get the content of data easily. This approach can make the program more suitable
for the current application in size and flexibility than the traditional form of a
fixed communication protocol analysis. Redundant structures does not need to
add into the programs, which is no longer processing data passively. Data trans‐
mission becomes more convenient, because the data format needn’t to be token
care when the administrator wants to transfer the data.
1 Instruction
form protocol analysis. Dynamic processing and analysis of data is very important for the
data processing flexibility, easy scalability and better fault tolerance.
The analysis framework of application layer communication protocol based on SQLite
[6], which is proposed in this paper, presents the format of message defined in the program
in the form of database table, users can configure the table flexibly according to needs, and
dynamically analyze it when in using. The whole framework contains protocol develop‐
ment process, dynamic object generation process, protocol analysis process, message
combination sending process, message receiving process, data stored procedure, commu‐
nication process. Dynamic is mainly reflected in the dynamic definition of communication
protocol, dynamic analysis and dynamic processing. In the dynamic analysis of the data, you
don’t need to set the specific format of the data in the program, only the framework of
dynamic analysis need to be constructed in the sending and receiving programs, the
programs can read the format of data automatically, and get the content of data easily. This
approach can make the program more suitable for the current application in size and flexi‐
bility than the traditional form of a fixed communication protocol analysis. Redundant
structures does not need to add into the programs, which is no longer processing data
passively. Data transmission becomes more convenient, because the data format needn’t to
be token care when the administrator wants to transfer the data.
In view of the SQLite memory database belongs to the lightweight database, with
superiorities such as less resource consumption, portability, security, reliability, zero
management costs and so on, we will use the SQLite memory data as the base carrier to
implement the message format storage and high-speed information buffer in this paper.
SQLite memory database is an embedded database engine. Specifically suitable for appro‐
priate data access on a variety of equipment with limited resources (such as mobile phones,
pads and other intelligent devices) [3, 7]. It follows the ACID relational database manage‐
ment system, its design goal is embedded, and now has been applied in many embedded
products, and its source code is abundant on the official website. SQLite memory database
takes up very few resources, the memory is only occupied in the order of 100 k bytes in the
embedded device [8]. It can be supported by mainstream desktop operating systems like
Windows/Linux/Unix and all the mobile operating system platforms, and can be combined
with plenty of programming languages, such as C#, C/C+, PHP, Java and ODBC inter‐
face, and it is faster than MySQL and PostgreSQL, which are two of the world famous open
source database management systems [4]. The SQLite supports the most of the SQL92
syntax, and allows developers to use SQL statements to manipulate data in the database, and
the SQLite is just a file, that doesn’t need to be installed or started the server processes as
databases like Oracle and MySQL. It has the following characteristics:
(1) Lightweight: The SQLite is a built-in database, all the database operations can be
completed by it with a dynamic library.
(2) Portability: The SQLite can be run in a variety of environments, we can see from
the source code provided by the official website, not only the desktop side, but also
the mobile side is covered by the mainstream platform.
152 W. Xu et al.
(3) Security and Reliability: When a certain data need to be accessed by multiple
processes simultaneously, data in the database can be read but cannot be written at
the same time by these processes, which ensures the reliability of the data.
(4) Green software: Another feature of SQLite is green: its core engine doesn’t rely on
any third-party software, and installation does not be required. So a lot of trouble
can be saved at the time of deployment.
(5) Easy to manage: The various data information in SQLite involve graphics, tables
and other files are isolated from each other, to ensure the mutual interference would
not take place, and also facilitate the operations on the database.
Figure 1 shows the architecture design of the application layer communication protocol
analysis framework. In this framework, the definition between the sender and receiver
is very vague, so we do not deliberately distinguish the sender or receiver.
Fig. 1. Overall structure of the application layer communication protocol analysis framework
Firstly the framework provides the user with a data table that can be created and
modified, and this table stipulates the format in which the protocol is created. The
protocol can be analyzed by software after it has been created, that is, getting into the
protocol analysis process. During the protocol analysis process, the object of each
Researches on the Analysis Framework 153
attribute in the protocol is generated by starting the dynamic object generation process.
The message combination sending process should be started if the simulation software
needed to send data after the protocol has been analyzed.
During the message sending process, the UDP or DDS [2] interface in communica‐
tion process will not be called directly for data transmission, but starts the data stored
procedure firstly, stores the data to be sent into the memory database, and then throw a
specific message to the communication process, which initiates the send function after
receiving this message, retrieves the data to be sent from the memory database and sends
it to the destination. If a message sent by the external software is received by the
communication process of the simulation software, this message would not be analyzed
but the raw string should be stored in the memory database firstly after the UDP or DDS
communication interface receives the information, and then a specific message should
be thrown by protocol analysis process which retrieves the raw string data from the
memory database and initiates the protocol analysis function to complete the analysis
of the message after receiving it. In the analysis process, message is associated with the
specific agreement according to the information identification, and finally completes the
analysis of information according to the protocol. The data which have been analyzed
would be stored in the memory database again after the analysis process, and then a
specific message should be sent to inform the business application layer to complete the
corresponding business functions, such as display, calculation and so on.
receiving is centered on SQLite memory database. Taking memory data as the center is
an another important feature of the framework proposed in this paper.
The whole system is divided into three modules by function in order to implement the
message buffer: first, Socket communication module, the module is mainly used to
accept and transmit messages which require to be buffered, such as string, structure, etc.;
second, SQLite message access module, the module is to access messages, which need
to be buffered, in the SQLite database; third, the forwarding module of Windows
message mechanism, the module generates a message reminder in the process and then
distributes it; The overall structure is shown in Fig. 3.
156 W. Xu et al.
6 Conclusions
in the form of database table, users can configure the table flexibly according to needs,
and dynamically analyze it when in using.
The framework that is proposed in this paper can make the program more suitable
for the current application in size and flexibility than the traditional form of a fixed
communication protocol analysis. Based on the framework we proposed, redundant
structures does not need to add into the programs, which is no longer processing data
passively. Data transmission becomes more convenient, because the data format needn’t
to be token care when the administrator wants to transfer the data.
However, this paper only uses one type of database, SQLite, and only implements
one kind of communication mode, UDP, so that the application scope of this framework
is still limited. In the future work, increasing the type of database and expanding the
communication model will be an important way to develop this framework.
References
1. Harrison, T., Pyarali, I.: An object behavioral pattern for demultiplexing and dispatching
handlers for asynchronous events. In: 4th Pattern Languages of Programming Conference in
Allerton Park, Illinois, 2–5 September 1997 (1997)
2. An, K., Gokhale, A.: A cloud-enabled coordination service for internet-scale OMG DDS
applications. In: Proceedings of the 8th ACM International Conference on Distributed Event-
Based Systems, New York (2014)
3. Hunt, P., Konar, M., Junqueira, F.P., Reed, B.: ZooKeeper: wait-free coordination for internet-
scalesystems. In: Proceedings of the 2010 USENIX Conference on USENIX Annual Technical
Conference, vol. 8, p. 11 (2010)
4. Li, M., Ye, F., Kim, M., Chen, H., Lei, H.: A scalable and elastic publish/subscribe service.
In: 2011 IEEE International Parallel and Distributed Processing Symposium (IPDPS),
pp. 1254–1265 (2011)
5. Mei, H., Shen, J.-R.: Progress of research on software architecture. J. Softw. 17(06), 1257–
1275 (2006)
6. Lin, M.: Design and Implementation of a Personal Address Book Management System Based
on SQLite. Jilin University (2015)
7. Tong, S.: Study on Application of Mobile Meter Reading Technology Based on SQLite. Jilin
University (2015)
8. Yang, X.: Design and Implementation of Beidou Navigation System Based on SQLite
Database. Lanzhou University (2015)
The Implementation of Growth Guidance Factor Diffusion
via Octree Spatial Structures for Neuronal Systems
Simulation
Abstract. The BioDynaMo project was created in CERN OpenLab V and aims
to become a general platform for computer simulations for biological research.
Important development stage was the code modernization by changing the archi‐
tecture in a way to utilize multiple levels of parallelism offered by todays hard‐
ware. Individual neurons are implemented as spherical (for the soma) and cylin‐
drical (for neurites) elements that have appropriate mechanical properties. The
extracellular space is discretized, and allows for the diffusion of extracellular
signaling molecules, as well as the physical interactions of the many developing
neurons. This paper describes methods of the real-time growing brain dynamics
simulation for BioDynaMo project, a specially the implementation of growth
guidance factor diffusion via octree spatial structures.
1 Introduction
The BioDynaMo project [2] aims at a general platform for computer simulations for
biological research. It is collaboration between CERN, Newcastle university, Innopolis
university and Kazan state university where is all sides take possible part in the project
according their professional skills. The main idea of the project was to close the gap
between very specialized applications and highly scalable systems to give life scientists
access to the rapidly growing computational resource [1]. Since the scientific investi‐
gations require extensive computer resources, this platform should be executable on
hybrid cloud computing systems, allowing for the efficient use of state-of-the-art
computing technology.
First development stage was the code modernization based by neurodevelopmental
simulation software Cortex3D [15] transforming the application from Java to C++ and
changing the architecture in a way to utilize multiple levels of parallelism offered by
todays hardware.
In the current state of the project distributed teams working on first iteration: devel‐
oping spatial organization layer [7] using Octree Encoding with simulation of neuronal
growing and extracellular diffusion and visualisation of these processes.
These tools designed for modeling the development of large realistic neural networks
like as the neocortex, in a physical 3D space. Neurons arise by the replication and
migration of precursors that mature into cells that can expand axons and dendrites. Indi‐
vidual neurons are represented through spherical (for the soma) and cylindrical (for
neurites) elements that have appropriate mechanical properties. The growth functions
of each neuron are encapsulated in set of predefined modules that are automatically
distributed across its segments during growth. The extracellular space is also discretized,
and allows for the diffusion of extracellular signaling molecules, as well as the physical
interactions of all developing neurons.
Typical approaches for space discretization include logically structured grids, block
structured and overlapping grids, unstructured grids, and octrees. Structured (regular)
grids are relatively easy to implement, have low memory requirements but limited
adaptivity, whereas unstructured grids can conform to complex geometries and enable
non-uniform discretization.
Unstructured grids also have the overhead by explicitly storing element-node
connectivity information and in general being cache inefficient because of random
access. Regular grids are easy to generate but can be quite expensive when the solution
of the PDE problem is highly localized, and localized solutions can be executed more
efficiently using unstructured grids. Octrees [12] offer us a good balance between adap‐
tivity and good performance. They are more flexible than regular grids, the overhead of
constructing element-to-node connectivity information is lower than in unstructured
grids, and they allow calculations without the use of matrix’s.
In this work we focus on reducing the time to build data octree structures, memory
overhead and time to perform numerical solution of diffusion equation using these data
structures.
An octree is a tree data structure that is used for spatial decomposition in which every
node (octant) has maximum eight children (see, Fig. 1). They are analogous to binary
trees (maximum 2 children per node) in 1D and quadtrees (maximum 4 children per
node) in 2D. The only octant with no parents is the root, and an octant with no children
is called a leaf. An octant with one or more children is called an interior octant, octants
with the same parents called siblings. Complete octrees are octrees in which every inte‐
rior octant has exactly eight children. The depth of an octant from the root is referred to
as its level. Octrees can be effectively used to partition cuboid regions, and these regions
are refereed to as the domain of the tree. Geometric modeling technique called Octree
Encoding firstly was presented in [10] at 1981.
160 A. Sabitov et al.
Ci,j,k
Fig. 1. Left: One large cell neighboring 8 smaller cells. The ci represent the x components of the
intermediate concentration c defined at the cell faces. Right: Zoom of one computational cell. The
concentration components are defined on the cell faces, while the pressure is defined at the center
of the cell.
𝜕ci ( )
− D2 ▵d ci + kci = Ji 𝐫j , t . (1)
𝜕t
Here D2 – the diffusion coefficient, k is the degradation coefficient, △d is the Laplace
operator in 3D space, and the quantity Ji(rj, t) means the source [6].
Figure 1 illustrates unrestricted octree data structure (see e.g. [12]) with a standard
grid arrangement [8]. This is convenient since interpolations are more difficult with cell
centered data (see e.g. [13]). Coarsening is performed from the smaller cells to the larger
The Implementation of Growth Guidance Factor Diffusion 161
cells, i.e. from the leaves to the root. This procedure shown by algorithmic pseudo-code
(see, Algorithm 1).
Fig. 2. The dependence of diffusion step solving time on number of octree nodes: The red line –
regular grid, blue line – octree structure.
162 A. Sabitov et al.
4 Results
We simulated the diffusion process with 100,000 nodes for a regular grid and using
octree spatial structures. Figure 2 demonstrates the time dependence of diffusion
problem solution from number of octree nodes. Comparing the results of these two
modes, one can observe the efficiency of using octree on large number of nodes. By
using octree spatial structures we get a significant reduction in computation time.
In Fig. 3 shows the results of simulation as a contour plot of substance concentration
for system containing two sources. The the 2D snapshots are shown for 4 time moments.
The diffusion process is implemented realistically what can be seen in the Fig. 3.
Fig. 3. Visualisation of the diffusion process of two sources for 4 different iterations a) 15, b) 26,
c) 41, d) 70.
5 Conclusion
In computational physics, there has been plenty of work using either multilevel grids or
adaptive mesh refinement to improve the computational efficiency.
Quadtree or octree-based adaptive refinement have also proposed in [3, 5, 11]. An
octree-based algorithm for diffusion process simulation, developed in this work, is used
for modeling growth and development in neural networks [15]. This method adaptively
subdivides the whole simulation volume into multiple subregions using an octree. Each
leaf node in the octree also holds a uniform subgrid which is the basic unit for simulation.
A novel node subdivision and merging scheme is developed to dynamically adjust the
octree during each iteration of the simulation.
Further development of the project involves the development of octree algorithms
to perform in the cloud using parallel computations. The next stage in the development
of diffusion of substances will be devoted to the realization of existing algorithms by
using methods of parallel calculations.
The Implementation of Growth Guidance Factor Diffusion 163
Acknowledgements. This work was funded by the subsidy of the Russian Government to support
the Program of competitive growth of Kazan Federal University among world class academic
centers and universities.
References
Xuejing Li ✉ , Yajuan Qin, Haohui Fu, Zhewei Zhang, and Xiaorong Lin
( )
Abstract. For the diversity of applications and traffic, the requirement of Quality
of Service (QoS) is various. In order to improve utilization of resource, perform‐
ance of the network and experience of users, the QoS aware routing approaches
have been researched. As the Software Defined Networking (SDN) emerging and
the machine learning developing, this paper proposes a mechanism based on flow
classification and routing strategy selection. This architecture is running as
modules on SDN controller. In Mininet platform, we conducted experiments on
POX controller to verify effective of our design and test the performance. The
measuring results show that throughput and delay of the network is distinct while
using different routing theory.
1 Introduction
For the rapid development trend, high bandwidth and low delay requirements are major
challenges with the emergence of new techniques like mobile communication and 5G.
Since the network traffic is increasing due to a wider spectrum of new applications, it is
important to address the essential trade-off problem of appropriate allocation of restrict
network resources and relatively efficient routing computing method. The QoS demands
required by different applications are diversity, and there are various flows existed in
same application. In addition, even in the similar types of applications there exists some
different QoS demand flows. For instance, real-time multimedia applications usually
require high QoS demands on bandwidth and delay but are tolerant to packet loss rate,
while the high definition video streaming and the online competitive gaming have high
data rate need and strict latency guarantee respectively. Furthermore, in the scenario of
multimedia delivered network, there are diverse specific types of flows such as, video,
audio, imaging, etc.
Custom route of path decision should be made for each flow according to its QoS
requirements and network conditions, which is conducive to the suitable assignment of
limited network resource. On the one hand, if every flow selects the shortest path
greedily, it would bring about local congestion so that the subsequent coming packets
may be dropped. On the other hand, more packets could be accepted with balancing
network traffic although it may consume excessive resource and derive long propagation
delay. Therefore, the routing theory needs to be customized for some special flows
according to the perception of QoS requirements and the detection of network conditions
in union, which is conducive to optimal QoS guarantee and network resource orches‐
tration [1, 2].
Moreover, there is no one algorithm can outperform all the remaining ones in every
case, while there are always some scenarios where an algorithm performs well.
However, the QoS aware theory makes the routing algorithm even more complex
considering with classifying various flows to match dynamic and adaptive strategy
respectively. Hence, the routing strategy should be trade-off between the optimality and
simplicity without too much processing overhead or memory consumption due to its
high complexity. Compared to some steadily and inflexible strategy aiming to realize
better average performance, we present a novel mechanism to improve the network
performance and reduce the computing complexity to some extent.
Considering traffic distribution state, QoS requirements of flows and dynamic avail‐
able bandwidth resource, the proposed strategy enhances the high utilization of the
overall resource at the cost of increasing resource consumed by individual flows. And
to find the minimum cost path, the evaluation results show that it alleviates the perform‐
ance degeneration and simultaneously improves the utilization of whole network
resource.
SDN architecture consists of three layers: forwarding layer, control layer and appli‐
cation layer, as shown in Fig. 1. Physical switch is the crucial element of the forwarding
layer, which takes on all the performance of forwarding data in accordance with the
forwarding rules in flow tables provided by the controller. Besides, the centralized
control plane is core layer of the network which takes the responsibility of topology
discovery, traffic classification, strategy formulation and network configuration. There‐
fore, the network traffic is allocated to the paths operated directly by OpenFlow switches,
where the flow-handling rules are installed by the central SDN controller [3].
The OpenFlow switch is composed by three components including flow table, secure
channel and OpenFlow protocol as illustrated in Fig. 2. Flow table contains the flow-
oriented forwarding rules issued by the relevant controller. Secure channel ensures the
166 X. Li et al.
interaction between the switch and the controller. And OpenFlow protocol regulates
messages in secure channel to realize functions, like traffic statistics collection, access
control and network configuration.
As the SDN architecture can provide global visibility of network resources and
inherent programmable interfaces, some explorations on the fusion of SDN and other
technology have been carried out. Although some concepts have been presented several
years ago, there are still many issues worth to be researching further, such as efficient
self-organized routing mechanism, network resource management, QoS perception and
environment forecasting [4–6].
SDN is a promising concept with the powerful advantage to introduce new dimen‐
sions of flexibility and adaptability in current communication networks. For instance,
the realization of QoS aware routing becomes possible in an agile and dynamic manner.
Although some routing algorithms have been well researched, they were not applied in
communication networks due to its high implementation complexity and realization
costs. We propose the QoS-aware resource management routing mechanism, which
seems to be promising to provide better network resource management, traffic control,
and application classification [7].
SDN paradigm radically transforms network architecture and provides some char‐
acteristics of programmable, efficient, fine-grained, dynamic, accurate, global-viewed,
and superior computational capacity. All of these features prompt the development of
network with using advanced optimization algorithm.
Computational intensive machine learning algorithms integrate several online
routing algorithms for the real-time control of new connection requests. Depending on
the available resource and status of the traffic, the cognitive computing could classify
the flows and make adaptive routing strategies. Since new variables and additional
constraints are emerging over time, the routing mechanism is supposed to utilize
machine learning techniques to select the best sequence of accommodative decisions,
while considering the anticipative future scenarios for improving the performance of
network. In this paper, the proposed dynamic adaptive QoS-aware routing mechanism
is constrained with available network resource and distinct flow requirements upon SDN
platform [8, 9].
The rest of this paper is organized as follows. Section 2 proposes implementation of
the QoS-aware routing mechanism with algorithms in detail. Section 3 describes the
simulations. Section 4 analyzes the experimental results and evaluates the performance.
Finally, Sect. 5 outlines the future researches and makes a conclusion.
Our QoS aware routing selection computing mechanism consists of five modules,
including Traffic Measurement Module, Flow Classification Module, Topology Detec‐
tion Module, Routing Selection Module, and Configuration Updating Module, which
are working collaboratively on SDN controller.
First, the module Traffic Measurement collects these information of flow statistics
from the forwarding layer. Then, Flow Classification module differentiates the QoS
Intelligent Perceptive QoS Based Centralized Admission Control 167
requirements like bandwidth of real-time traffic based on analyzing the flow statistics
information using data mining technique, such as K-means algorithm. Moreover,
Topology Detection module gathers all the up-to-date information of the link state from
switches, such as delay and bandwidth, with using traffic monitoring tools like sFlow
or NetFlow. Furthermore, Routing Selection module includes two parts. On the one part,
the smaller bandwidth demanding flows match to the routing algorithm of Bandwidth
Reserved Shortest Path (BRSP). On the other part, the larger bandwidth demanding
flows match to the routing algorithm of Load Balancing Considered Polling Path
(LBCPP). Finally, the Configuration Update module updates the strategies by rotating
above processes if the degree of the variation exceeds the setting threshold.
This module holds the purpose to find the reasonable and effective flow classification
theory according to the perception of flow’s QoS requirements and the detection of link
states from nodes. To trade off the complexity and the practicability of mechanism, the
module measures the bandwidth requirement of flows, and calculates a critical value to
divide the flows into two types.
In order to classify the flows rationally, the clustering algorithm of machine learning
is employed in this process. The K-means classification method make the previous traffic
flows to fall into categories, such as mice prior flows and inferior elephant flows. When
new flows reach the switches, the flows will be matched to the suitable class and
forwarded along the path based on the adaptive theory.
It is inevitable that the status of flows is changing ongoing. For instance, if the size
enlarges and the amount increases, the mechanism should reserve more bandwidth to
guarantee for the mice flows admission. Therefore, the residual network resource
remained for the elephant flows should be confined to release more resource for the mice
flows with priority.
required bandwidth of incoming flow. This principle ensures that the link with lower
utilization rate has smaller cost value. When link utilization comes near the link capacity,
it will become expensive to occupy this link.
Due to the instability and the resource restriction of network environment, more atten‐
tions should be paid to the continuously emerging variation. For accommodating the
variance with regard to the distribution of the flows and the condition of the network
over time, the routing theory should be dynamically adjusted the critical value of flow
Intelligent Perceptive QoS Based Centralized Admission Control 169
classification and the amount of reserved bandwidth. Meanwhile, the flow table installed
in the switches should be updated necessarily for efficient network usage.
In the module, the update process would be triggered if the classification metric or
the reserved bandwidth is out-of-date, which means that the variance is exceeding the
threshold of extent or time. If some flows matching the first routing algorithm cannot
find feasible path using the reserved bandwidth, the flow would use the other routing
theory to ensure forwarding temporarily. And the reserved bandwidth should be
increased to avoid packet losing with providing better QoS.
In addition, the traffic forwarding rules in flow table are generated by the central
controller and installed to the corresponding OpenFlow-enabled switches. As the
network traffic and topology are unpredictable, the flow table will be dynamically
updated to keep consistent with the employed strategy. On one hand, paths are timely
evaluated whether new flow arrives or network changes. On the other hand, the global
statistical information is timely updated and the flow table is set to delete overtime
entries. We employ suitable period and threshold of updating theory to trade off the
consistency of real-time state information and overhead of frequently updating commu‐
nication in the control plane.
as clients, while H3 and H4 as servers. Meanwhile, 9 Open vSwitches were built as the
OpenFlow-enabled switches. And the bandwidth of each link was given to 10 Mbps.
Under stationary condition, the flow distribution and the network state are invariable
over time. We conduct experiments on the condition to test algorithms of our theory.
Under dynamic condition, the moment of flows arrival processes could follow the distri‐
butions like Poisson. If the size of the flows is given, the departure time of flow related
to the transmission delay could be equivalent to the bandwidth demand which is set as
Normal distribution. This work conduct experiments on Mininet to evaluate our mech‐
anism and compare the results in custom dynamic condition, in which the UDP traffic
of 5 Mbps is emulated as background flow from H1 to H3 by using the command of
iperf. Using single routing algorithm, the throughput and delay are measured respec‐
tively when sudden flows are generated at some point.
In the case of the application of real time multimedia, QoS metrics with respect to band‐
width, delay, delay jitter, and packet loss ratio are usually taken as major concerns of
receivers, while the senders care more about the traffic transmission in high load and
high randomness. For the network management, the network performance depends on
the metrics, such as average link utilization, resource consuming, packet loss ratio,
throughput, device energy efficiency, and average end-to-end delay. Therefore, the QoS
aware routing problem could be converted to the multiple objective programming, which
is to find a minimum cost of object with satisfying some QoS constraints.
In this paper, we focus on two metrics namely throughput and delay. Under stationary
condition, intuitively, our classified synthesized routing QoS mechanism has larger
throughput than BRST and smaller delay than LBCPP. Under custom dynamic condition
with adding background flows through the shortest path at the time of 50 s and removing
the flows at the time of 110 s, we compare the throughput and delay between the two
unclassified single routings as shown in Figs. 4 and 5. When the background flows inject
in the network, we could find that the throughput with BRST collapsed obviously, while
the other algorithm of LBCPP shows a higher throughput at same time. Similarly, the
latency of the BRST increases more evidently than the other. Therefore, from H2 to H4,
the results show that the proposed routing mechanism could reduce the delay for prior
mice-flows and increase the throughput for inferior elephant-flows. In the future, we
will employ the other dynamic condition with injecting the background traffic subjected
to Poisson distribution or Normal distribution.
Intelligent Perceptive QoS Based Centralized Admission Control 171
Based on the fusion of techniques, this work presents an intelligent perceptive QoS
mechanism for forwarding the diverse flows with the adaptive routing. The architecture
consists of five modules which are collaborating together on SDN controller. First, it
finds the feasible classification among flows according to the requirements of the band‐
width. Second, for the delay sensitive flows, the theory is made with bandwidth reser‐
vation for finding the shortest path in order to guarantee the delay demand and reduce
packet loss rate caused by propagation in network. Third, for the bandwidth sensitive
flows, the multipath polling method of load balance considered multipath is applied to
improve the network performance. Finally, it updates an adjustable traffic classification
threshold so as to keep optimal for both network performance and service quality.
In the future work, we would employ and exploit other computing algorithms to
implement the proposed mechanism. In the paper, the machine learning technology
named K-means clustering is used to classify the flows by finding the most suitable
dividing metrics for the previous period. Since the flow classification named 2-means
may not be efficient, other skills like reinforcement learning would be used to find the
optimal adaptive dividing metrics with real-time feedback. Moreover, in the updating
module, the predicted method could be exploited to get the potential growing tendency
of traffic features for more appropriate classification. In addition, the machine learning
could also be applied in the controller analysis, admission control, and packet security.
References
1. Vissicchio, S., Tilmans, O., Vanbever, L., et al.: Central control over distributed routing. In:
SIGCOMM 2015, 17–21 August, vol. 45, no. 5, pp. 43–56 (2015)
2. Han, L., Sun, S., Joo, B., et al.: QoS-aware routing mechanism in OpenFlow-enabled wireless
multimedia sensor networks. Int. J. Distributed Sensor Netw. 12(7), 9378120 (2016)
3. OpenFlow Specification. Open Networking Foundation. https://www.opennetworking.org/
sdn-resources/onfspecifications/openflow
172 X. Li et al.
4. Liu, J., Li, J., Shou, G., et al.: SDN based load balancing mechanism for elephant flow in data
center networks. In: International Symposium on Wireless Personal Multimedia
Communications, pp. 486–490. IEEE (2014)
5. Craig, A., Nandy, B., Lambadaris, I., et al.: Load balancing for multicast traffic in SDN using
real-time link cost modification. In: IEEE International Conference on Communications, pp.
5789–5795. IEEE (2015)
6. Li, J., Chang, X., Ren, Y., et al.: An effective path load balancing mechanism based on SDN.
In: IEEE International Conference on Trust, Security and Privacy in Computing and
Communications, pp. 527–533. IEEE (2014)
7. Durner, R., Blenk, A., Kellerer, W.: Performance study of dynamic QoS management for
OpenFlow-enabled SDN switches. In: IEEE International Symposium on Quality of Service,
pp. 177–182. IEEE (2015)
8. Mizrahi, T., Saat, E., Moses, Y.: Timed consistent network updates in software-defined
networks. IEEE/ACM Trans. Netw. 24, 1–14 (2016)
9. Abu Alsheikh, M., Lin, S., Niyato, D., et al.: Machine learning in wireless sensor networks:
algorithms, strategies, and applications. IEEE Commun. Surv. Tutor. 16(4), 1996–2018
(2014)
10. Mininet. http://mininet.org/
11. POX. http://www.noxrepo.org/pox/about-pox/
Research on the Shortest Path Problem Based on Improved
Genetic Algorithm
1
Information and Network Center, Tianjin University, Tianjin 300072, China
{wbl,knlu,tdee79}@tju.edu.cn
2
School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
yaosusu@tju.edu.cn
1 Introduction
The shortest path problem has been widely employed in many fields, such as intelligent
transportation [1], computer networks [2], robot path planning [3, 4] and so on. For the
shortest path problem, there are many deterministic algorithms, such as Dijkstra algo‐
rithm [5] and Ford algorithm [6]. However, in the practical application, the size of the
shortest path problem is expanding, the constraint condition of the shortest path problem
is increasing, and the complexity of the influencing factors is also increasing. The above
leads to the difficulty of solving the shortest path is also increasing. In the traditional
algorithm, it is difficult to effectively find a perfect solution to quickly find the optimal
solution or suboptimal solution. For the shortest path problem of large scale, the intel‐
ligent algorithm can satisfy the user’s time and precision conditions to obtain the optimal
or suboptimal solution.
Genetic algorithm (GA) has good adaptability, robustness and flexibility [7]. In
this paper, we use genetic algorithm to solve the shortest path problem. But, there
are some problems in genetic algorithm, such as premature convergence and weak
local search ability, low efficiency and low precision. These problems are caused by
the irrational design of population scale, the irrational design of fitness function and
the irrational selection of cross probability and mutation probability. In order to solve
the above problems, two kinds of fitness function transformation methods are
designed, and the crossover probability and mutation probability selection strategies
are optimized. The effectiveness of the algorithm to deal with the shortest path
problem is verified by simulation experiments.
2 Genetic Algorithm
The shortest path genetic algorithm consists of the following five parts: (1) the genetic
representation of the shortest path problem, (2) the encoding and decoding of the shortest
path problem, (3) designing an fitness function according to the merits of each chromo‐
some, (4) designing the genetic operators which used to change the genetic structure of
offspring, usually include selection, crossover, mutation and so on, (5) parame‐
ters setting, including population size, probability of applying genetic operator, etc. The
keys to solve the shortest path problem based on GA are the design of the fitness function,
the setting of the crossover operator and the mutation operator.
Genetic algorithm [8, 9] is an algorithm to simulate the biological evolution process
to find the optimal solution. Its main advantage lies in its good adaptability, robustness
and flexibility. First, as a kind of intelligent search algorithm, the optimization process
of genetic algorithms do not have much mathematical requirements. The genetic algo‐
rithm is modeled according to the specific problem, and the solution is found by evolu‐
tion without having to consider the specific internal operation problem. So it has a good
adaptability. And it borrows the evolutionary idea of biogenetics can carry out global
search very efficiently. The computational complexity of genetic algorithm is small
Begin
Initial generate
population
gen=0
Calculate per
Individual
fitness value
Compared with the traditional algorithm. At the same time, the genetic algorithm can
be hybridized with other related heuristic algorithms, which can be flexible to deal with
practical problems.
The basic process of genetic algorithm as shown in Fig. 1, including initial popula‐
tion, crossover, mutation, selection and so on. Individuals evolve in the direction of
higher fitness by crossover, mutation and selection operations to find the optimal solution
or suboptimal solution [10].
Coding is a key factor to improve the efficiency and success rate of genetic algorithms.
In this paper, we use the priority coding method to represent the shortest path problem.
The advantage of the priority coding method is that any encoding has a corresponding
path; most crossover operations are very easy to implement; there is no loop after
crossing.
Encoding: Randomly generate sequences to form the original chromosomes. The
length is the total number of points in the graph. According to the coding pseudo-code,
we can get the random sequence which named “chromosome”. The node ID in Fig. 2 is
a random sequence containing 11 points. The specific encoding pseudo code is as
Table 1.
for(i=1:to n)
chromosome[i]=i;
for(i=1 to n/2)
repeat
q=random(1,n);
p=random(1,n);
until p!=q;
swap(chromosome[p],chromosome[q]);
end
176 B. Wang et al.
Decoding: The decoding process similar to the depth traversal algorithm. Take the
Fig. 2 as an example, firstly, according to the topology of the path, we can see that node
Initialize i=1;l=1;path[l]=1;
while S_1 !=
temp_j=max(chromosome[j],j∈S_i);
if(chromosome[temp_j]!=0)
path[l]=temp_j;
l=l+1;
chromosome[temp_j]=0;
i=temp_j;
else
chromosome[i]=0;
if(l<1)l=1;break;
i = chromosome[l];
end
Research on the Shortest Path Problem Based on Improved Genetic Algorithm 177
1’s son nodes are node 2, node 3 and node 4. In the node ID, We can get their value, ID
[2] = 1, ID [3] = 10, ID [4] = 3. Secondly, choose the maximum value from the son
nodes of node 1, and then change the path to 1-3. Thirdly, repeat the first two steps until
find the shortest path. So we can get path 1-3-6-5-8-11 based on the node ID. Decoding
method pseudo code is as Table 2.
logistic
As can be seen from Fig. 3 the logistic function gets a value of either 1 or a value of
0 when the value of x is outside the [−10, 10]. If the logistic function is used as fitness
function directly, the fitness value of many chromosomes will be close to 1 or 0, so that
majority of the chromosomes in the population are less competitive which will lead to
precocious phenomenon. In order to avoid the premature phenomenon, a new fitness
function is designed for the logistics function.
⎧ 1
g ≥ 30%n
∗ ⎪ 1 + exp((f − favg )∕c)
f (i) = ⎨ 1 (2)
⎪ g < 30%n
⎩ 1 + exp(f − favg )
where f ∗ (i) is a new fitness value of the ith chromosome obtained by logistic function
transformation, f (i) is the original fitness value of the ith chromosome, favg is the average
value of the current population f (i), g indicates the number of chromosomes where the
difference outside range (−10,10), c is the order of magnitude of the maximum value of
| |
|f − favg |.
| |
The above two kinds of fitness function to meet the general fitness function necessary
conditions. At the same time, relative to other fitness function, they have the following
properties: (1) the fitness function is a nonnegative number, so the objective function
can be regarded directly as the original fitness function, regardless of its positive or
negative; (2) the original fitness function value is inversely proportional to the value of
the new fitness function, that is, the greater the value of the original fitness function the
smaller the value of the new fitness function; (3) to ensure that most of the chromosomes
in the population are highly competitive, and the probability of premature appearance
is reduced.
chromosomes adopt the non-equal probability pairing strategy, we use the correlation
function in Formula 3 to calculate the correlation between chromosomes. A higher
crossover probability is set for two chromosomes with uncorrelated correlations.
First of all, we selected a chromosome X that had not been operated by crossover,
and then calculated the correlation index between X and other non-overlapping chro‐
mosomes. According to the correlation index, the probability of crossover between the
selected chromosome and X was calculated. The roulette wheel method is chosen as the
selection operator. So, we can get the chromosomes that cross with X .
a. Correlation index function
m {
∑ 0, xj = yj
r(X, Y) = xj ⊕ yj , xj ⊕ yj = (3)
j=1
1, xj ≠ yj
Where Yi is a chromosome that has not been crossed, m is the total number of genes
in the two chromosome which has the larger number of genes, r(X, Yi ) is indicated that
the number of genes is not the same between X and Yi, the greater the r(X, Yi ), the smaller
correlation between X , and Yi, and the smaller the probability of invalid crossover oper‐
ation will be reduced.
b. The crossover probability of chromosome Yi selected to cross with chromosome X
( )
1 r(X, Yi ) − ravg
p(Yi |X ) = 1+𝜆 , i ∈ [1, m] (4)
m rmax − ravg
1 ∑ m { }
Where ravg = r(X, Yi ), rmax = max r(X, Yi ), i = 1, 2, … , m ,
{ m i=1 }
rmin = min r(X, Yi ), i = 1, 2, … , m , 𝜆 is a constant, and 0 ≤ 𝜆 ≤ 1.
⎧p ,f < f
⎪ m1 (p avg− p )k (f
pm = ⎨ m2 1 max − f ) , f ≥ favg
p − m1 (5)
⎪ m1 fmax − favg
⎩
180 B. Wang et al.
Where pm1 = 0.1, pm2 = 0.001, k1 is a constant, and 0 < k1 < 1, fmax is the largest
fitness value in the population, favg is the average fitness value of population, f is the new
fitness value of the chromosome.
4 Experimental Analysis
In order to verify the influence of the crossover operator, the improved genetic algorithm
using the new crossover operator is compared with the traditional genetic algorithm.
The results in Table 3 are the average of 100 tests. According to Table 3, it can be seen
that the new crossover operator can improve the accuracy of the genetic algorithm, and
can effectively avoid the algorithm into the local optimal solution.
In order to verify the validity of the algorithm, the Dijstra algorithm is compared with
the improved genetic algorithm. In Table 5, the optimal solution and the suboptimal
solution represent the average of the success cases, and the time represents the average
time of the cases. The genetic algorithm parameter settings: population size of about
1000 which slightly fluctuate for specific cases, mutation probability is 15%.
As can be seen from Tables 4 and 5, the genetic algorithm has a better success rate
than Dijkstra algorithm, and is more effective under large-scale data. Although the
probability of genetic algorithm to get the optimal solution is 100% in theory. Because
it is difficult to meet the individual diversity in the population in practical application,
the success rate cannot reach 100%.
From the experimental results, we can see that the genetic algorithm has a high
computational cost compared with the traditional algorithm. The improved genetic
algorithm is much more expensive than Dijkstra algorithm, especially the optimal solu‐
tion cost. For the above-mentioned problem, we can adopt pruning strategy. First, we
cut the size of the graph and reduce the size of the graph. And then, we can use the
genetic algorithm mentioned in this paper in the results of the pruning. So that we can
reduce the computational cost in the case of guaranteed algorithm search capabilities.
5 Concluding
The shortest path problem is an active research area. In this paper, we modify the fitness
function of genetic algorithm, and then design a new selection method for crossover
probability and a new selection method for mutation probability. These improvements
make the genetic algorithm have a good success rate. At the same time, the algorithm
proposed in this paper can easily combine with the pruning algorithm. So that we can
effectively reduce the computational complexity and guarantee the search ability of the
shortest path problem. However, this paper only considers the constraint condition of
designated points, the algorithm also has room for improvement.
References
1. Ko, Y.D., Jang, Y.J., Min, S.L.: The optimal economic design of the wireless powered
intelligent transportation system using genetic algorithm considering nonlinear cost function.
Comput. Ind. Eng. 89(C), 67–79 (2015)
2. Chitra, C., Subbaraj, P.: A nondominated sorting genetic algorithm solution for shortest path
routing problem in computer networks. Expert Syst. Appl. 39(1), 1518–1525 (2012)
3. Cheng, H.: Obstacle avoidance shortest path algorithm and its application. Electron. Des.
Eng. 16, 56–60 (2013). (in Chinese)
4. Qi, X., Liu, S.: Selection algorithm for QoS routing based on k-shortest paths. J. Jilin Univ.
(Eng. Technol. Ed.) 05, 526–530 (2005). (in Chinese)
5. Dijkstra, E.W.: A note on two problems in connexion with graphs. Numer. Math. 1(1), 269–
271 (1959)
6. Bellman, R.: On a routing problem. Q. Appl. Math. 16, 87–90 (1958)
7. Lin, L., Gen, M.: Priority-based genetic algorithm for shortest path routing problem in OSPF.
Stud. Comput. Intell. 187, 91–103 (2009)
182 B. Wang et al.
8. Holland, J.H.: Adaptation in Natural Artificial Systems, pp. 1–17. MIT Press, Cambridge
(1975)
9. Cao, Z.: The study on exhaust algorithm, search algorithm, dynamic design for 0–1 Knapsack
problem. Comput. Knowl. Technol. 5(12), 3193–3198 (2009). (in Chinese)
10. Srinivas, M., Patnaik, L.: Adaptive probabilities of crossover and mutation in genetic
algorithm. IEEE Trans. Syst. Man Cybern. 24(4), 656–666 (1994)
Semantic Web Languages for Policy
Enforcement in the Internet of Things
1 Introduction
While challenges associated with timely processing of sensor data have been
relatively successfully tackled by the advances in networking and hardware tech-
nologies, the challenge of properly handling data representation and semantics of
IoT descriptions and sensor observations is still pressing. In the presence of mul-
tiple organisations for standardisation, as well as various hardware and software
vendors, overcoming the resulting heterogeneity remains one of the major con-
cerns for the IoT community. Moreover, apart from the syntactic heterogeneity
(i.e. heterogeneity in the data representation, such as, for example, differences in
data formats/encodings), we can also distinguish heterogeneity in the semantics
of the data [2]. For example, different units of measurement and metric systems
are common causes for incompatibility and integrity problems.
c Springer International Publishing AG 2018
Q. Zu and B. Hu (Eds.): HCC 2017, LNCS 10745, pp. 183–188, 2018.
https://doi.org/10.1007/978-3-319-74521-3_21
184 R. Dautov and S. Distefano
The Semantic Web [1] is the extension of the World Wide Web that enables peo-
ple to share content beyond the boundaries of applications and websites. This
is typically achieved through the inclusion of semantic content in web pages,
which thereby converts the existing Web, dominated by unstructured and semi-
structured documents, into a web of meaningful machine-readable information.
The Semantic Web is realised through the combination of certain key technolo-
gies [4], whereas the presented research specifically focuses on the Web Ontology
Language (OWL) and the Semantic Web Rule Languages (SWRL) as the two
potential ways of defining and enforcing policies in the context of the IoT.
Semantic Web Languages for Policy Enforcement in the Internet of Things 185
3 Sample Scenario
The following use case scenario demonstrates the proposed idea of leveraging
the idle potential of OWL ontologies and focuses on a complex IoT system com-
posed of multiple sensing devices, deployed both indoors and outdoors. Some
of these devices are temperature sensors installed in rooms within a building.
It is assumed that whenever any of these temperature sensors indicates a value
exceeding a dangerous level of 50◦ , the situation has to be classified as critical,
and thus needs taking reactive actions. A possible way to handle this scenario
would be to define explicit policies for every single temperature sensor within
the building. In the worst case, such policies would be either ‘hard-coded’ with
numerous if/then operators (i.e. any modifications would lead to the source
code recompilation), or defined declaratively (i.e. stored in some kind of con-
figuration file to be dynamically fetched by the analysis component). In both
cases, however, the resulting knowledge base would be saturated by the excessive
1
https://www.w3.org/Submission/iot-lite/.
186 R. Dautov and S. Distefano
— OWL subclasses: OWL also provides a simpler and more explicit way of
defining classes and subclass relationships. It supports multiple and transitive
inheritance, and, as in many other programming languages, subclasses inherit
all the properties of their parent classes. The code snippet in Listing 1.2 contains
two definitions. The first definition simply states that any temperature sensor
is a sensing device. The second definition illustrates the transitive inheritance
through a subclass hierarchy that states – in simple words – that if a device is
installed in a room, then it is automatically assumed to be installed in a building
as well, which in turn means it is an indoor device.
Listing 1.2. Defining OWL subclass relationships.
iot:TemperatureSensor IS A ssn:SensingDevice
iot:RoomDevice IS A iot:BuildingDevice IS A iot:IndoorDevice
Taking together all three definitions, we now assume that there is a temperature
sensor ts reporting a temperature level of 60◦ in its covered rectangular area.
The automated reasoner will follow the steps below:
This way, the reasoning engine is able to detect a critical situation by inferring
implicit facts from the limited explicit information. It is worth noting that using
generic rules for a wide range of devices, as in the example above, does not
affect the flexibility of the proposed approach and its ability to define fine-
grained, targeted policies for individual devices. Apart from inheritance, OWL
also supports overwriting parent properties by subclasses. This means that it
is possible to enforce device-specific policies, which will overwrite the default
behaviour and apply only to those specific devices. This way, flexible policy
enforcement at various granularity levels can be achieved.
ontologies are known to be used in a wide range of scientific domains (for exam-
ple, see [5] for an overview of biomedical ontologies), which are not necessarily
closely connected to computer science, and allows even for non-professional pro-
grammers (i.e. domain specialists) to design policies.
— Extensibility: IoT systems may be composed of an extreme number of smart
devices spread over a large area (e.g. traffic sensors distributed across a city-
wide road network) and have the capacity to be easily extended (as modern
cities continue to grow in size, more and more sensors are being deployed to
support their associated traffic surveillance requirements). To address this scal-
ability issue, the proposed semantic approach, suing the declarative definition,
can extend the knowledge base to integrate newly-added devices in a seamless,
transparent, and non-blocking manner. The same applies to the reverse process
– once old services are retired and do not need to be considered anymore, the
corresponding policies can be seamlessly removed from the knowledge base, so
as not to overload the reasoning processes.
— Increase in reuse, automation and reliability: policy enforcement mechanisms
already exist in the form of automated reasoners for OWL/SWRL languages,
and the proposed approach aims to build on these capabilities. Since the reason-
ing process is automated and performed by an existing reasoning engine, it is
expected to be free from so-called ‘human factors’ and more reliable, assuming,
of course, the validity of ontologies and policies. Arguably, as the policy base
grows in size and complexity, its accurate and prompt maintenance becomes a
pressing concern so as to avoid potential policy conflicts.
References
1. Berners-Lee, T., Hendler, J., Lassila, O., et al.: The semantic web. Sci. Am. 284(5),
28–37 (2001)
2. Dautov, R., Kourtesis, D., Paraskakis, I., Stannett, M.: Addressing self-management
in cloud platforms: a semantic sensor web approach. In: Proceedings of the 2013
International Workshop on Hot topics in Cloud Services, pp. 11–18. ACM (2013)
3. Dautov, R., Paraskakis, I., Stannett, M.: Towards a framework for monitoring cloud
application platforms as sensor networks. Cluster Comput. 17(4), 1203–1213 (2014)
4. Hitzler, P., Krötzsch, M., Rudolph, S.: Foundations of Semantic Web Technologies.
Chapman & Hall/CRC, Boca Raton (2009)
5. Rubin, D.L., Shah, N.H., Noy, N.F.: Biomedical ontologies: a functional perspective.
Briefings Bioinf. 9(1), 75–90 (2008)
6. Sheth, A., Henson, C., Sahoo, S.S.: Semantic sensor web. IEEE Internet Comput.
12(4), 78–83 (2008)
7. Studer, R., Benjamins, V.R., Fensel, D.: Knowledge engineering: Principles and
methods. Data Knowl. Eng. 25(1–2), 161–197 (1998)
Network Traffic Prediction Based on Wavelet Transform
and Genetic Algorithm
1 Introduction
Network traffic data is a kind of time-series data. The analysis of traffic data can help
people know the running status of the network, which can do it more reasonably in
bandwidth allocation, control flow, routing control, and error control [1]. Thus, it is an
effective way to improve QoS (Quality of Service).
Today’s network traffic presents some characteristics as nonlinear, abrupt and non-
stationary. The first-generation network traffic prediction models, such as Markov
model, Poisson model, AR model and ARMA model, can’t describe the non-stationary
characteristics of traffic. Meanwhile the prediction accuracy is too low to be suitable for
the current network traffic prediction [1].
Later, a second-generation traffic prediction model, such as FARIMA, gray model,
neural network model, wavelet model, and support vector machine (SVM), which can
capture long-run and non-linear flow characteristics [2]. But there are still some short‐
comings, for instance, FARIMA can’t describe the non-stationary characteristics of
traffic; neural network model requires more training samples, and the algorithm is
complex; gray model only applies in the occasions which original sequence changes not
fast, according to the exponential law.
In order to overcome the limitation of single model and describe the characteristics
of network data more accurately, some hybrid models have been proposed, such as gray
neural network model [3], wavelet combined with neural network model [4], wavelet
combined with time series model [5], and more complex combination of wavelet, neural
network and time series model [6] etc. In some cases, the complex combination of model
methods may not reduce the forecast error of component. Although the method improves
the accuracy of prediction, the complexity will increase at the same time. This model is
suitable for occasions that require higher prediction accuracy and less time.
In this paper, a traffic prediction model combining wavelet transform and neural
network is proposed and optimized by genetic algorithm. Firstly, process the traffic data
by the wavelet decomposition, then the time series data will be decomposed into rela‐
tively simple components, for smoothing the original signal. Secondly, the BP neural
network is used to predict the high frequency components and low frequency compo‐
nents respectively. Considering the shortcomings of slow convergence and local opti‐
mization, the genetic algorithm with good global search ability is used to optimize the
weights and thresholds. Finally, the component is reconstructed to obtain the final
prediction result. Experiments show that wavelet decomposition can effectively increase
the signal stability, and BP neural network for non-linear changes in traffic has a better
prediction effect. Meanwhile, the introduction of genetic algorithms accelerates the BP
neural network convergence rate, and improve the prediction accuracy.
2 Technical Overview
Wavelet decomposition is proposed by Meyer and Mallat. Time series data are decom‐
posed into two parts: low frequency and high frequency coefficient.
The decomposition of time series is achieved by Mallat algorithm, and the decom‐
position relationship is as follows:
{
aj + 1 = h0 ∗ aj
j = 0, 1, … (1)
dj + 1 = h1 ∗ dj
S = Aj + Dj (3)
Network Traffic Prediction Based on Wavelet Transform 191
Neur
ons
Expected Output
— + Vector
(Instructor
Signal)
The
√ number of hidden layer neurons can be determined by the formula
m = l + n + a, and a is an integer between [1, 10].
S=l×m+m×n+m+n (4)
In the formula, S is the code length, and l, m, n are the number of neurons in the
input layer, hidden layer, and output layer respectively.
2. fitness function
∑
n
F = k( abs(yi − oi )) (5)
i=1
192 X. Zhao et al.
In the formula, n is the number of output node in the network output layer, k is the
coefficient used to adjust the range of fitness values, yi is the expected output of the ith
node of BP neural network, oi is the ith node prediction output.
3. selecting
We use the roulette method based on individual fitness selection strategy.
k
Pi =
∑n
k (6)
Fi
j=1 Fi
In the formula, aij is to the jth gene in the ith individual, amax is the upper bound of
the gene aij, amin is the lower bound of the gene aij, r2 is a random number, g is the current
number of iterations, Gmax is the maximum evolution number, r is a random number
between [0,1].
The network traffic data is self-similar (single fractal) on a large time scale, while multi‐
fractals appears on a smaller scale [10]. The model in this paper firstly processes the
flow data by wavelet decomposition, and the prediction model is composed of BP neural
network, as shown in Fig. 2. Genetic algorithm is introduced to optimize the weights
and thresholds of neural networks [11].
Network Traffic Prediction Based on Wavelet Transform 193
comparing the actual results. Among them, the error of component prediction
is not only positive but negative. After combination, part of the error can be
offset.
In this paper, network traffic data comes from the CERNET backbone network center
in Northeast China, which is offered by China Education and Research Network. Using
the Shenyang-Changchun 30 days of data (from April 31, 2016 to May 29, 2016), as
shown in Fig. 4. Network traffic is in units of GB, and selecting 2 h as the acquisition
time granularity, that is, 12 points a day, 30 days a total of 360 points.
For the data flow in Fig. 4, wavelet transform is performed using the formula (1) (2).
Selecting db4 as Wavelet base, decomposition scale L = 5, the high-frequency and low-
frequency component are {D1 (k), D2 (k), D3 (k), D4 (k), D5 (k), A5 (k)} and the original
signal S = D1 + D2 + D3 + D4 + D5 + A5, as the results shown in Fig. 5.
Network Traffic Prediction Based on Wavelet Transform 195
In order to speed up the convergence rate of neural network and improve the prediction
accuracy, the normalized network flow data after wavelet decomposition is processed
[12]. The formula is:
[ ]
′ X − Xmin
X = (9)
Xmax − Xmin
The normalized data is within the range of [0,1], Xmax and Xmin are the maximum
and minimum values respectively in the network traffic data set, X’ is the normalized
value, X is the original value of the variable.
Xi,j represents the network traffic on the ith day at j time, Xi represents the network traffic
on the ith day. The network traffic data is X = {(X1,1, X1,2, …, X1,n), (X2,1, X2,2, …, X2,n),
…, (Xm,1, Xm,2,…, Xm,n), …}; The specified k learning samples is P = ((X1, X2,…, Xj);
(X2, X3,…, Xj+1);…; (Xk, Xk+1,…, Xk+j-1)); The corresponding k teaching samples
T = (Xj+1; Xj+2;…; Xj+k). The purpose of learning is to correct the weight, using the
error between the neural network output corresponding to the k learning samples P1, P2,
…, Pk and the corresponding teacher samples T1, T2,…, Tk, so that the output is close
196 X. Zhao et al.
to the expected teacher sample. Xi+j is a seasonal time series with j as the cycle, Ti is the
teacher sample, the relationship between the two is Ti = Xi+j.
In this paper, m = 30, n = 12 and n = 2, the input layer of network training uses 24
neurons, the flow of 2d is mapped respectively; the output layer uses 12 neurons,
mapping the future 1d flow.
The data of the first 20d(days) were selected for training, and the data of the last 10d(the
last 120 points) were used for testing. The flow time series of 30d is:
{ ( ) ( ) ( )}
X = t1 = t1,1 , t1,2 , … , t1,12 ;t2 = t2,1 , t2,2 , … , t2,12 ; … ;t30 = t30,1 , t30,2 , … , t30,12 ;
The learning samples: P = {(t1, t2);(t2, t3);…;(t18, t19)}, and the corresponding
teaching samples: T = {t3, t4,…, t20}. The network is predicted after learning, using t19
(network traffic on day 19) and t20 (network traffic on day 20) to predict traffic on day
21, and using t20 and t21 to predict traffic on day 22, and so on.
Figures 6 and 7 show the results of the short-term prediction of the last 120 data 1
day in advance. Figure 6 is a comparison of the new model without genetic algorithm
(GA) optimization, and Fig. 7 is the comparison chart of new model after GA optimi‐
zation. The abscissa is the detection point and the ordinate is network traffic (GB).
Network Traffic Prediction Based on Wavelet Transform 197
The prediction results of Fig. 6 shows that the new model has a higher prediction
accuracy and a better degree of fit. Figure 7 shows that the new model is satisfactory.
To clarify the performance of the new model, the results of the model performance
parameters using GA or not are listed in Table 1.
MSE is the mean square error, MAE is the absolute error, RMSE is the root mean
square error, and R-square is the determination coefficient.
From Table 1, we can see that the MSE of the new model is reduced by 20.21%,
comparing the new model without the optimization, and the prediction accuracy is
improved. Meanwhile, the coefficient of determination of the new model is increased
by 7.23%, which improves the fitting degree. This indicates that genetic algorithm is
feasible and effective to optimize the new model.
Although the genetic algorithm needs time to run, it speeds up the neural network
convergence rate, and improve the prediction accuracy. The comparison of training
times and running time are shown in Table 2.
The above results prove the validity of using genetic algorithm. The new model can
predict the traffic of t21 and t22 according to t19, t20 and so on. The prediction results are
shown in Fig. 8.
Using new model with GA optimization to realize the short-term forecast 3 days in
advance, as shown in Fig. 9.
Network Traffic Prediction Based on Wavelet Transform 199
The parameters of the new model performance with GA optimization are listed,
which is 2d and 3d predicted in advance, as shown in Table 3.
From Tables 2 and 3, we can see that the new model has higher accuracy and better
fitting degree in short-term prediction. Although the fitting of the 3d prediction curve of
Fig. 9 is not superior to Figs. 7 and 8, the trend of the flow rate can be roughly described,
with little effect.
6 Conclusion
The results of MATLAB simulation show that the network traffic prediction model based
on wavelet decomposition and genetic algorithm has good accuracy and it is an effective
prediction model, which is mainly embodied in:
1. After wavelet decomposition, the original signal becomes more simple and has good
stability, which provides the stability for BP neural network prediction;
2. The traditional time series analysis method is difficult to ensure high accuracy. For
non-linear traffic flow, BP neural network prediction is better;
3. Genetic algorithm can improve the defects of BP neural networks, which can easily
fall into the local optimal solution and the convergence rate.
200 X. Zhao et al.
However, the model in this paper is rather complicated. The high-frequency and low-
frequency components obtained by wavelet decomposition are predicted by neural
network, and the computational cost of prediction is large, which affects efficiency of
the model. Therefore, in the case of the high accuracy, to further improve the efficiency
of the model is the future research direction of this paper.
References
1. Wang, Z.X., Sun, Y.G., Chen, Z.Q., Yuan, Z.Z.: Study of predicting network traffic using
fuzzy neural networks. J. Chin. Inst. Commun. 3(26), 136–140 (2005)
2. Hou, J.L.: Prediction for network traffic based on Elman neural network. Comput. Simul.
28(7), 154 (2010)
3. Liu, X.W., Fang, X.M., Qin, Z.H., Ye, C., Xie, M.: A short-term forecasting algorithm for
network traffic based on chaos theory and SVM. J. Netw. Syst. Manag. 19, 427–447 (2011)
4. Cao, J.H., Liu, Y.: Network traffic prediction based on grey neural network integrated model.
Comput. Eng. Appl. 05, 155–157 (2008)
5. Chen, Z.W., Guo, Z.W.: Simulation realization of prediction model based on wavelet neural
network. Comput. Simul. 25(6), 147–150 (2008)
6. Zhang, H., Wang, X.: Modeling and forecasting for network traffic based on wavelet
decomposition. Appl. Res. Comput. 08, 3134–3136 (2012)
7. Li, D.D., Zhang, R.T., Wang, C.C., Xiao, D.P.: A new network traffic prediction model based
on ant colony algorithm in cognitive networks. ACTA ELECTRONICASINI -CA 10, 2245–
2250 (2011)
8. Bai, X.Y., Ye, X.M., Jiang, H.: Network traffic predicting based on wavelet transform and
autoregressive model. Comput. Sci. 07, 47–49 (2007)
9. Deng, K.L., Zhao, Z.Y.: Research and Simulation of stock price prediction mode based on
genetic algorithm BP neural network. Comput. Simul. 26(5), 317 (2009)
10. Boris, T., Nicolas, D.: Self-similar processes in communications networks. IEEE Trans. Inf.
Theory 44, 1713–1725 (1998)
11. Yao, M.H.: Application of improved genetic algorithm in optimizing BP neural networks
weights. Comput. Eng. Appl. 24, 49–54 (2013)
12. Dang, X.C., Hao, Z.J.: Prediction for network traffic based on modified Elman neural network.
J. Comput. Appl. 10, 2648–2652 (2010)
Research and Optimization of the Cluster
Server Load Balancing Technology
Based on Centos 7
1 Preface
Most web service involved internet I/O resource, which also reflected part of real load for
each server. We should consider all those factors together above and make a rational
division of upcoming task for load balancer to enhance performance of the whole system.
In next section, we introduce the traditional WLC algorithm and the improved version.
There are three ways such as VS/NAT, VS/TUN and VS/DR to realize LVS under IP
load balancing technique. VS/DR could avoid load balancer being the system bottle-
neck. It could also reduce the complexity of configuration sometimes. The following is
the VS/DR architecture diagram of this experiment, as shown in Fig. 1.
CðSmÞ=Csum CðSiÞ=Csum
¼ minf g ð1Þ
WðSmÞ WðSiÞ
WLC’s aim is to find the server Si in server pool with the minimal value of C/W,
and then assign the new task to it. However, it is not reliable by just using the
connections and weight as we discussed above. Administrator with much experience
must change the value of weight manually during the whole procedure. Therefore, we
introduce and adopt the improved WLC algorithm in next section.
The k, l, c represent the percentage of each item, whose range in (0, 1) and sum
is 1. We could assign them specific values into the above formulas to reduce the
complexity and benefit for control variables. Therefore, load weight of server could be
estimated as formula (6). W(Si) is proportional to L(Si) while it is in inverse propor-
tional to F(Si). Load balancer collects L(Si) and F(Si) of each server while new time
slice ΔT begins [8].
LðSiÞ
WðSiÞ ¼ ð6Þ
PðSiÞ FðSiÞ
In order to improve fault tolerance and avoid server becoming overload during
ΔT while a new task comes, we put forward the concept of load redundancy to indicate
extra load that server could hold [9, 10]. Load redundancy represents the max extra
capacity within a slice of Dt as formula (7) [5].
PðSiÞ FðSiÞ Dt
RðSiÞ ¼ ð7Þ
LðSiÞ
Dt’s range is expected to be [5, 10] seconds. On the one hand, Dt to be small would
bring extra computation frequently, on the other hand, it would be too large to reflect
the redundancy timely. Load balancer also needs to set an initial minimal redundancy
threshold Rmin and a binary sort tree, only the R(Si) which is equal or greater than
204 L. Ding et al.
threshold could be added in that data structure. Every node of the tree has chance to
receive new task, but we optimally choose the largest value of R(Si), which means we
choose the smallest W(Si) with less load pressure [5, 12].
P(Si) and F(Si) could be estimated by software in next section. L(Si) has three
factors which change as time goes on, here are intuitive ways to get these values
respectively below.
(1) Real-time CPU utilization rate considers differences within two close moment from
the file “/proc/stat” on a Linux server, the content shows many kinds of time slices
whose unit are jiffies. Formula (8) shows totalTime for CPU.
totalTime ¼ user þ nice þ system þ idle þ iowait þ irq þ softirq þ stealstolen þ guest ð8Þ
We set totalTime as T1, idle to I1. Then after a period of a very short time, we
calculate totalTime, idle again and set them to T2, I2 respectively. Finally, formula (9)
shows the CPU utilization rate.
MemTotal Mfree
LmemðSiÞ ¼ ð11Þ
MemTotal
(3) We can install “iotop” software to check whole I/O occupancy rate under Linux.
The command “iotop-o” shows all processes/threads which include I/O operations.
We focus the column named “IO>” and assign the sum to Ldisk(Si).
X
n
LdiskðSiÞ ¼ IOðtidÞ ð12Þ
tid¼1
The communication between load balancer and servers use TCP socket [11].
Server, a TCP client, sends message L(Si), R(Si) and F(Si) regularly but P(Si) only
once. Load balancer, a TCP server, receives values of each server and calculates new
W(Si) and R(Si). If R(Si) > Rmin, we put this server into the tree. If R(Si) < Rmin, then
we remove this server out of the tree [5]. Traversing the tree in order ensures the
smaller node we pick up can deal with a new task easier.
Research and Optimization of the Cluster Server Load Balancing Technology 205
Secondly, we discuss about the factors k, c in L(Si) and F(Si) with the same 800
requests respectively as Tables 1 and 2 below.
Table 1 shows changes of TTLB while providing k with different percent. k1 has
more influence than k2 and (1 − k1 − k2), so CPU may have the highest priority. It is
similar in Table 2 with c. Setting appropriate factors can improve the situation [13].
Boundary values could not generate reliable result.
Finally, according to these analyses above, the improved algorithm lifts about
15–18% of load balancing performance in cluster environment. It is benefit for
improving stability and efficiency of the whole system.
3 Conclusion
This paper introduces the approach to enhance performance of cluster system and offers
better concurrency service after improving Weighted Least-Connection Scheduling.
Estimating weights reasonably and filtrating statistical data for delivering task accel-
erate procedure of response smoothly. The improved one contributes to science
experiment and business such as neural network, big data analysis, e-commerce, virtual
host service, etc. However, there are still many problems to solve [1, 12], such as
understanding more influenced factors fully, improving speed, etc. We will break the
limitation and strong the algorithm with the development of technology in the future.
References
1. Semchedine, F., Bouallouche-Medjkoune, L., Aïssani, D.: Task assignment policies in
distributed server systems. J. Netw. Comput. Appl. 34(4), 1123–1130 (2011)
2. Cardellini, V., Colajanni, M., Yu, P.S.: Dynamic load balancing on web-server systems.
IEEE Internet Comput. 3(3), 28–39 (1999)
3. Zhang, W., Wu, T., Jin, S., Wu, Q.: Design and implementation of a virtual internet server.
J. Softw. (2001). (in Chinese)
4. Zhang, W.: Research and implementation of scalable network services. Comput. Eng. Sci.
(2001). (in Chinese)
Research and Optimization of the Cluster Server Load Balancing Technology 207
5. Bryhni, H.: A comparison of load balancing techniques for scalable web servers. IEEE
Network 14(4), 58–64 (2001)
6. Penmatsa, S., Cronopoulos, A.T.: Dynamic multi user load balancing in distributed systems.
In: IEEE International Parallel and Distributed Processing Symposium (2007)
7. Wang, H.: Research on Adaptive Load Balancing Scheduling Strategy in Web Server
Cluster System. Jilin University, Changchun (2013). (in Chinese)
8. Lu, J., Chen, Z., Liu, A.: A dynamic regulation weight value scheme for LVS. J. Comput.
Eng. 14, 038 (2006). (in Chinese)
9. Wang, J., Pan, L., Li, X.: Dynamic feedback scheduling algorithm in LVS. J. Comput. Eng.
19, 014 (2005). (in Chinese)
10. Sheng, G.Q., Wu, S.J., Ping, M.X., Chan, W.D., Min, Z.: Design and realization of dynamic
load balancing based on Linux Virtual System. J. Comput. Res. Dev. 6, 002 (2004). (in
Chinese)
11. Chun, W.: Core Python Applications Programming. 3rd edn. Prentice Hall (2012)
12. Wang, C., Dong, L., Jia, L.: Load balance algorithm for web cluster system. J. Comput. Eng.
36(2), 102–104 (2010). (in Chinese)
13. Guo, C., Yan, P.: A dynamic load balancing algorithm for heterogeneous web server cluster.
Chin. J. Comput. 2, 004 (2005). (in Chinese)
Base Station Location Optimization Based on Genetic
Algorithm in CAD System
College of Management Science and Engineering, Shandong Normal University, Jinan, China
15554130027@163.com, xls3366@163.com, sdxyliu@163.com
Abstract. A good base station deployment plan can help network operators save
cost and increase total revenue significantly under the premise of ensuring
network quality. But in the past, base station location planning is often manually
based on the engineer’s experience. It has a lower efficiency and very high error
rate. In this paper, a new method based on genetic algorithm is proposed to opti‐
mize base station location. In our work, a CAD system based Google Earth and
ACIS is designed to provide data for Genetic algorithm and display the location
of base station in the reconstructed terrain. This system which takes three-dimen‐
sional geographic coordinates as the input of the algorithm is advanced and
different from the traditional method which only uses two-dimensional coordi‐
nates, that is, this three-dimensional system can better display the base station
location and take the height into consideration. The proposed method is based on
a mathematical model of base station location. Genetic Algorithm is used to find
the solution of this model so that it can effectively reduce the error rate of base
station location.
1 Introduction
Excellent base station location plan for operators can help to get the best results in the
mobile communications market in the increasingly fierce competition [1]. Early plan
and design of the base station is mainly by the network optimization engineer based on
experience and field measurements to choose [3, 4]. Obviously, this method is not very
scientific. Over-reliance on subjective factors, led to the results often far from the actual
optimum configuration. Telecommunication base station automatic programming CAD
system based Genetic Algorithm proposed in this paper is based on Windows platform.
The process of solve the problem of base station planning as follows. First, use
Microsoft visual studio, Xtreme Toolkit Pro, ACIS/HOOPS to build the development
environment. Second, using com technology import Google earth into the development
environment and accessing and transforming Google earth coordinates which include
longitude, latitude, and elevation. Third, use ACIS/HOOPS technology to reconstruct
Three-Dimensional Terrain according to the transformed coordinates. At the same time,
establish a multi-objective mathematical model about base station planning according
to geographic information and produce base station primary scheme using genetic algo‐
rithm. At last, it uses ACIS/HOOPS modeling techniques to display the primary scheme
on the reconstructed terrain 1. Structure of the process is shown in Fig. 1:
Development
environment based on
VC++ and ACIS
Fig. 2. The extracted actual coordinates Fig. 3. The transformed screen coordinates
210 Y. Wang et al.
The transformed screen coordinates can be used to reconstruct the terrain by ACIS
technology with api_face_spl_apprx, api_face_plane and api_loft_faces interface. The
reconstructed terrain was shown in the Fig. 4, the actual terrain was shown in the Fig. 5.
The reconstructed terrain was used to show the plan of base station location in Three-
dimensional 3.
Coverage rate is calculated by using the number of demand points which are covered,
divided by the total number of the demand points of the terrain points [4]. To avoid a
repeat count of demand points, we just put the demand point included in the base station’s
coverage area which has the nearest distance from the demand point compared to other
base stations. To formulate coverage targets we defined as follows: U is the set of all
demand points; j is the demand points which j ∈ U , T is the set of all base stations; i is
the base station which i ∈ T . dij is the distance between base station i and demand point
j, ri is coverage radius of base station i, the coverage area of base station i, denoted by
Ci, is defined as the collection of demand points which is within the coverage area of i
and has the shortest distance from i among all the base stations. Formally stated, Ci is
given by:
{ }
Ci = j ∈ U|dij ≤ ri ∧ (dij ≤ di′ j , ∀i′ ∈ T ∧ i ≠ i′ ) (1)
The ultimate goal of optimizing the coverage is achieve full coverage of all user
points by using the minimum number of base stations.
In the actual experiment, we defined fCT as a function of the base station coverage
goals. The total number of all demand points covered by all base stations is given
∑
by i∈T Ci. The total number of all demand points deployed on the terrain is denoted
by |U|. When fCT = 1, all demand points in the terrain are fully covered. However, a
fCT = 0.95 is quite acceptable. Although a fCT = 1 is too costly to achieve, fCT is
maximized in the model. fCT is defined as follows:
Base Station Location Optimization 211
∑
i∈TCi
MaxfCT = (2)
|U|
Cost is a very important objective optimization goals, the investment of base station
planning is about two-thirds of the total investment of the whole network. Therefore, a
lower cost of base station planning can effectively reduce the total network investment
so that it can enhance the competitiveness of the enterprise communication. The cost of
the base station planning is mainly from the built of base stations. In condition of each
base station has a fixed cost, the number of base stations becomes the most important
optimization goals.
In order to prevent the waste of the base station, the number of base station we
calculated by divide the terrain area by the area of the base station. There are three
different radii of base stations, so the number of base stations in the interval [20, 30]
inner. By experiment, when the number is 25, the cost is optimal.
Begin
IniƟal populaƟon
Crossover operator
YES
MutaƟon operator
NO
Calculate populaƟon fitness, to
determine whether
meet preset requirements
YES
No
Whether up to the
maximum IteraƟons
YES
End
In the CAD system, we design a dialog to run the Genetic Algorithm which is shown
in Fig. 7. The maximum Iterations input box is used to set the maximum number of
iterations. Color setting input box is used to set the color of base station. First we design
the maximum iterations is 100. The Genetic Algorithm didn’t up to the optimal result
and stopped at the 100th generation. The result which the base station’s color is green
is shown in Fig. 8. Second we design the maximum iterations is 1000. The Genetic
Algorithm up to the optimal result and stopped at the 523th generation. The result which
the base station’s color is red is shown in Fig. 8. The cycle is the signal of each base
station.
Fig. 7. Genetic Algorithm dialog Fig. 8. The plan of base station location (Color
figure online)
The iterative process of the Genetic Algorithm is shown in Fig. 9, we use 25 base
station and iterated 300 times, the coverage rate is as shown in the Fig. 9.
Base Station Location Optimization 213
5 Conclusions
In this paper, we proposed a new method based genetic algorithm to make base station
location plan. At this time, we developed a CAD system based Google Earth and ACIS
to provide data for Genetic Algorithm and demonstrate base station location plan in the
reconstructed terrain which is Three-dimensional.
The future work will continue research on improving traditional intelligent optimi‐
zation algorithm and implement it on a three-dimensional CAD system. Meanwhile, the
mathematic model of base station plan will be further design. The genetic algorithm will
be improved.
References
1. Xie, Q., Liu, X., Yan, X.: Research on station location optimization CAD system based on the
cooperative mode. In: Zu, Q., Hu, B. (eds.) HCC 2016. LNCS, vol. 9567, pp. 930–935.
Springer, Cham (2016). https://doi.org/10.1007/978-3-319-31854-7_102
2. Xie, Q., Liu, X., Yan, X.: Base station location optimization based on the Google Earth and
ACIS. In: Zu, Q., Hu, B. (eds.) HCC 2016. LNCS, vol. 9567, pp. 487–496. Springer, Cham
(2016). https://doi.org/10.1007/978-3-319-31854-7_44
3. Ren, S., Li, X., Liu, X.: The 3D visual research of improved DEM data based on Google Earth
and ACIS. In: Zu, Q., Vargas-Vera, M., Hu, B. (eds.) ICPCA/SWS 2013. LNCS, vol. 8351,
pp. 497–507. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-09265-2_51
4. Tao, M., et al.: SA-PSO based optimizing reader deployment in large-scale RFID Systems. J.
Netw. Comput. Appl. 52, 90–100 (2015)
214 Y. Wang et al.
5. Fogel, L.J., Owens, A.J., Walsh, M.J.: Artificial Intelligence Through Simulated Evolution.
John Wiley & Sons, New York (1966)
6. Schwefel, H.-P.: Numerical Optimization of Computer Models. John Wiley & Sons Inc.,
New York (1981)
7. Golberg, D.E.: Genetic Algorithms in Search, Optimization, and Machine Learning. Addison
Wesley, Boston (1989)
A Study of Optimal Multi-server System
Configuration with Variate Deadlines
and Rental Prices in Cloud Computing
Abstract. Cloud computing is becoming more and more popular and attracts
considerable attention. In the there-tire cloud environment, an important prob-
lem is to determine the optimal multi-server system configuration so that the
profit of the service provider can be maximized. In related work, the maximum
allowed waiting time of service is assumed to be a constant, and the rental price
is also assumed to be constant for all servers despite the fact that different
servers have different execution speeds. These assumptions may not be valid in
realistic cloud environments. In this paper, we propose an optimization model to
determine the optimal configuration of the multi-server system. There are two
major differences of the proposed model with that of the existing work. First, the
maximum allowed waiting time is not a constant and may change with different
service requests. Second, the situation that the servers with different execution
speed may have different rental prices is taken into account. Experiments are
carried out to verify the performance of the proposed optimization model. The
results show that the proposed optimization model can help the service provider
gain more profit than the existing work.
1 Introduction
Cloud computing is becoming more and more popular and attracts considerable atten-
tion [5]. In a cloud computing environment, there are three tiers, i.e., infrastructure
providers, service providers, and consumers (see Fig. 1) [1, 2]. An infrastructure pro-
vider maintains the basic hardware and software resources. A service provider could rent
a certain scale of software and hardware resources from infrastructure providers and
provides services to consumers. Consumers can submit their service requests to the
service provider and pay them based on the quantity and the quality of the services. In
above cloud computing environment, the problem of optimal multi-server system
configuration for profit maximization is introduced as a significant, new research topic in
computer science [3–5]. The configuration of a multi-server system is characterized by
two basic features, i.e., the size of the multi-server system (the number of rented servers)
and the execution speed of the multi-server system (execution speed of the rented
servers) [3]. The key issue of the multi-server system configuration problem is to
determine the optimal size and execution speed of the multi-server system such that the
service provider can gain the maximum profit.
Like any other business, the profit is the most important issue to a service provider.
The service provider’s profit is determined by two parts, i.e., the gained revenue and
the corresponding cost, the revenue minus the cost is its profit. For a service provider,
the revenue is the sum of the charge to the consumers, and the cost is the rental cost
paid to the infrastructure providers plus the electricity cost caused by energy con-
sumption. The charge to a consumer is related to the quantity and the quality of the
service, that is, if the quality of the service is guaranteed, the service is fully charged,
otherwise, if the quality of the service request is lower than the promised Quality of
Service (QoS), the service provider serve the service request for free as a penalty of low
service quality [3, 6, 7]. The electricity cost is linearly proportional to the number of the
servers and to the square of the server speed [8, 9].
Many existing works have been done on the problem of multi-server system
configuration for profit maximization in the literature [3–5]. Cao et al. [3] proposed a
single renting scheme to configure the multi-server system. Using this scheme, the
servers in the multi-server system are all long-term rented servers, so the system is lack
of elasticity and easily leads to resource waste and consumer loss. To overcome this
weakness, Mei et al. [4] proposed a combined renting scheme. Using combined server
renting scheme, the main computing capacity is provided by long-term rented servers,
and the rest is provided by short-term rented servers. In this service system, when a
consumer submits a service request, the service request will be first put into a queue
and wait in the queue until it can be served by any server. But in order to satisfy the
QoS requirement, the waiting time of each service request in the queue must be limited
within a certain range, which is named the maximum allowed waiting time and is
determined by the service lever agreement (SLA). To guarantee the QoS, the service
request should be started no later than the maximum allowed waiting time. So, for a
service request in the queue, if its waiting time has reached the maximum allowed
waiting time, system temporarily rents a short-term server to provide service and the
short-term server is freed when the service is finished. But in Mei et al.’s study, the
maximum allowed waiting time for all service requests is deemed to be the same, that
is, the service requests in the multi-server system have equal maximum allowed waiting
A Study of Optimal Multi-server System Configuration 217
times. This assumption may not be realistic. Therefore, this paper takes into account the
condition that different service requests may have different maximum allowed waiting
time.
In Mei et al.’s work [4], the servers with different execution speed is deemed to
have the same rental price. Thus the rental cost in their work is only related to the
number of the rented servers. But more realistically, the rental price of the server may
depend on the execution speed. Hence, the rental cost is not only related to the number
of rented servers but also the execution speed of the rented servers, i.e., the more (less,
respectively) number of the rented server is, the more (less, respectively) rental cost is,
also the faster (slower, respectively) execution speed of rented server is, the more (less,
respectively) rental cost is.
In this paper, the problem of optimal multi-server system configuration for profit
maximization is studied as an optimization model. In this optimization model, the
conditions that the maximum allowed waiting times may be variables and the rental
price of the server may be changed with the execution speed of the rented server are
taken into account. The contribution of the paper is summarized as follows.
The maximum allowed waiting times of different service requests are deemed to be
no longer a constant but random variables.
The condition that the rental price for different execution speed server may be
different is taken into consideration. And a rental price model is proposed.
An optimization model is proposed to solve the problem of optimal multi-server
system configuration for profit maximization.
Experimental studies are conducted to verify the performance of the proposed
optimization model. The results show that the proposed optimization model proposed
in this paper can result in more profit than that of existing work.
The rest of the paper is organized as follows. Section 2 presents the relevant models
and rental scheme, including a three-tier cloud environment model, a multi-server
system model, an energy consumption model and a combined renting scheme. In
Sect. 3, an important probability used in this paper is calculated. Section 4 describes
the rental price model. Section 5 establishes an optimization model to solve the
problem of optimal multi-server system configuration for profit maximization.
Section 6 verifies the performance of the proposed model through comparison with that
of existing model. Finally, Sect. 7 concludes the work.
This section introduces three relevant models, namely a three-tier cloud environment
model, a multi-server system model and an energy consumption model, and a com-
bined renting scheme.
provider maintains the basic hardware and software resources. The service provider
rents a certain scale of software and hardware resources from infrastructure providers
and builds its service platform to provide services to consumers. Consumers submit their
service requests to the service provider and pay based on the quantity and the quality of
the services. The infrastructure provides two kinds of resources renting schemes, i.e.,
long-term renting and short-term renting, and the rental price per server per unit period
of time with long-term renting is much cheaper than that with short-term renting [4]. For
a service provider, the gained revenue is the sum of charge to consumers, and the cost
involved is the rental cost paid to the infrastructure providers plus the electricity cost
caused by energy consumption. Thus the profit is generated from the gap between the
revenue and the cost [3, 5].
kr k
q¼ ¼ ; ð1Þ
ms mls
where
" #1
ðmqÞm Xm1
ðmqÞk ðmqÞm
pm ¼ þ :
m! k¼0
k! m!ð1 qÞ
static power P* (unit: Watt) [12]. The dynamic power model used in the paper is given
by Eq. (3),which is also used in [3–5].
Pd ¼ nsa ð3Þ
When n ¼ 9:4192 and a ¼ 2:0, the value of power consumption is close to the value of
the Intel Pentium M processor [13]. Therefore, the power per unit period of time for a
busy server is
P ¼ Pd þ P : ð4Þ
Denote by Wim (unit: second) the maximum allowed waiting time of service request
i in the queue before it is served. When a consumer submits a service request i to the
system which is then put into the queue, the system records Wim and starts counting its
waiting time Wi. The requests are assigned and executed on the long-term rented servers
according FCFS discipline. Once Wi of service request i reaches Wim , a temporary
short-term server is rented from infrastructure provider to process the request [4]. The
time that spending on renting activity is very short, so this time is ignored in this paper.
A Study of Optimal Multi-server System Configuration 221
In related work [4, 5, 7], Wim for all i in the queue is assumed to be a constant. This
assumption may not be realistic, such as for a service request i, the maximum allowed
response time may be shorter than Wim , which will lead to lower QoS. So the condition
that the Wim may change with different service requests is considered.
In this paper, the QoS of service request i is reflected by the response time Ti (unit:
second). Denote by Tim (unit: second) the maximum allowed response time that service
request i can tolerate before it is completed. Tim is assumed to be linearly proportional
to the service size ri, namely Tim / ri , i.e., Tim ¼ d ri , where d is a positive constant
and determined by the service-level agreement (SLA). It can be noted that the waiting
time plus the execution time is the response time, i.e., Ti ¼ Wi þ xi . Then
where
ds 1
k¼ : ð6Þ
s
According to assumption (iii), the service sizes ri’s are i.i.d. exponential r.v.’s with
the parameter l, so Wim (i 2 {1,2,3,…}) are i.i.d. exponential r.v.’s with the parameter
l=k. The probability density function (p.d.f.) of Wim is
l l
fWim ðtÞ ¼ ek t ; ð7Þ
k
Substitute (2) and (7) into (8), and after some manipulation, we have
pm l
PrðWi \Wim Þ ¼ 1 : ð9Þ
ð1 qÞ½kmdð1 qÞ þ l
In PrðWi \Wim Þ, all service requests, in spite of exceeding their maximum allowed
waiting times, will be waiting in the queue. However, in the combined rented scheme
multi-server system, the service request whose waiting time reaches its maximum
allowed waiting time will be removed out of the queue and assigned to a temporary
short-term server, which will reduce the waiting time of the following requests.
Therefore, it will reduce the probability that the waiting time reaches maximum
allowed waiting time of the following service requests. According to [11],
peðWi \Wim Þ is as:
ð1 qÞ PrðWi \Wim Þ
pe ðWi \Wim Þ ¼ ; ð11Þ
1 q PrðWi \Wim Þ
0.09
0.08
0.07
0.06
p (W <W m)
i
0.05
i
e
0.04
0.03
0.02
0.01
0
0 5 10 15 20 25 30 35 40 45 50
d
In this section, a new rental price model is described. In related work [4], Mei et al.
deemed that the servers with different execution speed have the same rental price. Hence,
they assume all long-term servers have the same rental price b (unit: cents per second) for
A Study of Optimal Multi-server System Configuration 223
each server, and all short-term servers have the same rental price c (unit: cents per second)
for each server as well, where b\c. This assumption may not be realistic.
In practice, the rental price for a server normally changes with the execution speed
of the server, i.e., a server with higher execution speed normally has a higher rental
price. In this paper, this condition is taken into account. Denote by s0 (unit: billion
instructions per second) the baseline execution speed [3]. The rental price of a
long-term rented server with execution speed s0 is b0 (unit: cents per second), and the
rental price of a short-term rented server with execution speed s0 is c0 (unit: cents per
second), where b0 \c0 . The infrastructure provider maintains some kinds of servers
with different execution speed s. s can be either higher or lower than s0, but s 2 S,
where S is the set of the possible execution speed. The service provider can select any
kind of server with the speed limited in S to rent.
Assume that there are two relation expressions between execution speed and the rental
price for long-term rented server and short-term rented server, respectively, as follows:
s
b s
¼ ð12Þ
b0 s0
s
c s
¼ ð13Þ
c0 s0
Hence, the rental price of a long-term or short-term rented with the execution speed
s can get from (12) and (13), respectively, as follows:
s
s b
b ¼ b0 ¼ s0 ss ð14Þ
s0 s0
s
s c
c ¼ c0 ¼ s0 ss ð15Þ
s0 s0
Figure 4 shows the rental price b and c with different execution speed, where
s0 ¼ 1:0, b0 ¼ 1:5 and c0 ¼ 3:0. In Fig. 4, s ¼ 0:5 and s ¼ 1 for (a) and (b),
respectively.
τ =0.5 τ =1
6 8
β β
γ 7 γ
5
6
4
5
β and γ
β and γ
3 4
3
2
2
1
1
0 0
0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
s s
(a) (b)
To determine the optimal configuration (m, s), the optimal number m and execution
speed s of the servers in the multi-server system, such that the service provider can gain
the maximum profit, an optimization model will be established in this section.
Denote by pf(t) (unit: cents) and R(t) (unit: cents) the profit and the revenue of the
multi-server system during the time interval ð0; t, respectively. Denote by Cl(t) and
Cs(t) the cost of the multi-server system on long-term servers and the short-term servers
respective during the time interval ð0; t, including the rental cost and the electricity
cost. Then the profit of multi-server system during the time interval ð0; t can be
calculated as:
As the combined rented scheme can guarantee the QoS to all service requests, in
steady-state, the revenue during the time interval ð0; t can be represented as
XNðtÞ
RðtÞ ¼ i¼1
ari ; ð20Þ
where a is a positive constant, which indicates the price per billion instructions (unit:
cents per billion instructions).
According to [10], R(t) is a compound Poisson process, with the mean
akt
E½RðtÞ ¼ akt Eðri Þ ¼ : ð21Þ
l
A Study of Optimal Multi-server System Configuration 225
The cost on the short-term rented servers in steady-state, Cs(t), is calculated as:
XNs ðtÞ ri c þ wP XNs ðtÞ
Cs ðtÞ ¼ i¼1
ðc þ wPÞ ¼ i¼1 i
r; ð22Þ
s s
where P is given by (4), c is given by (15), w is introduced in Sect. 2.3.
According to (18) and (21), we have
c þ wP hXNs ðtÞ i c þ wP
E½Cs ðtÞ ¼ E i¼1 i
r ¼ kpe ðWi \Wim Þ t: ð23Þ
s sl
The cost on the long-term rented servers in steady-state, Cl(t), is calculated as:
where
PNl ðtÞ !
ri
Pl ðtÞ ¼ m i¼1
Pd þ P t ; ð25Þ
ms
and b is given by (14). In Eq. (25), Pd is given by (3), and P* is introduced in Sect. 2.3.
According to (19), we have
then
½1 pe ðWi \Wim Þkt
E½Pl ðtÞ ¼ m Pd þ P t
mls ð27Þ
¼ m f½1 pe ðWi \Wim ÞqPd þ P gt;
furthermore,
Therefore, substitute (3), (21), (23) and (28) into (17), we have
akt c þ cP
E½pf ðtÞ ¼ kpe ðWi \Wim Þt
l sl ð29Þ
m fb þ w½ð1 pe ðWi \Wim ÞÞqnsa þ P gt:
226 Z. Kang and B. Yang
Then
E½pf ðtÞ ak c þ cP
¼ kpe ðWi \Wim Þ
t l sl ð30Þ
m fb þ w½ð1 pe ðWi \Wim ÞÞqnsa þ P g:
The key issue of the multi-server system configuration problem is to determine the
optimal values of m and s of the servers so that the service provider can get the
maximum profit. As the profit can calculated as follows:
s1
ak c0 s cP
pf ðm; sÞ ¼ þ kpe ðWi \Wim Þ
l ss0 l sl
ð31Þ
b
m f s0 ss þ w½ð1 pe ðWi \Wim ÞÞqnsa þ P g:
s0
Subject to
k ð33Þ
ms;
l
s 2 S: ð35Þ
6 Experimental Study
The proposed optimization model can be solved by any optimization algorithms, such
as Genetic Algorithm, Ant Colony Algorithm, Simulated Annealing Algorithm, etc. In
this paper, we use Genetic Algorithm to solve the optimization model and compare the
results with those obtained from the optimization model in [4].
In [4], Wim , b and c are all deemed to be constant. In order to increase the prac-
ticality of the proposed optimization model than [4], the conditions that the maximum
allowed waiting times of services are r.v.’s and the rental price changed with the
A Study of Optimal Multi-server System Configuration 227
execution speed are taken into account. In order to distinguish the proposed opti-
mization model and the compared model, the proposed model is named as new model
and the compared model is named as old model in this paper.
In order to use optimal algorithm to solve the new model and old model, it is
necessary to get a closed-form expression of pm. In this paper, we use the same
P ðmqÞk
m1
k! e . This expression is accurate
mq
closed-form expression as [3, 4], which is
k¼0
when m is not too small and s is not to large [14]. According to Stirling’s approxi-
pffiffiffiffiffiffiffiffiffi m
mation of m!, i.e., 2pm me [16], one closed-form expression of pm is given by
pm ¼ pffiffiffiffiffiffi q m
1q
. In the following, we solve new model and old model based on
ð1qÞ 2pmðeeq Þ þ 1
above closed-form expression of pm.
Table 1. The optimal configuration and maximum profit obtained from old model.
m* s* pf*
6 1 57.9124
The QoS to services is guaranteed by the both models, so the revenue keeps same
constant in both models.
In the first experiment, we explore the difference between configuration (m, s) and
(m*, s*), and the difference between profit pf and pf’, with different s0 when parameter
s ¼ 1. The results are showed in Table 2. And Fig. 5a shows the changing trend of pf
and pf’ with different s0 when s ¼ 1. We can see that the figure shows the increasing
trend of pf and pf’ when s0 is increasing from 0.2 to 2. That is because with the
228 Z. Kang and B. Yang
Table 2. The difference between configurations (m, s) and (m*, s*), and the difference between
profits pf and pf’, with different s0 when s ¼ 1.
s0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2
(m, s) (10,0.6) (10,0.6) (10,0.6) (10,0.6) (10,0.6) (10,0.6) (10,0.6) (10,0.6) (10,0.6) (10,0.6)
pf 23.1808 46.8233 54.7041 58.6445 61.0087 62.5849 63.7107 64.5551 65.2118 65.7372
(m*, s*) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0)
pf’ 19.9803 43.6800 51.5799 55.5299 57.8999 59.4798 60.6084 61.4548 62.1131 62.6398
increment of s0, the rental prices b and c are both decreased. Hence the cost on
long-term servers and short-term servers are both decreased when the number and
execution speed of servers remain unchanged. In addition, the obtained profit pf to the
new model proposed by this paper is always greater than the pf’, which indicates that
the new model can gain more profit than old model by more reasonable configuration.
It is can be seen from Table 2, the optimal configuration obtained from new model is
10 servers with 0.6 execution speed, while that obtained from compared old model is 6
servers with 1.0 execution speed.
Table 3. The difference between configurations (m, s) and (m*, s*), and the difference between
profits pf and pf’, with different s0 when s ¼ 1:5.
s0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2
(m, s) (30,0.2) (15,0.4) (15,0.4) (15,0.4) (15,0.4) (10,0.6) (10,0.6) (10,0.6) (10,0.6) (10,0.6)
pf 11.8929 45.6617 56.4112 60.9118 63.2843 64.8931 66.0435 66.8462 67.4324 67.8758
(m*, s*) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0)
pf’ −38.608 29.9072 46.9823 54.1312 57.8999 60.1681 61.6569 62.6957 63.4542 64.0281
70
60
50
profit
40
30
20
pf
pf ’
10
0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
s0
In the second experiment, the configurations (m, s) and (m*, s*), the profits pf and
pf’ are compared, with different s0 when s ¼ 1:5, and the result is shown in Table 3.
Figure 5b shows the changing trends of pf and pf with different s0 when s ¼ 1:5. It can
A Study of Optimal Multi-server System Configuration 229
be seen that the changing trends of two curves are similar from Fig. 5a When s0 is 0.2,
the value of pf’ is a negative. Which indicates that the optimal configuration gained
from old model isn’t much truthful, when s ¼ 1:5, and s0 = 0.2.
In the third experiment, we explore the difference between configuration (m, s) and
(m*, s*), and the difference between profit pf and pf’, with different s when s0 ¼ 0:6.
And the results are showed in Table 4. We can see that the configuration (m, s) changes
from (6, 1) to (15, 0.4), when s changes from 0.5 to 1.5, then (m, s) remain unchanged
when s increases. Figure 6a shows the changing trends of profit pf and pf’ with
different s when s0 ¼ 0:6. It is can be seen from the figure, the two curves show
different changing trends, i.e., the pf shows increased trend, but pf’ shows decreased
trend. That is because when s is greater or less than 0.6, b and c show different trend
with the increasing of s, i.e. b and c show decreased trend when s is greater than 0.6,
but show increased trend when s is greater than 0.6. Furthermore, pf and pf’ shows
different changing trends.
Table 4. The difference between configurations (m, s) and (m*, s*), and the difference between
profits pf and pf’, with different s when s0 ¼ 0:6.
s 0.5 1 1.5 2 2.5
(m, s) (6,1) (10,0.6) (15,0.4) (15,0.4) (15,0.4)
pf 55.1413 54.7041 56.4112 58.7676 61.1497
(m*, s*) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0)
pf’ 55.1413 51.5799 46.9823 41.0467 33.3840
Table 5 shows the results of the difference between configuration (m, s) and (m*, s*),
and the difference between profit pf and pf’ are explored with a given s when s0 ¼ 1:2.
Figure 6b shows the changing trends of profit pf and pf’ with different s when s0 ¼ 1:2.
From the figure, we can see that the profit pf and pf’ show increasing trend when s is
increasing from 0.5 to 2.5. That is because with the increasing of s, b and c are both
decreased when s is not more than 1.0. It is shown in Table 4, s and s* are not more than
1, so b and c are both decreased with the increase of s in both configurations. Although
with the increase of the number of servers the rental cost shows increase trend, the
decrease trend is deeper with the increase of the execution speed of servers, so the rental
cost integrally shows increase trend. As a consequence, the profit of the multi-server
system increases. In addition, the profit obtained of our proposed new model is greater
than the profit of compared old model. And with the increasing of s from 0.5 to 2.5, the
gap is more and more great. It is can be seen from Table 4, the optimal configuration
obtained from this paper proposed new model is 10 servers with 0.6 execution speed
when s is 0.5 and 1, 15 servers with 0.4 execution speed when s is from 1.5 to 2.5, while
the counterpart obtained from compared old model is 6 servers with 1 execution speed.
In essence, the difference of configuration leads to the gap of profit.
230 Z. Kang and B. Yang
70 70
pf pf
65 pf ’ pf ’
68
60
66
55
profit
profit
50 64
45
62
40
60
35
30 58
0.5 1 1.5 2 2.5 0.5 1 1.5 2 2.5
Table 5. The difference between configurations (m, s) and (m*, s*) and the difference between
profits pf and pf’ with different s when s0 ¼ 1:2.
s 0.5 1 1.5 2 2.5
(m, s) (10,0.6) (10,0.6) (15,0.4) (15,0.4) (15,0.4)
pf 59.3206 62.5849 64.7123 66.6311 67.7390
(m*, s*) (6,1.0) (6,1.0) (6,1.0) (6,1.0) (6,1.0)
pf’ 58.7258 59.4798 60.1681 60.7965 61.3701
7 Conclusions
References
1. Lee, Y.C., Wang, C., Zomaya, A.Y., Zhou, B.B.: Profit-driven service request scheduling in
clouds. In: Proceedings of the 10th IEEE/ACM International Conference on Cluster, Cloud
and Grid Computing, pp. 15–24 (2010)
2. Chen, J., Wang, C., Zhou, B.B., Sun, L., Lee, Y.C., Zomaya, A.Y.: Tradeoffs between profit
and customer satisfaction for service provisioning in the cloud. In: Proceedings of the 20th
International Symposium on High Performance Distributed Computing, pp. 229–238. ACM
(2011)
3. Cao, J., Hwang, K., Li, K., Zomaya, A.Y.: Optimal multiserver configuration for profit
maximization in cloud computing. IEEE Trans. Parallel Distrib. Syst. 24(6), 1087–1096
(2013)
4. Mei, J., Li, K., Ouyang, A., Li, K.: A profit maximization scheme with guaranteed quality of
service in cloud computing. IEEE Trans. Comput. 64(11), 3064–3078 (2015)
5. Li, K., Mei, J., Li, K.: A fund-constrained investment scheme for profit maximization in
cloud computing. IEEE Trans. Serv. Comput. (2016). IEEE Early Access Articles
6. Ghamkhari, M., Mohsenian-Rad, H.: Energy and performance management of green data
centers: a profit maximization approach. IEEE Trans. Smart Grid 4(2), 1017–1025 (2013)
7. Liu, Z., Wang, S., Sun, Q., Zou, H., Yang, F.: Cost-aware cloud service request scheduling
for SaaS providers. Comput. J. 57, 291–301 (2013)
8. de Langen, P., Juurlink, B.: Leakage-aware multiprocessor scheduling. J. Sig. Process. Syst.
57(1), 73–88 (2009)
9. Mei, J., Li, K., Hu, J., Yin, S., Sha, E.H.-M.: Energy-aware preemptive scheduling algorithm
for sporadic tasks on DVS platform. Microprocess. Microsyst. 37(1), 99–112 (2013)
10. Ross, S.M.: Introduction to Probability Models, 11th edn. Elsevier, London (2014)
11. Boots, N.K., Tijms, H.: An M/M/c queue with impatient customers. Top 7(2), 213–220
(1999)
12. Ding, Y., Qin, X., Liu, L., Wang, T.: Energy efficient scheduling of virtual machines in
cloud with deadline constraint. Future Gener. Comput. Syst. 50, 62–74 (2015)
13. Enhanced Intel® SpeedStep® technology for the Intel® Pentium® M processor. White
Paper, Intel, March 2004
14. Li, K.: Optimal configuration of a multicore server processor for managing the power and
performance tradeoff. J. Supercomput. 61(1), 189–214 (2012)
15. Chen, J., Wang, C., Zhou, B.B., Lee, Y.C., Zomaya, A.Y.: Tradeoffs between profit and
customer satisfaction for service provisioning in the cloud (2011)
16. https://en.wikipedia.org/wiki/Stirling’s_approximation (2016)
17. Ghamkhari, M., Mohsenian-Rad, H.: Energy and performance management of green data
centers: a profit maximization approach. IEEE Trans. Smart Grid 4(2), 1017–1025 (2013)
18. Li, K., Liu, C., Zomaya, A.Y.: A framework of price bidding configurations for resource
usage in cloud computing. IEEE Trans. Parallel Distrib. Syst. 27(8), 2168–2181 (2016)
A Composite Anomaly Detection Method
for Identifying Network Element Hitches
1 Introduction
In the field of network management, fault warning is a very advanced subject. First of
all, the traditional method to deal with faults is to remedy the situation after mal-
function, which is neither predictable nor effective. In this situation, operation and
maintenance work will faces many difficulties. As a direct manifestation of devices
status, the data which comes from network element can reflect the state of the device. In
addition, it can helps operators to analyze and make decisions more accurately.
The data structure and data mining are carried out to explore the abnormal data in
the network equipment, which can helps user to quickly identify the possible failure
and achieve the target of early warning. Nevertheless, there are great challenges analyst
will meet in operator’s network management environment, because of the diversity and
complexity of the data comes from network elements. It’s also worth mentioning that
the method to get structured log data has objective impact on the result of anomaly
detection. Consequently, a comprehensive anomaly identification scheme for network
element log data is required for operation and maintenance work.
Taking into account the different needs of operator business analysis, we propose
diverse analytics solutions to match different business scenarios. In this paper, we
propose a time-series-based abnormity detection method for identify network element
hitches. Meanwhile, we propose a solution to meet the needs of high dimension
analysis. In the experimental phase, we use actual fault data provided by the operator to
evaluate the precision of our algorithm. The experiment results indicate that our
methods can find abnormal state of network element efficiently and accurately.
2 Related Work
The current anomaly detection algorithms can be divided into four categories: statistical
method, clustering-based method, distance-based method and density-based method.
Meanwhile, density-based anomaly detection is a hot topic in this field. Based on the
local density anomaly detection algorithm [1] proposed by Breunig in 2000. Following
researchers gradually improved and developed the algorithm to make it suitable for
different scenarios. Some researchers attempted to use inexpensive local statistics to
reduce the sensitivity of choosing parameter k [2]. In the paper of Lazarevic [3], Local
Outlier Factor (LOF) is used on multiple projections and combines the results for
improved detection qualities in high dimensions. [4, 5] presents effective methods to
measure the similarity between objects, which can increase the stability and precision
of outlier detection. But these algorithms is difficult to support large-scale data pro-
cessing needs, because of the complexity of the network element’s log data. Alterna-
tively, considering of the time characteristics of log data, timing analysis is another
feasible way to detect anomalies. In the field of time series analysis, moving average
model, linear regression model, polynomial regression model and exponential
smoothing are common algorithms for time series analysis. In statistics, a moving
average (MA) is a calculation to analyze data points by creating series of averages of
different subsets of the full data set [6]. The weighted moving average model (WMA) is
commonly used in the analysis of natural sciences and economics [7, 8]. Besides these,
linear regression was the first type of regression analysis to be studied rigorously, and
to be used extensively in practical applications [9]. The polynomial regression model
which is advanced by Gergonne [10, 11] had been applied extensively in a lot of
professions. Exponential smoothing was first suggested by Robert Goodell Brown in
1956 [12], and one of the commonly used model is known as “Brown’s simple
exponential smoothing” [13]. In exponential smoothing model, the current time can be
suspected by the past time, while the impact of recent time is stronger than remote time.
Another practical model to analysis time series is autoregressive integrated moving
average model (ARIMA) [14, 15], this model can predict the future value of a specified
time series on the basis of its past performance, but the basic ARIMA can’t retain the
seasonal characteristics of a time series.
All above models have their own advantages in time series analysis, and are
effective mediums for anomaly detection. In the process of analysis of network ele-
ment’s log data, both statistical method and density-based method can be reasonable
applied based on their properties.
234 D. Zhang et al.
If network management analysts would like to detect anomaly for a specific mode,
they can choose our time-series-based scheme. In the process of specific anomaly
detection, we combine qualitative analysis with quantitative analysis. On one hand,
generally determine the presence of abnormal period by creating ARIMA model. On
the other hand, use our comprehensive time series model to detect specific anomalies. If
network management analysts would like to contain the features of network element in
analysis process as much as possible, they may choose our LOF-based scheme.
Considering of high dimension input, we use principal component analysis (PCA) to
reduce the dimension and retain characteristics of raw data. In this scheme, analysts can
get global anomalies from log data.
ARIMAðp; d; qÞ ð1Þ
p denotes autoregressive lagged item, q denotes moving average lagged item, and d
denotes the order of difference. In our method we use ARIMA model to draw the curve
of specific mode which can help us find suspected anomalies qualitatively.
where yt is the measured value at time t and b denotes the weight coefficient which can
measure the impact of recent trends on current time.
– The seasonal average factor St is defined as
1X n
St ¼ Yt ¼ Yi ð3Þ
n i¼1
Yi denotes the time units which are at the same period of Yt , n denotes the number
of time units we concerned about.
– The predicted value of comprehensive time series model Pt is defined as
max½aCt þ ð1 aÞSt
a
ð4Þ
s:t: a 0:
In this manner, we can reduce the influence of single factor on predicting value,
also increase the tolerance of extremely value. Is easy to understand that Pt retains the
impact of recent trends and the relevance of data in the same period, which can predict
the value of specified mode’s time series at time t reasonably and effectively.
structured log data from network element has a large variety of characteristics, the
traditional LOF algorithm will meet the performance bottleneck in computing process.
Therefore, the PCA method is used to reduce the dimensions of input data, while
preserving the feature information of multi dimension data.
We use Euclidean distance to measure the difference between network element’s
log data in different periods.
sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
X n
dðxðt1 Þ xðt2 ÞÞ ¼ ðxi ðt1 ÞÞ xi ðt2 ÞÞ2 ð5Þ
i¼1
The abnormal degree of log data is measured by local outlier factor (LOF) [1]. And
the local outlier factor (LOF) is defined as
P lrdk ðoÞ
lrdk ðpÞ
o2Nk ðpÞ
LOFk ðpÞ ¼ ð6Þ
jNk ðpÞj
where Nk ðpÞ denotes the k-distance neighborhood of p. In other words, Nk ðpÞ contains
every element whose distance from p is not greater than the k-distance. The k-distance
of p is defined as the distance dðp; oÞ between p and an object o such that there are at
least k objects m 2 D dðp; mÞ dðp; oÞ, and there are at most k 1 objects
m 2 D dðp; mÞ\dðp; oÞ
The local reachability density of p is defined as
jNk ðpÞj
lrdk ðpÞ ¼ P ð7Þ
rdk ðp; oÞ
o2Nk ðpÞ
For a specified network element, the statistical distribution of log modes per hour
form snapshots which can reflect the status of the network element. According to the
definition of local outlier factor (LOF) [1], the higher the value of local outlier factor,
the more likely the object is an anomaly. Based on this, we abstract the status of
network element per hour as objects (denoted as xðtÞ) which can calculate their
abnormal degree by LOF. Getting the values of LOF for every xðtÞ, we can infer the
occurrence time of anomalies.
abnormal patterns of NE’s log data. Obviously, for network management analysts this
method is effective, accurate and targeted-oriented. And that is to say, this model can
effectively find the anomaly of specific log mode.
The global abnormity detection of network element considering of multiple attri-
butes of network logs, which can retain the comprehensive information of network
element’s log data. But the existence of relevance and the mutual influence between
attributes also has effects on detect results. Therefore, it’s suitable for network man-
agement personnel have a global perception of abnormity. And that is to say, network
management personnel can choose to find targeted anomaly and global abnormity
flexibly according to different scenarios.
4 Experiments
5 Conclusion
References
1. Breunig, M.: LOF: identifying density-based local outliers. ACM SIGMOD Rec. 29(2),
93–104 (2000)
2. Kriegel, H.P., Schubert, E., Zimek, A.: LoOP:local outlier probabilities. In: ACM
Conference on Information and Knowledge Management, CIKM 2009, Hong Kong, China,
DBLP, pp. 1649–1652, November 2009
3. Lazarevic, A., Kumar, V.: Feature bagging for outlier detection. In: Eleventh ACM SIGKDD
International Conference on Knowledge Discovery in Data Mining, pp. 157–166. ACM
(2005)
4. Schubert, E., Wojdanowski, R., Zimek, A., et al.: On evaluation of outlier rankings and
outlier scores. In: Proceedings of the 2012 SIAM International Conference on Data Mining,
pp. 1047–1058 (2012)
5. Guan, H., Li, Q., Yan, Z., et al.: SLOF: identify density-based local outliers in big data. In:
Web Information System and Application Conference, pp. 61–66. IEEE (2015)
244 D. Zhang et al.
6. Booth, E.G., Mount, J.F., Viers, J.H.: Hydrologic variability of the cosumnes river
floodplain. San Francisco Estuary Watershed Sci. 4(2), 1–19 (2006)
7. Domańska, J., Domański, A., Czachórski, T.: Fluid flow analysis of RED algorithm with
modified weighted moving average. In: Dudin, A., Klimenok, V., Tsarenkov, G., Dudin, S.
(eds.) BWWQT 2013. CCIS, vol. 356, pp. 50–58. Springer, Heidelberg (2013). https://doi.
org/10.1007/978-3-642-35980-4_7
8. Glabadanidis, P.: Market timing with moving averages. Int. Rev. Finance 15(3), 387–425
(2015)
9. Yan, X., Su, X., World Scientific: Linear Regression Analysis [Electronic Resource]: Theory
and Computing (2009)
10. Gergonne, J.D.: The application of the method of least squares to the interpolation of
sequences. Historia Mathematica 1(4), 439–447 (1974)
11. Stigler, S.M.: Gergonne’s 1815 paper on the design and analysis of polynomial regression
experiments. Historia Mathematica 1(4), 431–439 (1974)
12. Brown, R.G.: Exponential Smoothing for Predicting Demand, p. 15. Arthur D. Little Inc.,
Cambridge (1956)
13. Brown, R.G.: Smoothing, forecasting and prediction of discrete time series. J. R. Stat. Soc.
127(2) (1964)
14. Mills, T.C.: Time Series Techniques for Economists. Cambridge University Press,
Cambridge (1990)
15. Jenkins, G.M.: Autoregressive-Integrated Moving Average (ARIMA) Models. Encyclopedia
of Statistical Sciences. Wiley, New York (2004)
D-SVM Fusion Clustering Algorithm
Based on Indoor Location
1 Introduction
With the rapid development of mobile Internet and mobile terminal equipment, indoor
positioning has become the front of data information technology research [1]. Under the
support of the world’s four major satellite navigation systems, outdoor location services
have been widely into people’s lives. While 80% of the daily time in indoor activities, with
the increasing number of large buildings, indoor location services in commercial applica‐
tions, public safety and other aspects of the application of the most broad prospects. In
indoor environments, satellite system signals cannot be used due to building occlusion and
multipath [1]. At present, location fingerprint location based on signal strength (RSSI) is
widely used in indoor positioning. Fingerprint localization makes use of the fingerprint
feature of multipath non line of sight caused by the complex indoor architecture, and
improves the positioning accuracy under the traditional positioning problem.
Indoor positioning method (AOA, Arrival, of angle of arrival and time of arrival
(Angle) location TOA, Time of Arrive) location, time difference of arrival (TDOA Time,
Difference of, TDOA/AOA Arrival) positioning hybrid positioning, based on the
received signal strength indicator (RSSI, Received Signal Strength Indication) location.
TOA, TDOA positioning requires high-precision hardware synchronization, AOA
This method is generally divided into off-line training and online matching stage. The
off-line training phase of the mobile device to acquire the RSSI sequence, AP sequence
signal intensity and location information for the RSSI signal vector and stored in the
fingerprint database, the formation of space vector space according to a certain density
of fingerprint acquisition point position information (Fig. 1).
When the amount of fingerprint data is large, clustering algorithm is widely used to
optimize the structure of fingerprint database. Generally based on the data mean division
D-SVM Fusion Clustering Algorithm 247
method, density based density method, data model based machine learning methods.
This paper uses divisive hierarchical clustering algorithm, the entire data set as a cluster,
and then use the adaptive mesh split into multiple clusters in data modeling, the entire
data set model, using two different clustering algorithms are fused to play two kinds of
localization algorithm advantages, makes the use of the smallest the data space to get
the location information of the most useful, to improve the positioning accuracy.
⎧
⎪ (̂x, ŷ ) = 1 ∑ w (x , y )
K
⎪ K i=1 i i i
⎨∑ k (1)
⎪ wi = 1
⎪ i=1
⎩
(1) We need to give K neighbor points different weights to improve the system posi‐
tioning accuracy, learn from the classic signal positioning algorithm based on active
RFID system. The system selects some reference points and arrangement of refer‐
ence tags in the positioning area, the signal strength compared to the tag and refer‐
ence tag values, obtained several reference tags to be located closest to the label,
and then based on these coordinates weighted reference tags to estimate the coor‐
dinates positioning label. Using the weight design idea of the system, the weights
are redesigned.
(2) The size of the Euclidean distance D square reflects the weight change, the smaller
the D, the greater the weight, and directly depends on the current Euclidean distance
Di, enhanced the Euclidean distance Di weight ratio (Fig. 2).
Data point density and DBSCAN group connected to a cluster. The density of points
m/nV, M is the number of data points that N input falls in a small region around the
point and V is the area of the volume. The region is considered to be a hypersphere
radius, so the MinPts threshold density can be determined by specifying a parameter.
For given and MinPts values, DBSCAN finds a dense point input set and expands it
by merging adjacent dense regions. A point can be a dense (core) or non-dense point
(non core business point). Non-dense point is a boundary point dense region or noisy
pattern.
In the plane space data density space vector set as fingerprint clustering based on
density clustering algorithm, the number of clusters is not need to set in advance, so this
method is suitable for clustering point distribution of unknown data sets clustering.
Based on density of DBSCAN is a widely used mesh algorithm, the super ball algorithm
number of data samples within the region to measure the regional density of the fast.
DBSCAN algorithm can discover clusters of arbitrary shape, and effective identification
of singular points. The discovery of DBSCAN cluster grid data and characteristics, and
each cluster according to the data of different densities of different probability markers,
matching the priority of such arrangement has probability sequence cluster set the
maximum probability in the process of matching, the matching accuracy is very good
optimization (Fig. 3).
⎧ g(x) = 𝜔x + b = 0;
⎪
⎨ 𝜔xi + b ≥ 1(yi = 1); (4)
⎪ 𝜔xi + b ≤ −1(yi = −1).
⎩
When the training set is nonlinear, the input feature space by a nonlinear function is
mapped to a higher dimensional space m, X ∈ Rn → X ∈ Rm these classes can be used
to classify a hyperplane and get discriminant function.
𝜔Φ(x) + b = 0 (5)
Discriminant function:
In the formula C is the parameter, the objective function should look for the
maximum hyperplane as far as possible and the guarantee data point deviation quantity
is smallest, parameter C controls before two weights.
The original data points using density based DBSCAN select various density clusters.
After finding the clusters of various shapes at the fastest speed, the singular points are
excluded and the probability density of each cluster is sorted. In the matching process,
the cluster with higher probability is preferred (Fig. 4).
250 Z. Deng et al.
By means of formula (6), we can solve an optimal hyperplane of classifying the data
of the two kinds according to the mathematical model. Before the N data samples
obtained by the cluster, with permutations and combinations into the SVM classifier
training, N(N − 1)∕ 2 hyperplane model DBSCAN (Fig. 5).
⎧ 1 ∑
N
Fig. 5. Fingerprint database statistical positioning results and optimal classification plane sketch
map
D-SVM Fusion Clustering Algorithm 251
The penalty function can be described as the distance between the hyperplane model
and the measured point in order to avoid over fitting the correction factor to the distance.
5 Conclusion
References
1. Deng, Z., Yu, Y., Yuan, X., et al.: Indoor positioning status and development trend study.
China Commun. 10, 50–63 (2013)
2. Zheng, R., Deng, Z., et al.: Analysis of positioning error in multipath environment. In: China
Satellite Navigation Academic Annual Meeting (2015)
3. Ning, J.J.: An adaptive SA-DBSCAN: of density clustering algorithm. Univ. Acad. Sci. 26(4),
530–538 (2009)
4. Zhou, J., Li, W., Jin, L., et al.: Design of indoor positioning system based on KNN-SVM
algorithm. J. Huazhong Univ. Sci. Technol. Nat. Sci. Ed. S1, 517–520 (2015)
5. Zhang, B., He, Z.: New method of SVM multi classification based on support vector data
description. Comput. Appl. Res. 24, 46–48 (2007)
6. Li, S.: Application research on indoor location technology based on WiFi. Nanjing Normal
University (2014)
7. Zhou, F.: Design and implementation of positioning system based on WiFi technology. Beijing
University of Post and Telecommunication, p. 12 (2009)
Research on Evaluation of Sensor Deviations During Flight
1
System Engineering Research Institute of China State, Shipbuilding Corporation, Beijing, China
2
College of Computer Science and Technology, Jilin University, Changchun, China
13756651430@163.com
3
Chinese Journal of Electronics, Beijing 100036, China
Abstract. Aircrafts have been widely used in recent years with the rapid devel‐
opments in economies. In consideration of aircraft security and the importance
of navigation information, numerous sensors have been utilized for aircraft
tracking. Deviations are easily produced by sensors and have significant influence
on information accuracy. The existing literature on methods to monitor sensors
and evaluate their deviations is limited. In this paper, the information detected by
sensors is compared to the reference trajectory. Subsequently, a novel method to
calculate deviations and errors is proposed. To validate the reliability of the eval‐
uation results, the Chi Square test is integrated into the research method. A stat‐
istical method is also proposed to identify the existence of deviations between
two sensors.
1 Introduction
Aircrafts have been widely used in economies, and flight safety has gained increased
importance and attention. Many types of sensors have been deployed to acquire and
monitor location information in different domains, such as the military, medical care,
and environmental management. Multiple sensors are generally used to identify
geographical coordinates precisely. However, various environmental factors can cause
sensor deviation, which results in inaccurate data collection. These factors should be
considered when deploying sensor probes. Thus, the calculation and the analysis of
sensor deviations are crucial issues.
In most recent studies, sensors are considered as suitable tools to investigate aircraft
information. In [1], the airspeed sensors were used to reconstruct the approach-path
deviations through triangulation and the flow sensors controlled the touchdown phase.
The authors in [2] demonstrated a trajectory recommendation framework that relied on
various sensors. The authors of [3] reported that trajectories were difficult to predict the
precise location of thrown objects. A sensor system, which received data from distance
sensors, was responsible for detecting the deviations between the actual and predicted
values. Then an algorithm to measure the in-between points during flights was proposed.
The authors of [4] used inertial sensors to gather optical flow information and presented
a set of novel navigation techniques to estimate the ground velocity and altitude of
aircrafts. Two formulations were proposed to calculate aircraft velocity and altitude. A
novel method that used a ToF [5] camera to detect an aerial target was proposed in [6].
On the basis of a special requirement, the sky was considered as noise in the image and
the target could be recognized from a model of complete randomness, thus illustrating
the robustness of sensor detection with respect to the number of false alarms. In addition,
the authors conducted a study to analyze the detection performance. In [7], the accuracy
of sensors was noted. The authors proposed a linear covariance technique to predict the
accuracy of altitude determination systems for flights. The authors also carried out a
study to analyse the performance of the detection. In [8], a radar sensor was placed onto
a UAV, which calculated the position deviations of aircrafts from a theoretical trajectory.
A SAR processor and navigation system determined the difference between theoretical
and actual values.
To the best of our knowledge, the literature on the evaluation of sensor deviations
is limited. In the present research, we build the reference trajectory and propose a novel
method to calculate and evaluate the deviations for each sensor. The feasibility of the
obtained data is evaluated through residual analysis using the Chi Square test. We also
propose a statistical method to determine whether the deviations exist between two
sensors or not.
The rest of this paper is organized as follows. Section 1 discusses the potential
application of sensors in the aviation industry and the importance of evaluation methods.
Section 2 presents the proposed calculation and evaluation methods. Section 3 presents
the statistical method for detecting the deviations for each sensor. Finally, the study is
concluded in Sect. 4.
To obtain the deviation, we should build a reference trajectory first. The reference data
need to be transformed according to the local sensor framework. If the total number of
points tracked by two sensors is larger than 50, then high-accuracy sensors are used to
plan the trajectory. The reference time is calculated as follows:
( )
Min(time1 ) + Max(time2 )
Timeref = (1)
2
The reference trajectory of a sensor is built through the least square fitting of a poly‐
nomial. We sum up of the values in the set deviation. The sensor deviation is equal to
the average of values. In order to calculate the deviation between two sensors, the devi‐
ation for each sensor corresponding to the reference trajectory should be obtained. The
tracks of sensor Sx are recorded in the set 𝜉Sx. The deviation of sensor Sx can be defined
as follows:
Σsx
bSx |rt = 𝜉s− = (2)
x Nsx
254 Y. Xiong et al.
where bsx |rt and 𝜉̄sx are the deviation between Sx and the reference trajectory. The sum of
values in the set 𝜉Sx is expressed as Σsx, and variable NSx is the number of the points
tracked by sensor Sx. The deviation between sensor s1 and s2 is defined as follows:
The variable 𝜎x̄ denotes the average value of the standard deviations:
𝜎
𝜎x̄ = √ x (5)
Nx
The standard deviation for each sensor could be obtained easily. After acquiring the
standard deviations for sensors s1 and s2, the errors induced by the deviations between
these two sensors can be calculated as follows:
√
𝜎evaluate = 𝜎s2 + 𝜎s2 (6)
1 2
Subsequently, the methods to achieve the research objective are tested because the
sensor errors are obviously normally distributed. In this paper, residual analysis is first
utilized to identify the feasibility of the data. Then, the residual error is tested by Chi
Square test. The residual error must first satisfy the test above to obtain accurate data
and information. Otherwise, the deviation is discarded. Sensors have different preci‐
sions. Thus, the difference for each sensor should be normalized. The normalized differ‐
ence of sensor X is defined as follows:
𝜉x∗
𝜉x∗ = (7)
𝜎x′
where 𝜎x′ is the sample-standard-deviation. Normalization is applied on the two sets, and
the results are sorted and stored in the set 𝜎 ∗∗.
The normalized differences and their order are analyzed. The possible samples are
divided into discrete areas, which are used to calculate the expectation of samples. The
Chi Square test is conducted to identify the consistency of samples and expectation
distributions, that is:
( )2
∑
k
ne − ns
𝜒e2 = (8)
i=1
ne
Research on Evaluation of Sensor Deviations During Flight 255
where k is the number of discrete areas and ne is the distribution of the expectation. The
distribution of samples is expressed by ns. Subsequently, Xe2 is compared to the hypo‐
thetical threshold. If the threshold is not exceeded, the distributions of the samples are
considered consistent with the expected distributions. Otherwise, the residual error could
not be used further.
3 Surveillance of Deviations
Apart from evaluating the deviations, the results from calculations also can also help in
monitoring the state of sensors. In this study, a statistical method is employed to identify
whether the deviations exist between the two sensors. Two aspects are considered in
detecting deviations: the detection for a single sample and the detection for accumulated
samples. Both are described in the next sections.
The detection for a single sample transforms the reference trajectory and deviation
calculation data, such as distance, orientation, and elevation, into the same coordinate
systems. Unlike the deviation estimates, the result of deviation calculation constructs a
test statistic that is used to identify the consistency of the deviation for the reference
track in a statistically significant way. If the statistic is greater than the predefined
threshold value, the probability of significant deviation between the two sensors is
greater. Otherwise, the obvious deviation is not considered. Three steps are required to
detect a single sample: calculation of deviation, calculation of test statistics, and calcu‐
lation of examination. Each step is described in the following sub-sections.
where 𝜉px,s (i) indicates the difference of sensor s at point i . ps is the parameter (e.g.,
distance, orientation, and elevation). The reference value of parameter p is expressed
by p̄ (i), which can be formulated as follows:
∑
j=n
( )j
p̄ (i) = cj tj − tref (10)
j=0
where tj is the measured time for point j and cj is the polynomial coefficient calculated
by reference trajectory. The duration for constructing reference trajectory is expressed
by tref .
256 Y. Xiong et al.
The detected data, such as starts, stops, average difference are recorded. Then, the results
of the detection for each single sample are counted. For the two sensors, the detection
for accumulated samples can be defined as follows:
√( )
Σp,1 − Σp,2 N1 + N2 − 2 N1 N2
tc = √ ( ) 2 ( ) 2 (12)
N1 + N2
N1 − 1 𝜎p,1 + N2 − 1 𝜎p,2
where tc is the detection result of the accumulated samples. The standard deviation is
express by 𝜎p,s. The variable Ns is the number of records for parameter p. The total
number of data, which determines the existence of deviation, is stored in variable k1.
Similarly, the total number of data without deviation is k2. If k1 ⩽ 2 × k2, the accumulated
result is considered ambiguous and the accumulated samples will be detected in further
tests.
The multiple data calculation method is used to monitor the deviation between two
sensors. We combine the results of tests, and then achieve the cumulative tests statistic.
The multiple data calculation method is defined as follows:
Nm = n1 + n2 (13)
w1 a1 + w2 a2
Am = (14)
w1 + w2
√
𝜎m = w1 + w2 (15)
1
wx = (16)
𝜎x2
Research on Evaluation of Sensor Deviations During Flight 257
where ax is the average value of sensor x, and the variance is expressed by 𝜎x. nx are the
points of sensor x.
4 Conclusion
This paper discussed the importance of the track points during flight and presented
several practical methods to analyze the accuracy of the data obtained by sensors. We
first constructed the reference trajectory, and then transformed the tracked data into the
same coordinates system. The deviation between two sensors was evaluated, and then
the Chi Square method was used to investigate the correctness of the deviation evalua‐
tion. In addition, we proposed a statistical method to identify whether the deviation exists
between sensors or not. Future work may consider more realistic situations, more algo‐
rithms to calculate deviation, and more evaluation methods to determine the accuracy
of deviations.
References
1. Trittler, M., Rothermel, T., Fichter, W.: Autopilot for landing small fixed-wing unmanned
aerial vehicles with optical sensors. J. Guidance Control Dyn. 39(9), 2011–2021 (2016)
2. Yeung, S., Madria, S.K., Linderman, M.: A trajectory recommendation system via optimizing
sensors utilization in airborne systems (demo paper). In: Claramunt, C., Schneider, M., Wong,
R.C.-W., Xiong, L., Loh, W.-K., Shahabi, C., Li, K.-J. (eds.) SSTD 2015. LNCS, vol. 9239,
pp. 508–513. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-22363-6_31
3. Frank, T., Janoske, U., Schroedter, C.: Detection of position and orientation of flying cylinder
shaped objects by distance sensors. In: 2011 IEEE International Conference on Mechatronics
(2011)
4. Rhudy, M.B., Gu, Y.: Unmanned aerial vehicle navigation using wide-field optical flow and
inertial sensors. J. Robot. 2015 (2015)
5. Hussmann, S., Huhn, P., Hermanski, A.: Systematic distance deviation error compensation for
a ToF-camera in the close-up range. In: I2MTC - International Instrumentation and
Measurement Technology Conference (2012)
6. Rodríguez-Jiménez, S., Burrus, N., Abderrahim, M.: A-contrario detection of aerial target
using a Time-of-Flight camera. In: Sensor Signal Processing for Defence, SSPD 2012 (2012)
7. Blomqvist, J., Fullmerb, R.: The influence of uncertainties of attitude sensors on attitude
determination accuracy by linear covariance analysis. The International Society for Optical
Engineering (2010)
8. Labowski, M., Kaniewski, P., Konatowski, S.: Estimation of flight path deviations for sar radar
installed on UAV. Metrol. Meas. Syst. 23(3), 383–391 (2016)
A Distributed Routing Algorithm for LEO Satellite
Networks
1
Beijing Key Laboratory of Space-Ground Interconnection and Convergence, Beijing University
of Posts and Telecommunications (BUPT), Xitucheng Road No. 10, Beijing 100876, China
{dingjundong,yongzhang,liruonan}@bupt.edu.cn
2
China Electronic Appliance Corporation, Beijing, China
zhaol@ceac.com.cn
Abstract. Satellite networks support a wide range of service and provide global
coverage. Moreover, LEO satellites, which have global coverage potential, have
caused great attention in the field of satellite communication. Due to satellites’
mobility, an efficient and stable routing algorithm is necessary for transmission
in satellite networks. In this paper, a Distributed Multi-Path Routing algorithm
(DMPR) using Earth-fixed satellite systems is proposed. DMPR adopts the Polar
orbit constellation in which each satellite has four ISLs (inter satellite link) with
the neighbor satellites. A traffic-light is set for every ISL to judge the blocking
while avoiding some busy path and reducing the delay.
1 Introduction
Satellite networks are becoming important in the field of communication, especially the
LEO satellites. Compared to the terrestrial networks, the LEO satellites have global
coverage potential. Besides, they can also offer services with lower latency than the
geostationary satellites. Therefore, the LEO satellites communication has aroused great
interest in academia and industry.
The satellites’ topologies are dynamic with frequent link switching and interruption.
Since the satellite constellation is determined in advance, the topologies will be predict‐
able. These characteristics, different from terrestrial networks, bring many difficulties
in routing design. Many schemes have been proposed from researchers since the 1990s
[1]. Topological control strategy can generally be adopted to shield the topology’s
dynamic. There are two topology control strategies: Virtual topology [2, 3] and virtual
node [4, 5]. Virtual node strategy in which use the concept of satellite logical location
form a global virtual networks. Each node in the network is virtual node, which is serv‐
iced by the nearest satellite.
Werner [6] introduces the DT-DVTR routing algorithm (Discrete-Time Dynamic
Virtual Topology Routing) in the LEO satellites networks. They divided the routing
strategy into two processes: firstly, they set a virtual topology for all successive time
intervals of the system period, providing instantaneous sets of alternative paths between
all source destination node pairs; secondly path sequences over a series of time intervals
are chosen from that according to certain optimization procedures. Korçak [7] proposed
a multistate virtual network (MSVN) topology, provided formal mathematical model
for it and discussed its contribution to the overall system availability. Then Lu et al. [8]
formalized and optimized the virtual topology based on virtual node strategy while
elaborating its important features.
There are many distributed routing algorithms [9, 10] based on virtual nodes. A
datagram routing algorithm is proposed for LEO satellite networks where routing deci‐
sions are made on per-packet basis. Henderson and Kartz [11] has proposed a distributed
routing algorithm in which they flood a link state packet only as the routing radius for
a given satellite and then they choose the best path using Dijkstra algorithm.
This paper is organized as follows: Sect. 2 introduces the satellite network architec‐
ture that has the Earth-fixed footprints. Section 3 proposes the new routing strategy.
Section 4 presents the simulation result of the new routing algorithm and in the end
Sect. 5 summarizes this paper.
The satellite network adopts the Walker star [12] constellation which is composed of N
polar orbits (planes), each with M satellites at low distances from the Earth. The planes
are separated from each other with the same angular distance of 360°/(2 * N). They cross
each other only over the North and South poles. The satellites of each plane are separated
from each other with an angular distance of 360°/M. Due to the circular planes, the
satellites in the same plane have the same radius all the times. In this paper N is 6 and
M is 12.
Each satellite has four neighboring satellites: two in the same orbit and two in the
left and right planes. So each satellite has four transmission directions which means that
when deciding the next hop node each satellite has four choices.
In our analysis of virtual topology dynamics, we consider a satellite orbit with NSAT
satellites that have Earth-fixed footprints, and serve for total of NFP footprint areas along
the orbit. The footprint means the ground area is serviced by satellites. Firstly, average
number of satellites per footprint area (Nfsat ) is calculated as
In traditional VN (virtual node) concept, NSAT is equal to NFP, i.e., there exists a single
satellite for each footprint, namely Nfsat = 1. However, for the sake of increasing system
availability, this paper also considers the case where there is more than one satellite per
footprint. Target on improving the transmission success rate while simplifying the
handover process, we set Nfsat to 2 as shown in Fig. 1.
260 J. Ding et al.
3 Routing Strategy
to stay wrong status for Δt (checking interval). To ensure that BOQRn will not exceed
T during this checking interval, T should meet:·
Where L represents the size of buffer queue, APS represents average packets size,
and (I-O) max denotes the maximum difference between input and output traffic rates.
When the satellite moves to high latitudes, the distance between them is closer than
the low latitudes as shown in Fig. 2(b). The distance between satellites in orbit L2 can
be described by: d = 𝜋 × (R × sin𝛼 + h) ÷ N
(R means the Earth radius, h means the satellites height and N means the number of
orbits).
According to Shannon formula, the channel capacity can be described by:
S P ⋅h
Rd = B ⋅ log(1 + ) = B ⋅ log(1 + D 2 ss ) (4)
N 𝜎
hss means channel gain and hss ~ E(d−α). It can be seen satellites’ channel capacity
in high latitudes is larger than low latitudes. So the buffer size of inter-ISL can be set
higher when satellites moves to high latitudes. This means when satellites serving for
different footprint their inter-ISL buffer size change dynamically.
The routing algorithm begins with sending data from the terrestrial device to the
satellite. Each terrestrial device can send date to 2 different satellites, because every
footprint is served by two satellites. When the current satellite determines the next hop
node, it will compare the footprint’s coordinate to decide the direction.
<nd, md> is the destination footprint coordinate and <ni, mi> as the current satel‐
lite’s footprint coordinate. The current satellite will choose at most two directions to
transfer data. The vertical and horizontal movements are described by
262 J. Ding et al.
During a date receiving processing, firstly current satellite will check whether the
same data has been received. After checking, satellites can avoid transferring the same
data twice, so it can save source and improve the system’s performance. If not, it will
compare its own footprint coordinate with the destination coordinate to choose the
direction. For example, if mi > md, dv = +1 means the current satellite chooses upward
as the backup direction. If ni < nd, dh = +1 means the current satellite chooses right as
the backup direction (Table 1).
After choosing two backup directions, the satellite will enhance direction by taking
congestion into account. If the backup ISL’s traffic light is “GREEN”, the satellite will
transfer data through this ISL. If the backup ISL’s traffic light is “RED”, the satellite
will discard this ISL.
In this section, we mainly analyze the performance of satellite transmission delay and
success rate. In satellite communications, congestion and satellite failure are important
factors affecting data transmission. In this experiment, the average one-hop ISL propa‐
gation delay is set to 14 ms. When the ISL is set to “GREEN”, the maximum queuing
delay is set to 10 ms. When the ISL is set to “RED”, maximum queuing delay is set to
30 ms. In simulation, inter-orbit link’s buffer size is set dynamically. Satellites’ buffer
size (BQ(n)) is 20% smaller than higher latitudes, so does the threshold T.
Figure 3(a) shows the two routing algorithm’s latency increases with the proportion
of blocking ISL. DMPR’s latency is shorter than DT-DVTR’s obviously. When the link
is blocking, the satellite will wait which will increase queuing delay using DT-DVTR
routing. So the impact of blocking on the delay is small while DT-DVTR’s delay
increases rapidly with the increase in proportion.
In DMPR routing algorithm each satellite node has at most two choices when
selecting next node. Even if a node fails, it can also find an alternative node for trans‐
mission which can mitigate the adverse effects caused by node failures. So in the
Fig. 3(b), it shows the success rate of DMPR is higher than DT-DVTR’s especially when
the number of disabled satellite exceeds 5.
5 Conclusion
In this paper, we have proposed a new distributed routing algorithm for LEO satellite
networks. When choosing the next hop satellite node, the satellite will take congestion
into account. So our routing strategy has an advantage in reducing latency when encoun‐
tering congestion. Meanwhile, this routing algorithm can guarantee the success rate
when the node fails.
Acknowledgements. The author would like to thank to the reviewers for their detailed
suggestions. This work is supported by the National Natural Science Foundation of China under
Grant No. 61171097.
References
1. Taleb, T., Jamalipour, A., Kato, N., Nemoto, Y.: IP traffic load distribution in NGEO
broadband satellite networks – (invited paper). In: Yolum, Güngör, T., Gürgen, F., Özturan,
C. (eds.) ISCIS 2005. LNCS, vol. 3733, pp. 113–123. Springer, Heidelberg (2005). https://
doi.org/10.1007/11569596_14
2. Gounder, V.V., Prakash, R., Abu-Amara, H.: Routing in LEO-based satellite networks (1999)
3. Fischer, D., Basin, D., Engel, T.: Topology dynamics and routing for predictable mobile
networks. In: IEEE International Conference on Network Protocols, pp. 207–217. IEEE
Xplore (2008)
4. Mauger, R., Rosenberg, C.: QoS guarantees for multimedia services on a TDMA-based
satellite network. Commun. Mag. IEEE 35(7), 56–65 (1997)
5. Ekici, E., Akyildiz, I.F., Bender, M.D.: A distributed routing algorithm for datagram traffic
in LEO satellite networks. IEEE/ACM Trans. Netw. 9(2), 137–147 (2001)
6. Werner, M.: A dynamic routing concept for ATM-based satellite personal communication
networks. IEEE J. Sel. Areas Commun. 15(8), 1636–1648 (1997)
7. Korçak, Ö., Alagöz, F.: Virtual topology dynamics and handover mechanisms in Earth-fixed
LEO satellite systems. Comput. Netw. 53(9), 1497–1511 (2009). The International Journal
of Computer & Telecommunications Networking
264 J. Ding et al.
8. Lu, Y., Sun, F., Zhao, Y.: Virtual topology for LEO satellite networks based on earth-fixed
footprint mode. IEEE Commun. Lett. 17(2), 357–360 (2013)
9. Wang, K., Yi, K., et al.: Packet routing algorithm for polar orbit LEO satellite constellation
network. Sci. Chin. Inf. Sci. 49(1), 103–127 (2006)
10. Papapetrou, E., Pavlidou, F.N.: Distributed load-aware routing in LEO satellite networks. In:
2008 Global Telecommunications Conference, IEEE GLOBECOM, pp. 1–5. IEEE Xplore
(2008)
11. Henderson, T.R., Katz, R.H.: On distributed, geographic-based packet routing for LEO
satellite networks. In: 2000 Global Telecommunications Conference, IEEE GLOBECOM
2000, vol. 2, pp. 1119–1123. IEEE Xplore (2000)
12. Wang, J., Li, L., Zhou, M.: Topological dynamics characterization for LEO satellite networks.
Comput. Netw. 51(1), 43–53 (2007)
Optimization of Smart Home System Based
on Wireless Sensor Network
Abstract. In this paper, smart home is the research object. It analysed the intel‐
ligent acquisition for monitoring data, put forward a kind of solution for remote
monitoring system based on wireless sensor network, and designed a set of perfect
intelligent remote monitoring system by using the mobile terminal platform. This
system can achieve a series of basic functions, including monitoring, automatic
adjustment, alarm and control for the home appliances, which can provide
comprehensive and reliable home environment information for the home users
and meanwhile achieve remote control for home appliances. Finally, the smart
home monitoring system was deployed and tested, the test results show that the
system can achieve the basic functions of the modern smart home and has the
unique advantages in performance, cost, versatility, scalability and other
aspects.
1 Introduction
With the rapid development of wireless sensor network, smart phone communication
and Internet technology, the remote wireless mobile home monitoring system will
become one of the mainstream smart home network [1]. The smart home system needs
a lot of information acquisition and transmission. Introducing the wireless sensor
network technology into the construction of smart home can build a “home appliance
network” with adaptive control function for smart home [2]. This network can not only
make the smart appliances conduct mutual coordination, but also be interacted with the
external network. Therefore, users can achieve the remote control of the smart appliances
through this network.
At present, the domestic and foreign major operators are integrating resources and
carrying out innovative business to occupy the smart home market [3–5]. Although the
smart home industry is developing relatively quickly, the present level of development
is uneven. It is faced with many problems like the lack of unified industry standards,
expensive price, complex operation and the leakage of information which restrict the
development of the smart home industry [6, 7]. In this research, the field research was
conducted to find out the deficiencies of the current smart home system. And the envi‐
ronmental parameters required for the monitoring system were determined through the
2 Overall Design
The overall design is based on the research about the wireless sensor network tech‐
nology, wireless data communication technology, HTML5, WebSocket and other
related theoretical basis [8]. Combined with corporate field research and actual func‐
tional requirements analysis, the overall design of the smart home monitoring system
based on the wireless sensor network was determined generally, as shown in Fig. 1.
The system collects the required information of the family environment by the related
sensors and controls the switch of the household electrical appliances by the controller
modules. In this system, the sensor modules and the controller modules are connected
to the ZigBee terminal equipment. The data transmission is based on the ZigBee wireless
network and the ZigBee center coordinator. The ZigBee center coordinator is connected
to the server through the serial ports. It can save the data collected from the sensors into
the server database or sent the data to the client directly. According to the implementation
process of the system, the overall scheme includes the front-end sensors and the control‐
lers module design, the ZigBee wireless sensor network design, the data transmission
module design, the server program design, the data interaction module design, and the
client application software design.
Optimization of Smart Home System 267
The ZigBee wireless transmission module of this system used TI wireless RF chip called
CC2530F256 as the core chip. CC2530F256 is compatible with the ZigBee protocol
stack, providing a powerful and complete ZigBee solution. Because the transmission
distance is relatively short and the family housing area is limited, the system just needs
relatively small sensor nodes and a simple data transmission structure, which does not
need the router nodes to extend the network coverage area [9]. Meanwhile the position
and the number of the household appliances are easy to change, which will change the
network state. Therefore, under the premise of the communication requirements, consid‐
ering the cost of the equipment and the simple structure, this system used the star
topology to build the smart home wireless network. The star topology wireless network
consists of a ZigBee full-function central coordinator and several terminal functional
devices. And the data transmission is based on the principle of ZigBee wireless data
transmission.
In this paper, the DS18B20 temperature sensor module was selected as an example
to introduce the concrete design process of the data acquisition and transmission soft‐
ware. First, the system obtained the collected data from the terminal device. Then the
data was transmitted to the central coordinator. Finally, the coordinator sent the data to
the sever module through the serial ports. In order to observe the results of data acquis‐
ition expediently, the coordinator used LCD to display the collected data values. Mean‐
while the data in the terminal nodes and the coordinator node was sent to the computer
through the serial ports and shown by serial debugging assistant. In order to enhance the
system layout and installation convenience, meanwhile reducing energy consumption,
the data acquisition nodes all used the battery power [5, 10]. The node modules will go
into sleep mode in leisure time and be wake up when the system needs data acquisition
or transmission. The entire workflow is shown in Fig. 2.
According to the demand analysis of the system, the smart home monitoring system can
be divided into five main functional modules, including real-time monitoring module,
security alarm module, scene mode, basic information management and user manage‐
ment. These five functional modules can not only achieve the real-time monitoring data
display, alarm for abnormality, linkage control and historical data display, but also
achieve the statistical processing, analysis and diagnosis of the massive monitoring data
to search potential effective information intelligently. In the meantime, this system can
provide the mode selection, information management and scalability. The overall func‐
tional structure of the system is shown in Fig. 3.
Optimization of Smart Home System 269
The server used multithreading serial communication technology to obtain the induction
layer data. In the data communication process, a variety of serial operations were placed
in different threads, and every thread was only responsible for the corresponding serial
operation. The main thread was responsible for the coordination and management of the
auxiliary threads, which can use the CPU in fast switching between various threads.
Therefore, the system can achieve the concurrent execution and multi-task mechanism
to reduce the occupancy rate of the CPU. This mechanism can improve the utilization
of the serial ports and efficiency of data transmission, which can greatly improve the
data request and processing ability of the server.
In order to solve the common problems of the database like operating frequently and
low query efficiency, the system used the cache mechanism to save the collected data.
The cache mechanism saved the data in the cache memory. Returning data from the
cache memory is quicker than querying database, so it can greatly improve the perform‐
ance of the server application. In the process of operation, the database will start data
cache dependency. When the server requests the latest data collected by the sensor, it
will write the data to the cache and update the cache. And setting the key index
“CacheKey” when the system writes cache can improve query execution efficiency of
the data in the cache. The concrete process of data storage adopting data cache mecha‐
nism is shown in Fig. 4.
270 X. Guo and N. Hu
For the basic operations which don’t need high operation frequency and real-time
requirements, such as user login, obtaining the equipment information, querying the
historical data and other operations, the client used Ajax asynchronous loading tech‐
nology to invoke the Web Service from the server and obtain the basic information. In
order to make the system client and server can transmit data cross platform and cross
language, this system used lightweight data exchange format “Json” to achieve the data
exchange between them.
As for viewing and operating the data in real-time monitoring system, it needs client
and server keep persistent connection and constantly obtain the real-time updated data
from the sensor layer, which has high requirements for real-time communication, so this
system designed the WebSocket full duplex bidirectional communication technology to
achieve real-time monitoring. First, this system enabled a thread in the server to open
the WebSocket network service and used the delegation method to open or close the
WebSocket connection. Then it obtained the IP address of the machine and established
the monitoring ports. Finally, the system regarded the two values as the parameter values
of the WebSocket server address to transmit data. For the realization of the system client,
the system used the HTML user interface. After the WebSocket connection between the
client and the server, the two sides will be able to achieve continuous two-way commu‐
nication with a high utilization ratio.
transmission frequency was set as 3 s/times, and a total of 500 data was sent to the
coordinator. The average value of the test results was shown in Table 1.
Table 1. Communication distance and data packet loss rate of the ZigBee module
Test Communication Number of Number of Number of Average
environment distance(m) terminal data coordinator test packet loss
transmission data rate(%)
reception
Indoor 15 500 500 10 0
20 500 500 10 0
25 500 500 10 0
30 500 498 10 0.4
Outdoor 30 500 500 10 0
50 500 496 10 0.8
70 500 473 10 5.4
100 500 447 10 10.6
As we can see from the test results, by using the system based on the ZigBee wireless
sensor network, when the communication didn’t meet interference in the outdoor envi‐
ronment, the packets loss rate was very low within a 50 m range. When the node distance
was more than 70 m, the packet loss rate will increase a little. While in indoor environ‐
ment, when the transmission distance was less than 25 m, the packet loss rate was about
0%. The results show that the data transmission of the designed ZigBee network has
high reliability, and it can meet the requirements of data transmission in the general
family environment.
In order to examine the communication efficiency and real-time of the WebSocket tech‐
nology which was used in the smart home monitoring system, the Ajax polling based
on HTTP protocol and the real-time communication framework based on the WebSocket
were set up. In this test, the data request and response information of two kinds of scheme
in the process of communication were analyzed and compared through the Chrome
browser network tools. Then the test case was designed by using the Jmeter pressure
test tools, which simulated the communication between the client and the server to
compare the network throughput and delay of two methods.
throughput than the traditional Ajax polling scheme, which has great performance
advantages.
technology in the aspects of message delay time and network throughput. WebSocket
technology can decrease the network throughput greatly, and take advantage of the stable
long connection mode to reduce the communication delay time, which has a high effi‐
ciency in the real-time monitoring system of the smart home.
6 Conclusion
This paper introduced the wireless sensors into the system, which greatly simplified the
structure of the system and improve flexibility. It used the serial debugging assistant to
test the communication distance and the data packet loss rate of the wireless network
set up by ZigBee equipment, and completed the ZigBee communication module test and
WebSocket communication test in the smart home system. It verified the stability and
safety of the ZigBee wireless network. Then, the real-time communication framework
of Ajax polling and WebSocket were constructed respectively, and the network
throughput and the network delay of the two schemes were tested through Chrome
browser and Jmeter pressure test tool, which proved the advantages of WebSocket tech‐
nology in system real-time monitoring. Finally, the on-line overall test for the entire
smart home system was conducted through the different mobile terminals. The system
worked well and met the design requirements, which achieved the basic expected func‐
tions of the smart home system.
Acknowledgment. This paper was supported by the project in the Hubei Science and Technology
Pillar Program (No. 2015BKA222).
References
1. Lai, W.: Intelligent home system design. Comput. Knowl. Technol. 08, 23–25 (2015)
2. Xie, Y., Pei, L., Wu, T.: Internet of Things in the application of intelligent home. Instrum.
Technol. 03, 46–49 (2015)
3. Zhao, Y.: Intelligent home system in the Internet of Things. Internet Things 05, 32–35 (2013)
4. Alam, M., Reaz, M., Ali, M.: A review of smart homes-past, present, and future. IEEE Trans.
Syst. Man Cybern. 42(6), 12–16 (2012)
5. Zhu, M., Li, N.: Development status and future analysis of intelligent home. Telev. Technol.
04, 82–85, +96 (2015)
6. Bitterman, N., Shach-Pinsly, D.: Smart home - a challenge for architects and designers.
Architect. Sci. Rev. 58, 266–274 (2015)
7. Wu, L., Wei, Y., Yajie, M.: Development status analysis of domestic and international
intelligent home market. Mod. Telecommun. Technol. 12, 71–74 (2014)
8. Zou, S.: Study on the key technology of intelligent home system based on wireless sensor
network. South China University of Technology (2013)
9. Nagpal, P.: Key actors in the mobile phone industry: the smart phone wars. Acad. Inf. Manag.
Sci. J. 17(1), 87 (2014)
10. Liu, L., Hu, B., Li, L.: Energy conservation algorithms for maintaining coverage and
connectivity in wireless sensor networks. IET Commun. 4(7), 786–800 (2010)
The Energy Efficiency Research of Traffic Offloading
Mechanism for Green Heterogeneous Networks
1 Introduction
supplementary or alternative power sources, have received great attentions from both
academia and industry [6]. How to fully utilize the harvested energy to maximize network
energy efficiency while satisfying the quality of service (QoS) requirements is a critical issue.
Traffic offloading problem is one of the hot issues in heterogeneous network. [7, 8]
analyzes the relationship between the traffic offloading and customer service quality, and
puts forward the traffic offloading incentive mechanism based on reverse auction model,
can guarantee the quality of customer service, increase the amount of data offloading of
the system. Literature [9] puts forward offloading users which have high requirements
of service quality to femto to reduce the energy consumption of the macro, and analyzes
the relationship between the system energy efficiency and the distance between the
neighborhood. Literature [10] analyzes the system power consumption influence of
different traffic density by different amount of micro. But [9, 10] only consider the static
traffic loading scenario, without considering the dynamic changes of the business. [11]
considers the offloading when in the wifi enabled small cell case.
Although traffic offloading technology has been extensively investigated in on-grid
cellular networks, the conventional offloading methods cannot be applied when EH is
applied. Instead, traffic offloading technology of green heterogeneous cellular networks
needs to be devised, the operation of each cell is optimized individually based on their
renewable energy supply. Different from existing works, we focus on the design of traffic
offloading for green HCNs with multiple SBSs powered by diverse energy sources. We
aim to maximize the network energy efficiency while satisfying the QoS requirement of
every user. To this end, users are dynamically offloaded from the MBS to the SBSs,
based on the statistical information of traffic intensity and energy arrival rate, based on
which the optimal traffic offloading amount are obtained.
The reminder of the paper is organized as follows. System model is introduced in
Sect. 2. Section 3 analyzes the power demand and supply for SBSs with EH devices.
Then, the energy efficiency maximization problem for the single-SBS case is studied in
Sect. 4. Simulation results are presented in Sect. 5, followed by the conclusion in Sect. 6.
2 System Model
In this section, the details of the HCN with different energy supply are presented as
follows.
B = {BH , BR } be the set of all SBSs. Denote by Wa the number of wifi access points,
denote by Bw = {W1 , W2 , … , Wa } the set of wifi access points. SBSn serves a circular
area with radius Ds,n, and the small cells are assumed to have no overlaps with each
other. The MBS is always active to guarantee the basic coverage, whereas SBSs and
wifi access points can be dynamically activated for traffic offloading or deactivated for
energy saving, depending on the traffic and energy status.
where PTs is the transmit power level of the SBS, wu is the bandwidth allocated to user
u, dnu is the distance between user u and SBSn, 𝛼s is the path loss exponent of the SBS-
tier, hnu is an exponential random variable with unit mean reflecting the effect of Rayleigh
fading, 𝜃s is the ratio of inter-cell interference to noise among SBSs, and 𝜎 2 is the noise
power density. As each SBS allocates bandwidth equally to its associated users, the
achievable data rate of a generic user u is as follows:
where wss,nu denotes the bandwidth allocated to each user from SBSn.
Similarly, if user u is served by the MBS as a MMU or MSU, its received SINR is
given by
−𝛼
PTm wu dmum hmu
𝛾m,u = (3)
Wm (𝜃m + 1)𝜎 2 wu
where PTm is the transmit power level of the MBS, dmu is the distance between user u
and the MBS, 𝛼m is the path loss exponent of the MBS-tier, 𝜃m is the interference to noise
ratio from other MBSs, and hmu reflects Rayleigh fading with the same probability
distribution as hnu.
Then, the achievable data rate of user u is given by
where wmm,u and wms,nu denote the bandwidth allocated to each user from MBS.
If user u is served by wifi access points, its received SINR is given by
−𝛼
Pw wu dwu w hwu
𝛾w,nu = (6)
Ww (𝜃w + 1)𝜎 2 wu
As each wifi access point allocates bandwidth equally to its associated users, the
achievable data rate of a generic user u is as follows:
BSs can work in either active mode or sleep mode, with different power consumption
parameters. According to [13], the power consumption of a BS in active mode can be
modeled as:
w
PBS = PC + 𝛽P (8)
W T
where PC denotes static power consumption including the baseband processor, the
cooling system and etc., suppose the static power consumption of wifi access points is
0, coefficient 𝛽 is the inverse of power amplifier efficiency factor, W is the available
system bandwidth and w is the bandwidth of utilized subcarriers. PT is the transmit power
level and is treated as a system parameter, while we control the RF power by adjusting
the utilized bandwidth w.
1 wss,n (t)
UE,n (t) = (PCs,n + 𝛽 P ) (9)
E Ws s,n Ts,n
where PTs,n, PCs,n and 𝛽s,n are the transmit power, constant power and power amplifier
coefficient of SBSn, respectively. UE is called energy consumption rate in the rest of this
paper for simplicity.
For a SBS with EH, the variation of battery can be modeled as a M/D/1 queue with
arrival rate 𝜆E and service rate UE given by Eq. (9). In what follows, we analyze the stable
status of the energy queue. For the M/D/1 queue, the embedded Markov chain method
is usually applied to analyze the stable status [16]. When 𝜆E ∕UE ≥ 1, the queue is not
stable and the queue length goes to infinity, which means that the harvested energy is
always sufficient. When 𝜆E ∕UE < 1, the stationary probability distribution of L+ can be
derived by Pollaczek-Khinchin formula,
𝜆E
q0 = 1 − (10)
UE
𝜆E
𝜆E (11)
q1 = (1 − )(e UE − 1)
UE
𝜆E 𝜆E 𝜆 + 𝜆
𝜆E + (k E )L −k
+
(k E )b−L −1
L+ −1 k
𝜆E U
L
∑ U U
−k
UE UE (12)
qL + = (1 − ){e E + e E (−1) E [ + + ]} (L+ > 2)
UE k=1 (L − k)! (b − L+ − 1)!
Thus, the stationary probability distribution of the energy queue length (i.e., the
amount of available green energy) is derived.
where wss,nu is the bandwidth of SSU, 𝜏ss,nu denotes the spectrum efficiency of cell edge
users given by [13]:
PTs,n 𝛼s + 2 𝜂
𝜏ss,nu = log2 (1 + ) (14)
(𝜃s + 1) 2𝜎 2 Ws D𝛼ns
the physical meaning of Eq. (13) is that the average data rate of the non-cell-edge users
(with spectrum efficiency above 𝜏ss,nu) should be no smaller than RQ.
280 K. Wu et al.
Then
PTm 𝛼m + 2 𝜂
𝜏mm,u = log2 (1 + ) (16)
(𝜃m + 1) 2𝜎 2 Wm D𝛼mm
where wmm,u is the bandwidth of MMU, the SBSs are considered to be uniformly distrib‐
uted in the macro cell.
In addition, the area served by MBS except SBSn is
∑
𝜋D2m′ = 𝜋D2m − 𝜋D2s (17)
s∈B
Next, we consider an MSU. The outage probability constraint of the MSUs can be
written as
where
PTm 𝛼m + 2 𝜂
𝜏ms,nu = log2 (1 + ) (19)
(𝜃m + 1) 2𝜎 2 Wm D𝛼ms,n
m
wms,u is the bandwidth of MSU, and Dms,u denotes the distance between the MBS and
SBSn.
The outage probability constraint of the WWUs is
where
Pw 𝛼w + 2 𝜂
𝜏w,u = log2 (1 + ) (21)
(𝜃w + 1) 2𝜎 2 Ww D𝛼ww
In this section, we optimize the traffic offloading for the single small cell case, where
the HSBS and RSBS are analyzed respectively.
TH
EE = (22)
Psum
The Energy Efficiency Research of Traffic Offloading Mechanism 281
TH is the network throughput [17], the network throughput is defined as the total amount
of data that is successfully received by the users per unit time and area, and is measured
in bits/sec/m2, Psum is the total on-grid power consumption.
a a o o
rmm rss,n rms,n rw,n rmm rms,n rw,n
TH = IH ( + + + ) + (1 − IH )( + + ) (23)
𝜋D2m′ 𝜋D2s,n 𝜋D2ms,n 𝜋D2w,n 𝜋D2m′ 𝜋D2ms,n 𝜋D2w,n
where IH is a 0–1 indicator denoting whether SBSn is active or not, rmm is the total data
a
rate of MMUs, rss is the total data rate of SSUs, rms is the total data rate of MSUs when
o a
the SBS is active, rms is the total data rate of MSUs when the SBS is not active.rw,n is the
o
total data rate of WWUs when the SBS is active, rw,n is the total data rate of MMUs when
the SBS is not active. 𝜓1 is the traffic offloading ratio of the SBS. 𝜓2 is the traffic offloading
ratio of the MBS.
where PMBS, PHSBS and Pwifi denote the on-grid power consumptions of MBS, HSBS and
wifi access points respectively. PMBS,PHSBS and Pwifi can be derived based on Eq. (8).
While wmm is constrained by Eq. (12), wams and woms are the corresponding bandwidth
needed by the MSUs. Then we have
𝛽m PTm
PMBS = PCm + (wmm + IH wams + (1 − IH )woms ) (29)
Wm
In addition, as the HSBS consumes on-grid power only when the battery is empty, we
have
⎧ 𝛽s PTs
⎪ IH q0 (PCs + w ) , 𝜆E < UE
PHSBS =⎨ Ws ss (30)
⎪ , 𝜆E ≥ UE
⎩ 0
where q0 is the probability of empty energy queue, obtained from Eq. (10).
282 K. Wu et al.
Then
𝛽w PTw
Pwifi = w (31)
Ww ww
Then, the network energy efficiency maximization problem of the single-HSBS case
can be formulated as follows:
P1:
max EE (32)
IH ,UE
s.t.
rmm,u ≥ RQ , rss,u ≥ RQ , rms,u ≥ RQ , rw,u ≥ RQ (33)
0 ≤ wss ≤ Ws (35)
0 ≤ www ≤ Ww (36)
where Eq. (33) guarantees the QoS, Eqs. (34), (35) and (36) are due to the bandwidth
limitations of MBS, HSBS, and wifi access points, respectively. UE can be derived based
on the power consumption model of HSBS Eq. (9).
In practice, the HSBS usually provides higher energy efficiency compared with the
MBS, due to shorter transmission distance and lower path loss. As a result, more subcar‐
riers should be utilized to offload more users if the HSBS is active, Besides,
wmm + woms > Wm happens when the MBS is overloaded, in which case the HSBS should
be activated to relieve the burden of the MBS. But the SBS can only service the amount
while wss < Ws.
where PMBS is the power consumption of the MBS, and PHO reflects the additional power
consumed by SSU handover.
Denote by IR ∈ {0, 1} the ON-OFF state of the RSBS. If RSBn is active, handover
happens in the following cases: (1) RSBn is shut down when the battery runs out; (2)
RSBn is reactivated when new energy arrives. According to the energy queueing model,
the first case corresponds to the event when L+ transits from 1 to 0, with frequency of
The Energy Efficiency Research of Traffic Offloading Mechanism 283
q1 A21 UE after the energy queue becomes stable. Due to the duality between the two
cases, the additional handover power consumption is given by
⎧ 𝜆E
⎪ 𝜆 −
U
= ⎨ 2IR (1 − U )(1 − e E )UE CHO , 𝜆E < UE
E
PHO = IR ⋅ 2q1 A21 UE CHO (38)
⎪ E
⎩ 0 , 𝜆E ≥ UE
Note that SBSn may be shut down due to energy shortage even when its state is set
as on, in which the MBS has to utilize more bandwidth to serve SSUn with additional
bandwidth. Based on Eq. (8), the average on-grid power consumption of the MBS is
given as follows:
𝛽m PTm
PMBS = PCm + (wmm + IR ((1 − q0 )wams + q0 woms ) + (1 − IR )woms ) (39)
Wm
where wmm is constrained by Eq. (15), wams and woms denote the bandwidth needed by the
MBS to serve MSUs when the RSBS is active and shut down, respectively.
The average on-grid power consumption of the wifi access point is given as follows:
𝛽w PTw
Pwifi = w (40)
Ww ww
max EE (41)
IH ,UE
s.t.
rmm,u ≥ RQ , rss,u ≥ RQ , rms,u ≥ RQ , rw,u ≥ RQ (42)
0 ≤ wss ≤ Ws (44)
0 ≤ www ≤ Ww (45)
where the objective function is the network energy efficiency, Eq. (42) guarantees the
QoS, Eqs. (43), (44) and (45)account for the bandwidth limitation of MBS, RSBS and
wifi access points, respectively.
284 K. Wu et al.
In this section, we evaluate the energy efficiency of the optimal solution for the single-
SBS case. The SBSs are micro BSs, and the main simulation parameters can be found
in Tables 1and 2. Solar power harvesting devices are equipped at RSBS and HSBS.
Figure 2 shows the network energy efficiency of no offloading and offloading different
users with different user density under given system parameters.
Simulation shows with the amount of offloading traffic to wifi access points
increasing, the energy efficiency is decreasing, regardless of the energy arrival amount.
Because SBSs are powered by conventional energy and green energy, wifi access points
are powered only by conventional energy and the radio of output and power consumption
decreases quickly. So the following contents are not including wifi case.
The red line and black line with no point show the no offloading case. It can be seen
that energy efficiency generally increases with the increase of offloading ratio and with
the increase of users, and offloading’s is larger than no offloading. Besides, in the actual
situation, wss > Ws happens when more users are offloaded to the SBS, in which case
not all users can be offloaded to the SBS for satisfying the QoS of every user, so we can
choose some users whose energy efficiency is maximized to offload to the SBS. For
RSBS, because of difference of user density, the energy efficiency’s variation trend is
different. Figure 2(b) shows when the user density is large enough, energy efficiency is
maximized with the offloading radio is not until 1.
The Energy Efficiency Research of Traffic Offloading Mechanism 285
Fig. 2. The energy efficiency with different offloading ratio and different user density of single
SBS. (Color figure online)
5.2 Network Energy Efficiency of Single-SBS with Different Energy Arrival Rate
Figure 3 shows the network energy efficiency of no offloading and offloading different
users with different user density and different energy arrival rate under the same system
parameters. Figure 3(a) shows the energy efficiency for the HSBS, which increases with
the increase of energy arrival rate in the small cell and with the increase of user density.
Figure 3(b) shows the energy efficiency for the RSBS. It can be seen the energy effi‐
ciency’s variation trend is different with the increase of energy arrival rate, when energy
arrival rate is small enough, the energy efficiency is maximized with the offloading radio
not until 1.
286 K. Wu et al.
Fig. 3. The energy efficiency with different offloading ratio and different energy arrival rate of
single SBS.
In this paper, we have investigated the network energy efficiency through offloading
traffic for green heterogeneous networks. The analytical results reflect maximized
energy efficiency with different user density through traffic offloading, we also analyze
offloading case of different energy arrival rate, so we can choose how much traffic should
be offloaded to the SBS and wifi access points with different energy arrival rate. But we
do not consider the case that all SBSs coexist, which SBS should be activated and how
much traffic should be loaded, next we will achieve this with new limits.
Acknowledgment. This work was supported by the National Natural Science Foundation of
China. (No. 61372117), and the State Major Science and Technology Special Projects of China
under Grant 2016ZX03001017-004.
The Energy Efficiency Research of Traffic Offloading Mechanism 287
References
1. Hasan, Z., Boostanimehr, H., Bhargava, V.K.: Green cellular networks: a survey, some
research issues and challenges. IEEE Commun. Surv. Tutor. 13(4), 524–540 (2011)
2. Correia, L.M., Zeller, D., Blume, O., et al.: Challenges and enabling technologies for energy
aware mobile radio networks. IEEE Commun. Mag. 48(11), 66–72 (2010)
3. Damnjanovic, A., Montojo, J., Wei, Y., et al.: A survey on 3GPP heterogeneous networks.
IEEE Wirel. Commun. 18(3), 10–21 (2011)
4. Zhang, N., Cheng, N., Gamage, A., Zhang, K., Mark, J.W., Shen, X.: Cloud assisted HetNets
toward 5G wireless networks. IEEE Commun. Mag. 53(6), 59–65 (2015)
5. Ismail, M., Zhuang, W., Serpedin, E., Qaraqe, K.: A survey on green mobile networking:
From the perspectives of network operatorsand mobile users. IEEE Commun. Surv. Tutor.
17(3), 1535–1556 (2014)
6. Andrews, J., et al.: What will 5G be? IEEE J. Sel. Areas Commun. 32(6), 1065–1082 (2013)
7. Zhuo, X., Gao, W., Cao, G., et al.: Win-Coupon: an incentive framework for 3G traffic
offloading. In: IEEE International Conference on Network Protocols, pp. 206–215. IEEE
Computer Society (2011)
8. Zhuo, X., Gao, W., Cao, G., et al.: An incentive framework for cellular traffic offloading.
IEEE Trans. Mob. Comput. 13(3), 541–555 (2014)
9. Usman, M., Vastberg, A., Edler, T.: Energy efficient high capacity HETNET by offloading
high QoS users through femto. In: IEEE International Conference on Networks, pp. 19–24.
IEEE Computer Society (2011)
10. Arshad, M.W., Vastberg, A., Edler, T.: Energy efficiency gains through traffic offloading and
traffic expansion in joint macro pico deployment. In: 2012 IEEE Communications and
Networking Conference (WCNC), pp. 2203–2208. IEEE (2012)
11. Dhillon, H., Li, Y., Nuggehalli, P., Pi, Z., Andrews, J.: Fundamentals of heterogeneous
cellular networks with energy harvesting. IEEE Trans Wirel. Commun. 13(5), 2782–2797
(2014)
12. Zhang, S., Zhang, N., Zhou, S., et al.: Energy-aware traffic offloading for green heterogeneous
networks. IEEE J. Sel. Areas Commun. 34(5), 1116–1129 (2016)
13. Zhang, X., Zheng, Z., Shen, Q., Liu, J., Shen, X., Xie, L.: Optimizing network sustainability
and efficiency in green cellular networks. IEEE Trans. Wirel. Commun. 13(2), 1129–1139
(2014)
14. Yang, J., Ulukus, S.: Optimal packet scheduling in an energy harvesting communication
systems. IEEE Trans. Wirel. Commun. 60(1), 220–230 (2012)
15. Nakagawa, K.: On the series expansion for the stationary probabilities of an M/D/1 queue. J.
Oper. Res. Soc. Jpn 48(2), 111–122 (2005)
16. Yu, P.S., Lee, J., Quek, T.Q.S., et al.: Traffic offloading in heterogeneous networks with
energy harvesting personal cells - network throughput and energy efficiency. IEEE Trans.
Wirel. Commun. 15(2), 1146–1161 (2016)
17. Wei, Y., Ren, C., Song, M., Yu, R.: The offloading model for green base stations in hybrid
energy networks with multiple objectives. Int. J. Commun Syst 29(11), 1805–1816 (2016)
A Game-Theoretic Approach for Joint
Optimization of Sensing and Access
in Cognitive Cellular Heterogeneous Networks
1 Introduction
Due to the ability to bring about massive spatial reuse of frequency, the low-power,
short-range mini-base stations (BSs), referred to small-cells (femto-cell, pico-cell) are
increasingly important for improving network capacity. These small-cells coexist with
conventional macro-cell networks, giving rise to cellular heterogeneous networks
(CHNs). The structure of CHN causes interference problem to each layer. The inter-tier
(inner-tier) interference (ITI) from the Macro BSs (small-cell BSs) to the small-cell
users can be critical challenge to the CHN. Severely, the lower spectral efficiency
(SE) and higher interference often results in great loss in the capacity and coverage.
Thereby, it is necessary to improve SE and reduce interference.
2 System Model
2.1 Network Scenario
We consider a CCHN consisting one MBS and multiple cognitive FBSs, where
femto-cells coexist with macro-cell in overlay approach. Each FBS operates in the
closed access mode where only authorized UEs are allowed to access to it. The sets of
FBSs and MUEs are denoted by S F ¼ fi; i ¼ 1; 2; . . .; NF g and S M ¼ fj; j ¼
1; 2; . . .; NM g. The femto-cells share same spectrum while they adopt the OFDMA as
the spectrum allocation strategy in each cell. Specifically, the total spectrum band is
divided into NC orthogonal sub-channels, marked as S C ¼ fn; n ¼ 1; 2; . . .; NC g. We
focus on block transmissions over channels which are assumed to be perfectly syn-
chronized. With a fixed length T, the transmission frame of FBS i is divided in two
slots: sensing slot si and data slot as shown in Fig. 1. During sensing interval, FBSs
sense the environment, looking for the spectrum holes, whereas during the data slot
they possibly transmit over the licensed spectrum detected as available. The sensing
and transmission phases are described in 2.1 and 2.2 respectively.
The channel model follows Rayleigh fading and the large-scale path loss. The
additive complex white Gaussian noise is also assumed (with mean being zero and
variance being r2 ). The channel fading between FBS i and its authorized UEs on the nth
2 2
subcarrier is normally indicated as gi;n . Meanwhile, gM;j;n represents the channel
2
fading between MBS and MUE j on the nth subcarrier, and gi;j;n stands for that
between the FBS i and the MUE j. Besides, the channel gain between MBS and the
2
UEs in FBS i is denoted as gM;i;n . In CCHNs, the interference brought by the
mistaken sensing between the BSs in the different tier is not considered when
restricting sensing accuracy. Therefore, we just analyze the inner-tier which is depicted
in Fig. 2.
The Spectrum Sensing Problem
Assume FBSs sense the licensed channels using energy detection in non-cooperative
way. In sensing phase, we define the event of MBS’s occupation as follows:
pffiffiffiffiffiffi
ei;n r2 NF lsi
f
Pi;n ðsi ; ei;n Þ ¼Q pffiffiffiffiffiffiffiffiffiffiffiffiffiffi ð2Þ
r2 2NF NC
f
Deriving from (1), Pi;n ðsi ; ei;n Þ can be further expressed with respect to Pdi;n ðsi ; ei;n Þ
as (3), where Rn ¼ r2 I þ 2DiagðGn Þ.
0 pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi pffiffiffiffiffiffi1
Q1 Pdi;n ðsi ; ei;n Þ r 2kRn k1 NC þ kGn k1 lsi
Pi;n ðsi ; Pdi;n Þ ¼ Q@
f
pffiffiffiffiffiffiffiffiffiffiffiffiffiffi A ð3Þ
r2 2NF NC
Successful detection: The goal of each FBS is to optimize its opportunistic spectral
utilization of the licensed spectrum. According to this paradigm, each sub-channel n is
available for the transmission of FBS i if no MBS’s signal is detected over that
f
frequency band, which happens with probability 1 Pi;n ðsi ; ei;n Þ. This motivates the
use of the aggregate opportunistic throughput as a measure of the spectrum efficiency
of each FBS i. Given the power allocation profile p , ðpi ÞNi¼1F
of the FBSs, the detection
NC
thresholds ei , ei;n n¼1 , and sensing time si , the aggregate opportunistic throughput of
FBS i is defined as
si X
NC
f
Ri ðsi ; ei ; pi ; bi Þ ¼ ð1 Þ PrðHn0 Þ 1 Pi;n ðsi ; ei;n Þ ri;n ðpi;n ; bi;n Þ ð6Þ
T n¼1
where 1 si =Tðsi \TÞ is the portion of the frame duration available for opportunistic
f
transmissions, Pi;n ðsi ; ei;n Þ is the FP defined in (2), and ri;n ðpi;n ; bi;n Þ is the maximum
rate achievable of FBS i over channel n when no MBS’s signal is detected and the
power allocation of the FBSs is given. The maximum achievable rate ri;n ðpi;n ; bi;n Þ is
expressed as
2
pi;n gi;n
ri;n ðpi;n ; bi;n Þ ¼ bi;n logð1 þ P 2 Þ ð7Þ
r2 þ pj;n gj;i;n
j6¼i
where r2 is the variance of the background noise over channel n at the receiver UE
associated to FBS i (assumed to be Gaussian zero-mean distributed).
f
As mentioned above, Pi;n can be written as (3). According to [1], when
Pi;n ðsi ; ei;n Þ ¼ Pth , (6) can achieve maximal value. Thereby, (7) can be rewritten as the
d d
following one:
A Game-Theoretic Approach for Joint Optimization 293
si X
NC
f
Ri ðsi ; pi ;bi Þ ¼ ð1 Þ PrðHn0 Þ 1 Pi;n ðsi Þ ri;n ð8Þ
T n¼1
Here,
pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi pffiffiffiffiffiffi!
Q1 Pdth r 2kRn k1 NC þ kGn k1 lsi
f
Pi;n ðsi Þ ¼Q pffiffiffiffiffiffiffiffiffiffiffiffiffiffi ð9Þ
r2 2NF NC
Energy constraints: To control the total power cost by FBSs, we propose to impose
individual energy constraints. Individual constraints are provided at the level of each
FBS i to limit the energy constraint. Individual energy constraints:
X
NC X
NC
si psi;n þ ðT si Þ pi;n Eimax ð10Þ
n¼1 n¼1
where psi;n is the sensing power of FBS i on channel n, and Eimax is the maximum energy
limitation allowed to be generated by FBS i.
X
NC 2
i;n gM;i;n pi;n Ii
pmiss ð11Þ
max
n¼1
to detect the presence of the MBS on the sub-channel n and thus generating interfer-
ence against the MBS.
In Sect. 2, we have defined the opportunistic throughput Ri and the individual con-
straints (10)-(11). In the process, we assume that FBSs aim to maximize their own
throughput non-cooperatively, with the constraints holding. Therefore, a non-
cooperative game theory can be introduced to build an optimization model if we
regard each FBS as a game player and Ri as the player’s utility. Hence, in this section,
we focus on the formulation of the jointing optimization of the spectrum sensing and
resource allocation of the FBSs within the framework of game theory. We propose next
294 C. Pan et al.
equilibrium problem: games with individual constraints, and then provide a unified
analysis of the game.
As mentioned above, each FBS is modeled as a game player who aims to maximize
its own opportunistic throughput Ri ðsi ; pi ;bi Þ by choosing jointly a proper power
NC
NC
allocation strategy pi ¼ pi;n n¼1 , carrier allocation strategy bi ¼ bi;n n¼1 , and
sensing time si , subject to individual constraints (10).
In order to enforce the constraints (10)-(11) while keeping the optimization as
decentralized as possible, we propose the use of a pricing mechanism through a penalty
in the utility function of each player. To be simple, we set the price of the ith player is
denoted by ci . Assume price ci in (12) is given. Therefore, we have the formulation (12).
Player i’s optimization problem is to determine, for the given c, a tuple ðsi ; pi ;bi Þ,
!
X
NC
maximize Ui ðsi ; p;bi Þ ¼ Ri ðsi ; p;bi Þ ci ðsi psi;n þ ðT si Þpi;n Þ ð12Þ
si ;pi ;bi
n¼1
s.t.
X
NC X
NC
si psi;n þ ðT si Þ pi;n Eimax ð13Þ
n¼1 n¼1
X
NC 2
i;n gM;i;n pi;n Ii
pmiss ð15Þ
max
n¼1
Here, (14) is derived from (2), (3), (5). smin ¼ maxðs1min ; s2min ; . . .; sNmin
C
Þ,
pffiffiffiffiffiffiffiffiffiffi p ffiffiffiffiffiffiffiffiffiffiffiffiffiffi
ffi 2
ð ð th Þ
Q 1
P f
r 2
2N N Q 1
ð th Þ
P d
r 2 kR n 1 CÞ
k N
snmin ¼
F C
lk G k 2
.
n 1
For notational simplicity, we will use /i , ðsi ; pi ;bi Þ to denote the strategy tuple of
player i.
4 Solution Analysis
This section is devoted to the solution analysis of the game (12). We start our analysis
by introducing the definition of non-cooperative power allocation game (NPAG).
According to (12), we can determine the optimal sub-channel matrix b once the
network vector p is determined. In addition, to simplify the jointing problem, we
compute the optimal power allocation vector with different sensing time, and then
choose the optimal tuple ðsi ; pi ;bi Þ as the optimal solution. Therefore, the jointing
problem of sensing time and resource allocation turns into a power allocation problem,
and the game becomes a NPAG. G ¼ S F ; fpi gi2S F ; fUi gi2SF as follows:
A Game-Theoretic Approach for Joint Optimization 295
!!
X
NC
NPAG : max Ui ðsi ; p;bi Þ ¼ max Ri ðsi ; p;bi Þ ci ðsi psi;n þ ðT si Þpi;n Þ ð16Þ
pi
n¼1
st.
pi 2 Pi ð17Þ
Here,
( )
X
NC X
NC X
NC 2
Pi ¼ pi j s i psi;n þ ðT si Þ pi;n Eimax ; pmiss gM;i;n pi;n I max
i;n i ð18Þ
n¼1 n¼1 n¼1
where ki and li are the Lagrangian multipliers for the maximum energy constraint (10)
and interference constraint (11). ½ x þ ¼ x if x 0 and 0 otherwise.
Unfortunately, it is hard to find the max Lagrangian multiplier to solve the pi;n due
to two multipliers. To get around the difficulty, we consider an iterative water-filling
algorithm (IWF) mentioned in [10]. The optimal response of game can be found by
computing the multipliers iteratively until they reach a certain degree of accuracy.
The details of IWF algorithm is listed in Algorithm 1.
296 C. Pan et al.
A Game-Theoretic Approach for Joint Optimization 297
0.055
n=1 i=1
n=1 i=2
0.05 n=2 i=1
n=2 i=2
transmission power/w
0.045
0.04
0.035
0.03
0 10 20 30 40 50 60
iteration number
Figure 3 presents the convergence of the proposed power allocation, where n and i
denote the index of channel and FBS. It can be seen that the power of FBS on each
channel declines gradually, and then keeps smooth. Meanwhile, the system gains the
stable performance. In the proposed algorithm, the power converges within 10 times to
achieve d accuracy.
Fig. 4. The comparison of throughput ver- Fig. 5. The comparison of utility function
sus sensing time versus sensing time
From Figs. 4, 5 and 6, we provide the comparison of two schemes to show the
effect of sensing time on femto-cells’ throughput, utility function and the energy
efficiency (EE). As depicted in Fig. 4, the throughput firstly increases with the growth
of si, and then decreases to the zero for the proposed IWF algorithm. That is because
the higher si which means the lower FP, improves the availability of spectrum resource
and gains more throughput by making more FBSs access into the idle channels.
However, when the sensing time reaches to some threshold, it makes little effect on FP
and throughput. Contrarily, the shorter of transmission time reduces the transmission
efficiency and degrade the system performance. The same trend can be found in utility
function as illustrated in Fig. 5. It is not difficult to notice that the price factor ki will
execute self-regulation according to the total power, aiming to limit the power of each
FBS. In contrary to throughput, the utility function achieves a lower value due to price
mechanism.
In Fig. 6, curve of EE keeps the same trend as in Fig. 5. Furthermore, IWF
algorithm is better than APA algorithm. That is because, in context of guaranteeing
throughput, IWF algorithm strongly reduces the total value of power allocation.
Figure 7 shows the increasing trend of throughput with the growth of inner SNR.
Since the higher SNR brings higher power allocation, which improves throughput of
femto-cells. While, with the increasing SNR, the inner-tier interference also become
larger, which limits the improvement of throughput. Thereby, the throughput is no
longer varied by the inner SNR.
A Game-Theoretic Approach for Joint Optimization 299
Fig. 6. The comparison of EE versus sens- Fig. 7. The comparison of throughput versus
ing time inner SNR
6 Conclusion
In this paper, the problem of optimal multichannel sensing (with respect to sensing time
and threshold) and resource allocation (with respect to channel and power) in CCHN is
explored by formulating a game and the FBSs act as the players to maximize his own
opportunistic throughput by choosing jointly the sensing and access configurations.
A pricing mechanism that penalizes the FBSs in violating the power constraints is also
provided. Finally, we deal with the non-convexity of the game by utilizing the iterative
water-filling (IWF) algorithm. Simulation results show the proposed IWF algorithm
gains better performance in both femto-cell throughput and EE than traditional APA
algorithm.
Acknowledgement. This work was supported in part by the National Natural Science Foun-
dation of China under Grant No. 61372117, 61302081, 61421061, and the 863 project.under
Grant NO.2014AA01A701.
References
1. Fan, R., Jiang, H., Guo, Q., Zhang, Z.: Joint optimal cooperative sensing and resource
allocation in multichannel cognitive radio networks. IEEE Trans. Veh. Technol. 60(2), 722–
729 (2011)
2. Janatian, N., Sun, S., Modarres-Hashemi, M.: Joint optimal spectrum sensing and power
allocation in CDMA-based cognitive radio networks. IEEE Trans. Veh. Technol. 64(9),
3990–3998 (2015)
3. Zhang, N., Liang, H., Cheng, N., Tang, Y., Mark, J.W., Shen, X.S.: Dynamic spectrum
access in multi-channel cognitive radio networks. IEEE J. Sel. Areas Commun. 32(11),
2053–2064 (2014)
4. Wang, F., Wang, J., Li, W.W.: Game-theoretic analysis of opportunistic spectrum sharing
with imperfect sensing. EURASIP J. Wirel. Commun. Netw. (2016)
300 C. Pan et al.
5. Dai, Z., Wang, Z., Wong, V.W.S.: An overlapping coalitional game for cooperative
spectrum sensing and access in cognitive radio networks. IEEE Trans. Veh. Technol. 65(10),
8400–8413 (2016)
6. Zhang, S., Wang, H., Zhang, X., Cao, Z.: Optimal spectrum access algorithm based on
POMDP in cognitive networks. AEU Int. J. Electron. Commun. 69(6), 943–949 (2015)
7. Jiang, C., Chen, Y., Gao, Y., Liu, K.J.R.: Joint spectrum sensing and access evolutionary
game in cognitive radio networks. IEEE Trans. Wirel. Commun. 12(5), 2470–2483 (2013)
8. Bose, S., Natarajan, B.: Reliable spectrum sensing for resource allocation of cognitive radio
based WiMAX femtocells. In: The 9th Annual IEEE Consumer Communications and
Networking Conference-Wireless Consumer Communication and Networking, pp. 889–893,
14–17 January 2012
9. Cheng, S.-M., Ao, W.C., Tseng, F.-M., Chen, K.-C.: Design and analysis of downlink
spectrum sharing in two-tier cognitive femto networks. IEEE Trans. Veh. Technol. 61(5),
2194–2207 (2012)
10. Yu, W., Ginis, G., Cioffi, J.M.: Distributed multiuser power control for digital subscriber
lines. IEEE J. Sel. Areas Commun. 20, 1105–1115 (2002)
Energy Harvesting Relay Node Deployment
for Network Capacity Expansion
Zhiqiang Zhang1 ✉ , Yifei Wei1 ✉ , Bo Gu1, Xaojun Wang2, and Mei Song1
( ) ( )
1
School of Electronic Engineering, Beijing University of Posts and Telecommunications,
Beijing 100876, People’s Republic of China
{zhangzhiqiang,weiyifei,gubo,songm}@bupt.edu.cn
2
Dublin City University, Dublin, Ireland
xiaojun.wang@dcu.ie
Abstract. The increasing traffic demands in the past few years require further
expansion of cellular network capacity. Densely deployed base stations may cause
a large amount of energy consumption and lots of environmental problems.
Relaying is considered as a promising cost-efficient solution for capacity expan‐
sion and energy saving due to the physical characteristics and low power require‐
ments of the relay nodes. In this paper, we study the energy harvesting relay node
deployment problem in a cellular network where the user distribution may change
with time. We propose a greedy relay nodes deployment algorithm to satisfy the
network capacity requirements in different time and put forward an operation
optimization algorithm to make full use of renewable energy. Simulation results
show that the network capacity can be increased significantly and the energy
consumption can be reduced dramatically by the proposed algorithms.
1 Introduction
In the recent years, the rapid rise in the popularity of smart devices requires a fast and
ubiquitous wireless network, which makes it necessary to expand the original network
capacity. However, increasing the number of macro base stations (BSs) in the cellular
network significantly reduces the gain because of the elevated interference. Heteroge‐
neous networks (HetNets) provide services across both macro and small cells to optimize
the mix of capabilities and have attracted much attention in the recent years [1, 2]. Relay
networks, as one kind of HetNets, is a promising solution of performance enhancement
and energy saving. Relay is one of the proposed technologies in LTE-Advanced
networks for coverage expansion and capacity enhancement [3]. Relay Nodes (RNs) are
connected to the core network with wireless backhaul through an evolved Node B (eNB)
and they usually have much smaller energy consumption than BSs.
Over the years, cellular networks have become an important source of energy
consumption, reducing the network energy consumption has benefits in both economic
and environmental aspects, which encourages people to look forward alterative power
supply. The usage of renewable energy sources such as wind power and photo-voltaic
energy can significantly reduce traditional energy consumption. Many kinds of renew‐
able energy sources have been studied for green networks as introduced in [4].
The improvements of network performance have been widely studied. In the paper
[5], micro base stations are deployed to increase the area spectral efficiency (ASE) value.
The results show that the ASE can be improved significantly when the micro base
stations located at the edges of macro cells where the received signal is worst. The energy
efficiency problem of cellular networks is studied in [6], excessively increasing the
number of micro base stations may reduce the energy efficiency, and there exists an
optimum number of deployed BSs. The capital expenditure and the total operational
expenditure are considered as the base station costs in [7]. A hierarchical power manage‐
ment architecture is introduced in [8], the authors focus on effectively exploit the green
capabilities of networks. A green base station offloading model is proposed in [9], by
predicting the value of green energy collected and updating the residual energy of each
BS, the maximum number of users that each BS can offload can be calculated to achieve
different network performance. The article [10] mainly focuses on the energy supply
aspects in the network and investigates how to optimize the using of on-grid energy to
reduce the carbon emission rate. A green base station optimization algorithm is proposed
in [11] to reduce the traditional power consumption, the transmission power of BSs can
be adjusted to improve network performance. The deployment of RNs has been used to
improve network performances in the recent years such as connectivity by [12] and
lifetime maximization by [13]. The article [14] considered finite-state Markov channels
in the relay-selection problem to increase spectral efficiency, mitigate error propagation,
and maximize the network lifetime. These studies usually focus on the user distribution
in peak periods to ensure the network performance in the worst case while ignoring the
differences of traffic demands in different positions and time periods, which may cause
much resource waste in the low-load state of the network.
In this paper, a more practical condition is considered. We deploy RNs to improve
the network capacity where one BS has already been deployed and the changes of user
distributions among different time is full considered. In addition, The RNs are equipped
with energy harvesting devices so that they can harvest ambient energy. RNs still need
to be connected to the on-grid power network in case that the renewable energy sources
are inadequate.
The remainder of the article is organized as follows. In Sect. 2, the system model is
introduced. In Sect. 3, the problem, constraints and objective function are defined.
Section 4 describes the proposed algorithms for these problems. Section 5 gives the
simulation parameters and results. Section 6 concludes the paper.
2 System Model
Consider a wireless network in which all users are served by one BS b. The capacity of
this network needs to be expanded to satisfy the growing traffic demand by deploying
RNs. The deployed RN set is denoted by R. The users, denoted by U, always connect
Energy Harvesting Relay Node Deployment for Network Capacity Expansion 303
to the BS b or a RN r which can provide the highest signal strength to them. The signal-
to-interference-plus-noise ratio (SINR) of user u can be written as
Ptx (b)gu,b
Γu (b) = ∑ (1)
Ptx (r)gu,r + 𝜎 2
r∈R
Ptx (r)gu,r
Γu (r) = ∑ (2)
Ptx (b)gu,b + Ptx (r′ )gu,r′ + 𝜎 2
r′ ∈R,r′ ≠r
Γu (b) and Γu (r) denote the SINR of users connected to the BS b and the RN r, respec‐
tively. Ptx (b) and Ptx (r) denote the transmission power of BS b and RN r, respectively.
gu,b is the channel gain from BS b to user u and gu,r is the channel gain from RN r to user
u. Thermal noise is denoted by 𝜎 2. For simplicity, the shadow fading and small size
decline are ignored, only the path loss is considered. Then the network capacity of user
u can be calculated from SINR as
Cu = Bu log2 (1 + Γu ) (3)
Pc (b) is the static power of each BS and 𝛼b scales the transmission power depending
on the load.
The power consumption model of each RN is similar, however, unlike previous
studies, we consider their energy harvesting abilities. With the help of energy harvesting
devices, the RNs can produce energy from renewable sources such as sunlight or wind.
However, renewable energy sources may be limited and the generated energy may not
Phv
be enough to totally power the RNs [17]. So we define 𝛽 = as the energy
P(r)max
harvesting rate, where Phv and P(r)max denote the average harvesting power and the
maximum consumption power of each RN, respectively. The RNs can consume less on-
grid power as long as 𝛽 > 0 and can work for a whole day without consuming any on-
grid power when 𝛽 ≥ 1. Thus the consumption power of each RN can be written as
{
0 𝛽≥1
P(r) = (5)
Pc (r) + 𝛼r Ptx (r) − Phv 0 ≤ 𝛽 < 1
Pc (r) is the static power of each RN and 𝛼r scales the transmission power depending
on the load, the average harvesting power is denoted by Phv.
304 Z. Zhang et al.
It’s assumed that users connected to the network always have data to transmit with
full buffer, in addition, no power control algorithm is used. Therefore, BSs and RNs are
fully utilized, 𝛼b and 𝛼r are constant for all BSs and RNs.
3 Problem Definition
Deploying RNs can improve the network capacity to a higher level, but it also brings extra
energy consumption. By efficiently deploy RNs, we can minimize the extra energy
consumption while meeting the requirement of capacity. The problem can be formulated as
∑
min P(r)
∑
r∈R
(6)
s.t. Cu (B ∪ R) ≥ 𝜆 ⋅ Co
u∈U
Cu (B ∪ R) denotes the capacity of user u when both BSs and RNs are deployed. Co
is the original network capacity without deploying any RNs and the multiplier 𝜆 ≥ 1 is
the desired increasing times over Co after deploying RNs.
4 Solutions
Affected by user living and working conditions, the user distribution may always change
with time in a day [18]. The deployment of BSs generally aims at satisfying the capacity
demand in peak hours, while ignoring the resource wastes in the other hours. It has been
reasonable in the past few years for the expensive costs and deployment inconvenience
of traditional BSs. With the deployment of RNs, these problems matter less for low costs
and high flexibilities of RNs. A whole day can be divided into several parts, each of
them corresponding to a unique user distribution. Then the RNs can be deployed in each
time slice to meet the capacity requirements.
There can be lots of candidate locations for RN deploying in this area and we need
to select some of them. Furthermore, with the rising number of deployed RNs, the
network performance may not always increase due to the interference effect, especially
when the RNs are located close. The strong interference between each other even
decreases the network capacity. On the other hand, influenced by the user distribution,
capacity contributions from differently located RNs may vary dramatically. Therefore,
the RN deployment problem is actually a combinatorial problem and it quickly becomes
unsolvable when the number of candidate locations becomes large.
Thus, a greedy algorithm is proposed to select the relatively better locations of RNs.
In this algorithm, a RN will be deployed at the location which can provide the most
increase of network capacity. The RNs will be added iteratively until the capacity
requirement is satisfied.
Energy Harvesting Relay Node Deployment for Network Capacity Expansion 305
Rti,tj denotes the merged RN set and will work in both time slices ti and tj. dist(rx , ry )
gives the distance between rx and ry, getMidRN(rx , ry ) gives a new RN that located at the
middle position of rx and ry.
5 Simulation Results
The deployment area size is 3 km × 3 km and served by one BS. RNs need to be deployed
in this area to improve the capacity. More candidate locations can get more accurate
results but also brings extra calculation complexity. The number of candidate locations
is set to 1600 after a set of experiments. The other parameters used in simulations are
given in Table 1.
Energy Harvesting Relay Node Deployment for Network Capacity Expansion 307
3
Interference Considered
2.8 Interference Ignored
2.6
2.4
Normalized Network Capacity
2.2
1.8
1.6
1.4
1.2
1
0 5 10 15 20 25 30 35 40 45 50
Number of Deployed Relay Nodes
3000
Relay Nodes in Time i
Base Station
Relay Nodes in Time j
2500
2000
Position (m)
1500
1000
500
0
0 500 1000 1500 2000 2500 3000
Position (m)
10 7
2.5
1.5
0.5
0
1 2 3 4
Number of Time Slice Groups
6 Conclusions
Acknowledgments. This work was supported by the National Natural Science Foundation of
China (No. 61571059).
References
1. Damnjanovic, A., Montojo, J., Wei, Y., et al.: A survey on 3GPP heterogeneous networks.
IEEE Wirel. Commun. 18(3), 10–21 (2011)
2. Ghosh, A., Ratasuk, R.: Essentials of LTE and LTE-A. Cambridge University Press (2011)
3. Saleh, A.B., Redana, S., Hämäläinen, J., Raaf, B.: Comparison of relay and pico eNB
deployments in LTE-Advanced. In: IEEE 70th Vehicular Technology Conference (VTC
2009-Fall), USA, September 2009
4. Schmitt, G.: The green base station. In: International Conference on Telecommunication -
Energy Special Conference, VDE, pp. 1–6 (2009)
5. Richter, F., Fehske, A.J., Fettweis, G.P.: Energy efficiency aspects of base station deployment
strategies for cellular networks. In: Vehicular Technology Conference Fall, pp. 1–5. IEEE
Xplore (2009)
6. Coskun, C.C., Ayanoglu, E.: A greedy algorithm for energy-efficient base station deployment
in heterogeneous networks. In: IEEE International Conference on Communications, pp. 7–
12. IEEE (2015)
7. Chen, Y., Zhang, S., Xu, S.: Characterizing energy efficiency and deployment efficiency
relations for green architecture design. In: IEEE International Conference on
Communications Workshops, pp. 1–5. IEEE (2010)
8. Wei, Y., Wang, X., Fialho, L., Bruschi, R., Ormond, O., Collier, M.: Hierarchical power
management architecture and optimum local control policy for energy efficient networks. J.
Commun. Netw. 18(4), 540–550 (2016)
9. Wei, Y., Ren, C., Song, M., Yu, R.: The offloading model for green base stations in hybrid
energy networks with multiple objectives. Int. J. Commun. Syst. 29(11), 1805–1816 (2016)
10. Han, T., Ansari, N.: Optimizing cell size for energy saving in cellular networks with hybrid
energy supplies. In: Global Communications Conference, pp. 5189–5193. IEEE (2012)
11. Pamuklu, T., Ersoy, C.: Optimization of renewable green base station deployment. In: Green
Computing and Communications, pp. 59–63. IEEE (2013)
310 Z. Zhang et al.
12. Cheng, X., Du, D.Z., Wang, L., et al.: Relay sensor placement in wireless sensor networks.
Wirel. Netw. 14(3), 347–355 (2008)
13. Hou, Y.T., Shi, Y., Sherali, H.D., et al.: Prolonging sensor network lifetime with energy
provisioning and relay node placement. In: 2005 Second IEEE Communications Society
Conference on Sensor and Ad Hoc Communications and Networks, IEEE SECON 2005, pp.
295–304. IEEE Xplore (2005)
14. Wei, Y., Richard Yu, F., Song, M.: Distributed optimal relay selection in wireless cooperative
networks with finite state Markov channels. IEEE Trans. Veh. Technol. 59(5), 2149–2158
(2010)
15. GPP, TR 36.814: Further advancements for E-UTRA physical layer aspects (Release 9).
Technical report, March 2010
16. Richter, F., Fehske, A.J., Fettweis, G.P.: Energy efficiency aspects of base station deployment
strategies for cellular networks. In: Vehicular Technology Conference Fall, pp. 1–5. IEEE
Xplore (2009)
17. Han, T., Ansari, N.: ICE: intelligent cell brEathing to optimize the utilization of green energy.
IEEE Commun. Lett. 16(12), 866–869 (2012)
18. Nousiainen, S., Kordybach, K.: User distribution and mobility model framework for cellular
network simulations. In: VTT Information Technology, pp. 518–522 (2002)
Parameters Optimization for KFKM Clustering Algorithm
Based on WiFi Indoor Positioning
1
State Key Laboratory of Networking and Switching Technology, Beijing University of Posts
and Telecommunications, Beijing 100876, People’s Republic of China
Zhengying_hu@163.com
2
Hisense Co., Ltd, Qingdao 266061, People’s Republic of China
1 Introduction
The basic idea of the location fingerprinting method for WiFi positioning consists of
two steps. The first step is to build a database that stores pre-recorded received signal
strength (RSS), or “fingerprint”, from different WiFi access points (AP) at different
reference points (RP). Then, the second step is to estimate the target’s location by
matching real-time RSS with the offline-constructed database (Fig. 1).
The main limitation of the matching process is the high implementation cost when
the database is too large to traverse. Hence, some improved clustering methods were
proposed to manage the database such as K-means and fuzzy K-means (FKM) algo‐
rithms. However, non-linear data is beyond K-means method’s processing power since
the objective function is non-differentiable. Motivated by this, in [1], the authors came
up with fuzzy K-means (FKM) method, which aim to solve the non-differentiability of
the objective function; recently in [2], Saadi et al. used an un-supervised machine
learning K-means clustering based on two LEDs and achieved great accuracy. In this
paper, we propose a new scheme to strengthen the performance of the WiFi positioning.
We first amend the data acquisition process of the database; second, we use sample
density method to modify cluster amount to match with the characteristic of database;
last, we propose an ideal kernel matrix to determine the kernel parameter σ.
1
Q
∑ Di + 𝜀
(̂x, ŷ ) = × (xi , yi ) (1)
Q
i=1 ∑ 1
i=1 Di + 𝜀
Where (xi, yi) is the coordinate of the ith RP, set ε to 0.003 to prevent the “zero divisor”
condition. In order to evaluate the precision of positioning test, the root-mean-square-
error (RMSE) between the estimated position and the real position is proposed to
measure errors. What’s more, a RMSE drawn from one positioning test doesn’t represent
the performance of an algorithm. Hence, we collect RMSE from times of experiments
to evaluate the algorithm comprehensively. In general, cumulative probability curve of
RMSE is an intuitive representation for evaluation.
The Chemistry building in BNU was set to verify the optimizations of database
structure would improve localization accuracy. Considering storage capacity and calcu‐
lation complexity, set RP interval = 2 m and collecte RSS from 5 AP at all RP(black
points) in Fig. 2, we have the database of the on-site test.
Parameters Optimization for KFKM Clustering Algorithm 313
Figure 3 shows the δ-ρ situation of the on-site test database. There are 7 peaks in the
upper right corner of the figure. Thus, k = 7 is the optimal cluster amount.
Function_2: set the optimal kernel parameter σ to get better clustering result.
Gaussian kernel function is:
‖x− y‖2
k(x, y) = exp(− ) (3)
2𝜎 2
314 Z. Hu et al.
While ||·|| means the Euclidean distance of vectors. The experiential value of σ is
given by the reciprocal of sample hypersphere radius [5] or enumeration, none is optimal.
Therefore, the optimal σ is supposed to approximate the actual kernel matrix to the ideal
matrix [6]. K-means algorithm firstly divides the database into k clusters and gets a row
vector L with n dimensions. The element Li = k′ , 1 ≤ k′ ≤ k, i = 1, 2, ⋯ n means the
affiliation of corresponding sample.
Parameters Optimization for KFKM Clustering Algorithm 315
With the purpose of making the samples within the same cluster as similar as possible
and samples from different clusters as unlike as possible, the ideal kernel matrix should
meet the following condition:
{
′ 1, Li = Lj
kij = (4)
0, Li ≠ Lj
The kernel parameter that approximate the kernel matrix to the ideal one is optimal.
The similarity can be quantified by the difference between kij and kij′ :
n n
∑ ∑ ′
E(k, k′ ) = (kij − kij )2 (5)
i=1 j=1
dE(k, k′ )
The solution to = 0 (noted as σ0) minimizes E(k, k′ ).
d𝜎
After clustering, instead of matching all reference points in the database, it is capable
to complete matching process by comparing real-time RSS with 7 cluster centers and 9
316 Z. Hu et al.
RP (at most) in the nearest cluster for 16 times in total after POKFKM clustering.
Therefore, the computational load reduces by 74.6%.
After initialization, first position the test position RSS to the nearest cluster center,
then traverse other RP within the cluster using WKNN, estimated position is the result
of these two steps. We also use cumulative probability curves of RMSE of 70 times of
on-site tests to illustrate the positioning accuracy of the three different clustering results
above. The simulation result is illustrated in Fig. 5.
It’s shown that POKFKM have better performance than KFKM clustering. The
average RMSE of POKFKM is −23.48% narrowed than KFKM, which proves that
POKFKM method has superior positioning precision than the KFKM algorithm. Mean‐
while, POKFKM has less standard deviation of RMSE, which stands for higher robust‐
ness. KFKM Clustering method has 56.9% probability to derive expected error, while
POKFCN method is more reliable with the probability of 83.3%. Hence, POKFKM
clustering method largely increased the precision and the reliability of WiFi positioning
system.
5 Conclusion
In this paper, three optimizations for WiFi indoor positioning system were proposed to
enhance WiFi positioning. Simulation result has shown that the optimization of kennel
function and cluster amount performed better than original KFKM algorithm. During
on-site positioning experiments in the chemistry building in BNU, reasonable deploy‐
ment of RP and AP made the database both compendious and integrate. POKFKM
Parameters Optimization for KFKM Clustering Algorithm 317
algorithm was proved to be more reliable and robust and substantially increased the
efficiency of WIFI indoor positioning.
Acknowledgement. This work was supported by the National Natural Science foundation of
China (61421062) and Hisense Co., Ltd.
References
1. Suroso, D.J., Cherntanomwong, P., Sooraksa, P., et al.: Location fingerprint technique using
fuzzy C-means clustering algorithm for indoor localization. In: TENCON 2011 IEEE Region
10th Conference, pp. 88–92. IEEE, Bali (2011)
2. Saadi, M., Ahmad, T., Zhao, Y., et al.: An LED based indoor localization system using k-
means clustering. In: 2016 15th IEEE International Conference on Machine Learning and
Applications (ICMLA), pp. 246–252. IEEE (2016)
3. Sadeghi, H., et al.: A weighted KNN epipolar geometry-based approach for vision-based
indoor localization using smartphone cameras. In: Sensor Array and Multichannel Signal
Processing Workshop (2014)
4. Rodriguez, A., Laio, A.: Clustering by fast search and find of density peaks. Science 344(6191),
1492–1496 (2014)
5. Li, Z., Weida, Z.: Kernel clustering algorithm. Chin. J. Comput. 25(6), 587–590 (2002)
6. Yang, H.L., Huang, Z.G., Liu, J.W., et al.: WIFI fingerprint localization based on Kernel Fuzzy
C-means clustering. J. Zhejiang Univ. 50, 1126–1133 (2016)
Energy Harvesting Time Coefficient Analyze
for Cognitive Radio Sensor Network
Using Game Theory
Mengyu Zhao(&), Yifei Wei, Qiao Li, Mei Song, and Ningning Liu
Abstract. In this paper, the cognitive radio sensor node can harvest energy
from the radio frequency signal which is transmitted by the primary user. A time
switching protocol was used to divide cognitive users’ time into three phases:
spectrum sensing mode, energy harvesting mode and data transmission mode.
Therefore, the optimal energy harvesting mode time selection is a question to be
solved. We consider a non-cooperative game model in which cognitive users are
regarded as selfish players aiming to maximize their own energy efficiency.
Then we prove the existence and uniqueness of Nash Equilibrium. A distributed
best response algorithm is used to obtain the Nash Equilibrium. The simulation
results prove that this algorithm can converge to the same equilibrium from
different initial values. At last, we analysis the influence of various system
parameters on the results of Nash Equilibrium and energy efficiency.
1 Introduction
In recent years, with the increase of network energy consumption and the development
of green communication, energy harvesting has drawn more and more attention.
Energy harvesting is a technology to keep self-sustaining and prolong network lifetime
by harvesting the ambient energy (such as wind energy, solar energy, heat, etc.).
Moreover, the sensor network with the capacity of harvesting energy from the ambient
radio-frequency (RF) signals has received extensive research. Energy harvesting
wireless sensor network (EHWSN) can greatly prolong the lifetime of the sensor nodes,
which lays foundation for the development of emerging technologies such as big data
and Internet of Things (IoTs) [1]. But it is worth to note that the sensor node can only
work on the unlicensed spectrum band. And the unlicensed spectrum band are being
more and more crowded, which limit the development of the sensor network.
Meanwhile, the survey shows the spectrum efficiency of licensed spectrum is rel-
atively low. So energy harvesting cognitive radio sensor network (EHCRSN) has been
put forward. In EHCRSN, the second user is the sensor node equipped with the ability
© Springer International Publishing AG 2018
Q. Zu and B. Hu (Eds.): HCC 2017, LNCS 10745, pp. 318–329, 2018.
https://doi.org/10.1007/978-3-319-74521-3_35
Energy Harvesting Time Coefficient Analyze 319
of spectrum sensing, which can also harvest the RF energy. But energy harvesting and
data transmission of the sensor user can’t be implemented at the same time. Therefore,
two practical receiver methods have been proposed in [2], namely the time switching
protocol (TS) and power splitting protocol (PS). In the former, the receiver harvest
energy in the first period and then transmit the information in the remainder time. In PS
protocol, the receiver split the received signal power into two streams: energy har-
vesting and information transmission.
There have been several papers in the literature that focus on the TS and PS
protocol in the past years. In [3], the author research a decode-and-forward energy
harvesting relay cognitive network using TS protocol. However the scenario is too
specialized. [4] research the relay selection question using TS and PS protocol in
cooperative cognitive radio network. [5] investigates a distributed power splitting
architecture with the purpose of achieving the optimal sum-rate through selecting the
power splitting ratios. RF energy harvesting in DF relay network is studied in [6]. In
this paper, TS protocol, PS protocol and a hybrid TS-PS protocol are applied to
compare and analyze outage probabilities and throughput.
In the aspect of cognitive radio sensor network, the author study the power allo-
cation in cognitive sensor networks in [7]. However, no energy harvesting and TS
protocol are considered in it. [8] develop a framework and propose a low-complexity
algorithm in EHCRSN. [9] research the optimal mode selection question to maximize
the throughput of the sensor node in EHCRSN. But to the best of our knowledge, there
is seldom investigation of applying TS protocol in EHCRSN and using the best
response algorithm to solve it. The main contributions of this paper are summarized as
follows:
1. We propose a new cognitive radio sensor network architecture which discuss the
question of how to select the energy harvesting time to balance the energy con-
sumption and energy replenishment in EHCRSN.
2. We reply the time switching relaying (TSR) protocol into the cognitive users, which
divides the time T into three parts: spectrum sensing phase, energy harvesting phase
and data transmission phase. And we analysis the energy efficient maximization
question through selecting the optimal energy harvesting coefficient.
3. We formulate this problem as a non-cooperative game. Then we use the best
response algorithm to reach Nash Equilibrium and prove the good performance of
our algorithm. At last, we get the result that different parameters variations have the
different influence on the energy harvesting time and system utility.
The remainder of the paper is structured as follows. Section 2 describes the system
model of primary user and cognitive user. In Sect. 3 we formulate the maximization
question as a non-cooperative game and prove the existence and uniqueness of Nash
Equilibrium. Then best response algorithm is introduced in Sect. 4. Finally Sect. 5
presents the numerical results and Sect. 6 concludes this paper.
320 M. Zhao et al.
2 System Model
We consider a cognitive radio sensor network with one primary user (PU) and N
cognitive users (CUs) as illustrated in Fig. 1. The cognitive users, also called secondary
users (SUs), are sensor node equipped with the ability of spectrum sensing. Assume
that each cognitive radio sensor user can harvest energy and is equipped with the
energy storage device. Denote N = {1,2,…,N} the set of CUs.
Fig. 1. The system model of the primary user and cognitive users
Where gi is the channel gain from PU to the CUi. Ps and x1 are the fixed transmit
power and the transmission information of the PU. nai is the additive Gaussian noise
with zero mean and r2i; a variance. For simplicity, we ignore the influence of the
Gaussian noise. The energy CUi harvested during the duration abT can be expressed as
where g is the energy harvesting efficiency. Assume the energy Qi will be used all to
the next data transmission and spectrum sensing. The transmission power of CUi is
assumed to be so small that the harvested energy is enough to use.
In the rest time ð1 bi ÞaT, CUi will transmit data using the harvested energy to the
other CUs or the PU. Since we know the energy consumption during the sensing period
is Pa ð1 aÞT. We can calculate the transmission power of CUi as:
Qi Pses ð1 aÞT
Pi ¼
að1 bÞT
ð3Þ
1 1a
¼ ðg bi Ps jgi j2 Pses Þ
1 bi a
hii pi
SINRi ¼ P
n ð4Þ
hij pj þ a2c
j¼1;j6¼i
Where hij denote the channel gain between the CUi and CUj, a2c is the Gaussian
noise introduced by the environment. We konw the SINRi is in relation to the energy
harvesting coefficient vector bi ¼ ½b1 ; b2 ; . . .bN from the Eqs. (3) and (4).
To insure the data transmission of the PU would not be affected, the interference
power from CUs to PU must be limited in a tolerable range. Assume that the largest
transmission power of CUs is Pmax. Conversely, if the CUs wants to transmit infor-
mation normally, its transmission power must be larger than a fixed value Pmin.
Therefore, we get the range of CUi’s transmission power is
Where Pi is defined in Eq. (3), and we can get the inverse solution of bi .
In order to simplify the description, we define the minimum and maximum value of
b as bmin and bmax respectively.
On the basis of Shannon formula, we know the achievable transmission rate of CUi is
determined by the SINRi given by Eq. (4). Therefore the rate of CUi can be obtained by
ri B logð1 þ SINRi Þ
ui ðbÞ¼ ¼ ð8Þ
Ptotal Pses + Pi
In this paper, we express the time selection problem with the goal to maximize the
energy efficiency, which is a globally optimal network-wide performance problem. In
actuality, this is a tradeoff of time selection between energy harvesting mode and data
transmission mode. We need to solve the following network utility maximization problem
X
N
max ui ðbÞ
b
i¼1
aPmin þ Pses ð1 aÞ aPmax þ Pses ð1 aÞ ð9Þ
s:t: bi
aðPmin þ gPs jgi j Þ 2
aðPmax þ gPs jgi j Þ 2
0 bi 1
Energy Harvesting Time Coefficient Analyze 323
• Player set N ¼ f1; 2. . .; i; . . .Ng: the set of N cognitive radio sensor users.
• actions fSi g: Each player i selects its energy harvesting coefficient bi 2 Si to
Q
maximize its energy efficient. S ¼ Ni¼1 Si refer to the strategy space of N players
and strategy vector s ¼ fs1 ; s2 ; . . .; sN g is a subset of S.
• Utility function fui ðqi ; qi Þg: the utility of player i.
That is to say, a NE is a strategy vector with the property that any player won’t have
better utility by chancing its strategy. The following theorem tells us how to distinguish
whether a NE exists or not for a particular game.
Theorem 1. In the game g ¼ hN; fSi g; fui ðqi ; qi Þgi, a NE exists if it have the limited
players and the strategy sets Si are convex set, closed and bounded. Last the utility
functions ui ðqi ; qi Þ is continuous and quasi-concave for 8i 2 N.
Proposition 1. The formulated energy harvesting time selection game g exists at least
one NE.
324 M. Zhao et al.
Proof. Firstly, we have mentioned that there are N players in the game totally, which
meets the first condition in Theorem 1. Secondly, because the strategy sets of CUi are
segments in two dimensions. It is straightforward to know that Si is convex, closed and
bounded. Besides, the utility function ui ðbÞ is continuous in b. Thus, we only need to
focus on the quasi concavity of ui ðbÞ.
If f(x) is quasi-convex, –f(x) is quasi-concave. So as long as we proved ui ðbÞ is
quasi-convex, we can get the result that this game possesses at least one NE.
Theorem 2. A continuous function f: R ! R is quasiconvex if and only if at least one
of the following conditions is satisfied.
• f is nondecreasing.
• f is nonincreasing.
• there exists a point c 2 dom f which satisfy for t c (and t 2 dom f ), f is nonin-
creasing, and for t c (and t 2 dom f ), f is nondecreasing.
Now we perform the first-order derivative of the negative utility function,
B logð1 þ SINRi Þ
fi ðbÞ ¼ ui ðbÞ ¼ ð12Þ
Pses + Pi
ðpi þ Pses Þ2
In reality, we know the primary user’s transmission power is far bigger than the
sensing power of CUi. So the first half of Eq. (13) gPs jgi j2 1a
a
Pses [ 0 is always
@pi
established and the derivative @b [ 0 is true forever. The transmission power is
i
increasing with the increase of bi . When the value of bi is ba ¼ 1a a gPs jgi j2 , we have
Pses
i ðbÞ
pi ðba Þ ¼ 0 and @f@b \ 0 After analysis, we know there is only one constant bc
b¼ba
i
i ðbÞ
which satisfy @f@b ¼ 0. And fi ðbÞ is firstly decreasing and then increasing with the
i b¼bc
increasing of bi . Therefore, we can conclude that ui ðbÞ is quasi-concave and there exist
at least one NE in the game.
F. Uniqueness for the NE
@ 2 ui
0; 8i 6¼ j 2 N ð14Þ
@si @sj
@ 2 ui ðbÞ 1 a
¼ B logðeÞ ðgPs jgi j2 Pses Þ2
@bi @bj 2
ð1 bi Þ ð1 bj Þ 2 1 a
" #
Pn
hii hij pi hii pses ðac þ
2 2
hij pj Þ ð15Þ
j¼1;j6¼i
P
n P
n
ðpi þ pses Þ2 ða2c þ hij pj Þða2c þ hij pj Þ
j¼1;j6¼i j¼1
Because the sensing power is far smaller than the transmission power of cognitive
users. As a result, the ratio of sensing power to the transmission power is a pretty small
value and the above inequality is reasonable naturally. Finally, we can conclude that
this game is a super-modular game and only exist one NE.
In this section, we will introduce a distributed best response algorithm which can
achieve the unique NE from any initial state. The aim of this algorithm is to maximize
the energy utility through all the users changing their energy harvesting coefficient bi
one by one until a suitable termination condition is satisfied.
First, when t = 0, set each player an initial value bi ð0Þ from its domain and
compute their utility value ui ð0Þ. Then, each player will update its energy harvesting
time coefficient bi ðtÞ according to Eq. (17) in a fixed sequence while other players
bi ðtÞ remain unchanged.
Finally, the iterative process will stop until the following termination condition is
satisfied.
326 M. Zhao et al.
X
N
½bi ðt þ 1Þ bi ðtÞ2 g ð18Þ
i¼1
5 Numerical Results
In this section, we will present the numerical results to illustrate and demonstrate the
superiority of the algorithm. We select eight cognitive users to analyze and the channel
gain from PU to the CUs are from interval [0.3,0.6]. The channel gain hii and hij
between CUs vary from interval [0.8,1.2] and [0.05,1.15] independently. The recent
research shows that the achieved power of RF harvester can achieve 5 mW under
1000 MHz. And we set the shared spectrum is 1000 MHz. The minimum and maxi-
mum transmission power of CUi are Pmin ¼ 0:1 mW and Pmax ¼ 5 mW. For the fol-
lowing examples if not mentioned otherwise, we set Ps = 20 mW, Pses = 0.2 lW,
a = 0.2, g = 0.5 and a2c = 1 mW.
Figures 3 and 4 show that the best response algorithm can converge to a unique NE
since the initial value of CU1 choose different values. It is observed that the conver-
gence process of CU1’ energy harvesting coefficient b1 and the utility when the initial
condition is set in different values. It can be seen from Fig. 4 that the utility of CU1 is
rising constantly until reaching the NE.
Figure 5 depicts the utility comparison between the initial condition and the finial
condition achieving the NE. It shows the result that each player’s final utility is much
larger than its corresponding utility in initial condition, which illustrates the superiority
of our algorithm. Next, we will analysis the impacts of one parameter (e.g.: g, Pses and
N) change on the solution of NE and the utility value.
Energy Harvesting Time Coefficient Analyze 327
Fig. 5. Utility comparison between the initial condition and the finial condition
Figure 6 show the finial energy harvesting time coefficient bi and energy efficiency
ui versus the energy harvesting efficiency g. It can be observed from Fig. 6(a) that bi
decreases as g increases, because a larger g means less harvesting time to achieve the
Fig. 6. The impact of energy harvesting efficiency on (a) energy harvesting time coefficient and
(b) energy efficiency with N = 8, Pses = 0.2 lW and a = 0.2
328 M. Zhao et al.
needed transmission power. And we can see from Fig. 6(b) that g has no effect on the
finial energy efficiency.
Figure 7 show the influence of CUi’s sensing power on energy harvesting time
coefficient and energy efficiency. We can see that the energy harvesting time coefficient
bi keep unchanged first and then increasing with the growth of CUi’s sensing power.
Meanwhile, the energy efficiency keep unchanged first and then decreasing. This is
because when the sensing power is less than a certain threshold, the change of sensing
power is so small in relation to the transmission power that has no effect on the NE. But
with the growing of sensing power and when it is bigger enough, the CUi need more
harvesting energy to support it and the energy efficiency decrease based on the Eq. (8).
Fig. 7. The impact of CUi’s sensing power on (a) energy harvesting time coefficient and
(b) energy efficiency with N = 8, g = 0.5 and a = 0.2
P
N
At last, we will analysis to the finial total energy efficiency uall ¼ ui versus the
i¼1
number of cognitive users N. Figure 8 replies that the total energy efficiency of the two
carve is almost the same when there are two players and then have opposite tendency.
And we can see from the blue curve that with the increasing of N, the total utility after
game increases first and then decreases. Because when the user number is lower than a
certain value, the interference from other users is relatively small and the total utility
increase with increasing the number. However, with a further increase of N, the
interference become stronger, which causes the direct decreasing of the total utility.
Fig. 8. The total energy efficiency versus different number of players with Pses = 0.2 lW and
g = 0.5 (Color figure online)
Energy Harvesting Time Coefficient Analyze 329
6 Conclusion
In this paper, we have researched how to select the optimal energy harvesting time to
achieve the maximize energy efficiency using non-cooperative game. First of all we
proved the existence and uniqueness of Nash equilibrium, and then we got the con-
verging process to achieve the NE with the best response algorithm. The results
demonstrate that the best response algorithm can converge to the same equilibrium
from different initial values. Besides, non-cooperative game yields a good utility per-
formance to address this energy harvesting time coefficient selection question. At last
we studied the effect of various system parameters on the results of Nash equilibrium.
Acknowledgment. This work was supported by the National Natural Science Foundation of
China (No. 61571059), and the State Major Science and Technology Special Projects of China
under Grant 2016ZX03001017-004.
References
1. Ku, M.L., Li, W., Chen, Y., et al.: Advances in energy harvesting communications: past,
present, and future challenges. IEEE Commun. Surv. Tutorials 18(2), 1384–1412 (2016)
2. Nasir, A.A., Zhou, X., Durrani, S., et al.: Relaying protocols for wireless energy harvesting
and information processing. IEEE Trans. Wirel. Commun. 12(7), 3622–3636 (2012)
3. Mousavifar, S.A., Liu, Y., Leung, C., et al.: Wireless Energy Harvesting and Spectrum
Sharing in Cognitive Radio, pp. 1–5 (2014)
4. He, J., Guo, S., Wang, F., et al.: Relay selection and outage analysis in cooperative cognitive
radio networks with energy harvesting. In: ICC 2016 IEEE International Conference on
Communications, pp. 1–6. IEEE (2016)
5. Chen, H., Li, Y., Jiang, Y., et al.: Distributed power splitting for SWIPT in relay interference
channels using game theory. IEEE Trans. Wirel. Commun. 14(1), 410–420 (2015)
6. Elmorshedy, L., Leung, C., Mousavifar, S.A.: RF energy harvesting in DF relay networks in
the presence of an interfering signal. In: ICC 2016 IEEE International Conference on
Communications, pp. 1–6. IEEE (2016)
7. Chai, B., Deng, R., Cheng, P., et al.: Energy-efficient power allocation in cognitive sensor
networks: a game theoretic approach. In: Global Communications Conference, pp. 416–421.
IEEE (2012)
8. Zhang, D., Chen, Z., Awad, M.K., et al.: Utility-optimal resource management and
allocation algorithm for energy harvesting cognitive radio sensor networks. IEEE J. Sel.
Areas Commun. 34(12), 3552–3565 (2016)
9. Park, S., Heo, J., Kim, B., et al.: Optimal mode selection for cognitive radio sensor networks
with RF energy harvesting. In: IEEE, International Symposium on Personal Indoor and
Mobile Radio Communications, pp. 2155–2159. IEEE (2012)
10. Wei, Y., Ren, C., Song, M., Yu, R.: The offloading model for green base stations in hybrid
energy networks with multiple objectives. Int. J. Commun Syst. 29(11), 1805–1816 (2016)
11. Wei, Y., Wang, X., Fialho, L., Bruschi, R., Ormond, O., Collier, M.: Hierarchical power
management architecture and optimal local control policy for energy efficient networks.
J. Commun. Netw. 18(4), 540–550 (2016)
12. Lasaulce, S., Debbah, M., Altman, E.: Methodologies for analyzing equilibria in wireless
games. Signal Proc. Mag. IEEE 26(5), 41–52 (2009)
Traffic Paralysis Alarm System Based on Strong
Associated Subnet
1 Introduction
With the rapid development of urbanization and the remarkable improvement of living
quality in China, both the population and the number of private cars has increased
significantly in cities. It is reported that by the end of 2013 the urban population of China
accounted for 53.73% of the total population, and urbanization is expected to reach more
than 60% by 2020. In addition, the number of private cars rose from 18.48 million in
2005 to 123.39 million in 2014, according to the data from the National Bureau of
Statistics of China.
Urbanization means the movement of a large part of the rural population to cities.
Urbanization improves the living standards of many people, promotes economic
recover
1 2
unimpeded jam paralysis
recover
spaƟal-temporal locaƟon, road network
traĸc incident A topology, traĸc flow
These three factors determine whether a traffic jam can escalate to traffic paralysis,
and they provide the possibility for us to predict traffic paralysis.
We propose a traffic paralysis alarm system based on a strong associated subnet
which consists of three modules, namely, data preprocessing, strong associated subnet
mining, and traffic anomaly detection. The proposed system mainly depends on the
strong association between the road segments in a strongly associated subnet.
332 C. Yu et al.
2 Related Work
The proposed system is mainly related to three issues: map-matching, association rule
mining, and traffic anomaly detection.
Irregularities in GPS positioning systems prevent the direct use of GPS data. Map-
matching algorithms correct the deviation of GPS data and matches GPS data to road
segments by which to render the actual trajectories of vehicles.
Typical geometric algorithms include point-point, point-line, and line-line algo‐
rithms. A point-point algorithm matches GPS data to the nearest point on surrounding
road segments. The accuracy of such algorithms is unpredictable.
In [4, 5], topological algorithms consider the connectivity of the road, adjacency
relationships, and other attributes (such as one-way roads) to match. Paper [5] gives a
C-measure based algorithm, which uses distance, angle, and connectivity to calculate
match reliability.
Advanced algorithms [6–9] generally use more complex computational models, have
greater accuracy, and can meet some special needs. The ST-Matching algorithm
proposed for GPS trajectories with a low-sampling-rate (e.g., one point every 2–5 min)
[6], takes into account the geometric and topological structures of road networks and
the temporal/speed constraints of the trajectories. Paper [7] proposes a Hidden Markov
Model-based algorithm which accounts for measurement noise and the structure of the
road network. In our paper, we use the algorithm proposed in [7] to correct GPS data.
Association rule is one of the most active and widely used knowledge types in the field
of data mining. Now there are many association rule mining algorithms.
Apriority is the first typical rule mining algorithm, and many other algorithms
based on the idea of apriority after that. Paper [10] points out that the early itera‐
tions of Apriority-based algorithms is the dominating factor for the overall mining
performance, and proposes an effective hash-based algorithm that can generate
smaller candidates. FP-Growth in [11] does not generate candidates and only
requires two scans of transaction database. It compresses the transaction informa‐
tion to the FP-tree in which the support degree of item-sets is in descending order.
There are many algorithms for different scenarios on association rule mining in the
years of research and applications. For example, paper [12] presents an algorithm to
mine rules in databases which are partitioned among many computers that are distributed
over a wide area. Paper [13] summarizes the existing algorithms that address the issue
of mining association rules from data streams.
Traffic Paralysis Alarm System Based on Strong Associated Subnet 333
In this section, we will present the architecture and design of the proposed system. The
proposed system consists of three modules shown in Fig. 2: A. pre-processing module;
B. strong associated subnet mining module; C. traffic anomaly detection module.
Map-Matching
Digital Map Database
The module A mainly pre-processes digital map data, large text, and GPS data. First,
a digital map database is constructed by importing digital map data into a spatial database
and splitting each road into road segments. Second, the collected data must be
334 C. Yu et al.
preprocessed because every text file of stored one-day taxi GPS data of Zhuhai city is
bigger than 11 GB, and they are unordered and redundant. Therefore, we need to remove
redundancy and sort by taxi ID and timestamp. Finally, deploying the map-matching
server and importing GPS point data and corrected GPS point data into spatial database
as GPS point trajectory database.
The main purpose of module B is to mine strong associated subnet. First, road
segment trajectory database is constructed by the ways of transferring GPS point trajec‐
tories as {p1, p2, …, pn} to road segments trajectories as {r1, r2, …, rm}, pi represents a
GPS point and ri represents a road segments. Then, the most frequent passed is selected
as seeds through top-k road segments and these seeds are expanded into strong associated
subnets.
Module C is responsible for detecting traffic anomalies occurred at strong associated
subnets. The travel speed of road segments in strong associated subnets is monitored
and checked: if there occurs an extended low-speed situation, the proposed system will
regard this phenomenon as a traffic anomaly and send out a signal.
This module, combined with digital map database, expands the most frequent top-k road
segments into strong associated subnets. As summarized by the following steps.
Step 1: Load a corrected GPS point trajectory as pointTr = {p1, p2, …, pn}, for every
point pi of this trajectory, determine the road segment ri that point pi located at,
then we can gain the corresponding road segments trajectory of pointTr as
roadsegTr = {r1, r2, …, rm}. At the same time, update the number of trajectory
of ri.
Traffic Paralysis Alarm System Based on Strong Associated Subnet 335
In this section, we will introduce two key algorithms of traffic paralysis alarm system
based on strong associated subnet: strong associated subnet mining algorithm and traffic
anomaly detection algorithm.
We firstly define six operators used by strong associated subnet mining algorithm.
Preliminaries
Definition 1 (cntRoadseg): cntRoadseg(rx) is the number of the road segment trajectories
that contain road segment rx.
Definition 2 (cntShared): cntShared(rx, ry) represents the number of road segment
trajectories that contain both road segment rx and ry. It is the quantized form of the
association between road segment rx and ry. Under the same conditions, the bigger the
score of this operator, the stronger the association between road segment rx and ry.
Definition 3 (cntU): cntU(rx, ry) represents the sum of these road segment trajectories
that contain any one of road segment rx and ry.
Definition 4 (support): this operator is defined as the following:
( ) ( ) ( )
support rx , ry = cntShared rx , ry ∕cntU rx , ry (1)
336 C. Yu et al.
In the process of mining strong associated subnet, we check the possibility of road
segment ry which is not in a strong associated subnet becoming a part of strong associated
subnet. If road segment rx is adjacent to ry and is a part of strong associated subnet,
support(rx, ry) is the first we check, it must satisfy the following:
( )
support rx , ry ≥ 𝜆 (2)
Definition 6 (cluster): The cluster degree of subnet V is the average of the correlation
among these adjacent road segments in subnet V. This operator is defined as the
following:
Algorithm 1: cluster(r, V)
Output: the cluster deg ree of subnet V after r added
into
1: clus= 0.0;
2: for rx in V loop
3: listRx=loadAdjacent(rx, V);
4: for ry in listRx loop
5: clus+=correlation(rx,ry);
6: end loop;
7: end loop;
8: clus /= 2;
9: listR=loadAdjacent(r, V);
10: for re in listR loop
11: clus +=correlation(r, re);
12: end loop;
13: return clus /(V.size + 1);
cluster(r, V) ≥ 𝜇 (4)
where μ is the predefined threshold and it will be analyzed in Sect. 5.1.
Traffic Paralysis Alarm System Based on Strong Associated Subnet 337
Details of Algorithm
Strong associated subnet mining algorithm mainly consists of two steps. First, selecting
the most frequent top-k road segments as strong associated subnet seeds. Second,
checking if there exists a road segment in strong associated subnets that adjacent to one
or more road segments which are not in strong associated subnets and meet the thresholds
for expansion: if there exists one, we call this one as scalable road segment, and add
those meeting thresholds into strong associated subnets. This process is iterative until
there does not exist one meeting the requirements. More precisely, the algorithm process
is the following:
In this section, we use the flow chart to present our successive low-speed based traffic
anomaly detection algorithm. The proposed algorithm takes 30 s as a calculation cycle
and computes average travel speed for every road segment in strong associated subnets
in every cycle. This algorithm runs forever once started. It iteratively computes travel
time and checks if there occurs a traffic anomaly.
338 C. Yu et al.
In this section, we first analyze two parameters, λ and μ, of strong associated subnet
mining algorithm. Then, a simulation experiment based on VISSIM is carried out for
explaining the theory of traffic paralysis alarm system based on strong associated subnet
(Fig. 3).
START
Y
is there sƟll a
trajectory that pass
Y
through roadseg in tseg
i<4
N N
Next iteratoraƟon
In the system testing process, we find that the scores of these operators, support and
correlation, are low in most of the time, and the low score of operator correlation directly
causes the score of operator cluster to be also low. Therefore, we perform an analysis
based on the GPS data of Zhuhai, and get the conclusion that the smaller the sampling
rate is, the smaller these operators’ scores are.
The original GPS data of Zhuhai is sampled every 10 s. We manually dilute the
original data and get another three datasets with sampling intervals of 20 s, 40 s, and
80 s. For each of the most frequent top-200 road segments, calculating separately average
score of the operator support based on four datasets. The statistical distribution of the
Traffic Paralysis Alarm System Based on Strong Associated Subnet 339
average score of operator support is shown in Fig. 4. Similarly, the statistical distribution
of the average score of operator correlation is shown in Fig. 5.
It is easy to observe from the above two figures that the sampling rate is higher, the
distribution peaks are at further to the right positions.
The idea of our proposed system is that when a traffic anomaly is detected on a strong
associated subnet, alarming immediately for this subnet to reduce the traffic entering
this subnet, to avoid the transformation from traffic jam to traffic paralysis. We want to
discover whether reducing the traffic entering this subnet helps avoid the transformation.
We perform an experiment with VISSIM and the result can be concluded as: when a
traffic anomaly occurs on a strong associated subnet, reducing the traffic entering this
subnet can decrease the influence of the traffic anomaly.
We simulate a traffic incident occurring at the 300th s and resolved at the 1500th s
and collected a delay time dataset and a queue length dataset under different traffic
volumes shown in Figs. 6 and 7.
Fig. 6. Delay time under different traffic Fig. 7. Queue length under different traffic
volume volume
340 C. Yu et al.
It is obvious that the delay time and queue length increase remarkably as the traffic
volume rises. The delay time and queue length clearly show the traffic anomaly’s influ‐
ence. Therefore, we can get the conclusion declared above: reducing the traffic entering
this subnet can decrease the influence of the traffic anomaly.
In this paper, we propose a traffic paralysis alarm system based on strong associated
subnet. The system mainly designs and implements strong associated subnet mining
algorithm and traffic anomaly detection algorithm based on the taxicabs GPS data of
Zhuhai. The former is used to mining strong associated subnets with trajectories on road
network, and the latter calculates travel speed of interesting road segments periodically
and checks if there exists a successive low speed situation on interesting road segments:
if there exists one, the algorithm sends out a signal.
Two experiments are carried out in this paper. One examines the case of low score
of operators and concludes that the smaller the sampling rate is, the smaller these oper‐
ators’ scores are. The second experiment uses VISSIM to simulate traffic situations when
a traffic incident occurred. Under different traffic volume, we analyze delay time and
queue length, concluding that reducing the traffic entering this subnet can decrease the
influence of the traffic anomaly.
In future work, the proposed system should be improved to solve real-time GPS data
stream, how we store, pre-process, and analyze the data stream may need further
research.
Acknowledgements. The work is partly supported by NSFC (No. 61472149), the Fundamental
Research Funds for the Central Universities (2015QN67) and the Wuhan Youth Science and
Technology Plan (2016070204010132).
References
1. Pan, B., Zheng, Y., Wilkie, D., Shahabi, C.: Crowd sensing of traffic anomalies based on
human mobility and social media. In: Proceedings of the 21st ACM International Conference
on Advances in Geographic Information Systems, pp. 334–343 (2013)
2. Chen, S., Wang, W., Zuylen, H.V.: Construct support vector machine ensemble to detect
traffic incident. Expert Syst. Appl. 36(8), 10976–10986 (2009)
3. Tang, S., Gao, H.: Traffic-incident detection-algorithm based on nonparametric regression.
IEEE Trans. Intell. Transp. Syst. 6(1), 38–42 (2005)
4. Velaga, N.R., Quddus, M.A., Bristow, A.L.: Developing an enhanced weight-based
topological map-matching algorithm for intelligent transport systems. Transp. Res. Part C
Emerg. Technol. 17(6), 672–683 (2009)
5. Xu, Y., Wang, Z.: Improvement and implement of map matching algorithm based on C-
measure. In: Proceedings of the 2010 2nd IEEE International Conference on Information
Management and Engineering, pp. 284–287 (2010)
Traffic Paralysis Alarm System Based on Strong Associated Subnet 341
6. Lou, Y., Zhang, C., Zheng, Y., Xie, X., Wang, W.: Map-matching for low-sampling-rate GPS
trajectories. In: Proceedings of the 17th ACM International Symposium on Advances in
Geographic Information Systems, pp. 352–361 (2009)
7. Newson, P., Krumm, J.: Hidden Markov map matching through noise and sparseness. In:
Proceedings of the 17th ACM International Symposium on Advances in Geographic
Information Systems, pp. 336–343 (2009)
8. Yuan, J., Zheng, Y., Zhang, C., Xie, X., Sun, G.Z.: An interactive-voting based map matching
algorithm. In: Proceedings of the Eleventh International Conference on Mobile Data
Management, pp. 43–52 (2010)
9. Li, Y., Huang, Q., Kerber, M., Zhang, L., Guibas, L.: Large-scale joint map matching of GPS
traces. In: Proceedings of the 21st ACM International Conference on Advances in Geographic
Information Systems, pp. 214–223 (2013)
10. Park, J.S., Chen, M.S., Yu, P.S.: An effective hash-based algorithm for mining association
rules. In: Proceedings of the 1995 ACM International Conference on Management of Data,
pp. 175–186 (1995)
11. Han, J., Fu, Y.: Discovery of multiple-level association rules from large databases. In:
Proceedings of the 21th International Conference on Very Large Data Bases, pp. 420–431
(1995)
12. Wolff, R., Schuster, A.: Association rule mining in peer-to-peer systems. In: Proceedings of
the 3rd IEEE International Conference on Data Mining, pp. 363–370 (2003)
13. Nan, J., Le, G.: Research issues in data stream association rule mining. ACM SIGMOD
Record 35(1), 14–19 (2006)
Fault Recovery Algorithm Based on SDN Network
Yi Zhang1,2 ✉ , Yifei Wei1, Ruqin Huang2, Bo Gu1, Yue Ma1, and Mei Song1
( )
1
School of Electronic Engineering, Beijing University of Posts and Telecommunications,
Beijing 100876, People’s Republic of China
{zzyy,weiyifei,gubo,mayue,songm}@bupt.edu.cn
2
Institute for Network Sciences and Cyberspace, Tsinghua University, Beijing, China
15652964297@163.com
1 Introduction
Network virtualization is a new technology which can solve the problem of the internet
and support the development of new technology. Important advantages of this technique
is different network can share the underlying physical infrastructure. The virtual network
is above the physical layer. Virtual nodes and virtual links constitute a virtual network.
Network topology, the use of technology, the provision of services for different virtual
networks are different.
The purpose of network virtualization is to facilitate the configuration and manage‐
ment and to facilitate the construction of new network technology. Resource manage‐
ment has become the key to achieve the advantages of the network virtualization tech‐
nology. The resource management in the network virtualized environment should
rationally design the management structure and the resource scheduling algorithm to
realize the efficient sharing of the underlying physical network resources. In ensuring
the virtual network user resource request conditions, greatly improve the network
resource utilization.
The traditional network of network construction cannot meet the diverse needs of
different business. Then SDN network is proposed, control and forwarding separation,
to achieve the underlying hardware abstraction. SDN is a new network architecture
proposed by Stanford University and is the main application architecture of this paper.
With network virtualization technology, each experimental user can customize a virtual
network and construct a different topology for each virtual network. The underlying
physical network forms a network resource pool through node virtualization and link
virtualization and allocates resources for the virtual network to improve network
resource utilization.
Based on the exchange equipment, through the definition of a common standard
interface to achieve data plane and control plane separation, to facilitate the researchers
in the real network environment for technical research. The SDN forwarding plane is
separated from the control plane. The experimenter can access the underlying physical
device with a defined common protocol and programmatically control the forwarding
path of the data in the network. Network virtualization technology can divide the phys‐
ical network into multiple virtual networks. Each virtual network can run different
routing protocols and provide different services.
The actual situation of the physical network often causes the failure due to natural
causes or malicious attacks and other non-natural, affecting the normal operation of the
user’s business. And the previous algorithm is to enhance the physical network resource
utilization, there is no study of network failure. Aiming at the reliability problem of virtual
network, there is a remapping mechanism of virtual node and virtual link, which not only
can improve the receiving rate of virtual network but also benefit the load balance of
network. Migrating a failed virtual link to a backup path by pre-building the backup path
is one of the methods. Another way, building a unified backup resource pool which can
dynamically allocate backup resources for virtual links, and improve the utilization of
physical resources. These algorithms use different mechanisms to improve the reliability
of virtual networks. The cost of reserving backup resources in advance is too high in the
first method. In the second method, a new path is found when the fault occurs, resulting in
a low failure rate. In this paper, we propose a fault recovery algorithm based on the number
of users for the virtual network mapping, and compare the network performance with the
link remapping algorithm and the backup path construction algorithm.
Virtual network requests are mapped to the underlying physical network that is made
up of different devices. The management creates a virtual network mapping request
based on the needs of the user and sends the request to the physical layer. Finally, the
virtual network provider can provide the customized virtual network service to the user
(Table 1).
344 Y. Zhang et al.
∑
oij ≤ 1 ∀njs ∈ N′s (1d)
niv ∈Nv
Link mapping
Find all nodes that meet the requirements, select one from the node that satisfies the
requirement as an endpoint. Use the Dijkstra algorithm to find the shortest path to meet
the bandwidth. Finally, link mapping is implemented.
( ) ( ) ( )
Br lm
s
≥ R liv , ∀lm
s
∈ PhS,D PhS,D = M liv (2a)
( ) ( ) ∑ ( )
Br lm = B lm − R liv (2b)
s s m
ls ∈M(liv )
Fault Recovery Algorithm Based on SDN Network 345
( ) ( )
j
Nfwd
r
ns represent the remaining forwarding ctrl
( m ) resources of the node, Nr gk is the
remaining control resource for the area, Br ls A represents the remaining bandwidth
resources of the link, α, β( is )resource conversion weight between node CPU resource
and the link bandwidth. S Gs denote the physical resources available, which is in direct
proportion to the( number
) of virtual networks supported. Therefore, set the optimization
target to max{S Gs }.
In this paper, the ant colony algorithm is used to solve the virtual network mapping
problem in the software defined network model.
1. Initialization();
2. VN_Creation();
3. For i=1 to N;
4. { Update_probability();
5. For ant=1 to M
6. { Node_Map();
7. Link_Map();
8. if(S(local_solution)>S(global_solution))
9. global_solution=local_solution;
10. }
11. Update_info();
12. }
346 Y. Zhang et al.
According to the above algorithm, in each iteration process, select the network
resource surplus value
( )of the largest feasible solution as the current cycle optimal solu‐
tion according to S Gs . Then update the pheromone of the virtual node to the physical
node according to the optimal solution. After N iterations, the M feasible solutions
generated by the initialization can eventually converge to the approximate global
optimal solution.
j
The pheromone update to the physical node ns is as follows:
1
Δ= 𝜎 (5)
S(pbest )
( )
pbest is the optimal solution of this cycle, S pbest is the objective function value of
the optimal solution, 𝜎 is the influence factor of the optimal solution on pheromone.
This paper presents a new fault recovery mechanism based on the previous network
mapping model. The number of users is different during different period. So we can
implement the strategy according to the number of users. When the number of users is
low, there are more free link resources. We use the backup path construction algorithm
for most users to ensure the reliability of the network as far as possible. The remaining
users take the faulty link remapping algorithm when a failure occurs. When the number
of users is large, the free link resources are less. So leave backup path for a small number
of users. Most failed users are remapped. Most failed users should take the faulty link
remapping algorithm. For each time slice, we will update the strategy based on the
number of users. Change the proportion of pre-backup users to normal users. In this
paper, we believe that the failure process is subject to uniform distribution.
The free link resources vary with the number of users. A fault recovery algorithm
based on the number of users is proposed, that is, the traffic of the pre-backup users
affected by the fault is migrated to the backup path. And the common user’s faulty link
is remapped. The algorithm can improve the fault repair rate and network resource
utilization rate.
Mapping rules
Node mapping
( j) ( )
Nfwd
r
ns ≥ Rfwd niv oij = 1 (6a)
( ) ( )
Nctrl
r
gk ≥ Rctrl niv oij = 1, njs ∈ gk (6b)
Fault Recovery Algorithm Based on SDN Network 347
∑
′ oij = 1 ∀niv ∈ Nv (6c)
ns ∈Ns
j
∑ ′
oij ≤ 1 ∀njs ∈ Ns (6d)
niv ∈Nv
Link mapping
For low level users, just meet
( ) ( ) ( )
Br lm
s
≥ R liv , ∀lm
s
∈ PhS,D PhS,D = M liv (7a)
( ) ( ) ∑ ( )
Br lm = B lm − R liv (7b)
s s m
ls ∈M(liv )
For pre-backup users, link mapping is not only required to meet bandwidth require‐
ments, but also to ensure that there is no overlap with the backup link.
( ) ( )
Br lm
s
≥ R liv , ∀lns ∈ Ph′SD (8a)
The res-backup algorithm is to find all the backup paths in advance and enable the backup
path when the network fails. If the backup path is available, the backup path is returned.
If the backup path is not available, the virtual network recovery fails. This algorithm
has a high recovery rate, but the backup path takes up too much resource, resulting in
reduced network utilization and degraded network performance.
348 Y. Zhang et al.
( )
S and D represent the two endpoints of the link; Wb lSD indicates the bandwidth
occupied by the backup resource on the link; Gl is represented as a collection of backup
links. The algorithm is described as follows:
( )
Input: lSD, Wb lSD , Gl
Output: PSD
(1) Remove the faulty link in Gs;
(2) Determine if the backup path is available. If it is not available, set PSD to NULL
and skip to 4);
( ) ( ) ( )
(3) Br lms
= Br lm
s
− Wb lSD , Update the remaining bandwidth resources of the link;
(4) If PSD is NULL, the virtual network recovery failed. Otherwise, return the recovery
path;
The ResRemap algorithm is a timely repair when the link fails. When the network fails,
first remove the fault link from the set, calculate whether to find the link which has the
same endpoint with the faulty link. If it can be found, update the resource, restore the
path, and if it was not found, the virtual network recovery fails.
( )
Input: lSD, Wb lSD , Gl
Output: PSD
(1) Remove the faulty link in Gs;
(2) Using the shortest path algorithm to find an alternative path PSD. Alternative path
and fault path have the same endpoint. If it is not found, set PSD to NULL and skip
to 4);
( ) ( ) ( )
(3) Br lms
= Br lm
s
− Wb lSD , Update the remaining bandwidth resources of the link;
(4) If PSD is NULL, the virtual network recovery failed. Otherwise, return the recovery
path;
5 Simulation Parameters
In this paper, we use MATLAB to generate the underlying physical network and virtual
network request topology. 25 nodes are generated in the space of 200 * 200, and some
nodes are generated according to the random function, which are connected to each other
by 0.5 probability (Table 2).
Fault Recovery Algorithm Based on SDN Network 349
200
180
160
140
120
100
80
60
40
20
0
0 20 40 60 80 100 120 140 160 180 200
Nacreq
Rac = (9)
Nalreq
Nsf is the number of failures that were successfully repaired. Nalf is the number of
all failure. Rfr is the fault recovery rate
Nsf
Rfr = (10)
Nalf
Nnorm
𝜂lr = (11)
Ntotal
6 Simulation Results
𝜆f
When γ = 0.1(γ = ), the request strength is not high, because the faulty link remap‐
ping algorithm does𝜆notr have a backup link, so there are more free network resources,
the virtual network has the highest success rate. The backup link of the ResBackup
algorithm takes up many link resources, so the network running the lowest success rate.
The NumMap algorithm backs up some users’ links and takes up part of the additional
link resources, so the Rac is between the two algorithms.
The user’s requests get more when γ = 0.5. The success rate of ResRemap algorithm
is still the highest. But the Rac decreased from 67% to 55%. And The ResBackup algo‐
rithm back up link in advance, taking up a lot of link resources, so the the change of
running success rate is not obvious for virtual network. The Rac of the virtual network
operation of the fault recovery algorithm is slightly lower, but still between the two
algorithms (Fig. 2).
Fault Recovery Algorithm Based on SDN Network 351
1 1
ResRemap ResRemap
Nummap Nummap
0.9 ResBackup 0.9 ResBackup
0.8 0.8
Acceptance ratio(γ=0.5)
Acceptance ratio(γ=0.1)
0.7 0.7
0.6 0.6
0.5 0.5
0.4 0.4
0 5 10 15 20 25 30 35 40 0 5 10 15 20 25 30 35 40
Time Time
1 1
0.95
0.9
0.9
Fault recovery rate(γ=0.1)
0.8
0.85
0.8 0.7
0.75
0.6
0.7 ResBackup ResBackup
Nummap(high-level) Nummap(high-level)
0.5
0.65 ResRemap ResRemap
Nummap(low-level) Nummap(low-level)
0.6 0.4
0 5 10 15 20 25 30 35 40 0 5 10 15 20 25 30 35 40
Time Time
The number of users is large when γ = 0.1.Because it was when the link malfunctioned
that ResRemap started to repair, there is no backup link to seize resources, so 𝜂lr is the
highest, reaching 50%. ResBackup takes up a lot of link resources, which backup link,
352 Y. Zhang et al.
so 𝜂lr is the lowest. Fault recovery algorithm reached 45%, based on user number,
backuped part of the user’s link according to the number of users, which link resource
rate is between the two algorithms above. In the case of γ = 0.1, link resource utilization
declined (Fig. 4).
0.5 0.5
Link resource utilization(γ=0.5)
0.3 0.3
0.2 0.2
ResRemap ResRemap
0.1 0.1
Nummap Nummap
ResBackup ResBackup
0 0
0 5 10 15 20 25 30 35 40 0 5 10 15 20 25 30 35 40
Time Time
7 Conclusion
The virtualized network represented by SDN network can meet the diversified business
needs of different users, support multiple routing protocols, protect the security of user
information, and promote the evolution of traditional Internet architecture to the next
generation architecture. In this research, based on the premise of single link failure,
which is the most prone to network in the network, based on Network reliability require‐
ments diversity from users, we designed a fault recovery algorithm based on number.
Finally, we proved the superiority of the algorithm from the aspects of virtual network
fault repair rate, successful operation rate of virtual network and utilization of work link
resource.
Acknowledgements. The authors would like to thank the reviewers for their detailed reviews
and constructive comments, which have helped improve the quality of this paper. This work is
supported by the National Natural Science Foundation of China under (Grant No. 61471055).
References
1. Karakus, M., Durresi, A.: A survey: control plane scalability issues and approaches in
Software-Defined Networking (SDN). Comput. Netw. 112, 279–293 (2017)
2. Bardgett, J.A., Miguez, M.A., Wierzbowski, P.: Network management system generating
virtual network map and related methods. US2015043378, 12 February 2015
3. Tootoonchian, A., Ganjali, Y.: HyperFlow: a distributed control plane for OpenFlow. In:
Proceedings of the USENIX INM Workshop on WREN (2010)
4. Luo, Y., Chen, X., Wang, J.: A virtual network embedding algorithm for providing stronger
connectivity in the residual networks. J. Netw. 8(4), 779–786 (2013)
Fault Recovery Algorithm Based on SDN Network 353
5. Yeganeh, S.H., Ganjali, Y.: Kandoo: a framework for efficient and scalable offloading of
control applications. In: Proceedings of the ACM SIGCOMM Workshop on HotSDN, pp.
19–24 (2012)
6. Chai, R., Dong, X.Y., Ma, J., et al.: An optimal IASA load balancing scheme in heterogeneous
wireless networks. In: 6th International ICST Conference on Communications and
Networking in China, pp. 714–719 (2011)
7. Huang, T., Liu, J., Chen, J., Liu, Y.: A topology-cognitive algorithm framework for virtual
network embedding problem. China Commun. 4, 73–84 (2014)
8. Mijumbi, R., Serrat, J., Rubio-Loyola, J., et al.: Dynamic resource management in SDN-based
virtualized networks. In: 10th International Conference on Network and Service Management
(CNSM), 17–21 November 2014, Rio de Janeiro, Brazil, 1st International Workshop on
Management of SDN and NFV Systems (ManSDN/NFV), 21 November 2014, Rio de Janeiro,
Brazil, pp. 412–417 (2014)
9. Bays, L.R., Oliveira, R.R., Buriol, L.S., Barcellos, M., Gaspary, L.P.: A toolset for efficient
privacy-oriented virtual network embedding and its instantiation on SDN/OpenFlow-based
substrates. Comput. Commun. 82, 13–27 (2016)
10. Khan, A., An, X., Iwashina, S.: Virtual network embedding for telco-grade network protection
and service availability. Comput. Commun. 84, 25–38 (2016)
11. Li, Y., Zhao, Y., Zhang, J., Yu, X., Chen, H., Zhu, R., Zhou, Q., Yu, C., Cui, R.: First field
trial of Virtual Network Operator oriented network on demand (NoD) service provisioning
over software defined multi-vendor OTN networks. Optical Fiber Technol. 33, 22–29 (2017)
12. Xing, C., Lan, J., Hu, Y.: Virtual network with security guarantee embedding algorithms. J.
Comput. 8(11), 2782–2788 (2013)
13. Arakawa, S.I., Minami, Y., Koizumi, Y., Miyamura, T., Shiomoto, K., Murata, M.: A
managed self-organization of virtual network topology controls in WDM-based optical
networks. J. Optical Commun. 32(4), 233–242 (2011)
14. Sun, G., Yu, H., Li, L., Anand, V., Di, H.: The framework and algorithms for the survivable
mapping of virtual network onto a substrate network. IETE Tech. Rev. 28(5), 381–391 (2011)
15. Heisswolf, J., Zaib, A., Weichslgartner, A., König, R., Wild, T., Teich, J., Herkersdorf, A.,
Becker, J.: Virtual networks – distributed communication resource management. ACM Trans.
Reconfigurable Technol. Syst. (TRETS) 6(2), 8 (2013)
Sensor Location Verification Scheme Based on Range-Free
Localizations in WSNs
1
College of Xingzhi, Zhejiang Normal University, Jinhua, China
netmcy@zjnu.cn
2
College of Computer, Zhejiang University of Technology, Hangzhou, China
1 Introduction
Wireless sensor networks (WSNs), which consist of a large number of sensor nodes,
have been widely used in military and human daily life, e.g. surveillance, environmental
monitoring system and medical health [1]. Localization is one of the most essential
research issues in WSNs because the sensed information without location is insignifi‐
cant, in some scenarios such as environment monitoring, target tracking, and geograph‐
ical routing [2]. To acquire the locations of sensor nodes, we can either mount nodes
with GPS receiver or predefine nodes’ positions manually in deployment. Because of
relatively high price and energy-extensive consumption, GPS receivers may not be
available for power-limited and low-price WSN, and the second method is not available
for large scale WSN. As a result, we always predefine a small part of nodes’ locations
manually in deployment, which are called beacon nodes. And other nodes are called
unknown nodes. Take the location of beacons as reference, the normal sensor nodes can
estimate their locations using some certain localization algorithms. These localization
algorithms can be divided into Range-Based localization algorithms and Range-Free
localization algorithms [3]. The former assumes the distances between sensors and
beacons can be estimated by using different measurements, such as TDoA, ToA, AoA
and RSSI [4–7]. These algorithms can usually provide higher positioning accuracy with
higher hardware cost. While the latter estimates location of normal nodes based on the
features of networks, such as hop counts [8], topology of network [9], etc.
In the study of localization of static WSN, we usually assume that beacons’ locations
are reliable. Nevertheless, for many WSN systems deployed in unstable environments,
all nodes may be moved unexpectedly and Beacons have hardware failures or be
captured to provide false locations. The drifting beacons are called unreliable beacons.
The locations of drifting normal nodes can be re-localized by re-invoking localization
algorithms. But when unreliable beacons’ locations are used in re-localize process, it
will seriously degrade the accuracy of re-localization and affect the quality of network
service. So, recognizing and filtering out these unreliable beacons becomes one of the
most important research issues in node localization of WSNs.
To solve these problems mentioned above, we proposed a distributed and lightweight
beacons locations verification algorithm based on neighborhood-similarity, which use
the similarity of the beacons’ neighborhood in different time slot to verify which beacons
are drifting.
2 Related Works
At present, the existing study of reliable localization in WSNs is divided into robust
localization algorithms [10] and unreliable beacons filter algorithms [11]. The former
is applicable in some scenes, where there exist the interference of ranging information
or small movement of beacons. The main idea of these algorithms is reducing the envi‐
ronment interference of localization. The latter algorithms verify beacons locations to
recognize and filter out the unreliable beacons. As a result, unreliable beacons filter
algorithms are more universal and can be used in localization algorithms to ensure
localization accuracy.
In [12], it proposed a point to point location verification algorithm, which can be
applied to range-based algorithm. But the nodes are equipped with GPS receiver. DJ.
He et al. proposed a location verification algorithm based on TOA to eliminate abnormal
range value, which can be applied to Range-Based localization algorithms [13]. Kuo
et al. proposed beacon movement detection (BMD) algorithm to detect the unexpected
movements of beacons [14]. The basic idea of BMD algorithm is to let the beacons
record the variances of RSSI measurement results between each other and report to a
calculation center to determine the moved beacons. Like all other centralized algorithms,
BMD algorithms will bring a heavy communication burden and need a sink node or a
computer with a strong ability to calculate, which is not fit for WSN. There are also
some related works, which use the location of the hidden checked nodes to verify the
locations of beacons, which are also a centralized algorithm and needs involvement of
additional nodes [15]. In [16], by using rigid theory to exclude the abnormal location of
beacons, it can provide reliable localization results. However, the rigid theory relies on
high accuracy of range and the algorithm’s computation is too heavy. Ravi Garg et al.
eliminated the beacons, which provide a larger descent gradient during the localization,
to improve the credibility of localization. The algorithm did not consider the reference
356 C. Miao et al.
of the Locations of normal nodes so that it can use in beacon-sparse networks [17].
Yawen Wei et al. presented a location verification probabilistic model based on the
mutual observation between neighbor nodes, which achieved a better result [18]. In [19],
it developed a distributed neighbor nodes score mechanism based on variation of RSSI
to identify the moved beacons. However, all of above methods can’t be suitable for
Range-Free algorithms and solve single problem.
3 Problem Statement
We assume that there are two types of nodes deployed in the network: beacons and
sensors. The beacons know their locations in advance. The sensors do not know their
locations, which are also called unknown nodes. But they can estimate locations by using
localization algorithms with beacons’ locations. All nodes are mounted with RSSI
transceiver. All nodes can be moved unexpectedly. Since drifting unknown nodes are
always existed, we need to re-localized the network.
Besides, we assume all nodes’ communication ranges have the same radius. Notice
that environmental interruptions and permutations exist, so that neighborhood observa‐
tion is not always symmetric. Our algorithm also can solve such asymmetry problem.
In the initialized localization, we assume that beacons are reliable. And each node can
participate in the algorithm calculation. During the calculation, nodes are static. In the
most cases, the drifting nodes are 10%–20% of the total nodes. Unreliable beacons’
percentage is not over that 50% of beacons [20]. For conveniently introducing algorithm,
our notations are introduced in Table 1.
Table 1. Notations
Si Node i
Bi Beacon i
ID(i)t At t time, the set of node i’s neighbors
Nei(i, j)t At t time,the common set of node i’s and node j’s neighbors
DisRank(i, j)t At t time, the actual distance coefficient between node i and node j
EstDisRank(i, j)t At t time, the estimated distance coefficient between node i and node j
EstDis(i, j)t At t time, the estimated distance between node i and node j
IDSame(i)(t, t+1) In the (t, t + 1), the change of ID(i)t and ID(i)t+1
neiv(i)t At t time, a vector to store DisRank(i, j)t , j ∈ ID(i)t
Relation(i, j)t At t time, the relationship of between node i and node j
nodes, beacons are called drifting beacons. As an example, Fig. 1 shows a scenario where
beacon B2 is moved, whose broadcast location doesn’t match its actual location. So
during re-localization, if re-invoking localization algorithms utilize B2’s location, it can
degrade the accuracy of re-localization.
S3(t2)
B1
S2 B4 S3(t1)
B2(t1)
B5 B7 B6
Beacon
S1 S4 Sensor
Drifted Beacon
B2(t2) B7'(t2) Malicious Beacon
Min||Ad − A′d ||
In this section, we describe the DV-Hop localization algorithm and how to correct hop
count by using RSSI technology.
Niculescu et al. [21] proposed the DV-Hop localization algorithm by utilizing distance
vector routing mechanism. It has three phases [22]:
In the first phase, the network employ a typical distance vector routing mechanism.
Beacons flood their locations throughout the network with the initial hop-count of 0.
Each node that relays the message increase the hop-count by one. After the flooding
procedure, every node can obtain the minimum hop-count to each other.
In the second phase, after obtaining the locations and hop-count information to all
other beacons, each beacon estimates the average distance per hop. Bi calculates the
358 C. Miao et al.
average distance per hop, called Hi, using the formula (1). Where (xi, yi) is the coordinate
of Bi. hij is the hop-count from Bi to Bj. Then, Hi will also be flooded to sensors near to Bi.
∑ √
i≠j (xi − xj )2 + (yi − yj )2
Hi = ∑ (1)
i≠j hij
In the last phase, before conducting the self-localization, each sensor estimates the
distance to each beacon based on its hop-count and the H to this beacon. After obtaining
all the distance information, each sensor conducts the triangulation or maximum like‐
lihood estimation to estimate its own location.
In this section, we utilize the RSSI technology to correct the hop count between nodes,
which reflects the distance proximity between nodes. In the recent study of DV-Hop
localization algorithm, many papers utilize the RSSI technology to correct the hop count
between nodes [22, 23]. In this paper, we use this idea to correct the hop count to help
nodes get the distance proximity with their neighbors.
The coverage area of WSNs is large and the environments of different regions are
different, but in some scenarios, the environment is relatively stable, such as the range
of node and its neighbors. Based on this assumption, we can utilize the RSSI technology
to correct the hop count.
At first, each node transmits the information of RSSI to its neighbors. So each node
gets the RSSI values of its neighbors. Secondly, each node normalize the RSSI values,
using the formula (2). Rssi(i, j)t represents the RSSI value between Si and Sj, and
( )
min Rssi(i, m)t , m ∈ ID(i)t is the minimum of these RSSI values. For RSSI value is form
–95 dbm to –55 dbm, the rssi is a constant and set as –50 dbm. It is convenient for the
proposed algorithms to calculate similarity. At last, each node gets the distance coeffi‐
cients with its neighbors. Besides, if Si can’t communicate with Sj, DisRank(i, j)t = 0.
5 BLVNS Algorithm
B2(t1) S5
S4(t0)
B2(t0)
S4(t1) S6(t0)
S3(t0)
S3(t1)
S1
B1
S6(t1)
S2
In the (t, t + 1), At time t, Si gets ID(i)t and DisRank(i, j)t , j ∈ ID(i)t. At time t + 1, Si
gets ID(i)t+1 and DisRank(i, j)t+1 , j ∈ ID(i)t+1. IDSame(i)(t,t+1) is computed by the formula
(3). Neighborhood relationship of Si and Sj’s at time t is defined as Relation(i, j)t. Rela‐
tion(i, j)t and Relation(i, j)t+1 are computed by
⎧ DisRank(i, j) ⋅ IDSame(j) ( )
⎪ t ( (t,t+1) ) j ∈ (ID(i)t ∩ ID(i)t+1)
Relation(i, j)t = ⎨ DisRank(i, j)t ⋅ 1 − IDSame(i)(t,t+1) j ∈ (ID(i)t ∕ID(i)t+1 ) (4)
⎪0 j ∈ ID(i)t+1 ∕ID(i)t
⎩
360 C. Miao et al.
{
DisRank(i, j)t ⋅ IDSame(j)(t,t+1) j ∈ (ID(i)t+1 )
Relation(i, j)t+1 = (5)
0 j ∈ ID(i)t ∕ID(i)t+1
In the formula (4) and (5), to minimize the influences of drifting neighbors,
DisRank(i, j)t is multiplied by IDSame(j)(t,t+1). When Si is static but Sj is moved, the
change of DisRank(i, j)t and DisRank(i, j)t+1 is caused by Sj. Because of Sj’s movement,
IDSame(j)(t,t+1) is closed to 0, which makes Relation(i, j)t and Relation(i, j)t close to 0.
( )
Besides, for j ∈ ID(i)t ∕ID(i)t+1 , at time t + 1, Sj is not Si’s neighbor, so Si can’t get
Sj’s IDSame(j)(t,t+1). IDSame(j)(t,t+1) is replaced by 1 − IDSame(i)(t,t+1). If Si is moved,
1 − IDSame(i)(t,t+1) is closed to 1. If not, 1 − IDSame(i)(t,t+1) is closed to 0. These opera‐
tions can minimize the influences of drifting neighbors.
After Si calculates Relation(i, j)t and Relation(i, j)t+1 of each neighbor, We define
vectors RelationV(i)t and RelationV(i)t+1 to store these data. Finally, we use cosine law
to calculate the similarity of RelationV(i)t and RelationV(i)t+1 by
RelationV(i)t ⋅ RelationV(i)t+1
NeiSim(i) = (6)
‖RelationV(i)t ‖ ⋅ ‖RelationV(i)t+1 ‖
‖ ‖ ‖ ‖
In this section, firstly, the simulation results are presented to validate the performance
and robustness of our proposed algorithms. Then, the algorithms are applied in the
dynamic WSN to improve the accuracy of re-localization.
BLVNS is based on neighborhood mutual observation, so the average connectivity
degree of network is more than 15. The network configuration of our simulation is set
as follows: 150 nodes, including 15 beacons and 135 sensors, are deployed randomly in
a 150 m × 150 m region. The transmission range of each nodes equals to 30 m. We use
the signal attenuation model to simulate the RSSI value between nodes, by using the
formula (7). Where pkl is the path dissipation function mattered with nodes’ distance.
dkl is the distance between sender Sk and receiver Sl. d0 is a reference distance and equals
to 1 m. np is an exponent of path loss. ε is an error coefficient.
( ( ))
d
pkl = p0 − 10np lg kl ⋅ (1 − 𝜀) (7)
d0
We use success detection rate Rs and error rate Re to evaluate the detection perform‐
ance. The calculation of Rs and Re is given in (8) and (9), in which Bu is the set of
unreliable beacons, Bdu is the set of unreliable beacons detected by algorithms.
Num(Bu ∩ Bdu )
Rs = (8)
Num(Bu )
Sensor Location Verification Scheme 361
( )
Num (B − Bu ) ∩ Bdu
Re = (9)
Num(Bu )
The goal to this experiment is to study how the number of drifting nodes impact the
performance and analyze the environmental adaptability of the verification algorithms.
At first, the measurement of RSSI does not exist any error, 𝜀 = 0. And the thresholds
of BLANS and BLATM are 0.6 and 0.45 respectively. Figure 3 shows the performance
of the verification algorithms with the drifting beacons. For example, when exponent of
path loss is 1.5, 2, and 2.5, respectively, Rs of two algorithms is maintained at about 95%
and Re is under the 20% with the drifting nodes increasing. So the influences of drifting
sensors are small. When the measurements of RSSI exist errors, we set 𝜀 = 0,
𝜀 ∈ (−0.05, 0.05) and 𝜀 ∈ (−0.1, 0.1) respectively. Figure 4 shows the performances of
the verification algorithms with measuring errors. Compare Fig. 4 with Fig. 3, we can
find their performances are similar so that the influences of RSSI measuring errors are
small. The experiment shows that BLVNS is robust.
0.9
0.8
0.7
np=1.5 Rs
0.6 np=2 Rs
BLVNS Rs/Re
np=2.5 Rs
0.5 np=1.5 Re
np=2 Re
0.4
np=2.5 Re
0.3
0.2
0.1
0
5 10 15 20 25 30
Number of drifting nodes
0.9
0.8
0.7
=0 Rs
0.6 (-0.05,0.05) Rs
BLVNS Rs/Re
(-0.1,0.1) Rs
0.5 =0 Re
(-0.05,0.05) Re
0.4 (-0.1,0.1) Re
0.3
0.2
0.1
0
5 10 15 20 25 30
Number of drifting nodes
Fig. 4. Performance of BLVNS with different drifting nodes and measuring errors
362 C. Miao et al.
7 Conclusion
In this paper, we analyze the severe impacts of the unreliable beacons on Range-Free
localization algorithms, which reflects the importance of beacon location verification.
To eliminate the influences of localization arise from these unreliable beacons, we
propose the algorithms BLVNS which can efficiently recognize and filter out the drifting
beacons. BLVNS can minimize the re-localization error and have strong anti-jamming
capacity in different environments and networks. Future study will extend the location
verification model to real-world experiments.
Acknowledgment. This work is supported by the National Natural Science Foundation of China
under Grant (No: 61502413; 61379023), and Supported by Zhejiang provincial Top Discipline
of Cyber Security at Zhejiang Normal University.
References
1. Akyildiz, F., Su, W., Sankarasubramaniam, Y., Cayirci, E.: Wireless sensor Networks: a
survey. Comput. Netw. 38(4), 393–422 (2002)
2. Jang, W.S., Lee, D.E., Choi, J.H.: Ad-hoc performance of wireless sensor network for large
scale civil and construction engineering applications. Autom. Constr. 26(10), 32–45 (2012)
3. Safa, H.: A novel localization algorithm for large scale wireless sensor networks. Comput.
Commun. 45, 32–46 (2014)
4. Gezici, S., Tian, Z., Giannakis, G.B., et al.: Localization via ultra-wideband radios: a look at
positioning aspects of future sensor networks. IEEE Sign. Process. Mag. 22(4), 70–84 (2005)
5. Patwari, N., Ash, J.N., Kyperountas, S., et al.: Locating the nodes: cooperative localization
in wireless sensor networks. IEEE Sign. Process. Mag. 22(4), 54–69 (2005)
6. Vempaty, A., Ozdemir, O., Agrawal, K., et al.: Localization in wireless sensor networks:
Byzantines and Mitigation techniques. IEEE Trans. Sign. Process. 61(6), 1495–1508 (2013)
7. Shao, H.J., Zhang, X.P., Wang, Z.: Efficient closed-form algorithms for AOA based self-
localization of sensor nodes using auxiliary variables. IEEE Trans. Sign. Process. 62(10),
2580–2594 (2014)
8. Wang, Y., Wang, X., Wang, D., et al.: Range-free localization using expected hop progress
in wireless sensor networks. IEEE Trans. Parallel Distrib. Syst. 20(10), 1540–1552 (2009)
9. Zhou, Y., Xia, S., Ding, S., et al.: An improved APIT node self-localization algorithm in
WSN based on triangle-center scan. J. Comput. Res. Dev. 46(4), 566–574 (2009)
10. Zhong, S., Jadliwala, M., Upadhyaya, S., et al.: Towards a theory of robust localization against
malicious beacon nodes. In: INFOCOM, pp. 2065–2073 (2008)
11. Hwang, J., He, T., Kim, Y.: Detecting phantom nodes in wireless sensor networks. In:
INFOCOM, pp. 2391–2395 (2007)
12. Liu, D., Lee, M.C., Wu, D.: A node-to-node location verification method. IEEE Trans. Ind.
Electron. 57(5), 1526–1537 (2010)
13. He, D., Cui, L., Huang, H., et al.: Design and verification of enhanced secure localization
scheme in wireless sensor networks. IEEE Trans. Parallel Distrib. Syst. 20(7), 1050–1058
(2009)
14. Kuo, S.P., Kuo, H.J., Tseng, Y.C.: The beacon movement detection problem in wireless
sensor networks for localization applications. IEEE Trans. Mob. Comput. 8(10), 1326–1338
(2009)
Sensor Location Verification Scheme 363
15. Yang, Z., Jian, L., Wu, C., et al.: Beyond triangle inequality: sifting noisy and outlier distance
measurements for localization. ACM Trans. Sens. Netw. (TOSN) 9(2), 26 (2013)
16. Yang, Z., Wu, C., Chen, T., et al.: Detecting outlier measurements based on graph rigidity
for wireless sensor network localization. IEEE Trans. Veh. Technol. 62(1), 374–383 (2013)
17. Garg, R., Varna, A.L., Wu, M.: An efficient gradient descent approach to secure localization
in resource constrained wireless sensor networks. IEEE Trans. Inf. Forensics Secur. 7(2),
717–730 (2012)
18. Wei, Y., Guan, Y.: Lightweight location verification algorithms for wireless sensor networks.
IEEE Trans. Parallel Distrib. Syst. 24(5), 938–950 (2013)
19. Xia, M., Sun, P., Wang, X., et al.: Distributed beacon drifting detection for localization in
unstable environments. Math. Prob. Eng. (2013)
20. Alfaro, J.G., Barbeau, M., Kranakis, E.: Secure localization of nodes in wireless sensor
networks with limited number of truth tellers. In: 2009 Seventh Annual Communication
Networks and Services Research Conference, CNSR 2009, pp. 86–93. IEEE (2009)
21. Niculescu, D., Nath, B.: DV based positioning in Ad Hoc networks. Netw. IEEE 22(1–4),
267–280 (2003)
22. Wu, J., Chen, H., Lou, W., et al.: Label-based DV-Hop localization against wormhole attacks
in wireless sensor networks. Pervasive Mob. Comput. 16, 22–35 (2014)
23. Guo, Z., Min, L., Li, H., Wu, W.: Improved DV-hop localization algorithm based on RSSI
value and hop correction. In: Wang, R., Xiao, F. (eds.) CWSN 2012. CCIS, vol. 334, pp. 97–
102. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-36252-1_10
Design and Development of Parametric System
for Planetary Reducer
1 Introduction
Most components of the mechanical product have certain characteristics and structure.
Applying the parametric design method, it is helpful to improve the design efficiency
and simplify the design work by using the established 3-D model library, and generate
3-D models of these components with same characteristics and structure only different
in size [1].
SolidWorks is a kind of powerful 3-D modeling software [2]. And it is able to meet
the specific needs of enterprises in the integration of the secondary development and the
parametric design. Variables are used to replace fixed parameters of components through
the parametric design in the physical modeling. The parametric design of similar compo‐
nents is completed through the modification of variables.
Enterprises and designers always pay attention to design, development and manu‐
facture of planetary reducers. Especially for the aspects of standardization, serialization
and generalization of planetary reducer components, it has become a general trend to
achieve parametric design of reducer components [3]. The rapid development of the
computer technology and CAD technology provides software with realizable conditions
for parametric design [4]. To design the planetary reducers, the mathematical model and
product structure are kept the same while the parameters in size should be modified [5].
Therefore, the use of digital technology to achieve parametric design of planetary
reducers helps to improve efficiency and correctness of the serial design.
ActiveX Automation
SolidWorks2013(Sever)
Help
Finite element
analysis
Automatic assembly
Parametric Design System
of Planetary Reducer General parts and
standard parts
Parametric
operation
Overall design
In Fig. 2, the parametric operations (3-D modeling and engineering drawing), auto‐
matic assembly, and finite element analysis in the component parametric design are key
modules of the digital design system.
The block diagram of the planetary reducer system is shown in Fig. 3.
Enter parameters of
overall design
Check of motion
Structural analysis
N intervention
Generate engineering
drawings
Users are able to select first and second transmission system through modifying
parameters of the gear to generate 3-D models and engineering drawings of new gears,
and the parametric design of components is achieved as shown in Fig. 5.
After completing the selection of the gear type, the remaining components are able
to be designed accordingly as shown in Fig. 6.
After modifying parameters, the new 3-D model and corresponding engineering
drawing will be automatically saved to a new folder, so that the product will be viewed
again and corresponding changes can be made by the user.
Start
Call the API object and assign a new value to the variable
Save
Mechanism of
parametric drive
Generate components
User
In general, the common steps for the parametric establishment of a 3-D model are
shown in Fig. 9.
Start
N
modify
Y
Develop parametric program
End
In this paper, 3-D models of main components of the two-stage NGW planetary
reducer – including input shaft, sun gear, planetary gear, ring gear, planetary carrier,
pin and output shaft are modeled.
Taking the output shaft of the reducer as an example, parametric modeling is
performed according to Fig. 9.
The 3-D model template of the output shaft is shown in Fig. 10.
The parametric model interface of the output shaft is set up with VB to develop
SolidWorks API function as shown in Fig. 11.
Design and Development of Parametric System 371
After modifying the parameters of the output shaft, according to the modification
process, new parameters are transferred to corresponding parameter names of the output
shaft. Then corresponding parameters in the template of the output shaft are modified
and the modification of size parameters and custom properties is completed. Finally, a
new 3-D model and corresponding engineering drawing of the output shaft can be
generated, and the parametric operation of components is achieved [13].
After building a 3-D model, it is convenient to generate 2-D engineering drawings. But
the parametric ratio of the pattern, position of view, size and notes are not processed
automatically. These problems will result in defects of chaotic layouts and uncoordi‐
nated proportions, so it is not suitable for guiding the production directly [14]. Therefore,
it is necessary to carry out automatic adjustments for engineering drawings from Solid‐
Works, so that output engineering drawings can meet the requirements of enterprise
standards and facilitate rapid productions.
Parametrically drive
Create models of components
Generate new 3-D models
and assemblies
National
standard Open an existing
Engineering drawings of the engineering drawing
enterprise
Reality of
enterprises
Attributes added
New engineering automatically
Generate engineering Update
drawings of
drawings of components and
components and
assemblies
assemblies
Adjust the Scale of
view
Coordinates of the
position of view
Save the new
engineering drawing
Coordinates of
annotation
Fig. 12. The process for the automatic adjustment of engineering drawings
Design and Development of Parametric System 373
When the new 3-D model is generated, the functions of the adjustment of engineering
drawings are able to be achieved with VB [14]. The process for automatic adjustments
of engineering drawings is shown in Fig. 12.
Figure 14 is the dialog box of custom properties in SolidWorks after the custom
properties of the planetary carrier is assigned.
5 Conclusions
Based on the understanding of the structure and design steps of the planetary
reducer, the general framework of the digital design system for the planetary reducer
is planned by the parametric design method. The development of the parametric
operation module in the system is completed by using SolidWorks as the develop‐
ment platform and Visual Basic as the development tool, and the automatic output
of 3-D models and engineering drawings of the same planetary reducer is realized.
The system is designed to provide enterprises and designers with a way to get serial
products rapidly in order to improve the efficiency and quality of design and make
the product more standardized and versatile.
References
1. Hua, Z., Dingfang, C.: Application and secondary development of SolidWorks based on
Visual Basic. J. Hubei Univ. Technol. 4, 36–38 (2010)
2. Xiuzi, Y., Chaoxiang, C.: SolidWorks Advanced Course. Secondary Development and API.
Mechanical Industry Press, Beijing (2007)
3. Xinhua, Z.: Modularization and Serialization of Planetary Gear Reducer. Dalian University
of Technology (2014)
4. Fengqin, Z., Chun, Z.: Parametric programming of NGW planetary gear reducer. Mech.
Transm. 05, 36–38, 88 (2005)
5. Yanli, C.: Research on Parametric Design of CAD System for NGW-S Planetary Gear
Reducer. Liaoning Technical University (2005)
6. Weiliang, L., Yixiang, W.: Parameterized design of hydraulic cylinder based on secondary
development of SolidWorks. Mach. Manufact. Autom. 01, 74–77 (2017)
Design and Development of Parametric System 375
7. Yi, W.: Parametric Design and Application of Wind Power Converter. Dalian University of
Technology (2014)
8. Zhaobin, H., Lin, X., Peng, H.: Parameterization design of hydraulic cylinder based on VB
and Solidworks. Coal Mine Mach. 12, 254–256 (2014)
9. Junliang, Z.: Study on Three-Dimensional Optimization Design of Planetary Gear Reducer
of Shearer Cutting Part. Harbin University of Science and Technology (2007)
10. Zhigeng, L.: Study on Key Technology of Parametric Design of Bridge Crane Bridge. North
University of China (2007)
11. Bo, P., Yue, Y., Chunzhu, S.: Parameterization modeling of centrifugal impeller using
SolidWorks API. J. Eng. Graph. 05, 1–7 (2009)
12. Ke, W., Zongyan, W., Chunyue, L., Bing, Z., Zhenyu, C.: Template-based parametric design
system of reducer. Mech. Transm. 05, 50–52 (2014)
13. Chao, S., Yuxiang, L.: Secondary development of SolidWorks for parametric design based
on VB language. Manufact. Autom. 15, 137–140 (2013)
14. Yuhua, P., Zongyan, W.: Study on key technology of automatic adjustment of parametric
design drawings. Hoisting Transp. Mach. 06, 58–61 (2012)
15. Yuhua, P., Zhigeng, L., Zhili, H., Jingang, Y., Zongyan, W.: Three-dimensional parametric
design system of main girder of bridge crane. Mech. Eng. Autom. 06, 45–47 (2006)
A Dynamic Double Threshold Based Cooperative
Spectrum Sensing Strategy in Heterogeneous Cognitive
Radio Networks
Chongxiao Peng1 ✉ , Yifei Wei1 ✉ , Bo Gu1, Ligang Ren2, and Mei Song1
( ) ( )
1
School of Electronic Engineering, Beijing University of Posts and Telecommunications,
Beijing 100876, People’s Republic of China
{pengchongxiao,weiyifei,gubo,songm}@bupt.edu.cn
2
China Unicom System Integration Limited Corporation Beijing Branch, Beijing, China
15611030998@wo.com.cn
1 Introduction
The cognitive radio technologies adopt strategy which shares the frequency spectrum
among the users, and the secondary user (SU) opportunistically occupy the spectrum
only if spectrum hole has been detected, which means the primary user (PU) doesn’t
occupy the spectrum band at a certain period of time [1]. Before using shared spectrum
among users, spectrum sensing is necessary to determine if there is a PU currently
occupying the licensed spectrum band [2].
In this paper, we adopt dynamic double threshold method which is set dynamically
according to the instantaneous SNR, so that further improve the performance of double
threshold energy detection based CSS scheme.
The rest of the paper is organized as following: Sect. 2 describes the proposed system
model. Section 3 demonstrates the single threshold energy detection based CSS scheme.
Section 4 proposes a dynamic double threshold energy detection based CSS scheme.
The simulation results and analysis are presented in Sect. 5 and the conclusion from the
paper is drawn in Sect. 6.
2 System Model
There are three phases in the sensing system: sensing, reporting and broadcast. The SUs
receive the signal transmitted from PU and transfer the signal into the form of energy
value [3], then make the local decision or remain the original energy value. The local
sensing results would be reported to the fusion center at reporting phase. This process
would require some bits of data as communication overhead [4]. The finally global
decision would be made in the fusion center, then all the SUs would get the final results
whether PU is occupying the licensed spectrum band.
{ } Γ(M, 𝜆∕2)
Pfi = P Ei > 𝜆||H0 = (2)
Γ(M)
where QM (..) is the generalized Marcum Q function [6] and Γ(..) is upper incomplete
gamma function. H0 denotes that the PU is absent of licensed spectrum band, while H1
denotes that the PU is present of licensed spectrum band.
K out of N rule is belong to the hard fusion decision scheme, which can also be called
voting rule or majority rule [7]. The fusion center deal with the local sensing results
reported by each SU. Assuming that the number of SU processing spectrum sensing is
N. The fusion center makes the final decision that PU is present of the licensed spectrum
band only if there are at least K SU indicating the licensed spectrum band being occupied
by PU. The finally global probability of detection and false alarm can be formulated as
following,
( ) k
∑
N
N ∏ ∏N
( )
PdKoutN = Pdi 1 − Pdj (3)
k=K
k i=1 j=k+1
( ) k
∑
N
N ∏ ∏(
N
)
PfKoutN = Pfi 1 − Pfj (4)
k=K
k i=1 j=k+1
378 C. Peng et al.
The Maximal Ratio Combining (MRC) [8] method adopts strategy that getting the
weighted summation, in which the normalized weight factor is set according to instan‐
taneous SNR. The method can be formulated as following,
{ } (√ √ )
PdMRC = P EMRC > 𝜆MRC ||H1 = QM 2𝛾MRC , 𝜆MRC (5)
( / )
{ } Γ M, 𝜆MRC 2
PfMRC = P EMRC > 𝜆MRC ||H0 = (6)
Γ(M)
∑
N
∑
N
𝛾i 𝛾i
where 𝛾MRC = 𝛾i, EMRC = wi Ei, wi = = .
i=1 i=1 ∑
N 𝛾MRC
𝛾i
i=1
𝜆0i = wi × 𝜆 (7)
𝜆1i = (2 − wi ) × 𝜆 (8)
𝛾i
wi = ( ) (9)
max 𝛾1 , 𝛾2 , ⋯ ⋯ , 𝛾N
The fusion center makes the global decision GD1 based on the local binary decision,
and makes the global decision GD2 based on the original energy value. Finally, we can
get the final decision which is expressed as following.
{
0, GD1 + GD2 < 1
FD = (10)
1, otherwise
{
0, L1 < K
GD1 = (11)
1, otherwise
⎧ ∑G
⎪ 0, EMRC = wj Ej ≤ 𝜆MRC
GD2 = ⎨ j=1 (12)
⎪ 1, otherwise
⎩
where L1 is the number of reporting binary decision which indicating the present state
of the PU.
A Dynamic Double Threshold Based Cooperative Spectrum 379
The final decision is made in the fusion center. We could compute the final proba‐
bility of detection and false alarm which is formulated as following.
( )
( ) ∏
G
Pd = 1 − 1 − PdB 1 − PdMRC × P1i (13)
i=1
( )
( ) ∏
G
Pf = 1 − 1 − PfB 1 − PfMRC × P0i (14)
i=1
where PdB and PfB is computed by using K out of N rule, PdMRC and PfMRC is computed
by using MRC, G is the number of original energy value reported from SU, the proba‐
bility of energy located between double thresholds must be taken account, which is
expressed as following.
( ) ( )
Γ M, 𝜆0i ∕2 Γ M, 𝜆1i ∕2
P0i = − (15)
Γ(M) Γ(M)
(√ √ ) (√ √ )
P1i = QM 2𝛾i , 𝜆0i − QM 2𝛾i , 𝜆1i (16)
The simulation is carried out in the assumption that channel is along with AWGN. The
time bandwidth product M = 5. We would compare the dynamic double threshold based
Fig. 1. Detection probability curve for single threshold based CSS using K out of N
380 C. Peng et al.
CSS scheme with the fixed double threshold scheme in which it is set as
λ0 = 0.8λ, λ1 = 1.2λ.
Figure 1 shows the relationship curve of probability of detection and false alarm
using single threshold based CSS scheme, in which the final probability is computed by
using the K out of N rule. Where the SNR is randomly generated between (−15 ~ 10)
dB. We can find that the more SUs sensing spectrum is, the higher the probability is.
Meanwhile the smaller the K is, the higher the probability is as expected. Figure 2 shows
the changing tendency of detection probability along with false alarm probability which
Fig. 2. Detection probability curve for single threshold based CSS using MRC
Fig. 3. Detection probability curve for dynamic double threshold based CSS
A Dynamic Double Threshold Based Cooperative Spectrum 381
utilizes the MRC method to complete the final decision. Apparently to achieve higher
detection probability it requires the number of original energy value as many as possible.
Figure 3 shows the proposed dynamic double threshold based CSS scheme under
the condition (N = 10, K = 10), meanwhile compares with the scheme which ignores
the energy value between the double thresholds and the scheme which deals with the
original energy reported from the SUs further. The proposed scheme presents a better
performance due to computing with the energy value between the double thresholds and
adjusting the threshold according to the instantaneous SNR appropriately.
6 Conclusion
The proposed dynamic double threshold based CSS scheme could realize better perform‐
ance by using MRC to deal with the original energy value between the double thresholds
and utilizing K out of N rule to deal with the local binary decision, the critical factor is
dynamic threshold adjusted according to instantaneous SNR.
Acknowledgement. This work was supported by the National Natural Science Foundation of
China (No. 61571059).
References
1. Ghasemi, A., Sousa, E.S.: Opportunistic spectrum access in fading channels through
collaborative sensing. J. Commun. 2(2), 71–82 (2007)
2. Yifei, W., Yinglei, T., Li, W., Mei, S., Xiaojun, W.: QoS provisioning energy saving dynamic
access policy for overlay cognitive radio networks with hidden markov channels. China
commun. 10(12), 92–101 (2013)
3. Urkowitz, H.: Energy detection of unknown deterministic signals. Proc. IEEE 55(4), 523–531
(2005)
4. Vien, Q.T., et al.: A hybrid double-threshold based cooperative spectrum sensing over fading
channels. IEEE Trans. Wirel. Commun. 15(3), 1821–1834 (2015)
5. Digham, F.F., Alouini, M.S., Simon, M.K.: On the energy detection of unknown signals over
fading channels. IEEE Trans. Commun. 55(1), 21–24 (2007)
6. Nuttall, A.H.: Some integrals involving the Q-function. IEEE Trans. Inf. Theor. 21(1), 95–96
(1972)
7. Mustapha, I., et al.: A weighted hard combination scheme for cooperative spectrum sensing in
cognitive radio sensor networks. In: IEEE Malaysia International Conference on
Communications, pp. 12–17. IEEE (2015)
8. Teguig, D., Scheers, B., Nir, V.L.: Data fusion schemes for cooperative spectrum sensing in
cognitive radio networks. In: Communications and Information Systems Conference IEEE,
pp. 1–7 (2012)
A Distributed Self-adaption Cube Building Model
Based on Query Log
Abstract. Among the diverse distributed query and analysis engine, Kylin have
gained wide adoption since its various strengths. By using Kylin, users can
interact with Hadoop data at sub-second latency. However, it still has some
disadvantages. One representative disadvantage is the exponential growth of
cuboids along with the growth of dimensions. In this paper, we optimize the
cuboid materialization strategy of Kylin by reducing the number of cuboids based
on the traditional OLAP optimization method. We optimize the strategy mainly
from two aspects. Firstly, we propose Lazy-Building strategy to delay the
construction of nonessential cuboid and shorten the time of cuboid initialization.
Secondly, we adopt Materialized View Self-adjusting Algorithm to eliminate the
cuboids which are not in use for a long period. Experimental results demonstrate
the efficacy of the proposed Distributed Self-Adaption Cube Building Model.
Specifically, by using our model, cube initialization speed has increased by 28.5%
points and 65.8% points space are saved, comparing with the cube building model
of Kylin.
1 Introduction
In the era of big data, many modern companies produce huge amounts of data in their
service lines. These data are used to conduct report analysis based on OLAP analysis.
In order to conduct report analysis, companies need a system which can response to the
query of thousands of data analysts at the same time. That requires high scalability,
stability, accuracy and speed of the system. In fact, there doesn’t exist a widely-accepted
method in distributed OLAP field. Many query engines can also conduct report analysis,
such as Presto [4], Impala [2], Spark SQL [14] or Elasticsearch [10], but they are more
emphasis on data query and analysis. As a matter of fact, Kylin [7] is the specialized
tool in Distributed OLAP field which is used often.
Kylin is originally developed by eBay, and is now a project of the Apache Software
Foundation. It is designed to accelerate analysis on Hadoop and allow the use of SQL-
compatible tools. It also provides a SQL interface and supports multidimensional anal‐
ysis on Hadoop for extremely large datasets. Kylin can reach the scale of one million or
2 Background
There are several strategies of data cube materialization [11] to reduce the cost of aggre‐
gation calculation and increase the query processing efficiency including iceberg cube
calculation Algorithm [3], condensed cube calculation Algorithm [15], shell fragment
cube calculation Algorithm [13], approximate cube calculation Algorithm [17], and
time-series data stream cube calculation Algorithm [6]. They are all based on Partial
Materialization [16], which means that a data sub-cube is selected and pre-calculated
according to specific methods. Partial Materialization is a compromise between storage
space, cost of maintenance and query processing efficiency.
In the process of iceberg cube calculation, sub-cubes which are higher than the
minimum threshold are aggregated and materialized. Beyer proposed BUC algorithm
[12] for iceberg cube calculation, which is widely-accepted.
According to the order of cuboid calculation, the methodologies of aggregation
calculation can be divided into two categories: top-down and bottom-up.
384 M. Song et al.
1. Top-Down: Firstly, calculate the metric of the whole data cube, and then the recur‐
sive search is performed along each dimension. Secondly, check the conditions of
the iceberg, prune branches that do not meet the condition. The most typical algo‐
rithm is BUC algorithm, it perform best on sparse data cube.
2. Bottom-Up: Starting from the base cuboids, compute high level cuboid from the low
level cuboid in the search grid according the parents-children relationship. Typical
algorithms are Pipesort algorithm, pipehash algorithm, overlap algorithm and
Multiway aggregation algorithm [18].
However, Kylin doesn’t follow the principle of partial materialization. In order to
reduce unnecessary redundant calculation and shorten the cube construction time, Kylin
adopts a Method called By Layer Cubing, which is a distributed version of the Pipesort
algorithm, a kind of bottom-up algorithm [1].
Fig. 1. By layer cubing. Each level of computation is a MapReduce task, and serial execution.
A N dimensional Cube needs N times MapReduce Job at least.
A Distributed Self-adaption Cube Building Model Based on Query Log 385
Compared with By Layer Cubing, the By Segment Cubing has two main differences:
1. The mapper will do pre-aggregation, this will reduce the number of records that the
mapper output to Hadoop, and also reduce the number that reducer need to aggregate;
2. One round MR can calculate all cuboids;
Based on the work mentioned above, we take advantage of both two Algorithm, and
optimize cuboid materialization strategy.
386 M. Song et al.
In this section, we first introduce the architecture of Self-adaption Cube Building Model
(SCBM) and the overall workflow. Then we explain cuboids Lazy-Building and the
cuboid spanning tree. Finally, we describe the implementation details of the Materialized
View Self-Adjusting Algorithm.
Self-adaption cube building model takes fact table [5] as input of the overall system,
usually the fact table is managed by distributed Data Warehouse Hive. We first set the
parameters of the cube model, such as the filed for analysis, the base cuboid level and
so on, then we build base cuboids in a mapper-reduce. After the construction of the base
cuboids, the system can support query request, Query execution engine [7] resolves the
query to find the required cuboids. If the cuboid has been generated, the query will be
executed; if the cuboid is missing, the Lazy building module will be triggered to build
the cuboid using the method in Sect. 3.2. When the query result returns, the system
records the query log and waits for the adjustment of cube launched by Self-Adaption
module according to the Materialized View Self-Adjusting Algorithm explained in
Sect. 3.3. At the same time, the system maintains a dynamic cube spanning tree to store
the metadata of cuboids.
A Distributed Self-adaption Cube Building Model Based on Query Log 387
1. User set a base-layer parameter in cube model info to control the scale of base
cuboids layer. If this parameter is not set, it will use default value log
(dimensions) + 1
2. Base cuboids building module import data from fact table and build the base cuboids
by Cube Build Engine in Kylin.
3. Update the Cuboid Spanning Tree and save the metadata.
4. Client launch a query select avg (measure i) from table group by C which hit the
missing cuboid [C]. Then, lazy building module receive request to build cuboid [C].
5. Lazy building module find a cuboid generation path according to Ancestor Cuboids
Selection and build the missing cuboid to response the query as soon as possible.
6. Record the path and determine whether to build all the cuboid on the path at low
load according to Materialized View Self-Adjustment Algorithm.
Ancestor Cuboids Selection. When the needed cuboids is missing, we should select
an ancestor cuboid and a cuboid generation path. The basic principle is to choose the
ancestor cuboid whose measures are the least to aggregate, which means we can get the
minimum amount of computation and time to generate the missing cuboid. After that,
we need to find a path P from ancestor cuboid to the missing cuboid in compliance with
the Minimum cardinality principle.
For example, in the Fig. 4. In order to generate the missing cuboid [C], we firstly
find all the candidate cuboids [A B C] [A C D] [B C D]. Then, we compare the size of
the three candidate cuboids. Assuming [B C D] is selected, we generate [C] by aggregate
[B C D] on dimension B, D. The cube is enough to response the query. However, for
the sake of maintenance of cube according to By Layer Cubing, we need to find a path
from [B C D] to [C].
When aggregating from parent to a child cuboid, assuming that from base cuboid [B
C D] to 1-dimension cuboid [C], There are two paths: [B C D] [B C] [C] and [B C D]
[C D] [C]. We assume Card(D) > Card(B) and the dimension A is independent with
other dimensions, after aggregation, the cuboid [BCD]’s size will be about 1/Card(D)
or 1/Card(B) of the size of base cuboid; So the output will be reduced to 1/Card(D) or
1/Card(B) of the original one in this step. So we choose the first path, the records that
written from mapper to reducer can be reduced to 1/Card(D) of original size; The less
output to Hadoop, which means less I/O and computing and the model can attain better
performance.
Self-adaption module adjusts the cube according to the Materialized View Self-
Adjusting Algorithm. This chapter proposes a query statistics method which takes fixed
times of queries as a statistical period, and this method updates the corresponding query
statistics. This method adjusts materialized views set according to the threshold of elim‐
ination and generation, stabilizes the query efficiency, and minimizes the shake of mate‐
rialized view.
A Distributed Self-adaption Cube Building Model Based on Query Log 389
In the above algorithm, from line 1 to line 8, it scans the query log in a statistical
period Tn, and update the query task set Q during the scanning. From line 9 to line 17, it
iterate around query in Q, and determine whether eliminate or materialize corresponding
( ( ))
views according to the comparison of threshold and the E Tn qi calculated by
formula 2. Suppose query task set Q contains k different query, then the time complexity
( )
of the algorithm is O Tn + k .
4 Experimental Evaluation
4.1 Dataset
To test performance, we use the standard weather dataset from the China Meteorological
Data network. The dataset contains 4726499 weather records from China’s 2170 distinct
counties started from January 1, 2011 to January 1, 2017. The original dataset is too
complicated. In order to better conduct the experiment, we select eight dimensions:
Province, city, county, date, weather, wind direction, wind speed, air quality level and
two measures: Maximum temperature, Minimum temperature.
A Distributed Self-adaption Cube Building Model Based on Query Log 391
1 ∑n ∑mTi
Average Query Time = response time(qi ) (2)
NUM(Q) i=1 j=1
Cube Size refers to the disk allocation that the whole cube takes up.
We first compare the metric of cube first construction time. Because the parameter of
base-layer has a great impact on this metric, in order to reflect the average condition,
we use the default value log (dimensions) + 1. We test the model 5 times and the results
were aggregated to calculate averages which can avoid the impact of MapReduce failure.
Results can be seen from Table 1, the time consumption of the new model is reduced
by 28.5% (Fig. 5).
In query time, we set Materialized view adjustment cycle Tn 50 and test 30 cycles T1
–T30. We can observe that Cuboid hit rate and query response time significantly increased
and improved along with the increase of query requests. Finally, the query efficiency of
the two models are almost on a par.
In cube size, we see that the curve that represents this metric tends to be stable after
vibration in prophase from Fig. 6. Finally, the spaces consumption of the proposed model
was reduced by 65.83%.
5 Conclusion
Acknowledgement. This work is supported by the National Key project of Scientific and
Technical Supporting Programs of China (Grant No. 2015BAH07F01); Engineering Research
Center of Information Networks, Ministry of Education.
A Distributed Self-adaption Cube Building Model Based on Query Log 393
References
1. Chen, Y., Dehne, F., Eavis, T., Rau-Chaplin, A.: Parallel ROLAP data cube construction on
shared-nothing multiprocessors. Distribut. Parallel Databases 15(3), 219–236 (2004)
2. Impala (2017). http://impala.apache.org/. Accessed 13 Apr 2017
3. Deshpande, P.M., Gupta, R., Gupta, A.: Distributed iceberg cubing over ordered dimensions,
March 2015. US Patent App. 14/658,542
4. Presto (2017). https://prestodb.io/. Accessed 13 Apr 2017
5. Gray, J., Chaudhuri, S., Bosworth, A., Layman, A., Reichart, D., Venkatrao, M., Pellow, F.,
Pirahesh, H.: Data cube: a relational aggregation operator generalizing group-by, cross-tab,
and sub-totals. Data Mining Knowl. Disc. 1(1), 29–53 (1997)
6. Kalisch, M., Michalak, M., Przystałka, P., Sikora, M., Wróbel, Ł.: Outlier detection and
elimination in stream data – an experimental approach. In: Flores, V., et al. (eds.) IJCRS 2016.
LNCS (LNAI), vol. 9920, pp. 416–426. Springer, Cham (2016). https://doi.org/
10.1007/978-3-319-47160-0_38
7. Kylin (2017). http://kylin.apache.org/. Accessed 13 Apr 2017
8. Lee, S., Kim, J., Moon, Y.-S., Lee, W.: Efficient distributed parallel top-down computation
of ROLAP data cube using MapReduce. In: Cuzzocrea, A., Dayal, U. (eds.) DaWaK 2012.
LNCS, vol. 7448, pp. 168–179. Springer, Heidelberg (2012). https://doi.org/
10.1007/978-3-642-32584-7_14
9. Li, F., Ozsu, M.T., Chen, G., Ooi, B.C.: R-store: a scalable distributed system for supporting
real-time analytics. In: 2014 IEEE 30th International Conference on Data Engineering
(ICDE), pp. 40–51. IEEE (2014)
10. Elasticsearch (2017). https://www.elastic.co/products/elasticsearch. Accessed 13 Apr 2017
11. Nandi, A., Yu, C., Bohannon, P., Ramakrishnan, R.: Distributed cube materialization on
holistic measures. In: 2011 IEEE 27th International Conference on Data Engineering (ICDE),
pp. 183–194. IEEE (2011)
12. Shi, Y., Zhou, Y.: An improved apriori algorithm. In: Granular Computing (GrC), pp. 759–
762. IEEE (2010)
13. Silva, R.R., Hirata, C.M., de Castro Lima, J.: Computing big data cubes with hybrid memory.
J. Convergence Inf. Technol. 11(1), 13 (2016)
14. Spark SQL (2017). http://spark.apache.org/sql/. Accessed 13 Apr 2017
15. Wang, W., Feng, J., Lu, H., Yu, J.X.: Condensed cube: an effective approach to reducing data
cube size. In: 18th International Conference on Data Engineering, Proceedings, pp. 155–165.
IEEE (2002)
16. Xia, Y., Luo, T.T., Zhang, X., Bae, H.Y.: A parallel adaptive partial materialization method
of data cube based on genetic algorithm (2016)
17. Yin, D., Gao, H., Zou, Z., Li, J., Cai, Z.: Approximate iceberg cube on heterogeneous
dimensions. In: Navathe, S.B., Wu, W., Shekhar, S., Du, X., Wang, X.S., Xiong, H. (eds.)
DASFAA 2016. LNCS, vol. 9643, pp. 82–97. Springer, Cham (2016). https://doi.org/
10.1007/978-3-319-32049-6_6
18. Zhao, Y., Deshpande, P.M., Naughton, J.F.: An array-based algorithm for simultaneous
multidimensional aggregates. ACM SIGMOD Rec. 26, 159–170 (1997)
Property-Based Network Discovery
of IoT Nodes Using Bloom Filters
1 Introduction
In the IoT context, a user (or an application) using hundreds or thousands
of devices has to deal with considerably long addresses to uniquely identify
and refer to network devices. This exponential growth is expected to introduce
new challenges to traditional computer network protocols, such as, for example,
(i) efficient access to a huge number of devices; (ii) security and privacy; (iii)
interoperability and standardisation; (iv) efficient energy consumption. More-
over, given the extreme amounts of heterogeneous devices constituting the IoT
ecosystem, timely and accurate device discovery based on some specific parame-
ters such as device type, sensing/actuating capabilities, status, powering options
is also seen as a pressing and challenging issue. In this light, a potential way to
enable device discovery in the IoT, taking into account the size and complex-
ity of underlying networks, could be to include additional parameters in the
c Springer International Publishing AG 2018
Q. Zu and B. Hu (Eds.): HCC 2017, LNCS 10745, pp. 394–399, 2018.
https://doi.org/10.1007/978-3-319-74521-3_42
Property-Based Network Discovery of IoT Nodes using Bloom Filters 395
routing procedure with a goal to limit the search space. More specifically, a
potential solution would be able address IoT network devices not only through
their IP addresses, but also through a combination of device properties, such
as their type, location, sensing/actuating capabilities, available resources and so
on. From this perspective, an envisaged solution could implement some kind of
selective routing algorithm, which would facilitate time- and network-efficient
device discovery in the IoT context. For instance, this will allow collecting infor-
mation from specific sensing devices (e.g. within a single building) and apply
actuation commands (e.g. turn on heating in rooms with low temperature) in a
selective manner. Similarly, remote device management and maintenance would
also become feasible, as users would be able to diagnose errors and patch required
devices with corresponding software updates (e.g. keeping a surveillance system
up to date with most recent system security updates).
Taking into considerations these desired features of a possible solution, this
paper presents an approach to facilitate property-based device search and discov-
ery in complex IoT networks using counting Bloom filters. As it will be explained
below, the proposed approach benefits from the space-efficient way of storing
information about devices and their properties, as well as fast calculation times
when deciding whether a matching device is present in the network. Moreover,
with property-based search using Bloom filters, it becomes possible to perform
device discovery at various granularity levels.
3 Proposed Approach
The proposed approach is based on a two-step procedure. First, all the interme-
diate nodes in the network hierarchy are populated with information about edge
nodes. Second, once the network is populated, an IoT device can be discovered
using a corresponding query.
At the device discovery step, Bloom filters are used to represent correspond-
ing discovery queries. More specifically, each query is represented by a tuple
Q = (ID, P rop), where ID is a unique identifier of this query, and P rop is a
set of device properties, which are expected to be discovered within the given
network. These properties can be seen as search parameters, as typically used in
traditional searching. Each query is then represented by a corresponding Bloom
filter, using the same hash functions. Device discovery can be seen as a reverse
process of populating the network hierarchy, executed in a top-down manner
starting from the very top of the network topology. By performing the bitwise
AND operation, the top-level server first checks with its own Bloom filter whether
Property-Based Network Discovery of IoT Nodes using Bloom Filters 397
The simplified use case scenario assumes each device (and its properties) is rep-
resented by a 6-bit Bloom filter. From left to right, these bits indicate whether a
device is: a camera (1), an environmental sensor (2), a smartphone (3), battery-
powered (4), powered by a solar panel (5), or powered by a cord (6). It is also
assumed that there are three subnets in the network, each containing three
devices (Fig. 1). The three network gateways contain combined Bloom filters of
their respective subnets, and the server contains the overall Bloom filter rep-
resentation of the network. The goal of this scenario is to discover a camera,
powered by a cord. Accordingly, the query is represented by the following Bloom
filter BF = (1, 0, 0, 0, 0, 1).
At the first step, the server evaluates the query against its own Bloom filter,
and decides that there is indeed a matching device present somewhere down the
network. Next, the query is propagated down to three subnets. The respective
gateways start evaluating the query against their own Bloom filters. As it is seen
from the diagram, Subnet A contains only smartphones, and Subnet B contains
only environmental sensors. The query evaluation returns false, and, as a result,
the network call is not propagated down the first two subnets. Gateway C, by
evaluating the query, understands that there is a matching device in its subnet
and sends the query to all three nodes. Two of these nodes are cameras, but
only one of them – Device C3 – is actually powered by a cord. By evaluating
the incoming query, it realises that it matches query parameters, and replies
back with its ID and network location. The reply is then sent back to the server
through intermediate hops.
The presented property-based search for IoT devices enables flexible, fine-
grained discovery of IoT nodes. The more properties are specified in the query,
the more precise the search is and less matching devices are discovered. On
contrary, for a single specified property, the search space is expected to be wider,
since more devices might satisfy the search query parameter. For example, in the
simplified scenario above, searching for a solar panel-powered device will yield 6
results.
5 Conclusions
The presented solution enables flexible property-based device discovery in the
context of complex IoT networks using Bloom filters. As opposed to the IPv6
routing, which requires 128 bits to encode an address, the proposed approach
benefits from the space-efficient way of representing and storing data using a
Bloom filter. This also contributes to decreased traffic and network latency, as
the device discovery duration depends on how narrow-focused a search query is
(i.e. the less devices matching the query, the less network traffic is generated).
The Bloom filter decides whether a device belongs to a subnet branch or not,
and can ‘cut off’ the entire branch before actually checking it. As a result, this
considerably reduces the amount of network traffic, especially when compared
to broadcast and multicast routing techniques.
References
1. Bloom, B.H.: Space/time trade-offs in hash coding with allowable errors. Commun.
ACM 13(7), 422–426 (1970)
2. Broder, A., Mitzenmacher, M.: Network applications of Bloom filters: a survey. Int.
Math. 1(4), 485–509 (2004)
3. Fan, L., Cao, P., Almeida, J., Broder, A.Z.: Summary cache: a scalable wide-area
web cache sharing protocol. IEEE/ACM Trans. Netw. (TON) 8(3), 281–293 (2000)
4. Foley, S.N., Navarro-Arribas, G.: A Bloom filter based model for decentralized
authorization. Int. J. Intell. Syst. 28(6), 565–582 (2013)
5. Snoeren, A.C., Partridge, C., Sanchez, L.A., Jones, C.E., Tchakountio, F.,
Kent, S.T., Strayer, W.T.: Hash-based IP traceback. In: ACM SIGCOMM Com-
puter Communication Review, vol. 31, pp. 3–14. ACM (2001)
6. Tarkoma, S., Rothenberg, C.E., Lagerspetz, E., et al.: Theory and practice of Bloom
filters for distributed systems. IEEE Commun. Surv. Tutorials 14(1), 131–155 (2012)
The Personality Analysis of Characters in Vernacular
Novels by SC-LIWC
1
School of Computer and Control, University of Chinese Academy of Sciences,
Beijing 100190, China
libb@ucas.ac.cn
2
Institute of Psychology Chinese Academy of Sciences, Beijing 100101, China
tszhu@psych.ac.cn
Abstract. There are many researches on psychological text analysis, and it has
been proved that the words people use can reflect their emotional states. In this
paper, we introduce how to analyze the psychology of the characters in vernacular
novels automatically. First, we process the dialogs with word segmentation, and
analyze the segmented text with SC-LIWC. Then, a vector reflecting the
psychology of the character is obtained and we map it to the big five. Finally,
taking the dialogs of the Journey to the West as corpus, We have got the person‐
alities of four main characters which are verified to be same as some famous
comments of the Journey to the West, which shows that our work is effective.
Keywords: LIWC · Vernacular · The Journey to the West · The big five
Text analysis
1 Introduction
The use of words in the text can reflect the individual’s psychological state and person‐
ality [1]. Linguistic Inquiry and Word Count is a tool which we can use to analyze text.
The way that the Linguistic Inquiry and Word Count (LIWC) program works is fairly
simple. Basically, it reads a given text and counts the percentage of words that reflect
different emotions, thinking styles, social concerns, and even parts of speech.
To date, LIWC has been applied to many psychology research. It is often used to
examine suicide writings in order to characterize the quantitative linguistic features of
suicidal texts, in [2], the authors analyze texts compiled in Marilyn Monroe’s Fragments
using LIWC, in order to explore the contact between the use of different linguistic cate‐
gories over the years and her suiside. The result is coincide with different theories of
suicide. López-López et al. [3] analyzed the StackOverflow’s answers and questions to
explore the users’ personality traits. They found that the top reputed authors are more
extroverted than general users. Moreover, authors who got more votes express signifi‐
cantly less negative emotions than those who got less votes. Markovikj et al. [4] explored
the modeling feasibility of user personality based on the features extracted from Face‐
book. In [5], they collected a sample of 363 participants, including their written self-
introductions and final course performance, the result shows that course performance
could indeed be predicted by the word usage of linguistic categories.
LIWC for Traditional Chinese, TC-LIWC, is published with the authorization of
Pennebaker by Huang et al. (2012). After that, SC-LIWC for Simplified Chinese [6, 7]
is published on the basis of TC-LIWC, which lays the foundation for the following
research [10, 12].
Recently, some researchers are concerned about automatic personalistic prediction
using liwc. Personality is stable in a period of time, so a collected corpus from several
months is suitable for this research. Gao [12] selects 1766 participants, first make them
fill in a big five inventory for comparison, and then collect their weibo through the API
of Sina. 90% of the samples are trained using liwc and the rest act as test set. In the
training stage, they compute the Pearson’s coefficient between the inventory and the
training results, then choose features which behave well in the training. At last, the
features are composed to predict personality of test set. They compute the Pearson’s
coefficient as before, and the results are between 0.3~0.4. While the coefficient between
self-rating and rating by observers is about 0.5, hence the method has prediction ability
to some extent.
We will seek method to predict the personality of characters in novels written in
vernacular. Vernacular is a written language with some artistic processing. It is easy to
read, but still have some features of ancient Chinese. Vernacular is generally used for
literacy, especially in the novels. Vernacular novels are very popular from the beginning
of Ming Dynasty. Three of the four famous Chinese novels were accomplished in Ming
Dynasty. After that, vernacular novels were more and more popular. There are many
excellent ancient books in China, which created numerous virtual characters, a book
named A Dream of Red Mansions only, contains hundreds of characters. We will pay
a lot of time to read books, look up in the library, to understand these figures and step
into the author’s inner world. If we can analyze the characters of the books automatically,
it will save us much time and help us follow the books.
In this study, we use LIWC to analyze the personality of the characters in vernacular
novels. The process of personality analysis of the characters in the vernacular novels is
shown in Fig. 1. The rest of this paper is organized as follows. Section 2 will discuss
the preprocessing work for the novel. And then in Sect. 3, we will split the dialogs
obtained from Sect. 2, which will be used to analysis the personality of the characters
and get the big five of characters in Sect. 4. Finally, we also present the personality
change of Sun Wukong before and after the three strikes of White Bone Demon.
402 Y. Yuan et al.
The Journey to the West is the first ancient Chinese romantic novel. The book deeply
depicts the social reality of the time, mainly describes the origin of Tang priest, Sun
Wukong, Zhu Bajie, Sha monk, and together with the story of pilgrimage to the west.
After the spread of centuries, the Journey to the West has been translated into many
languages, and a number of relevant research monographs have been published, which
made a high evaluation of the novel. The Journey to the West is known as one of the
four famous Chinese classics. There are four main characters in the Journey to the
West. They are Sun Wukong, Sha monk, Zhu Bajie and their master, Tang priest.
James W. Pennebaker proved that words used in their daily lives could contains
important information of psychological information [8]. Especially, he proved that not
only nouns and verbs serve as markers of emotional state, social identity, and cognitive
styles, particles, serve as the glue that holds nouns and regular verbs together, can also
do the same things. That means we can study the particles instead of more complex
methods. On the basis of his work, we decide to use the dialogs of the characters to study
their personality. We select the dialogues and their inner monologues for each character
respectively, and put them into 5 different files. There are a lot of descriptive verses in
the text, we should delete them because they would interface the process of participation,
and these verses are often said by other people, not the roles themselves, so we believe
that the descriptive verse have little influence on the personality. Notice that it shouldn’t
include quotes on the end of the sentence. But other interval will be retained for the next
step. At last we get four files.
3 Segmentation
In order to get the particles for analysis, we should first do the text segmentation. As we
know, the current methods of Chinese word segmentation can be divided into three
kinds: the method based on lexicons, the method based on statistics and one based on
The Personality Analysis of Characters in Vernacular Novels 403
semantics. Also there are methods mixing two or three of them in order to improve the
accurate. There have been many kinds of word segmentation systems for modern
Chinese. For example, LTP-CLOUD, NLPIR, jieba, and so on. These systems all have
good results in Chinese word segmentation. Among these tools, LTP and NLPIR are
systematic tools, while jieba only contains segmentation function. Moreover, LTP and
jieba are open source tools, but NLPIR is not.
Few studies are involved in ancient Chinese segmentation. First, there is little corpus
for ancient Chinese. As we all know, segmentation need a lot of corpus which has been
marked manual to improve the accurate, but no one have done the work for ancient
Chinese. Second, it seems that we cannot obtain any economic benefits from it. However,
Hou, et al. have studied ancient Chinese segmentation [9], but the corpus is still very
small, which couldn’t be generalized easily.
Vernacular has the features of both ancient Chinese and modern Chinese. Therefore,
we can refer to the methods for modern Chinese segmentation.
The punctuation and function words are not changed much over the years. They are
also used in ancient Chinese. Besides, there are a lot of words that still exist in ancient
Chinese. In addition, the Chinese word segmentation methods are able to identify new
words according to statistical methods.
In order to simplify the segmentation procession, we make a simple test on the
segmentation of vernacular and find, most words are segmented correctly by applying
the LTP directly. But there are still many mistakes; notional words are not distinguished
from others, idioms are segmented wrong, and there are other mistakes generated in the
algorithm. That is mainly because of the difference between vernacular and modern
Chinese. There are many words which are not used now, especially those appear only
several times in the text.
A simpler and more efficient method we used to solve above questions is add a
manual dictionary to the LTP. The dictionary includes notional words and some idioms.
Notional words include place name, monster name, Tang priest and his three appren‐
tices’ name and nickname, gods’ name and their nickname, the particular items and some
words about the emperor and the dynasty.
• Place name. There are a lot of places made up by the author, such as the monster’s
cave, the god’s mountain, the monkey’s birthplace, and so on. Take these words into
dictionary will make sure they are segmented correctly.
• Monster’s name. Tang priest and his four apprentices meet many big monsters on
the Journey to the West, and each of them get a nickname, even some of small monster
under them get one, too. To split them correctly, we had better put them into the
dictionary.
• Four characters’ name and their nickname. Though only four people, each of them
have many nickname. Only Tang priest has more than 5 nicknames, for example,
priest, Tang priest, Xuanzang, elder, Tang elder, master, and so on. Especially they
often emerge in the dialogs. Therefore, split them from other words are important.
• Gods’ name and their nickname. When walking through the long way, the four meet
a lot of gods. Each of them have several nickname, especially Guanyin, his nicknames
even catch up with Tang priest.
404 Y. Yuan et al.
• The particular items. There are many particular things in the Journey to the West,
such as kinds of weapons, treasures, and so on. Much of them are not usually used
in modesty Chinese.
• Some idioms. Idioms are fixed phrases in the ancient Chinese.
• Some words about the emperor and the dynasty.
We select 1000 characters randomly from the segmentation results for each of
the four, five of our workmates check it respectively. When finished all, they vote
on the contradicted ones. Table 1 shows that the error rate is about 2%~3%, in other
words, the accurate rate is about 97%~98%. Though the punctuation is count into the
words, the error rate will not exceed 1.2 times of the existing data. As we can see,
this method is quite effective.
In [12], five of six feature classes are properties of Weibo, so in our work, we
choose only the features of liwc. Due to the differences between ancient Chinese and
modern Chinese, liwc dictionary will have corresponding change, so will the features
selected based on liwc. To solve it, we get rid of the features which are not consis‐
tent with ancient Chinese. The rest features will serve as input of the model which
have been trained in [12]. In order to inspect the personality more intuitive, we adopt
a commonly used quantitative method in psychology—the big five personality traits
[11]. In the big five traits model, the user’s personality is abstracted into five dimen‐
sions, which are shown in Tables 2 and 3. The big five score of Tang priest and his
apprentice is shown in Table 4.
Table 4. The big five of Tang priest and his three apprentice.
Agree. Cons. Extra. Open. Neur.
Tang priest 13.06 11.26 7.94 3.50 25.42
Sun Wukong 9.63 4.97 1.34 0.92 1.93
Zhu Bajie 9.25 5.02 18.54 15.08 18.27
Sha monk 15.24 6.54 14.15 1.10 26.31
4.1 Agreeableness
As an eminent monk from Tang dynasty, Tang priest is very kind and compassionate
[13]. On the Journey to the West, he tries to help others, though when he is in danger.
He is modest and subject to authority, such as emperor of Tang Dynasty, and all the
gods they meet on the Journey to the West. But sometimes he is egoistic, and often shifts
responsibility. His agreeableness is relatively high.
Sha monk is very careful and slavish since he was surrendered by Sun Wukong. He
has never been egoistic, doing his best to serve the master and help his brothers [14].
Tang priest and Sun Wukong all have deep trust on him. The agreeableness of him is
highest.
Sun Wukong is capable, but his master does not trust him. He has helped many
people, but that does not mean he is willing to sacrifice. If someone harms his interests,
he will not hesitate to teach him a lesson. Sun Wukong is also an arrogance role, never
understanding what modesty is. He is not gentle too. In conclude, his agreeableness is
lowest.
4.2 Conscientiousness
Tang priest has no ability to protect himself, and has no experience to deal with monsters.
He strictly abides by the doctrine, trying his best to protect it. And he has a strong will,
which makes him go to the west firmly to obtain Mahayana Buddhism [13]. All in all,
his conscientiousness is relatively high.
406 Y. Yuan et al.
Though exiled from heaven just because knocked over a glass made of colored glaze,
Sha monk not only gets no angry, but also accepts the destiny to atone for his sin. In this
point, he is something like Tang priest [14]. Compared with his brothers, he seems plain,
but he is better than his brothers in human nature. So his conscientiousness is relatively
higher.
Sun Wukong is strongest among his brothers. He pursues the idea that the stronger
should hold the power, never yielding to authority [15]. Therefore, many gods serve him
as a servant. He tries his best to protect the master, not for any benefits, but for returning
Tang priest’s salvation. He could not restrain his aggressive instincts, and that makes
Tang priest most unhappy. It is not strange that his conscientiousness is lower than his
brothers.
Zhu Bajie is similar with Sun Wukong in agreeableness and conscientiousness. There
is also difference: Zhu Bajie gets more score in tender mindedness, while Sun Wukong
gets more in altruism [16, 17]. Zhu Bajie obeys the laws. Sun Wukong is a king of
monkey before he follows Tang priest, so he has no knowledge of it. Thus in order, Zhu
Bajie gets more score. But he always declares to go back to Gao Village when he
encounters danger. Sun Wukong loves battle while Zhu Bajie loves women. In total,
they go head in head with each other in agreeableness and conscientiousness.
4.3 Extraversion
In the point of extraversion, Zhu Bajie gets first without question. He is very lustful,
showing great enthusiasm for women. He loves to eat, too [18]. Each time when they
arrive to a new town, he is happiest because he can eat a lot. Zhu Bajie is outgoing
compared with other people.
Sha monk is upright and honest [19]. He talks little, but what he said is very useful.
When there is contradiction among the group, Sha monk is the one who tries to solve it.
Tang priest is kind and friendly when dealing with people, but not enthusiastic. He
prefers quiet than noisy. So his extraversion is low.
In Table 4, Sun Wukong is lowest in extraversion. Though his warmth is not as good
as his brothers, but the rest features should be better than others. The possible reason
may be the dictionary we use may not be so fit with his dialogs.
4.4 Openness
Openness is an indicator of the level of intelligence. From Table 4 we can observe that
Zhu Bajie gets the highest score. Zhu Bajie is always being called “fool”, but that is not
the case [18]. He is fond of eating and sleeping, and good at flattering in front of master,
so Tang priest trusts him very much. In general, he always makes the best decision for
himself.
As a leader of the group, Tang priest is well learned and behaved. But he is often
cheated by monsters and confused by Zhu Bajie, and getting rid of Sun Wukong several
times, who tries to protect him.
Sha monk actually is a servant of Tang priest. He is responsible for all trifles, but
never complaining about it [20].
The Personality Analysis of Characters in Vernacular Novels 407
Sun Wukong is powerful and good at dealing with enemies, however, he is often
fooled by Zhu Bajie. So in fact, Zhu Bajie is the most intelligent person among the group.
4.5 Neuroticism
In the term of neuroticism, Tang priest and Sha monk are similar to each other. When
facing the danger, Tang priest is anxious and scared, and often bursts into tears [13].
Sha monk is puzzled, and ‘What should we do?’ is his pet phrase. Zhu Bajie helps Sun
Wukong a lot in fighting, but once failed the first thing he thinks of is escaping. Sun
Wukong never gives up, even if he was alone, he will fights until success.
We also did a research on the personality change of Sun Wukong before and after the
three strikes of White Bone Demon with this model. The results are shown in Table 5.
Table 5. The big five of Sun Wukong before and after his beating the White Bone Demon for
three times.
Agree. Cons. Extra. Open. Neur.
Before 6.28 3.63 2.83 19.27 6.62
After 10.56 4.99 0.39 1.03 1.19
Beating the White Bone Demon for three times was a turning point of Sun Wukong.
The author explained the mind change of Sun Wukong by the words of Zhu Bajie.
Before that, Sun Wukong was very irritable, but under the constraint of Tang priest,
he changed gradually. He became no more impulsive at all. It is corresponding to the
decrease of neuroticism.
In the early stage, Sun Wukong despised the authority, refused to obey the discipline,
and showed a clear sense of rebellion. While later he was influenced by Tang priest, no
longer having a strong sense of resistance. As we can see in Table 5, his openness is
greatly changed before and after the three beats of White Bone Demon.
The agreeableness is higher after the three strikes. In the respect of getting along
with people, Sun Wukong changed obviously, especially to Bodhisattva, Buddha, and
some others who is venerable. He was more and more courteous and no longer as arro‐
gant as before.
In the early stage, the responsibility of Sun Wukong is not clear, he even tried killing
Tang priest and attacked the Bodhisattva at first. However, he accepted his duty in the
late, becoming a qualified defender.
6 Conclusion
observe. We compare the results with many famous reviews, and compare the results
between characters, and compare the results back and forward. These comparisons all
show that automatic personality analysis of characters with LIWC is feasible.
References
1. Tausczik, Y.R., Pennebaker, J.W.: The psychological meaning of words: LIWC and
computerized text analysis methods. J. Lang. Soc. Psychol. 29, 24–54 (2010)
2. Fernándezcabana, M., Garcíacaballero, A., Alvespérez, M.T., Garcíagarcía, M.J., Mateos, R.:
Suicidal traits in Marilyn Monroe’s fragments: an LIWC analysis. Crisis 34, 124–130 (2013)
3. López-López, E., del Pilar Salas-Zárate, M., Almela, Á., Rodríguez-García, M.Á., Valencia-
García, R., Alor-Hernández, G.: LIWC-based sentiment analysis in Spanish product reviews.
In: Omatu, S., Bersini, H., Corchado, J.M., Rodríguez, S., Pawlewski, P., Bucciarelli, E. (eds.)
Distributed Computing and Artificial Intelligence, 11th International Conference. AISC, vol.
290, pp. 379–386. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-07593-8_44
4. Markovikj, D., Gievska, S., Kosinski, M., Stillwell, D.J.: Mining Facebook data for predictive
personality modeling. In: AAAI International Conference on Weblogs and Social Media
(2013)
5. Robinson, R.L., Navea, R., Ickes, W.: Predicting final course performance from students’
written self-introductions: a LIWC analysis. J. Lang. Soc. Psychol. 32, 469–479 (2013)
6. Gao, R., Hao, B., Li, H., Gao, Y., Zhu, T.: Developing simplified Chinese psychological
linguistic analysis dictionary for microblog. In: Imamura, K., Usui, S., Shirao, T., Kasamatsu,
T., Schwabe, L., Zhong, N. (eds.) BHI 2013. LNCS (LNAI), vol. 8211, pp. 359–368. Springer,
Cham (2013). https://doi.org/10.1007/978-3-319-02753-1_36
7. Zhao, N., Jiao, D.D., Bai, S.T., Zhu, T.S.: Evaluating the Validity of Simplified Chinese
Version of LIWC in Detecting Psychological Expressions in Short Texts on Social Network
Services, vol. 11, p. e0157947 (2016)
8. Pennebaker, J.W., Mehl, M.R., Niederhoffer, K.G.: Psychological aspects of natural language
use: our words, our selves. Annu. Rev. Psychol. 54, 547–577 (2003)
9. Zeng, Y., Hou, H.: Research on the extraction of ancient texts. J. China Soc. Sci. Tech. Inf.
30, 132–135 (2008)
10. Gao, R., Hao, B.B., Li, L., Bai, S., Zhu, T.S.: The establishment of the software system of
Chinese language psychological analysis. In: Psychology and the Promotion of Innovation
Ability: the Sixteenth National Conference on Psychology, p. 3 (2013)
11. de Raad, B.: The Big Five Personality Factors: The psycholexical Approach to Personality,
pp. 309–311. Hogrefe & Huber Publishers, Ashland (2000)
12. Gao, R.: The Research and Application of the Psychological Analysis Technology of Weibo
Content, University of the Chinese Academy of Sciences, vol. Master. University of the
Chinese Academy of Sciences (2014)
13. Cao, B.J.: The reflection and critique of pure confucianism personality: a new comment of
Tang priest. Acad. J. Zhongzhou 4, 110–112 (1999)
14. Jiang, K., Huang., L.J.: Compromise and secularization: an image evolution of Sha monk. J.
Zibo Normal Coll. 4, 58–62 (2007)
15. Zhou, X.S.: The time spirit and cultural connotation of Sun Wukong. J. Southeast Univ. 8,
63–73 (2006). (Philosophy and Social Science)
16. Cao, B.: Secularized Comic Image and Civil Personae 1: A New Reading of Ba Jie in A
Journey to the West. J. Huaihai Inst. Technol. 5, 23–27 (2007). (Social Science Edition)
The Personality Analysis of Characters in Vernacular Novels 409
17. Cao, B.: Secularized Comic Image and Civil Personae 2: A New Reading of Ba Jie in A
Journey to the West. J. Huaihai Inst. Technol. 5, 27–30 (2007). (Social Science Edition)
18. Liu, W.L., Qiu, H.C.: The Empiricism of Zhu Bajie. J. Keshan Teachers Coll. 2, 42–46 (2002)
19. Cai, S.: Discussion on Sha monk in “Journey to the West”. J. Educ. Inst. Jilin Province 30,
114–115 (2014)
20. Lu, L.: Industrious, duty, firm and persistent: a comment of Sha monk. South J. 11, 98–99,
110 (2014)
Digging Deep Inside: An Extended Analysis
of SCHOLAT E-Learning Data
Aftab Akram1,2, Chengzhou Fu1 ✉ , Yong Tang1, Yuncheng Jiang1, and Kun Guo3,4
( )
1
School of Computer Science, South China Normal University, Guangzhou 510631,
Guangdong Province, China
aftabit39@gmail.com, fucz@m.scnu.edu.cn,
{ytang,ycjiang}@scnu.edu.cn
2
University of Education, Lahore 54000, Pakistan
3
College of Mathematics and Computer Sciences, Fuzhou University, Fuzhou 350116,
Fujian Province, China
gukn@fzu.edu.cn
4
Fujian Provincial Key Laboratory of Network Computing and Intelligent
Information Processing, Fuzhou University, Fuzhou 350116, Fujian Province, China
Abstract. More and more higher education institutions are adopting computer
based learning management system to boost learning of the students. Networking
and collaboration through social media platforms are vital realities. Learning is
not merely limited to class rooms, which is now independent of location and time.
Understanding how students learn in this realm is a mighty challenge for teaching
professionals. Fortunately, data is abundantly available through learning manage‐
ment systems and social media platforms. Analyzing this vast data could give an
insight into how learning is happening in these days. Data mining techniques are
vastly being used for this purpose. In this paper, we present a statistical analysis
of e-leaning data obtained from SCHOLAT, a scholar oriented social networking
system. The analysis aims at getting data oriented perspectives of learning, e.g.,
which factor to what extent impacts learning. The analysis revealed factors which
positively or negatively affect learning achievement of the students, i.e., course
final scores.
1 Introduction
2 Literature Review
of other variables. The distributional shape were skewed and outliers were present in
the data.
However, previous analysis was carried out on a single class and only four variables
were used (3 independent and 1 dependent). In this paper, we present an extended and
diverse analysis of SCHOLAT e-learning data. We include four class in this analysis to
study behavior across different students’ cohorts. Further, we use 11 variable (10 inde‐
pendent and 1 depended). These variables represent administrative and collaborative
use of learning management system. By use of statistical and correlation analysis tech‐
niques, the analysis reveals common trends of system usage among different groups of
students. We use course final scores as criteria of successful learning. We plan to use
this analysis as base work for developing machine learning models representing
students’ learning in blended social learning environment.
SCHOLAT is an emerging vertical social networking system designed and built specif‐
ically for scholars, learners and course instructors. It uses un-directed graphs to represent
social network structure. It is bi-lingual, i.e., supports English and Chinese. The main
goal of SCHOLAT is to enhance collaboration and social interactions focused around
scholarly and learning discourses among community of scholars. In addition to social
networking capabilities, SCHOLAT incorporates various modules to encourage collab‐
orative and interactive discussions, for example, chat, email, events, news post, etc.
Table 1 shows a comparison of SCHOLAT with other similar social networking system.
SCHOLAT definitely has an advantage over them.
study is being conducted, 32758 students have been enrolled in 1785 classes of 760
courses.
SCHOLAT-course is still developing. Being a part of SCHOLAT scholar social
networking system opens vast horizons for its future development. We plan to work in
dimensions of social collaborative learning. We plan to study the factors affecting the
learning in collaborative social learning environment. This study is start of a journey
towards this destination.
In this study, we explore the statistically supported answers of four research ques‐
tions: (1) what is the general pattern of system usage among four classes, (2) course
administration and collaboration, what type of course support is mostly used by students,
(3) is the data is normally distributed or not?, (4) which variable(s) is (are) most corre‐
lated with the course final scores.
In previous study [20], the analysis was performed on one class. In present study, we
include four classes. We also increase number of variables from four to eleven. These
four classes were selected from three different courses, course ID165 (C Language
Programming) class ID279 taught in autumn 2014, course ID520 (Introduction to
Learning Sciences) class ID1149 taught in autumn 2015, and course ID206 (Software
Requirement) classes ID1563 and ID1729 taught in spring 2016. The number of students
enrolled in four classes was 216 out of which 188 students passed the course or took the
final exam.
Table 2 shows the variable included in previous and current studies. All of these
variables are quantitative values. The variables are categorized in to two types: student
level data for which observations for each student was available, and course level data
for which observations was only available at course level. Ten out of eleven variables
are independent variable, one variable (course final score) is dependent variables. These
variables were extracted from course records stored in SCHOLAT database. The course
final scores were not present in the database, so the relevant course instructor was asked
to provide. The main cause of selection of these four classes was availability of course
final scores.
Earlier, we used two types of descriptive statistical techniques: univariate and
bivariate [20]. The univariate technique was used to uncover the properties of individual
variables, whereas bivariate analysis was used to discover relationship between inde‐
pendent and dependent variables. We intend to use same statistical techniques in this
study. This study has an added advantage that results will be compared among four
different classes.
The results of the analysis are presented in following three subsections.
Table 5 shows the obtained figures for course level variables. We did not perform any
calculations on these variables like we did on student level variables, since these figures
are available only at course level. However, we present a discussion on its implications
in next section. The class ID1563 and ID1729 belong to same course, so in Table 5 the
figures for these two classes are presented jointly.
5 Discussion
students using system not at all while some other students usage is very high. A similar
trend is observed in class ID1563, i.e., median 47 (range: 0−17053). Earlier high average
values for these two classes were observed. The other two classes show relatively
balanced usage, i.e., median is 19 (range: 5−46) for class ID1149 and median is 41
(range: 11−1059) for class ID1729. From above discussion it can concluded that classes
ID279 and ID1563 show high and unbalanced usage while classes ID1149 and ID1729
show low but balanced usage. We also note presence of outliers in classes ID279 and
ID1563. The high values of standard deviation also signifies presence of varied usage
behavior. The classes ID279, ID1563 and ID1729 shows high variability in system
usage, i.e., 1667.57, 2695.07 and 168.10. The class ID1149 has relatively consistent
usage pattern and low value for standard deviation, i.e., 10.59.
The variables number of question and number of reply are indication of level of
collaboration, since asking questions and getting replies promotes collaboration and
interaction among students. The descriptive statistics shown in Table 3 indicates that
these activities are highly overlooked by students. The mean, mode, median and standard
deviation values for these two variables for classes ID1563 and ID1729 are zero. For
other two classes these values are very low. The data indicates very low collaboration
among the students.
The SCHOLAT course module has comprehensive facility for instructors to upload
homework and for students to submit homework. Submitting homework on time is
indication of good behavior whereas late submission or not submitting at all indicates
procrastinating behavior which is alarming for students’ learning. The three variables
number of homework on-time, number of homework late and number of homework not
hand-in are related to this key aspect of course administration. The statistics provided
in Table 3 rather provide a satisfactory view, as mean, mode and median values for all
classes indicates a healthy trend of submitting assignments on-time and to avoid procras‐
tination. For class ID279 the mean value is 85.4% of maximum values of 105. Whereas
for other classes ID1149, ID1563 and ID1729 the mean values are 28.6%, 41% and
52.3%, although lower than class ID279 but still satisfactory. The standard deviation
values for all four classes are not high indicating consistent behavior of submitting
homework timely. The mean, mode and median values for variables number of home‐
work late and number of homework not hand-in indicate that the students tend not to
procrastinate because it can lower their final scores. In Table 3, mostly zero and low
values for mean, mode and median are observed for these variables. Further, low
standard deviation values indicate that avoiding procrastination is universal trend among
all four classes.
The final score variable is indication of successful learning. The scores have
maximum value of 100. The mean, mode and median values in Table 3 indicate that
despite all odds students manage to get good scores. The low standard deviation speaks
of healthy trend of getting good scores after all among all classes.
However, the distribution of data is not normal and we mostly see skewed distribu‐
tions. The skewness is measure of how much a frequency distribution is asymmetric
[23]. A normal distribution has skewness values to zero, whereas positive or negative
non zero values indicates a positively or negatively skewed distribution. Table 3 indi‐
cates that most distribution are positively or negatively skewed. There are also outliers
Digging Deep Inside: An Extended Analysis of SCHOLAT E-Learning Data 419
present in data which make distribution further skewed. The skewed distribution speaks
of extreme behavior by students, i.e., against a normal distribution in which an average
behavior is mostly seen.
Table 4 shows the results of bivariate correlation coefficient analysis. The analysis
was done to find out the strength of relation between each of independent variables in
each course and the final scores. The values of correlation coefficient r illustrates the
strength of positive or negative relationship, with two extreme values +1 or −1 (strong
positive or negative relation) and value of 0 for no relation at all [24]. The variables have
weak to moderate relationships. But very few significant correlations (p-value < 0.05)
are observed. The course ID279 has only one significant correlation, i.e., number of
homework not hand is negatively correlated with final scores (r = −0.35). In course
ID1149, the number of logins has positive significant correlation with final scores
(r = 0.44) and number of homework late has negative significant correlation (r = −0.39).
In courses ID1563, the number of homework on time has positive moderate significant
correlation with final scores (r = 0.41) and number of homework not hand-in has nega‐
tive moderate significant correlation (r = −0.50). Finally, in course ID1729 the number
of homework on-time has positive moderate significant correlation with final scores
(r = 0.42) and number of homework late has negative moderate significant correlation
(r = −0.50).
Next we examine the course level data. This is data either same for entire group or
figures were not available for individual students. For example, total number of home‐
work is same for all students, i.e., 105, 8 and 13 for classes ID279, ID1149 and ID1563/
ID1729 respectively. For other three variables, number of hits course notice, number of
downloads course resources and number of comments, the figures were not available
for individual students. However, these figures are helpful to find out the extent of use
of beyond class room on-line services. Since, the figures are not available for each
student, we cannot use them to correlate to student’s achievement.
First, we look at variable number of hits course notice. The figures of this variable
for classes under study, i.e., ID279, ID1149 and ID1563/ID1729 are 6041, 553 and 54,
where number of notices issued by course instructors were 30, 18 and 1 respectively.
As such, we see more activity in class ID279 reading and staying in touch with course
instructor. Similarly, we can view number of downloads of course resources as another
indicator of on-line interactive activities. Table 5 indicates that in class ID279 students
downloaded course resources 5969 times, in ID1149 9375 times, and in ID1563/ID1729
2176 times, whereas 61 resources were uploaded in ID279, 63 in ID1149 and only 18
resources were uploaded in classes ID1563/ID1729. We see rather healthy trend that the
students are benefited from this facility effectively, which is indicative by the number
of downloads in each class.
We observe a disappointing activity in terms of comments posted by students. For
our classes, there are only 30 comments for ID279, 3 comments for ID1149 and no
comment for other two classes combine. These comments can be useful for course
instructors and other students willing to join the course.
420 A. Akram et al.
6 Conclusion
7 Future Work
In future, we intend to extend this work for building machine learning algorithm to
predict students’ current learning and future achievement. This would enable course
instructors to take timely actions to avoid students’ failures. We also intend to increase
data collection ability of the system.
Acknowledgment. This work is supported by the National Nature Science Foundation of China
(Grant Nos. 61272066, 61272067, 61300104), the Applied Technology Research and
Development Foundation of Guangdong Province (2016B010124008), the Technology
Innovation Platform Project of Fujian Province (Grant Nos. 2009J1007, 2014H2005), the Fujian
Collaborative Innovation Center for Big Data Applications in Governments.
References
1. Chen, B., Breyer, T.: Investigating instructional strategies for using social media in formal
and informal learning. Int. Rev. Res. Open Distance Learn. 13(1), 87–104 (2012)
2. Hrastinski, S.: A theory of on-line learning as on-line participation. Comput. Educ. 52, 78–
82 (2009)
3. Junco, R., Helbergert, G., Loken, E.: The effect of Twitter on college student engagement
and grades. J. Comput. Assist. Learn. 27, 119–132 (2011)
4. Hurt, N.E., Moss, G.S., Bradley, C.L., Larson, L.R., Lovelace, M.D., Prevost, L.B., et al.:
The ‘Facebook’ effect: college students’ perceptions of online discussions in the age of social
networking. Int. J. Sch. Teach. Learn. 6(2), 2–14 (2012)
5. Fewkes, M., McCabe, M.: Facebook: learning tool or distraction? J. Digital Learn. Teacher
Educ. 28(3), 92–98 (2012)
6. Mostow, J., Beck, J., Cen, H., Cuneo, A., Gouvea, E., Heiner, C.: An educational data mining
tool to browse tutor-student interactions: Time will tell!. In: Proceedings of the Workshop on
Educational Data Mining, pp. 15–22, Pittsburgh, USA (2005)
7. Dewey, J.: Experience and education, New York: Collier, 1938/1963
8. Vygotsky, L.S.: Mind in society: the development of higher psychological processes. Harvard
University Press, Cambridge Mass (1978)
9. Cole, J.: Using Moodle. O’Reilly, Sebastopol (2005)
Digging Deep Inside: An Extended Analysis of SCHOLAT E-Learning Data 421
10. Dawson, S.: A study of the relationship between student social networks and sense of
community. Educ. Technol. Soc. 11(3), 224–238 (2008)
11. Yu, A.Y., Tian, S.W., Vogel, D., Kwok, R.C.: Can learning be virtually boosted? an
investigation of on-line social networking impacts. Comput. Educ. 55, 1494–1503 (2010)
12. Junco, R.: Too much face and not enough book: the relationship between multiple indices of
Facebook use and academic performance. Comput. Hum. Behav. 28, 187–198 (2012)
13. Lamb, L., Johnson, L.: Bring back the joy: creative teaching, learning, and librarian-ship.
Teach. Librarian 38(2), 61–66 (2010)
14. Klosgen, W., Zyttow, J.: Handbook of Data Mining And Knowledge Discovery. Oxford
University Press, New York (2002)
15. Romero, C., Ventura, S., Garcia, E.: Data mining in course management systems: moodle
case study and tutorial. Comput. Educ. 51(1), 368–384 (2008)
16. Romero, C., Ventura, S.: Data Mining in E-Learning, Southampton. Wit Press, UK (2006)
17. Mazza, R., Milani, C.: Exploring usage analysis in learning systems: gaining insights from
visualization. In: Workshop on Usage Analysis in Learning Systems at 12th International
Conference on Artificial Intelligence in Education, pp. 1–6, New York, USA (2005)
18. Mor, E., Minguillon, J.: E-learning personalization based on itineraries and long term
navigational behavior. In: Proceedings of the 13th International World Wide Web
Conference, pp. 264–265 (2004)
19. Talavera, L., Gaudioso, E.: Mining students data to characterize similar behavior groups in
unstructured collaboration spaces. In: Workshop on Artificial Intelligence in CSCL, pp. 17–
23, Valencia, Spain (2004)
20. Akram, A., Chengzhou, F., Yong, T., Yuncheng, J., Xueqin, L.: Exposing the hidden to the
eyes: analysis of SCHOLAT E-learning data. In: Proceedings of the 2016 IEEE 20th
International Conference on Computer Supported Cooperative Work in Design, pp. 693–698
(2016)
21. Romero, C., Ventura, S.: Data mining in education. WIREs Data Min. Knowl. Disc. 3, 12–
27 (2013). https://doi.org/10.1002/widm.1075
22. Huck, S.W.: Reading Statistics and Research, vol. 2(28). Pearson Education, Boston (2012)
23. Huck, S.W.: Reading Statistics and Research, vol. 2(24). Pearson Education, Boston (2012)
24. Huck, S.W.: Reading Statistics and Research, vol. 3(49). Pearson Education, Boston (2012)
Social Network Analysis of China Computer Federation
Co-author Network
Abstract. We extract the data set from 2010 to 2014 of the China Computer
Federation (CCF (http://www.ccf.org.cn)) by the distributed web crawler system
and build the co-author network with R language. In this co-author network,
authors represent nodes and a pair of authors is connected by an edge if they have
co-authored at least one paper over the entire duration. We analyze this network
with social-centric and ego-centric methods to study the situation of co-author
network of CCF and visualize the analysis results. Social-centric measure reveals
that the co-author network density, the usage of key words, the average number
of authors per article, and confirms that most authors in computer science field
publish a very small number of papers but have higher collaborators than those
of other fields. Ego-centric analysis discovers betweenness centrality, closeness
centrality, and degree centrality, indicates that only a small percentage of the
authors locate in the center of the coauthor network. Finally, we pick out eight
key persons and point out the group teams where the key persons are respectively.
Based on these findings, we suggest that computer science field should promote
wider collaboration, encourage more authors to publish their papers.
CCF aims at bringing scholars together in computer research. CCF was founded in 1962,
member of China Association of Science and Technology. It contains 13 kinds of journals,
namely Chinese Journal of Computers, Journal of Computer-Aided Design & Computer
Graphics, Journal of Computer Science & Technology, Journal of Computer Research and
Development, Journal of Software, Journal of Computer Applications, Computer Engi‐
neering and Applications, Computer Technology and Development, Computer Science,
Journal of Chinese Computer Systems, Journal of Frontiers of Computer Science and Tech‐
nology, Computer Engineering and Science, and Computer Engineering and Design.
Social network analysis (SNA) methods have been used to study co-author network
in various fields including physics, library and information service (LIS), biology, and
computer science. Scholars in LIS field need strengthen communication with each other
after studying the co-author network of LIS with betweenness centrality, degree
centrality, and average degree features [1]. Zhu et al. [2] discovered the present situation
and pattern of Chinese-foreign cooperation, simultaneously offered suggestions to the
international collaboration through studying the co-author network of information
system field. Sun et al. [3] figured out the most influential author and paper in the co-
author network with network density, betweenness centrality methods. Shen et al. [4]
applied the vector space model into the identification of scientific research teams within
the co-author network, and revealed the scientific research co-author relationship by
degree of similarity between author vectors. Du [5] extracted the co-author network over
the past 26 years of USIT (User Interface Software and Technology), which is generally
recognized as the top conferences in the Human-Computer Interaction field, and
analyzed the co-author network with SNA methods. El Kharboutly and Gokhale [6]
revealed the collaborative pattern of co-author from the co-author network that extracted
over the entire history of the SEKE conference.
After retrieving and reading paper about social network analysis, we can deeply
understand the collaboration, communication among authors. Thus, we can offer sugges‐
tions on how to improve and strengthen the collaboration and communication among
authors, but no one has investigated into CCF co-author network before. Therefore, it
is necessary for us to study the CCF co-author network. In this paper, the main of
contribution of our work is summarized as follows:
1. Design and implement the distributed crawler system; collect the data set from CCF
through the distributed crawler system; build the co-author network of CCF.
2. Discover the co-author network with social-centric and ego-centric. Social-centric
analyzes the major metrics of the data set and compares the metrics with other field.
Ego-centric reveals the betweenness centrality, closeness centrality, degree
centrality of the co-author network.
3. According to the analysis of key words, we point out the research hotspots and
research trendy, find out the key persons with the higher degree centrality, and gain
the group teams where the key persons are from the co-author network.
The rest of paper is organized as follows: Sect. 2 describes data collection and pre-
processing. Sections 3 and 4 discuss socio-centric and ego-centric analysis respectively.
Section 5 points out key person and the group team. Section 6 concludes the paper and
offers directions for future work.
We extract the data set from the last recent 5 years of CCF by the distributed crawler
system. Table 1 shows the major metrics of CCF co-author network. The distributed
crawler system consists of a master node and several slave nodes. Master node is the
core of the crawler system, and it is responsible for scheduling task, managing process
424 C. Fu et al.
and crawling log, while every slave node is a performer of the task. Furthermore, slave
node is a plug-in base on the Google browser Chrome, thus it can run on any computer
that has chrome and run independently with other programs [7]. Architecture of the
distributed crawler system is presented in the Fig. 1.
Since data obtained from the web by the crawler system are JSON (JavaScript Object
Notation) format, we parsed and formatted data into XML format by program. Further‐
more, in pre-processing we found that some (only 10–20 instances) authors share their
first name and last name. We disambiguated between such authors by their emails,
assuming that authors who share a name but not email represent different individuals.
After pre-processing, we build the CCF co-author network, which authors represent
nodes, the pairwise of authors is connected by an edge if they have co-author at least
Social Network Analysis of China Computer Federation 425
one paper. In addition, this analysis of CCF co-author network is an unweighted graph.
That is, although authors A and B have multiple co-authored relationship, there is only
one edge connect author A with author B. Figure 2 shows the overview of the co-author
network graph in CCF.
3 Social-Centric Analysis
In this section, we discuss social-centric metrics that we computed overall nodes in the
CCF co-author network. We compare these metrics, shown in the Table 1, with those
of other fields.
articles. Authors with 40 or more articles are very rare, with 87 being the maximum. In
order to analyze the distribution of articles per author, we extract the authors data over
the entire data set where per author has 10 or more articles. Figure 4, which shows the
distribution of number articles per author who has 10 or more articles further confirms
that author with 40 or more articles are rare.
Fig. 3. Distribution of number of articles per author with less than 10 articles.
Research hotspot is the trend in special field. In CCF co-author data set, there are 70631
key words. In order to point out the research trend of computer science, all those key
words, that have been use more than 1000, are referred to as Main Key Words (MKWs).
Figure 5, which shows MKWs of CCF co-author, indicates that seven MKWs occupy
31.02%, which is a larger proportion than any other key words. Seven MKWs are data
Social Network Analysis of China Computer Federation 427
mining, support vector machine (SVM), clustering, wireless sensor network, genetic
algorithm (GA), segmentation, and cloud computing. Apparently, they are consistent
with big data, Internet of Things (IoT), image processing, community discovery and
cloud computing research major trend.
The network density (D) is defined as the number of edges E to the number of possible
edges and is given by Eq. (1) [10]. The density of the CCF co-author network is 0.0037,
indicating an overall very sparsely connected network.
2∗E
D= (1)
N(N − 1)
4 Ego-Centric Analysis
Betweenness centrality is used to answer the question of who controls knowledge flows.
Betweenness centrality is defined as the number of the shortest paths from all authors
that pass through the given author [11]. It is an indicator of an author’s centrality in a
428 C. Fu et al.
network or an author’s ability to control the knowledge flows, resources and information
[12]. Betweenness centrality of author v is given by Eq. (2), where Pi,j is the total number
of geodesic linking author i and author j, and Pi,j (v) is all the geodesics linking author i
and author j which pass through author v. Authors with high betweenness centrality can
play a role of “bridge” or “middleman” in the co-author network, also the author with
high betweenness centrality can obtain resources, information and knowledge efficiently
from other authors in the co-author network [13]. Table 2 shows the top 10 authors with
betweenness centrality in CCF co-author network.
∑ Pij (v)
Cb (v) = (2)
i,j≠v Pij
Closeness centrality is used to answer the question of who has the shortest distance to
other authors. Closeness Centrality is defined as the mean length of all shortest paths
from a node to other nodes in the network [14]. It is measured as the average of the
reciprocal distance of an author from others. Closeness centrality of an author v is given
by Eq. (3), where d(v, j) is the distance between authors v and j, while N is the total
number of authors where author v can reach in the co-author network. Closeness
centrality judges how important an author is. As we know, the higher closeness centrality
is, the more important the author is. Moreover, an author with higher closeness centrality
could access or obtain resources in the co-author network more efficiently than others
with lower closeness centrality [15]. Additionally, an author with higher closeness
centrality also indicates that an author can communicate efficiently than those with lower
closeness centrality [12]. Table 3 shows the top 10 authors with closeness centrality in
CCF co-author network.
∑N 1
Cc (v) = (3)
n=1 d(v, j)
Social Network Analysis of China Computer Federation 429
Degree centrality is used to answer the question of who knows the most authors. It is
measured as the tie of author with others. Degree centrality is defined as the number of
links incident upon a node [16]. Degree centrality of author v is given by Eq. (4). Authors
with higher degree centrality represent they are more central to the co-author network.
Table 4 shows the top 10 authors with ten degree centrality in CCF co-author network.
Above the discussion, we try to discover key persons and group teams. According to
the value of centrality, we regard the author as a key person who is in the central position
of CCF co-author network. Moreover, we point out group team members who have a
certain count of collaboration with the key person.
of publishes in the CCF co-author network confirms that key persons not only have
higher degree centrality, but also have higher number of publications.
Table 5. Key persons with degree centrality and number of publications in CCF co-author
network.
Author alias Degree centrality Number of publish
28452 73 85
13772 85 87
20252 71 51
851 58 47
14929 55 57
30312 52 42
28719 59 42
15748 53 56
As usual, key person has a group team or laboratory as a support, furthermore, key person
is also the soul in the group team. Therefore, we point out group teams through collab‐
oration with key person. Firstly, group team appears as a cluster in the co-author network,
that is, a key person has much collaboration with others in the group team and key person
is in the center of group team co-author network. Secondly, those authors who have less
than 3 co-authored relationships with key person may represent collaborations across
institutions but not the member in the group team where key person is. Therefore, while
selecting the group team members, we filter those who have less than 3 co-authored with
the corresponding key person. Figure 6 shows eight key persons and their group team
clusters in the co-author network. The bigger sizes nodes stand for the key persons, other
sizes nodes represent other authors that have collaboration with key persons. From
Fig. 6, apparently, a large proportion of collaboration exists in the group team, but there
is rare collaboration across group teams. Base on the above finding, there will be fruitful
produce if authors collaborate frequently with others across group teams.
In this paper, we describe the process of extracting the data set from last resent 5 years
CCF paper websites and building the co-author network of CCF. We analyze CCF co-
author network using social-centric and ego-centric social network analysis methods to
understand the pattern of author collaboration and communication. Otherwise, we
analyze the MKWs to confirm the research trend in computer science field. Moreover,
we identify the key person in the CCF co-author network through the authors’ degree
centrality. Finally, according to the collaboration with key person, we point out eight
group teams where eight key persons are respectively.
Our future research involves co-authorship order to understand how the pattern of
collaboration has been influenced. Moreover, we will deeply study group team research
interest by analyzing the key words of group team members.
Acknowledgement. This work is supported by the National Nature Science Foundation of China
(Grant No. 61272067), the Applied Technology Research and Development Foundation of
Guangdong Province (2016B010124008).
References
1. Rong, X., Tao, Q.: Social network analysis in LIS field co-relation of empirical research. Libr.
World 01 (2010)
2. Zhu, Q., Fan, Z., Shi, W.: Information systems and foreign co-authored research network.
Inf. Stud. Theory Appl. 11 (2011)
3. Sun, N., Zhu, J., Cheng, H., Wu, Y.: Study co-author relationship based on social network
analysis papers. J. Suzhou Univ. 29(09) (2014)
4. Shen, G., Huang, S., Wang, D.: On the scientific research teams identification method taking
co-authorship of collaboration as the source data. New Technol. Libr. Inf. Serv. 1, 57–62
(2013)
5. Du, R.: UistViz: 26 Years of UIST Coauthor Network Visualization. CMSC 734 (2013)
6. El Kharboutly, R., Gokhale, S.S.: Social analysis of the SEKE co-author network. SEKE
2015-092 (2015)
7. Yang, Z., Cai, Z., Chen, G., Tang, Y., Zhang, L.: Design and implementation of distributed
web crawler for open access journal. J. Front. Comput. Sci. Technol. 10 (2014)
8. Sarigol, E., Pfitzner, R., Scholtes, I., Garas, A., Schweitzer, F.: Predicting scientific success
based on coauthorship networks. arXiv:1402.7268 (2014)
9. Newman, M.E.J.: Coauthorship network and patterns of scientific collaboration. Proc. Natl.
Acad. Sci. U.S.A. 101(1), 5200–5205 (2001)
10. Abbasi, A., Cheung, K.S.K., Hossain, L.: Egocentric analysis of co-authorship network
structure, position and performance. Inf. Process. Manag. 45, 671–679 (2012)
11. Borgatti, S.P.: Centrality and network flow. Soc. Netw. 27(1), 55–71 (2015)
432 C. Fu et al.
12. Freeman, L.C.: Centrality in social networks: conceptual clarifications. Soc. Netw. 1(3), 215–
239 (1979)
13. Jamali, M., Abolhassani, H.: Different aspects of social network analysis. In:
IEEE/WIC/ACM International Conference on Web Intelligence (2006)
14. Lu, H., Feng, Y.: A measure of authors’ centrality in co-authorship networks based on the
distribution of collaborative relationships. Scientometrics 81(2), 499–511 (2009)
15. Borgatti, S.P.: Centrality and network flow. Soc. Netw. 27(1), 55071 (2005)
16. Wasseman, S., Faust, K.: Social Network Analysis: Methods and Applications, vol. 8.
Cambridge University Press, Cambridge (1994)
Detecting Postpartum Depression in Depressed
People by Speech Features
1 Introduction
2 Related Work
Voice is one way of emotional expression. Speech features have been found to be able
to identify different emotions. Nwe et al. [8] reported that classifying voice as different
emotions based on HMM (hidden markov model) had a higher accuracy rate (average
7.7%) than artificial judging, the average rate was up to 78%. Wu et al. [9] used
prosodic and spectral features to identify seven emotions, with the best precision as
91.6%. From the above studies, we know emotional can be predicted based on speech
features. It motivates us to identify mental illness like PPD using speech features. There
have been few studies about PPD patients’ voices. We think those studies about
phonetic changes of depression probably can be generalized to PPD.
The sounds of depressed patients have significant changes because of their illness
[6]. The diagnostic speech features of depression in DSM-5 are described as slow,
volume sank, variation of tone lessen, pause duration increase [4] (P163). Experiments
revealed some specific vocal indicators in depression, such as speech speed slow down
[10], increased pause duration and times [10, 11], shortened duration of utterance [12],
longer initiative time latency [13]. Speech features in depression express with changes
of F0 such as the decreasing of bandwidth, amplitude, energy [10, 14], shrunken F0
range (DF0) [14], weakened intensity [15], variation of frequency spectrum like
shrunken second formant transition [12] and shrunken spectral energy distribution [16],
and so on.
Detecting Postpartum Depression in Depressed People by Speech Features 435
Reviewing recent findings, Cohen and Elvevåg [17] believed that computer-based
assessments of natural language has the potential for measuring speech disturbances in
people with severe mental illnesses. Some researchers attempted to predict depression
via patients’ speech. Mundt et al. [11] stated that the regression model consisted of F0,
pause duration, speech speed, speech duration, etc. could predict depression, and the
explanatory power of model to depression reaching 79.2%. Emerging evidence sug-
gested that speech features have a strong performance in predicting depression, which
obtained a RMSE of 10.17 well below the baseline of 14.12 [18]. The study of Cohn
et al. shown that the accuracy in detecting depression was 79% for vocal prosody [19].
In our study, we choose the depressed voices collected from natural circumstance to
improve ecological validity, differing from the controlled experimental environments in
the previous studies. In clinical diagnosis, it is more crucial and harder to make a
distinction between different mental illnesses than distinguish healthy people from
psychiatric patients. To improve the differential diagnosis among depressive spectrum
disorders, our detective aim is detecting PPD within depression.
3 Methods
3.1 Participants
In this study, patients’ voices were secondary data which acquired from CONVERGE
(China, Oxford and VCU Experimental Research on Genetic Epidemiology) project of
MDD which recorded during interviewing. Our analyses were based on a total of 740
depressed patients recruited from 58 provincial mental health centers and psychiatric
departments of general medical hospitals in 45 cities of 23 China provinces. All
patients were female. They were excluded if they had bipolar disorder, intelligence
deficiency or any type of psychosis. Patients were aged from 30 to 60, the mean age
(standard deviation) of them was 44.4 (8.9). More details of this research include
diagnosis and measures were described in [20].
The first step was matching and exclusion. First of all, we exported 11875 MP3
files from CONVERGE database. We need to match recordings with the results of
psychological scales. We only left those patients who have both voice recording and
questionnaire results. After matching, a total of 4243 patients were remained. The next
step is exclusion. To make sure that enough recordings with enough length were used
in the following steps, the short recordings should be excluded. On the basis of our
experiences, we excluded those recordings which less than half an hour. Finally, a total
of 3964 patients remained in this step.
The second step was screening. There were all kinds of noises in these recordings.
We selected high-quality recordings to avoid the impact of noises on the predictive
results. We divided all recordings into different levels according to the certain evalu-
ation criteria (see Table 1). Finally, 774 recordings of level A were labelled, which
were used for the further process.
In the last and most important step, our mission was cutting and denoising. We
needed to separate the voices of patients from interviewers, and wiped out other noises.
We recruited a few workers to engage in this part of work. The requirements of this
work including: (1) the voice clips should be longer than 5 s; (2) the noises need to be
cut include but not limited to ring, telephone ring, click, voices of other people, and so
on. There were 740 patients remained after denoised, including 21459 voice clips. Each
voice clip is equal to one answer of one patient.
including intensity, loudness, zero cross rate, voicing probability, F0, F0 envelope,
eight linear spectral pair frequencies (LSP) and twelve Mel-frequency cepstral coeffi-
cients (MFCC). Secondly, to investigate the variability of voices, 26 features were
turned into their first-order derivatives. Thirdly, we calculated 19 statistics of those
features mentioned above, such as mean, standard deviation, range, etc. At last, we
acquired 988 (= (26 + 26) 19) speech features.
It is expected that there are some irrelevant and redundant data will weaken the
prediction performance. Therefore, it is not a good idea to input all speech features for
prediction. Speech features should be selected before prediction. For choosing relevant
features and achieving dimension reduction of speech features, we used the Sequential
Floating Forward Selection (SFFS) algorithm.
Data analyses mainly includes classification and correlation analysis. Patients were
divided into two groups in the light of whether they have been diagnosed as PPD or
not. The group labels were considered as golden standard in classification: patients with
PPD were labeled 1, without PPD were labeled 0. As classification, we implemented
SVM and 5-fold cross validation for training and testing different models. To figure out
whether there are salient relationships between the independent variables “number of
features” and “sample size” and the predictive effects, we used partial correlation
analyses to test them. In addition, paired-sample t-test was used in trying out the
impacts of dimension reduction and the content of question on the predictive effects.
4 Results
4.1 Prediction
The rate of PPD group and non-PPD group was kept 1:1 to ensure that sample size has
no obvious impact on predictive results. In order to directly observe the effect of
dimension reduction, we used all speech features and features after dimension reduc-
tion to predict PPD, respectively. The results are respectively shown in Tables 2 and 3.
We list the results of top ten best-performing questions, and order by the sample size
from small to large.
In Table 2, the classification was based on 988 speech features. The best predictive
result of F-measure is 65%, which is the reply to one question of depression scale.
Observation of row 4–9, we found that different questions with same sample size had
different predictive powers, considering the speech features used in these ten questions
were the same. In addition, the predictive effects of demographic answers were com-
mon lower than the other questions.
In Table 3, the number of features were dramatically reduced after dimension
reduction. The best predictive result 69% of the selected features is PSY.3, which is the
reply to one question of psychosis scale. The average value of F-measure was increased
5.2%. By looking into row 4–9, we found that the differences of predictive effects of
different questions with same sample size decreased.
438 J. Wang et al.
PPD and non-PPD Confusion matrixes were shown in Table 4. The precision of
PPD was markedly improved after dimension reduction, reaching 75%. In contrast, the
precision of non-PPD had slightly decreased after dimension reduction.
Detecting Postpartum Depression in Depressed People by Speech Features 439
Fig. 2. The impact of sample size on predictive indexes (dr, dimension reduction)
To make it clear that if dimension reduction can evidently improve the predictive
effect or not, we compared the predictive results of total features with reduced features
by using paired-sample t test (Table 5 lists the means and standard deviations of
440 J. Wang et al.
predictive indexes). The results indicated that the predictive results improved signifi-
cantly after dimension reduction (precision: t = −4.763, p < .01; recall: t = −4.31,
p < .01; F-measure: t = −4.061, p < .01). Further analysis, considering the impact of
question, the contrast of three pairs questions with same sample size was ran by
paired-sample t test. The results shown that the differences between different questions
with same sample size had saliently shrank after dimension reduction (120 sample size:
t = −1, p = .42; 170 sample size: t = 7, p < .05). The results of the sample size of 130
cannot test t value because their SD is zero. But their difference value’s change is the
largest after dimension reduction.
5 Discussion
The purpose of this study is to detect PPD in depressed patients via speech features. We
used openSMILE for feature extraction, selected SFFS for feature selection, tried
different numbers of features, set 5-fold cross validation for strengthening generaliz-
ability of model and applied SVM in Weka for training and testing different models.
The best performance of F-measure reaching 69%, comparing with the random pre-
dictive effect 50%, which suggests that voice could be used as a potential behavioral
indicator to identify depression disorder’ subtypes.
We speculate there may be important influences of the number of features, sample
size and the content of question on the predictive effect. Our results indicated that the
number of features has no significant relationship with the prediction, but the predictive
effect is dramatically improved after dimension reduction. The number of features
among different questions are different after dimension reduction, so we think the
positive impact of dimension reduction on predictive results is a combined result of
number of features and content of question. It is unexpected that there is a negative
correlation between sample size and predictive effect. The probable cause is demo-
graphic questions lack of the ability of emotional induction, which results in the
undistinguishable neutral emotion in all patients’ voices. Just as it is shown in Fig. 2,
questions D2.B and D4.A make significant contributions to obvious dents of curves.
Different question has different predictive effect. We can find some clues by the
most effective questions in Tables 2 and 3. The most effective question is DEP.E24.F
before dimension reduction. This question asked patients to recall the state in their
severest depressive episode. The most effective question is PSY.3 after dimension
reduction. This question asks, “have you ever taken medicine for your nerves or the
way you were feeling or acting?”. All patients are recurrent depressive sufferers, they
Detecting Postpartum Depression in Depressed People by Speech Features 441
6 Conclusion
In this study, the best predictive performance of our speech-based models is F-measure
69%, which suggests that the speech features could be used as a potential behavioral
indicator to identify PPD in depressed patients. A combined impact of features and
question contribute to the improvement of predictive effect. After dimension reduction,
the average value of F-measure was increased 5.2%, and the precision of PPD was rose
to 75%. Compared with neutral demographic questions, the features of emotional
induced questions have better predictive effects.
Acknowledgments. This work was supported by the National Basic Research Program of China
(973 Program) (No. 2014CB744603), and Natural Science Foundation of Hubei Province
(2016CFB208).
References
1. Pitt, B.: ‘Atypical’ depression following childbirth. Br. J. Psychiatry 114(516), 1325–1335
(1968)
2. Burke, L.: The impact of maternal depression on familial relationships. Int. Rev. Psychiatry
15(3), 243–255 (2003)
3. American College of Obstetricians and Gynecologists. Committee on Obstetric Practice,
Committee opinion no. 453: Screening for depression during and after pregnancy. Obstet.
Gynecol. 115(2 Pt 1), 394–395 (2010)
4. Accounts Payable Association: Diagnostic and Statistical Manual of Mental Disorders
(DSM-5®). American Psychiatric Publishing (2013)
5. Kramer, E.: Elimination of verbal cues in judgments of emotion from voice. J. Abnorm. Soc.
Psychol. 68(4), 390–396 (1964)
6. Cannizzaro, M., Harel, B., Reilly, N., Chappell, P., Snyder, P.J.: Voice acoustical
measurement of the severity of major depression. Brain Cognit. 56(1), 30–35 (2004)
7. Sobin, C., Sackeim, H.A.: Psychomotor symptoms of depression. Am. J. Psychiatry 154(1),
4–17 (1997)
8. Nwe, T.L., Foo, S.W., De Silva, L.C.: Speech emotion recognition using hidden Markov
models. Speech Commun. 41(4), 603–623 (2003)
9. Wu, S., Falk, T.H., Chan, W.-Y.: Automatic speech emotion recognition using modulation
spectral features. Speech Commun. 53(5), 768–785 (2011)
10. Ellgring, H., Scherer, P.K.R.: Vocal indicators of mood change in depression. J. Nonverbal
Behav. 20(2), 83–110 (1996)
11. Mundt, J.C., Vogel, A.P., Feltner, D.E., Lenderking, W.R.: Vocal acoustic biomarkers of
depression severity and treatment response. Biol. Psychiatry 72(7), 580–587 (2012)
442 J. Wang et al.
12. Flint, A.J., Black, S.E., Campbell-Taylor, I., Gailey, G.F., Levinton, C.: Abnormal speech
articulation, psychomotor retardation, and subcortical dysfunction in major depression.
J. Psychiatr. Res. 27(3), 309–319 (1993)
13. Mandal, M.K., Srivastava, P., Singh, S.K.: Paralinguistic characteristics of speech in
schizophrenics and depressives. J. Psychiatr. Res. 24(2), 191–196 (1990)
14. Porritt, L.L., Zinser, M.C., Bachorowski, J.-A., Kaplan, P.S.: Depression diagnoses and
fundamental frequency-based acoustic cues in maternal infant-directed speech. Lang. Learn.
Dev. 10(1), 51–67 (2014)
15. Cohen, A.S., Kim, Y., Najolia, G.M.: Psychiatric symptom versus neurocognitive correlates
of diminished expressivity in schizophrenia and mood disorders. Schizophr. Res. 146(1–3),
249–253 (2013)
16. Tolkmitt, F., Helfrich, H., Standke, R., Scherer, K.R.: Vocal indicators of psychiatric
treatment effects in depressives and schizophrenics. J. Commun. Disord. 15(3), 209–222
(1982)
17. Cohen, A.S., Elvevåg, B.: Automated computerized analysis of speech in psychiatric
disorders. Curr. Opin. Psychiatry 27(3), 203–209 (2014)
18. Cummins, N., Joshi, J., Dhall, A., Sethu, V., Goecke, R., Epps, J.: Diagnosis of depression
by behavioural signals: a multimodal approach. In: Proceedings of the 3rd ACM
International Workshop on Audio/Visual Emotion Challenge, New York, NY, USA,
pp. 11–20 (2013)
19. Cohn, J.F., et al.: Detecting depression from facial actions and vocal prosody. In: 2009 3rd
International Conference on Affective Computing and Intelligent Interaction and Work-
shops, pp. 1–7 (2009)
20. Yang, F., et al.: Clinical features of and risk factors for major depression with history of
postpartum episodes in Han Chinese women: a retrospective study. J. Affect. Disord. 183,
339–346 (2015)
21. Eyben, F., Weninger, F., Gross, F., Schuller, B.: Recent developments in openSMILE, the
Munich open-source multimedia feature extractor. In: Proceedings of the 21st ACM
International Conference on Multimedia, New York, NY, USA, pp. 835–838 (2013)
Research on Network Public Opinion
Propagation Mechanism Based
on Sina Micro-blog
Abstract. In the era of micro-blog, network public opinion has become the
main expression of public opinion. Network public opinion can quickly form a
network of public opinion, and promote the dissemination of information fis-
sion, with a strong interaction and effectiveness. In the case of “A girl suffered
attacks in a Yitel”, we construct network propagation diagram by ucinet soft-
ware and research the structural characteristics of the network public opinion
propagation network based on Sina micro-blog, and the whole structure of the
propagation network and the key nodes are measured. The results show that the
key nodes in public opinion network propagation has a high ability to spread, so
we can control the velocity of micro-blog public opinion through affecting these
key nodes.
1 Introduction
opinion in micro-blog, and it has very important reference significance to the monitor
and guide of network public opinion.
The information of Sina micro-blog spread mainly in the following ways: the release
function, @ function, forwarding function and comment function [1]. These four ways
can quickly and effectively release information, so that this information can spread in
the Sina micro-blog and can be concerned by more users. The propagation of
micro-blog’s network public opinion is a very complex process, and it is not a single,
linear propagation mode. However, based on the sub-public communication formed by
the users’ concern function, it develops into mass communication by the first level
information receiver who forward information through forwarding function for the
secondary spread, and then, it turns into a multi-level propagation mode with the
traditional mass media platform. In this paper, we selected “A girl suffered attacks in a
Yitel” as a case, and we collected data and tried to build the propagation analysis
framework of micro-blog public opinion events to study the evolutionary mechanism of
network public opinion.
propagation relationship between each node, and constructed the propagation network
matrix of micro-blog public opinion of this event, this network matrix was
anon-symmetric matrix of 502*502, and the data indicated the nodes’ frequency of
emergence in the sample data. Based on the matrix, using UCINET software for
visualization processing of the propagation network for this event (Fig. 1).
Fig. 1. Network propagation diagram of “A girl suffered attacks in a Yitel” (Numbered Edition)
structural holes in the propagation network can be analyzed [2]. By using UCINET
software, we analyzed part nodes of structural holes in micro-blog public opinion
propagation network of “A girl suffered attacks in a Yitel”, and we sort out the effective
scale to get the Table 1. We funded that No. 1 “Wanwan_2016”, No. 2, “Ctrip”,
No. 3 “Safety Beijing”, No. 5, “Homeinns Co., Ltd” and No. 4 “and “Yitel” ranked the
top five. The value of effective scale reflects the position of nodes in the propagation
network, and the bigger the value is, the more core the position becomes [2]. In
addition to the effective scale, the limiting degree of these five nodes are relatively
small, they are less than 0.2, and it also reflects that the five nodes are not easy to be
controlled by other nodes and easier to access and spread public opinion.
Table 1. Analysis of propagation network structural holes of “A girl suffered attacks in a Yitel”
(part)
Nodes Degree Effective Efficiency Limiting Grade
scale degree degree
1Wanwan_2016 199 152.078 0.764 0.176 0.904
2Ctrip 100 76.185 0.762 0.166 0.795
3Safety Beijing 87 64.389 0.740 0.196 0.810
5Homeinns Co., Ltd 82 60.809 0.742 0.190 0.795
4Yitel 77 56.717 0.737 0.198 0.787
6Reporter 28 19.286 0.689 0.223 0.451
hong-taoXue
10 Police Wang 15 9.090 0.606 0.281 0.195
7 People outside the 15 8.825 0.588 0.306 0.273
city
27 Update step by step 11 5.583 0.508 0.327 0.065
Table 4. Measurement results of network density and distance between nodes of “A girl
suffered attacks in a Yitel”
Density Standard Average Distance-based Distance-weighted
deviation distance cohesion fragmentation
0.0102 0.1018 3.193 0.344 0.656
According to the statistical results in Table 4, the network density of this public
opinion event is only 0.0102, the result shows that the network density is very small in
the propagation process of the public opinion events, the links between nodes are more
dispersed, and the exchange of information is not frequent. The distance between nodes
of the propagation network is 3.193, cohesion distance-based is 0.344, and fragmen-
tation distance-weighted is 0.656. This result shows that the propagation ability of the
450 W. Huang et al.
public opinion information in data samples is general and cohesion is not strong. The
propagation probabilities of distance from 1 to 4 are 1%, 26.1%, 25.4% and 47.4%, so,
the propagation ability of sub-public propagation is general, however, the relationship
links whose distance is 4 are nearly half, and it also shows that the mass communi-
cation has a huge effect in this event.
The clustering coefficient CC of the whole network is the average of the clustering
coefficients of all nodes j:
P
N
CCðjÞ
j¼1
CC ¼ ð3Þ
N
The value of clustering coefficient ranged from 0 to 1, the greater the number of
clusters is, the stronger the cohesive force of the whole propagation network is, and the
links between nodes are more closely. The measure results in Table 5 show that
efficient clustering coefficient is 0.1, in propagation network in the public opinion event
of “A girl suffered attacks in a Yitel”. The clustering coefficient is too small, and it
means that there is a relatively low level of information communication between nodes.
At the same time, the network structure is relatively loose, the community structure and
the state of the internal substructure are not obvious, and the connection between the
nodes is relatively weak.
In this paper, we use three indexes, the network density, the distance between nodes
and the clustering coefficient [7–9], in order to analyze the overall network structure of
the micro-blog public opinion of “A girl suffered attacks in a Yitel”. According to
measure results, the network density is very small in the propagation network of the
public opinion events, it is only 0.0102, the links between nodes are more dispersed,
and the exchange of information is not frequent. The propagation ability of sub-public
communication is general; however, the relationship links whose distance is 4 are
nearly half, so it also shows that the mass communication has a huge effect in this
event. In spite of this, the cohesive force of propagation network is weak and lack of
communication between nodes and nodes, and the forwarding relationship does exist,
but it did not communicate with other nodes for effective communication after the
forwarding, making public opinion propagation effect cease to advance. For this
micro-blog public opinion event, information is more concentrated on the party and the
company involved, “Wanwan_2016” and “Yitel”, and they are the key communicators
of this event. For “Safety Beijing”, “Reporter_ hongtaoXue” and other key nodes, etc.,
they only have single forward relationship rather than mutual forwarding relationship,
and the connectivity between nodes is poor, so it is difficult to form a small group.
4 Conclusion
In this paper, we measured the overall network structure that are composed of nodes and
the key nodes of the propagation network through statistics and arrangement on the
relationship of micro-blog public opinion propagation node. Through the analysis of the
measurement results, it is found that it is difficult to form a small group because of the
lack of cohesion between the nodes of this event. Investigating its deep reason, the event
of “A girl suffered attacks in a Yitel” is a vicious incident caused by the individual party,
and in the information propagation of micro-blog public opinion, it is focused on the
hotel’s indifference to violence that complained by the party, companies involved also
issued a statement and it said they would investigate it to the end, in order to give the
party and the public an account. Although the forwarding quantity in the short term of
this incident is very high, and it also appeared the “bridge” node as the “opinion leader”,
and there is a lack of cohesion between the nodes of the propagation network, and there
is no small group, and the communication between the nodes is weak.
Therefore, the parties of public opinion events should actively enhance commu-
nication with other users, and it needs more media, celebrities and government
micro-blog to pay attention to public opinion events so as to promote the transfer of
public opinion information through their own influence. On the one hand, the parties
should strengthen the communication with other users and select the representative
users to forward their questions for each other after forwarding their own micro-blog. It
will not only form a small group based on forwarding, but also can enhance the
cohesion between the user nodes, so that public opinion information can spread rapidly.
On the other hand, we can develop opinion leaders to achieve the effect of guiding
public opinion. The “opinion leader” in Sina micro-blog refers to a kind of commu-
nicator that has appeal and influence in the propagation process of network public
opinion. This speech of information communicator is very easy to cause the recognition
452 W. Huang et al.
of other users, and the users use comment and forwarding function to make it spread in
the micro-blog, resulting in a greater impact on public opinion, which has a greater
impact on public opinion. Therefore, we need to have the opinion leaders to exert their
influence and spread positive energy, in order to achieve the purpose of infects and
affects other user groups.
Acknowledgment. Work described in this paper was funded by the National Natural Science
Foundation of China under Grant No. 71671093. The authors would like to thank other
researchers at the Nanjing University of Posts and Telecommunications.
References
1. Russo, T.C., Koesten, J.: Prestige centrality and learning: a social network analysis of an
online class. Commun. Educ. 54(3), 254–261 (2005)
2. Jinlou, Z., Junhui, C.: Analysis of micro-blog public opinion diffusion based on SNA: an
empirical study on April 20 Ya’an Earthquake in Sichuan. Manag. Rev. 01, 148–157 (2015)
3. Wei, K.: Analysis of the key nodes in public opinion spread during emergencies based on
social network theory—a case study of the 723 Wenzhou high-speed train collision. J. Public
Manag. 03, 101–111, 127–128 (2012)
4. Xiaojuan, H., Nan, J., Jinjin, X.: Study on micro-blog rumor based on social network
analysis—a case study on the micro-blog about food security. Intell. J. 08, 161–167 (2014)
5. Wei, J.C., Bu, B., Liang, L.: Estimating the diffusion models of crisis information in
micro-blog. J. Inform. 6(4), 600–610 (2012)
6. Lee, C.H.: Mining spatio-temporal information on microblogging streams using a
density-based online clustering method. Expert Syst. Appl. 39(10), 9623–9641 (2012)
7. Nair, H.S., Manchanda, P., Bhatia, T.: Asymmetric social interactions in physician
prescription behavior: the role of opinion leaders. J. Mark. Res. 47(5), 883–895 (2010)
8. Xin, M.J., Wu, H.X., Niu, Z.H.: A quick emergency response model for micro-blog public
opinion crisis based on text sentiment intensity. J. Softw. 7(6), 1413–1420 (2012)
9. Xia, Z.Y., Yu, Q., Wang, L.: The public crisis management in micro-bloging environment:
take the case of dealing with governmental affairs via micro-blogs in China. Adv. Intell. Soft
Comput. 141, 627–633 (2012)
Scholar Recommendation Model in Large Scale Academic
Social Networking Platform
Ming Chen1, Chunying Li2, Jiwei Liu1, Dejie Meng1, and Yong Tang1 ✉
( )
1
School of Computer Science, South China Normal University, Guangzhou, China
mchencf@163.com, {cocoliu,2016022321,ytang}@m.scnu.edu.cn
2
School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
zqxylcy@163.com
1 Introduction
In recent years, with the advent of WeChat, micro-Bo, Facebook and other social
networking platform, large-scale social networks have developed rapidly. The greatest
charm of a social network is Social, and each social networking platform creates many
explicit or hidden circle of friends [1]. With the rapid development of intelligent termi‐
nals, a rapid growth of network data has been seen in social networking platform, so
people often encounter the problem of information overload when they find interested
friends. Therefore, the provision of friends recommended in social networking platform
can help users more quickly and more accurately find their potential friends. As a kind
of social platform, academic social networking platform also need a scholars recommend
system to help users brush selected interested scholars.
At present, the domestic and foreign scholars have proposed many friend recom‐
mendation algorithm to solve the mentioned above problems, which can be divided into
two categories. One is the friend recommendation algorithm based on the user’s existing
information. In the literature [2], a friend recommendation model is established by
obtaining user’s preferences and information that already exists on the social network
platform, and the model explores and recommends friends who may have intersects with
the user on the view of user’s existing information. The algorithm in [3] analyzes the
microblog content published by the user to discover the interest of the user, and then
according to the user’s interest to recommend it may be interested friends. The paper [4]
aimed at some friends recommend algorithms, but ignored the relationship between
individuals and timing factors, this paper proposed a method to model the temporal
behavior among users based on the extraction of the user’s existing information, so they
could found the neighbour collection that have the greatest influence on the current user,
and then merged the collection in-to a coordinated ltering recommendation algorithm
which based on probability matrix decomposition. In the literature [5], the non-topology
information were used to calculate the similarity among users, and then recommended
the potential friends who were similar to the target user according to the similarity
between users. Another friends recommendation algorithm was based on social network
topology modeling. In the literature [6], the separation degree of each node was calcu‐
lated by dividing the topological graph of the buddy relationship in the social network.
Based on the degree of separation, the algorithm divided the nodes with the same degree
of dissimilarity into the same community group, and then recommended nodes within
the same group reciprocally.
Considering the unique social nature of academic social networks [7–9], we propose
a scholars recommendation algorithm based on label community discovery, which is a
friend recommendation algorithm based on social network topology modeling. The
Label Propagation Algorithm (LPA) [10] proposed by Raghavan et al. in 2007 firstly,
which is a community discovery algorithm based on label propagation. LPA is relatively
simple. The algorithm first assigns a unique label to all the nodes, and then flushes the
labels of all the nodes until the convergence requirements are reached, and finally get
the non-overlapping community. In recent years, many scholars [11–13] have improved
the LPA algorithm from different angles in view of its simple and high efficiency. For
the LPA can only be applied to the discovery of non-overlapping community problems,
Gregory proposed the Community Overlap Propagation Algorithm (COPRA) [14] algo‐
rithm in 2009, which was applicable to overlapping communities. Concerning the
problem that community detection algorithm based on label propagation in complex
networks has a pre-parameter limit in the real network and redundant labels, LI Chunyin
proposed the Adaptive Label Propagation Algorithm (ALPA) [15]. The algorithm used
the adaptive threshold to eliminate the unreasonable label in the iterative process, and
finally classified the node with the same label into a community. ALPA algorithm has
achieved good results, especially in the academic social network. But it is not ideal when
dealing with social networks with large user population and complex user relationships,
because the time complexity of the algorithm is relatively high in this case. To solve
this problem, this paper proposes Community Detection Based on GraphChi (CDBG)
algorithm which is a community discovery scholars recommendation algorithm based
on GraphChi system. CDBG uses the characteristics of GraphChi system to achieve
ALPA algorithm in parallel, and then uses the community clustering results to achieve
friend recommendations. This algorithm improves the response speed of friends recom‐
mendation.
Scholar Recommendation Model 455
2 Recommended Approach
The main idea of the model is to explore the community according to the topology
diagram of the user relationship in the social network, and then recommend a friend in
the community. Finding out core networks and then building the community by tag
propagation is a better-performing and easy-to-implement approach. However, it is very
time-consuming to find the core network in the massive graph data, which makes the
recommendation system unable to meet the user’s psychological response time. There‐
fore, it is very important to improve the search strategy of the core network. The strategy
of this paper that to use the concurrent mechanism of Graphchi system to split the
topography of the buddy relationship, and then find out the core network from each part.
Each local topology to find out the core network. As the number of local nodes is much
smaller than the whole, it improves the speed of finding the core network, thus improving
the response time of the whole recommendation system.
2.1 GraphChi
The Graphchi system is a stand-alone version of the large-scale graphing system intro‐
duced by GraphLab Labs. Although it is not based on the popular distributed architec‐
ture, the efficiency of this system is very high, and it is not inferior to the distributed
computing system [16] in dealing with large-scale data. The ingenious design allows
Graphchi to efficiently process large-scale map data. In order to speed up the efficiency
of the system, Graphchi load the data into memory. This approach is similar to other
large scale computing systems. Since the memory is very limited in stand-alone envi‐
ronment, it is impossible to load all the data into memory when it encounters large graph
data. So Graphchi first cut into a number of small map, and then load the sub-graph data
into the memory for each calculation. Loading each subgraph one after the other can
complete an iteration. After several rounds of iterations, you can complete the task of
graph calculation. The above process is the main idea of Parallel Sliding Windows
(PSW).
The running process of PSW has completed the update of the sub graph data. Each
round diagram update process includes three main steps: the loading of sub-graph data,
the concurrent implementation of node update function, write back the updated sub-map
data. When Graphchi completes the three sequential steps mentioned above, it completes
the update of a subgraph. When all the P intervals are updated, a round of iteration of
the entire graph is completed.
As can be seen from the above algorithmic flow, the key of Craphchi to handle
massive amounts of data in a single-machine environment is the data is sliced and then
stored in memory. Because only need to read a piece of information at a time, so long
as the external memory space is large enough to be able to adjust the value of P to make
the memory to meet the computing requirements. Graphchi uses a vertex-centric
programming model in which adjacent vertices pass messages between edges. Devel‐
opers only need to consider a single vertex update function, and Graphchi framework
to solve specific details such as vertex-parallel.
456 M. Chen et al.
Definition 2: Core network. In graph G = {V, E}, assume that A and B are any two
nodes in V that are not tagged. If A and B are the maximum degree and unmarked nodes
of each other, then the edges A to B are used to find the complete graph for the initial
edge. If Gs ∈ G and there is no any complete graph Gt ∈ G making Gs ∈ Gt, then Gs is a
core network of graph G.
(c) Normalize the labels in node v, and finally make the sum of the weights in each
node equals to 1.
3. Repeat steps 2–4 until all nodes have at least one label;
4. Traverse all nodes and classify nodes with the same label as the same community.
If a node has more than one label, it is categorized into multiple communities;
5. Community consolidation based on Definition 3.
According to the above rules, now take the Fig. 1 as an example to update the label.
Eventually generated two communities, namely C1 = {1, 2, 3, 4, 5}, C2 = {6, 7, 8}. The
results of community detection are shown in Fig. 2.
In a social network, users in the same community tend to have similar interests or areas
of work, and users in the same community are more likely to accept it. Therefore, the
recommendations in this paper are carried out in the divided communities. The CDBG
algorithm is used here to classify communities. After the community division, this model
uses three methods to make personalized recommendation, namely: Community Weight
Recommendation (CWR), Acquaintance Community Recommendation (ACR) and
Community Random Recommendation (CRR).
After the establishment of the scholar community using the CDBG algorithm, each
scholar may belong to several different communities. In the CWR recommendation
mode, all the scholars who are in the same community as the target scholar A and not
the friends of A are sorted by weight from high to low. Then the former N highest weight
scholars recommended to the scholar A, which is a CWR recommendation. This recom‐
mendation is suitable for recommending scholars to Ph.D. and professors because we
assume that highly educated user groups are more likely to accept influential scholars
in the same community. The ACR recommended method selects the community with
the least number of nodes among the communities to which the target scholar A belongs,
and ranks the users of the community in descending order of weight, and then recom‐
mends the first N highest weight scholars to the scholar A. This recommendation is
recommended for scholarships for Masters and below, as these users have limited
communication and academic skills and are more willing to accept scholars in acquaint‐
ance communities. For example, a college counselor is more willing to accept the college
Scholar Recommendation Model 459
clerk and other college counselors in her acquaintance community than a well-known
scholar in a field. CRR recommended method is the most simple. The CRR randomly
recommends nodes that are in the same community as Target A and whose weights are
higher than the specified value. This approach is suitable for those who have just joined
the academic social networking platform and friends and team information is scarce. It
is a cold start recommended way. Take the CWR method as an example, the process of
Scholar Recommendation is illustrated below:
460 M. Chen et al.
3 Experimental Results
In this part, we first introduce the data set used in the experiment, and then explain the
evaluation method of the experiment. Finally, we give the experimental results of CWR,
CRR and ACR, and analyze the experimental results.
The data set used in the experiment is the friend relationship data set of the academic
social network platform (SCHOLAT) on October 12, 2016. The data set records the
relationship between scholars and friends on the social network platform. After
denoising the data set, there are 5168 users whose information is disclosed and 22284
user relations.
In the experiment, we used four general indicators [17] to evaluate the recommended
results, namely: Precision, Recall, F-measure and MAP. These four methods are the
standard to measure the accuracy of the recommendation system, and reflect the accu‐
racy of the recommendation system to the specified users, where Precision is the accu‐
racy of the recommended system recommendations, and it can be defined as follows:
1 ∑ Nt
P= (1)
T u L
Where T is the number of experimental test samples, and Nt is the number of recom‐
mended objects the user likes in each recommendation, and L is the recommended list
length in a recommendation.
Recall expresses the possibility that the user’s favorite object is recommended by
the recommended system, and it can be defined as follows:
1 ∑ Nt
R= (2)
T u Au
Where Au represents the total number of objects in the test set that are accepted by
the recommended user.
F-measure is the weighted harmonic average of Precision and Recall, and when the
value is high, the test result is more effective. It can be defined as:
( )
a2 + 1 P ∗ R
F1 = (3)
a2 (P + R)
When a = 1, the formula is the most common F1-measure, and it defined as:
Scholar Recommendation Model 461
2∗P∗R
F1 = (4)
P+R
The final evaluation criterion is MAP, which reflects the average accuracy of the
recommended system and is the probability of recommending an object to be accepted.
it can be defined as:
( L ( ))
1 ∑ ∑ p Ljk
T
MAP = (5)
T j=1 k=1 mj
Where p(Ljk) represents the ratio of the number of users preferred in the first K
recommended objects in the jth recommendation to k. j represents the number of objects
that the user likes in the jth recommendation. L represents the length of the recommended
list, and we get L = 10 and L = 5.
In the data set described above, we use the community recommendation model proposed
in this paper to generate a total of 442 communities, with the largest community having
1986 user nodes and the smallest community having 3 nodes. The total time it takes to
generate the community is 1.508 s. In order to verify the validity of this model and
compare the advantages and disadvantages of CWR, CRR and ACR recommendations,
we calculate the Precision, Recall, F-measure and MAP by questionnaire survey. These
questionnaires are divided into two categories, one with a recommended length of 10
and another of 5; Table 1 gives the experimental results for L = 10 and L = 5. From
Table 1, we can see that no matter which kind of recommendation method, the recom‐
mended accuracy rate is above 55%, and the average accuracy MAP can be maintained
above 55%. Additionally, when the ACR recommended method is adopted, the accuracy
can reach 77.68% and the MAP can reach 81.63% under the recommended length L = 10.
In this way, when the recommended length L = 5, the accuracy is 84%, MAP is 85.84%.
This is a very good result, and at the same time, it proves the effectiveness of the proposed
model in this paper.
By comparing the results in Table 1, it can be seen that the CWR recommendation
method has the worst performance among the three methods. There are two reasons for
this situation. First of all, in the scholar network social network platform, the weight of
a scholar is not exactly the same as the scholar has great influence, because some scholars
often use the scholar network platform for teaching, thus accumulating a large number
of students friends to increase their weight, but this weight does not affect the choice of
other people; In addition, a highly qualified scholar does not guarantee that he is well
known by other scholars, nor can he guarantee that the direction of his research is to
accept the recommendation of users interested in. And the majority of users of social
networking platform is a master’s degree, so the user’s academic social and research
areas are relatively limited. So the CWR method is more suitable to recommend scholars
to doctoral and professors and other highly educated users. As can be clearly seen in
Table 1, the ACR method is most effective. The reason for this result is that there is a
high possibility that users have a connection in acquaintance communities, so the user
is more inclined to accept a scholar in the community. In addition, by comparing the
two results of L = 10 and L = 5, it can be seen that there is no significant difference in
the effect of CRR between the two lengths. However, CWR perform better at L = 5, but
ACR at L = 10 gives better results. This just verifies the above analysis.
Figure 3 shows the change in the recommended results with the length of the recom‐
mended list. As can be seen from the figure, Recall and F1-measure are more sensitive
to the recommended list length, while Precision and MAP are less affected by the
recommended length. In addition, it can be seen from the diagram that the ACR method
can achieve better results under different recommended lengths. It can be seen that the
use of ACR recommended in many cases can get the best results. But the ACR method
Scholar Recommendation Model 463
References
1. Zhang, Z., Li, Q.: Questionholic: hot topic discovery and trend analysis in community
question answering systems. Expert Syst. Appl. 38, 6848–6855 (2011)
2. Colace, F., Santo, M.D., Greco, L., Moscato, V., Picariello, A.: A collaborative user-centered
framework for recommending items in online social networks. Comput. Hum. Behav. 51,
694–704 (2015)
3. Armentano, M.G., Godoy, D., Amandi, A.A.: Followee recommendation based on text
analysis of micro-blogging activity. Inf. Syst. 38, 1116–1127 (2013)
4. Sun, G.F., Wu, L., Liu, Q., Zhu, C., Chen, E.H.: Recommendations based on collaborative
filtering by exploiting sequential behaviors. J. Softw. 24, 2721–2733 (2013)
5. Debnath, S., Ganguly, N., Mitra, P.: Feature weighting in content based recommendation
system using social network analysis. In: International World Wide Web Conference, pp.
1041–1042 (2008)
464 M. Chen et al.
6. Silva, N.B., Tsang, I.R., Cavalcanti, G.D.C., Tsang, I.J.: A graph-based friend
recommendation system using genetic algorithm. In: Evolutionary Computation, pp. 1–7
(2010)
7. Gruzd, A., Staves, K., Wilk, A.: Connected scholars: examining the role of social media in
research practices of faculty using the UTAUT model. Comput. Hum. Behav. 28, 2340–2350
(2012)
8. Nández, G., Borrego, Á.: Use of social networks for academic purposes: a case study.
Electron. Libr. 31, 781–791 (2013)
9. Gu, F., Widén-Wulff, G.: Scholarly communication and possible changes in the context of
social media: a finnish case study. Electron. Libr. 29, 762–776 (2011)
10. Raghavan, U.N., Albert, R., Kumara, S.: Near linear time algorithm to detect community
structures in large-scale networks. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 76(3 Pt 2),
036106 (2007)
11. Liu, X., Murata, T.: Advanced modularity-specialized label propagation algorithm for
detecting communities in networks. Physica A Stat. Mech. Appl. 389, 1493–1500 (2010)
12. Cordasco, G., Gargano, L.: Community detection via semi-synchronous label propagation
algorithms, pp. 1–8 (2011)
13. Xie, J., Szymanski, B.K.: Community detection using a neighborhood strength driven label
propagation algorithm. In: IEEE NSW, pp. 188–195 (2011)
14. Gregory, S.: Finding overlapping communities in networks by label propagation. New J. Phys.
12, 2011–2024 (2009)
15. Li, C., Huang, Y., Tang, Z., Tang, Y., Zhao, J.: Adaptive label propagation algorithm to detect
overlapping community in complex networks. Int. J. Future Gener. Commun. Netw. 9, 317–
326 (2016)
16. Kyrola, A., Blelloch, G., Guestrin, C.: GraphChi: large-scale graph computation on just a PC.
In: Usenix Conference on Operating Systems Design and Implementation, pp. 31–46 (2012)
17. Lichtenwalter, R.N., Lussier, J.T., Chawla, N.V.: New perspectives and methods in link
prediction. In: ACM SIGKDD International Conference on Knowledge Discovery and Data
Mining, Washington, DC, USA, July, pp. 243–252 (2010)
Research on Simulation for Fuel Consumption of UAV
1
College of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, China
jiazongpu@126.com
2
China Shipbuilding Information Technology Co., Ltd., Beijing, China
shiyh@csit.net.cn
3
College of Computer Science and Technology, Jilin University, Changchun, China
gusongyuan614@163.com, liusf@mail.jlu.edu.cn
1 Introduction
Aerodynamic parameters contain lift coefficient, drag coefficient, lift-drag ratio, etc.
Fundamental aerodynamic parameters of UAVs with different types or different
performance would vary as temperatures and speeds differ. However, the correlations
among those fundamental aerodynamic parameters are complex and could hardly be
described in functions, and thus are measured by wind tunnel test or test flight, which
are shown in Figs. 1 and 2.
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1 1.2
The performance of engine is linear with respect to the fuel consumption; the fuel
consumption of engine is determined by UAV type, speed and altitude. One of the most
important indexes is the Thrust specific fuel consumption (TSFC) [3]. The TSFC of
Research on Simulation for Fuel Consumption of UAV 467
2.3.1 Speed
The speed to be considered here is the level flight speed of UAV. There are two speed
units in aviation field: the commonly used km/h, and the Mach number. Speed relates
directly to fuel consumption of UAV.
circling when the UAV makes a turn. Figure 3 shows a series of roll angles and deflection
angles sampled in a turning process.
2.3.4 Acceleration
The acceleration of an aircraft consists of horizontal acceleration and vertical acceler‐
ation. Figure 4 shows scatters of vertical acceleration and fuel consumption sampled
during UAV missions.
2.4 Weight
The weight of a UAV is the sum of fuselage weight and load weight. The heavier the
UAV weighs, the greater thrust is needed, accordingly the fuel consumption increases.
When in a UAV mission, the UAV weight would reduce as the fuel consumption
increases. In fuel consumption calculation, therefore, the flight process should be divided
into several stages, weight deduction in each stage is neglected, and thus simplified
calculation is proceeded minimizing calculation error in accordance with aforemen‐
tioned assumption.
dv ( )
m× = T × cos 𝛼 − f − m1 + m2 g × sin 𝛾 (1)
dt
470 Z. Jia et al.
Engine performance
Flight trajectory
parameters
Motion equation
Mechanical energy
d𝛾 ( )
m×v× = T × sin 𝛼 + L − m1 + m2 g × cos 𝛾 (2)
dt
dy
= v × sin 𝛾 (3)
dt
dx
= v × cos 𝛾 (4)
dt
The energy height of the UAV is following
1
E=h+ × v2 (5)
2g
The total energy equals to the product of the energy height and the UAV mass, thus
the state model based on principle of energy balance could be obtained as follows
⎧ 1 2
⎪E = h+ v
⎪ 2g
⎨ E = v × (T × cos 𝛼 − f ) (6)
⎪ mg
⎪ x = v × cos 𝛾
⎩
dE dE∕dt dE∕dt
= = (8)
dMf dMf ∕dt Ce P
Research on Simulation for Fuel Consumption of UAV 471
where the TSFC of engines varies according to aircraft types: the TSFC of an airliner is
generally 0.1–0.5; the TSFC of a fighter is generally 0.7–1.5; and the TSFC of the UAV
in this paper is about 0.1. Fuel consumption of a UAV in cruising stage could be derived
from formula 3.6 and formula 3.8.
4 Conclusion
Factors influencing fuel consumption of UAV are analyzed in detail above all in this
paper, and then on the basis of these factors, a fuel consumption model of UAV is
presented. Employing the presented model, fuel consumption simulation in UAV simu‐
lation control system could be implemented.
References
1. Huang, K.M., Zhang, M.Y., Wang, T.: The design of UAV simulation model system based on
component modeling technology. Ship Electron. Eng. 12, 86–89 (2015)
2. He, Y.C., Liu, K., Shen, X.Y., et al.: Simulation study of aircraft fuel consumption estimates
model. Comput. Simul. 32(5), 33–36 (2015)
3. Shirley, C.M., Schetz, J.A., Kapania, R.K., et al.: Tradeoffs of wing weight and lift/drag in
design of medium-range transport aircraft. J. Aircraft 51(3), 904–912 (2014)
4. Perez-Rodriguez, D., Maza, I., Caballero, F., et al.: A ground control station for a multi-UAV
surveillance system: design and validation in field experiments. J. Intell. Robot. Syst. 69(1–4),
119–130 (2013)
5. Zhang, Y.X.: Research on Mission Flight Path Planning and Optimization Method for Fixed-
wing UAV. Zhejiang University (2016)
6. Xu, M.X., Zhu, X.P., Zhou, Z., et al.: Exploring an effective method of thrust allocation for
solar-powered UAV with multiple propellers. J. Northwest. Polytechnical Univ. 4, 505–510
(2013)
Research on Meteorological Data Simulation
Abstract. This paper proceeded with simulating the meteorological data. Based
on the meteorological data that has been collected, this paper puts forward the
simulation method where the meteorological data can be classified into discrete
data and continuous data which can be simulated respectively. By establishing the
mathematical model of meteorological variables and the employment of the rele-
vant knowledge, the simulation of meteorological data can be achieved. The
simulated relationship among meteorological data accord with the relationship
among the real data. In other words, the simulated result approach the real mete-
orological data to a certain extent. Moreover, this result guarantees the application
of the unmanned aerial vehicle in logistics, disaster relief, medical care etc.
In UAV simulation training, the weather is a very important factor, Changes in weather
will greatly affect the operation and flight of UAV in simulation environment. In order to
improve simulation level of the simulation training system, the simulation of meteoro-
logical data needs to be studied. The objective of Ref. [1] was to develop and validate a
simulation model of the evaporation rate of a Class A evaporimeter pan. Daily weather
data and three climatic variables for four cities in China were gathered, investigated and
analyzed, and the sensitivity of climatic variables on building energy consumption was
discussed [2]. An experimental embankment was constructed, and the first objective is to
investigate the influence of climatic changes on the soil response such as changes in water
content and temperature [3]. The main contribution of Ref. [4] is an analysis of
requirements for a wind power infeed model used in a power system simulator from a
meteorological viewpoint. The WARMF model was applied to the Catawba River
watershed of North and South Carolina to simulate flow and water quality in rivers and a
series of 11 reservoirs [5]. The data generated in Ref. [6], can be directly applied to the
EIA prediction model and serve for EIA. Edgar et al. [7] analyze the performance of the
physically based snow model SNOWPACK to calculate the snow cover evolution with
input data commonly available from automatic weather stations. A system for observing
meteorological data based on a wireless sensor network is designed to fulfill the business
requirements for meteorological data observation in unattended areas [8]. This paper puts
forward the simulation method where the meteorological data can be classified into
discrete data and continuous data which can be simulated respectively. This paper pro-
ceeded with simulating the meteorological data. Based on the meteorological data that
has been collected. Task deducing and simulated training are conducted by coordinating
with the comprehensive task control system and the result is satisfactory.
Under the condition that the way of establishing statistic model has been determined,
appropriate meteorological data can be collected according to the types of meteoro-
logical data which is required to simulate. Table 1 shows the daily meteorological data
of one of the districts in Changchun.
Meteorological data variable mainly consists of six types which are meteorology,
temperature, relative humidity, air pressure, wind power, wind direction. According to
the meteorological knowledge, initial analysis will be conducted on the types of
variable as well as the value range. Atmospheric phenomenon generally includes the
sky condition, rainfall, snowfall which are usually described as fine, rain, fog, snow.
Meteorology names physical quantity that measures the degree of the hot and cold as
474 Z. Wu et al.
Markov process is such a kind of process: In the process of x(t), every transfer of the
state is related to the state of the previous moment but not the past state. In other words,
there is no after effect during the state transfer process, hence, such a state transfer
process is called Markov process. Markov chain is a discrete Markov process in the
time discretization state.
Markov chain is a sequence of random variables like X1, X2, X3, X4. The range of
these variables called state space is the set of all their possible values and the value of
Xn represents the state in time n. If Xn + 1 is a function of Xn with respect to the
conditional probability distribution of the past state, then
This identical equation represents the characteristic of the Markov chain. In the
Markov chain, n-step transition probability has the following characteristic:
ðnÞ
X ðlÞ ðnlÞ
pij ¼ pik pkj
k2I
Table 2. The frequency and probability of each kind of weather condition table
Weather condition Fine Partly cloudy Cloudy Overcast Rainy
Frequency 83 20 7 5 3
Probability 69.1% 16.6% 5.8% 4.1% 2.7%
Divide the value range of the random number into corresponding interval according
to the probability and then stochastic number p is generated. Then, the interval of the
stochastic number p can be estimated while the initial value of the meteorology can also
be determined. For example, stochastic number range [1–100] can be divided into fine
[1,69], partly cloudy [70,86], cloudy [87,93], overcast [93,97], rainy [98,100]. The
generated stochastic number 19 suggests that the initial value of the weather condition
is fine.
For the equations above, we can use the least square method to find the solutions.
The fundamental principle of the least square method is to find b0, b1, … bn which can
minimize the residual sum of square between observation and regression value. It is
equivalent to find the solutions of the equations:
@ 0
^ ¼0
^ Y Xb
@ ^0 X 0 Y Y 0 X b
^þb
^0 X 0 X b
^ ¼0 @ 0 ^þb
^0 X0 Xb
^ ¼0
Y Xb Y 0Y b Y Y 2Y 0 X b
^
@b ^
@b ^
@b
^¼0
X 0 Y þ X0 X b ^
X0 Y ¼ X0 Xb ^ ¼ ðX 0 X Þ1 X 0 Y
b
We can obtain the estimations b0, b1, b2 of b0, b1, b2 which are 42.9, −1.08, −0.85.
Therefore, the regression equation of temperature is: y = 42.9 − 1.08x1 − 0.85x2.
Then, it is required to proceed with the test of significance of the regression equation.
478 Z. Wu et al.
From the diagram, it can be known that the coefficient of R is 0.905, the coefficient
of R2 is 0.819, the coefficient of adjusted R2 is 0.813, we have to preferentially consider
adjusting the coefficient of R2. Since 0.813 is bigger than 0.05 and approaches 1, the
explanation degree of the independent variables against dependent variables is high.
H0 : b1 ¼ b2 ¼ 0
Assume:
H1 : b1 ; b2 at least one nonzero
We can identify significance level a = 0.05 and find the rejection region. We find
F0.05(2, 60) = 3.15 by referring to the table,
By rejecting H0 and assuming to accept H1, we can believe that the linear rela-
tionship between independent variables and dependent variables is significant.
Research on Meteorological Data Simulation 479
t1 = −16.698, t2 = −3.477.
ta/2(n−m−1) = ta/2(62) = 1.999, |t1| > ta/2(62), |t2| > ta/2(62). Reject null hypothe-
sis, then linear relationship between independent variable and explained variable is
significant which should be in the equation.
5 Conclusion
This paper mainly focuses on the simulation method of the meteorological data. These
meteorological data can be classified into discrete variable and continuous variable
according to their own characteristics and respectively simulated. This paper begins by
collecting and analyzing the data. Then, Markov model is introduced to generate
discrete data. Finally, multiple regression model is required to e analyze both generated
variables and processing continuous variables. This paper provides a reliable data
simulation method for the application of UAV in logistics, disaster relief, medical care.
480 Z. Wu et al.
References
1. Molina Martínez, J.M., Martínez, A.V., González-Real, M.M., Baille, A.: A simulation model
for predicting hourly pan evaporation from meteorological data. J. Hydrol. 318(1), 250–261
(2006)
2. Gao, Q., Liu, J., Yang, L.: Sensitivity studies on elements of meteorological data for building
energy simulation in China. In: International Building Performance Simulation Association,
pp. 217–222 (2007)
3. Cui, Y.J., Gao, Y.B., Ferber, V.: Simulating the water content and temperature changes in an
experimental embankment using meteorological data. Eng. Geol. 114(3), 456–471 (2010)
4. Brose, N.: Specification of meteorological data requirements for a wind power infeed model
used in power system simulator. In: 12th International Conference on Environment and
Electrical Engineering, pp. 140–144 (2013)
5. Herr, J.W., Vijayaraghavan, K., Knipping, E.: Comparison of measured and MM5 modeled
meteorology data for simulating flow in a mountain watershed. J. Am. Water Resour. Assoc.
46(6), 1255–1263 (2010)
6. Wang, Q., Li, S., Ding, F., Zhao, X.: Simulation of high-altitude meteorological data used to
environment impact assessment by MM5 model. Procedia Environ. Sci. 2, 1713–1716 (2010)
7. Edgar, S., Christoph, M., Charles, F., Michael, L.: Evaluation of modelled snow depth and
snow water equivalent at three contrasting sites in Switzerland using SNOWPACK simulations
driven by different meteorological data input. Cold Reg. Sci. Technol. 99, 27–37 (2014)
8. Reza, A., Zhiliang, Z., Shuang, Z.: Design and simulation of a meteorological data monitoring
system based on a wireless sensor. Int. J. Online Eng. 12(5), 27–32 (2016)
Anti-data Mining on Group Privacy Information
Abstract. In the big data era, privacy preserving is a vital security challenge for
data mining. Common object of privacy preserving is personal privacy, which
should be kept unrevealed while data mining on group information. However, for
a few sensitive groups, such as suffering from some particular disease, engaging
in some special occupation or having some peculiar hobby, even if every personal
data is processed for privacy preserving, group specificity can be still exposed.
Therefore, we propose the concept and method of anti-data mining on group
privacy information. By adding, swapping data according to our rules, the minable
characteristic and group specificity of original data is destroyed and eliminated
to prevent group privacy from data mining.
1 Introduction
The concept of “big data” has appeared in the field of physics, biology, environment
ecology, finance and communication for several years. In 2011, the well-known
consulting company Mckinsey predicted that “the era of big data” was coming for the
first time [1]. Big data is a term for data sets that are so large or complex that traditional
data processing application software is inadequate to deal with them [2]. The lifecycle
of big data includes data extraction and integration, data analysis and interpretation,
among which data analysis is the core [3]. Different from conventional data, big data
possesses the features of “4V”, which are Volume, Velocity, Variety and Value [4].
Consequently, the common ways of data analysis aren’t suitable for big data anymore.
Data mining is defined as the procedure of extracting or excavating useful knowledge
from vast data stored in database [5]. Pattern and feature contained in big data is so
valuable that almost all industries, such as enterprises, telecom operators and govern‐
ments, are engaged in data mining. Sometimes, science and technology is a double-
edged sword. Data mining on big data brings forth both mass valuable information and
huge privacy leakage risk. The data mining algorithms are bound to collect abundant
users’ data for a long term to conclude the behavioral habits behind [6].
The concept of privacy preserving data mining (for short, PPDM) was firstly
presented in [7] to resolve the conflict between the precise excavation of knowledge rule
and the privacy protection of original information. Lindell indicated the need to protect
data privacy while ensuring data accuracy [8]. Evfimeievski used randomization to
establish an accurate data mining model for aggregated data [9]. In 2004, Vaidya intro‐
duced how data mining should be changed to accommodate the attacks of privacy advo‐
cates [10]. To solve the problem of constructing aggregated data without accurate data,
Zhang Nan describes the related technologies such as reconstruction [11]. By 2005,
Vaidya has proposed a relatively systematic idea and solution in [12], including a variety
of corresponding models. Lindell focused on the decision tree algorithm, especially the
ID3 algorithm [13]. Most recently, Saranya conducted extensive surveys of different
PPDM algorithms and analyzed the representative technologies [14]. In addition, since
PPDM is a highly integrated cross-cutting topic, there have been significant progress
for PPDM in statistics, machine learning, etc. [15–17]
Common object of PPDM is personal privacy, which should be kept unrevealed
while data mining on group information. But for a few sensitive groups, such as suffering
from some particular disease, engaging in some special occupation or having some
peculiar hobby, even if every personal data is processed for privacy preserving, group
specificity can be still exposed. Group privacy refers to the private information shared
inside the group, but unwilling to reveal to ones outside the group. Anti-data mining
indicates destroying the minable specialty of raw data and invalidating data mining to
protect the covert information contained in big data.
The rest of this paper is formed as following. Section 2 lists the technological clas‐
sification of PPDM, compares existing corporate privacy and anti-data mining to ours.
Section 3 explains our concrete algorithms. Section 4 conducts a series of experiments
to state the validity of above algorithms. Section 5 draws a conclusion.
According to the mainstream technologies of the raw data transformation, PPDM can
be classified into five dimensions [18]:
(i) Data distribution
(ii) Data modification
(iii) Data mining algorithms
(iv) Data or rule hiding
(v) Privacy preservation
The first dimension of data distribution can be classified as centralized data and
distributed data scenarios [19]. Distributed data scenarios can also be divided into hori‐
zontal data distribution [20–22] and vertical data distribution [23–25]. Generally, the
raw data needs to be modified before releasing to the public for privacy. The data
modification of second dimension is divided into perturbation [26, 27], blocking [28],
aggregation/merging, swapping [29] and sampling [30]. The third dimension refers to
Anti-data Mining on Group Privacy Information 483
the data mining algorithms, among which relatively important ones are decision tree
inducers [31], association rule mining [32, 33], clustering algorithms [34], rough sets
[35] and Bayesian networks [36]. The fourth dimension refers to whether raw data or
aggregated data should be hidden. To achieve higher utility and protect privacy, selective
modification is essential. The last dimension of privacy preservation to modify data
includes heuristic-based techniques [37], cryptography-based technique [38], recon‐
struction-based techniques [39] and so on.
We present anti-data mining on group privacy to solve the dilemma from the following
scenario. There are many active social networks based on common hobby, belief or
anything else. For example, there exists a network community among which are all AIDS
patients, meeting the social need for this special group. But this community is quite
vigilant against outside for the real world still discriminates them. Existing data mining
tools may collect the google searching records to find out the keyword (AIDS) of this
community. Here comes the requirement that protects the common feature of this group.
Our work pre-processes the searching records so that both the needs of acquiring infor‐
mation and protecting group privacy are met.
484 F. Yang et al.
Clustering is a division of data into groups of similar objects. Each group, called cluster,
consists of objects that are similar between themselves and dissimilar to objects of other
groups [45]. The similarity is measured by the distances between the described objects.
As searching record is the main study object of our thesis, complex text clustering need
calculation of nonmetric similarity function. So, we introduce Jaccard index. Given two
sets, A and B, the Jaccard index is defined as the ratio of intersection and union about
A and B.
Definition 1. Point of Searching Record. Point of searching record refers to the set of
searching keywords that the individual of special group submits in Internet, denoted as
PSR as below. There are n searching records in PSR, denoted as ri.
{ }
PSR = r1 , r2 , ⋯ , rn (1)
Definition 2. Point Similarity. Point similarity of two PSRs are defined as the Jaccard
index of two sets. For set A and B,
|A ∩ B| |A ∩ B|
PS(A, B) = = (2)
|A ∪ B| |A| + |B| − |A ∩ B|
⎧ n
− ra
⎪ s n
when ≺ ra
⎪ ra s
CS = ⎨ n (3)
⎪ − ra
s n
⎪ when ≥ ra
⎩ 1 − ra s
Definition 4. Average Clustering Score. Average clustering score indicates the average
of absolute value sum about all result clusters. Assume that there are m final clusters,
then ACS is defined as below.
|CS1 | + |CS2 | + ⋯ + |CSm |
ACS = | | | | | | (4)
m
Different to the common data mining, anti-data mining makes the result of clustering
unapparent by variation process. Our method follows the general approach of data
mining except the process of variation, indicating the core of group privacy preservation.
By different clustering algorithms, patients of same kind tend to gather together into one
or a few clusters unequally. We grade the clustering result, rank and vary it, after which
it’s graded again to judge whether the mixing is uniform. If not, we loop to vary and
cluster again, or else output the result. In ideal condition, the target patients spread evenly
in all result clusters, indicating the data mining on group privacy is invalid.
3.4.1 Clustering
We use clustering algorithm to testify the effect of anti-data mining. The distance calcu‐
lation of two points is the proportion of common entries to total entries in the clustering
procedure as the Jaccard index in Eq. (2). The central point is a virtual and auxiliary one
containing the most frequently used entries, whose length is the average of all points
inside a cluster. This parameter is used frequently in the following steps.
3.4.2 Grading
Grading refers to the hidden extent of target expressed by the uniformity degree distrib‐
uted in every cluster. Based on the expectation distribution of targets in the cluster,
calculate the deviation coefficient to the expectation. The calculation follows Eq. (3),
where the result CS falls into the interval of [−1,1]. The value is closer to 1, the better
the clustering effect is. The result of 1 means all records in the cluster are target. The
value is closer to −1, the worse the clustering effect is. The result of −1 means no records
in the cluster are target. But for anti-data mining, the perfect result is 0, meaning final
target after clustering approximates to the random distribution.
486 F. Yang et al.
3.4.3 Variation
Before variation, rank the clusters in descending order according to the above grades.
The variation consists of two situations: adding and swapping. In the former adding,
match a pair clusters of the correspondent high and low grades every time. Then,
randomly select a keyword of central point from the other to add into one of its own real
records. If the keyword number of the added record is more than the maximum allowable
value, then replace the newly added one. By adding uncorrelated noise to change the
clustering characteristics of original records, the records belonging to one or a few clus‐
ters intensively can be randomly scattered into clusters as many as possible. In the latter
swapping, exchange a keyword pair from two neighboring clusters every time, where
the two keywords both come from the central point sets of each cluster. By reducing the
occurrence of high-frequency words, the clustering results can be altered.
4 Experiments
At the beginning, two dictionaries are built artificially. One includes 100 entries about
AIDS keywords, such as HIV, homosexuality, incubation period, sex and so on, named
as target dictionary. The other includes 400 entries selected from everyday vocabulary
randomly, named as common dictionary. Next, we should construct ten original clusters,
where the first cluster is all about AIDS and the other nine are non-AIDS. The AIDS
cluster contains 1000 records, where each record consists of one to twenty entries picked
from the target dictionary. Other nine original clusters are built as above from the
common dictionary.
Using k-means algorithm on 10000 records from above ten original clusters, in which
k (k = 5) indicates the number of result clusters.
condition of ACS ≤ 0.1. In this experiment, we test our method on two variation ways:
adding only vs both adding and swapping. Tables 1 and 2 list the clustering conditions
before and after variation (adding only). Figures 1 and 2 illustrate the target distributions
in clusters before and after variation (adding only).
Fig. 1. Distribution of target before variation Fig. 2. Distribution of target after variation
Comparing Tables 1 and 2, before variation, the clustering gathers the target records
into one or two clusters intensively and the ACS approximates 0.9 indicating the distri‐
bution is uneven. After variation, every cluster has target records and the ACS approx‐
imates 0.1. As our experimental data includes 1000 target records and 9000 normal
records, 0.1 implies so random distribution that data mining is without effect on group
privacy. We choose the data of last line in Tables 1 and 2, showing the distributions in
histograms of Figs. 1 and 2. From these two figures, we find before variation, the sizes
between clusters are quite different. While after variation, every cluster is more similar
in scale than before and the target records have uniform distributions.
Table 3 lists the clustering condition after variation (both adding and swapping). For
this variation includes adding and swapping, the effect is more obvious than Table 2 in
that the ACS and running generations are less. In particular, adding only is suitable for
the relative centralization of target data in early variation, while adding and swapping
is appropriate for the relative uniformity of target data in late variation.
488 F. Yang et al.
5 Conclusion
This paper puts forward the concept and realization of anti-data mining on group privacy
information. Transforming the clustering process by adding and swapping, the group
specificity is altered and the data mining becomes ineffective. The validity of our idea
is testified by a series of experiments.
In the era of big data, anti-data mining is of practical significance for privacy
preserving and data security. We will research further on this theme in the near future.
Acknowledgement. The project was supported by the National Natural Science Foundation of
China under Grant 61502440 and the Open Research Project of The Hubei Key Laboratory of
Intelligent Geo-Information Processing under Grant KLIGIP1610.
References
1. Brown, B., Chui, M., Manyika, J.: Are you ready for the era of “big data”? http://
www.mckinsey.com/business-functions/strategy-and-corporate-finance/our-insights/are-
you-ready-for-the-era-of-big-data
2. https://en.wikipedia.org/wiki/Big_data
3. Meng, X., Ci, X.: Big data management: concepts, techniques and challenges. J. Comput.
Res. Dev. 50(1), 146–169 (2013)
4. Mauro, A.D., Greco, M., Grimaldi, M.: A formal definition of big data based on its essential
features. Libr. Rev. 65(3), 122–135 (2016)
5. Han, J., Kamber, M.: Data Mining: Concepts and Techniques, 2nd edn. Morgan Kaufmann
Publishers, San Francisco (2006)
6. Zetlin, M.: The Latest Privacy Invasion: Retailer Tracking (2012)
7. Agrawal, R., Srikant, R.: Privacy preserving data mining. In: Proceedings of SIGMOD 2000,
New York, pp. 439–450. ACM (2000)
490 F. Yang et al.
8. Lindell, P.: Privacy preserving data mining. J. Cryptol. 15(3), 177–206 (2002)
9. Evfimievski, A., Gehrke, J., Srikant, R.: Limiting privacy breaches in privacy preserving data
mining. In: PODS, pp. 211–222 (2004)
10. Vaidya, J., Clifton, C.: Privacy-preserving data mining: why, how, and when. IEEE Secur.
Priv. Mag. 2(6), 19–27 (2004)
11. Zhang, N.: Privacy-preserving data mining. Texas A&M University, pp. 439–450 (2006)
12. Vaidya, J., Zhu, Y.M., Clifton, C.W.: Privacy Preserving Data Mining. Advances in
Information Security. Springer, New York (2005)
13. Lindell, P.: Privacy Preserving Data Mining. Springer, New York (2006)
14. Saranya, K., Premalatha, K., Rajasekar, S.S.: A survey on privacy preserving data mining.
In: International Conference on Electronics and Communication Systems, pp. 1740–1744.
IEEE (2015)
15. Aggarwal, C.C., Yu, P.S.: A general survey of privacy-preserving data mining models and
algorithms. In: Aggarwal, C.C., Yu, P.S. (eds.) Privacy-Preserving Data Mining. ADBS, vol.
34, pp. 11–52. Springer, Boston (2008). https://doi.org/10.1007/978-0-387-70992-5_2
16. Matwin, S.: Privacy-preserving data mining techniques: survey and challenges. In: Custers,
B., Calders, T., Schermer, B., Zarsky, T. (eds.) Discrimination and Privacy in the Information
Society. SAPERE, vol. 3, pp. 209–221. Springer, Heidelberg (2013). https://doi.org/
10.1007/978-3-642-30487-3_11
17. Xu, L., Jiang, C., Wang, J., et al.: Information security in big data: privacy and data mining.
Access IEEE 2, 1149–1176 (2014)
18. Verykios, V.S., Bertino, E., Fovino, I.N., et al.: State-of-the-art in privacy preserving data
mining. ACM Sigmod Rec. 33(1), 50–57 (2004)
19. Li, F., Ma, J., Li, J.H.: Distributed anonymous data perturbation method for privacy-
preserving data mining. J. Zhejiang Univ. Sci. A 10(7), 952–963 (2009)
20. Kantarcioglu, M., Clifton, C.: Privacy-preserving distributed mining of association rules on
horizontally partitioned data. IEEE Trans. Knowl. Data Eng. 16(9), 1026–1037 (2004)
21. Inan, A., Sayg, Y., Savas, E., et al.: Privacy preserving clustering on horizontally partitioned
data. Data Knowl. Eng. 63(3), 646–666 (2007)
22. Ouda, M.A., Salem, S.A., Ali, I.A., et al.: Privacy-preserving data mining (PPDM) method
for horizontally partitioned data. Int. J. Comput. Sci. 9(5) (2012)
23. Dwork, C., Nissim, K.: Privacy-preserving datamining on vertically partitioned databases.
In: Franklin, M. (ed.) CRYPTO 2004. LNCS, vol. 3152, pp. 528–544. Springer, Heidelberg
(2004). https://doi.org/10.1007/978-3-540-28628-8_32
24. Muthulakshmi, N.V., Sandhya Rani, K.: Privacy preserving association rule mining in
vertically partitioned databases. Int. J. Comput. Appl. 39(13), 29–35 (2012)
25. Malik, M.B., Ghazi, M.A., Ali, R.: Privacy preserving data mining techniques: current
scenario and future prospects. In: Third International Conference on Computer and
Communication Technology, pp. 26–32. IEEE Computer Society (2012)
26. Liu, K., Kargupta, H., Ryan, J.: Random projection-based multiplicative data perturbation
for privacy preserving distributed data mining. IEEE Trans. Knowl. Data Eng. 18(1), 92–106
(2006)
27. Kargupta, H., Datta, S., Wang, Q., et al.: On the privacy preserving properties of random data
perturbation techniques. In: IEEE International Conference on Data Mining, p. 99. IEEE
Computer Society (2003)
28. Sugumar, R., Jayakumar, C., Rengarajan, A.: An efficient blocking algorithm for privacy
preserving data mining. J. Comput. (2011)
Anti-data Mining on Group Privacy Information 491
29. Fienberg, S.E., McIntyre, J.: Data swapping: variations on a theme by Dalenius and Reiss.
In: Domingo-Ferrer, J., Torra, V. (eds.) PSD 2004. LNCS, vol. 3050, pp. 14–29. Springer,
Heidelberg (2004). https://doi.org/10.1007/978-3-540-25955-8_2
30. Li, G., Wang, Y.: Privacy-preserving data mining based on sample selection and singular
value decomposition. In: International Conference on Internet Computing & Information
Services, pp. 298–301. IEEE (2011)
31. Fang, W.W., Yang, B.R., Yang, J., et al.: Decision-tree model research based on privacy-
preserving. Pattern Recogn. Artif. Intell. 23(6), 776–780 (2010)
32. Oliveira, S.R.M., Zaïane, O.R., Saygin, Y.: Secure association rule sharing. In: Dai, H.,
Srikant, R., Zhang, C. (eds.) PAKDD 2004. LNCS (LNAI), vol. 3056, pp. 74–85. Springer,
Heidelberg (2004). https://doi.org/10.1007/978-3-540-24775-3_10
33. Agrawal, R., Srikant, R.: Fast algorithms for mining association rules in large databases. In:
International Conference on Very Large Data Bases, pp. 487–499. Morgan Kaufmann
Publishers Inc. (2000)
34. Kumar, P., Varma, K.I., Sureka, A.: Fuzzy based clustering algorithm for privacy preserving
data mining. Int. J. Bus. Inf. Syst. 7(1), 27–40 (2011)
35. Pawlak, Z.: Rough Sets: Theoretical Aspects of Reasoning about Data. Kluwer Academic
Publishers, Dordrecht (1991). https://doi.org/10.1007/978-94-011-3534-4
36. Iqbal, K., Asghar, S., Fong, S.: A PPDM model using Bayesian Network for hiding sensitive
XML Association Rules. In: IEEE International Conference on Digital Information
Management, ICDIM 2011, Melbourne, Australia, September, pp. 30–35. DBLP (2011)
37. Ferguson, D., Likhachev, M., Stentz, A.: A guide to heuristic-based path planning. Comput.
Knowl. Technol. (2005)
38. Pinkas, B.: Cryptographic techniques for privacy-preserving data mining. ACM Sigkdd
Explor. Newsl. 4(2), 12–19 (2002)
39. Agrawal, R., Srikant, R.: Privacy-Preserving Data Mining. Foundations and Advances in Data
Mining, pp. 36–54. Springer, Berlin Heidelberg (2005)
40. http://news.china.com/domestic/945/20160718/23079010.html. [DB/OL]
41. Fang, B., Jia, Y., Aiping, L.I., et al.: Privacy preservation in big data: a survey. Big Data Res.
(2016)
42. Brugger, S.T., Kelley, M., Sumikawa, K., et al.: Defining privacy for data mining. In: National
Science Foundation Workshop on Next Generation Data Mining, pp. 126–133 (2002)
43. Chen, T.S., Chen, J., Kao, Y.H., et al.: A novel anti-data mining technique based on
hierarchical anti-clustering (HAC). In: Eighth International Conference on Intelligent
Systems Design and Applications, pp. 426–430. IEEE Computer Society (2008)
44. Chen, T.S., Chen, J., Kao, Y.H.: A novel hybrid protection technique of privacy-preserving
data mining and anti-data mining. Inf. Technol. J. 9(3), 500–505 (2010)
45. Berkhin, P.: A survey of clustering data mining techniques. Grouping Multidimension. Data
43(1), 25–71 (2006)
Big Data Analysis of Reviews on E-commerce
Based on Hadoop
Abstract. With the increasing popularity of online shopping, it has brought with
its massive online consumers and the growth of merchandise information data.
In order to deal with the demand for big data processing, building an analysis
system of e-commerce reviews base on Hadoop software framework. The reviews
of Internet commodity are chosen to be the samples of study. Choosing Navie
Bayesian classification to analyze the attributed values are discrete. The classifi‐
cation algorithms in accordance with MapReduce parallel computing theory
designed and run on Hadoop platform. Constructing the Naive Bayesian senti‐
ment classifier, and make the classifiers on the Hadoop platform to achieve
commodity reviews mining job. Result shows that it can improve the efficiency
of the commodity reviews analysis by using the Hadoop distributed platform.
1 Introduction
According to Chinese Online Shopping Market Research Report in 2015 [1], which
CNNIC published in 2016, China’s online sales continue to maintain the high growth
rate. With online shopping becoming more and more popular, it has also brought the
explosion of commodity review texts, produced huge amounts of data information, so
the demand for the analysis of the Internet commodity reviews is higher efficiency. For
automatic text sentiment analysis, traditional single machine has some limitations. This
project takes advantages of the MapReduce programming model of distributed
computing, based on Hadoop, compared with single machine, it has more CPU kernel
number and bigger RAMs. Under the lots of data reviews text information, with Hadoop
distributed computing framework in the review of text information has important signif‐
icance.
The text emotion analysis is also called opinion mining, it refers to the machine learning,
statistics, natural language processing and other techniques to automatically extract,
P(X | H)P(H)
P(H | X) = (1)
P(X)
The above pilot knowledge can be used to illustrate a simple Bayesian classification
problem. Here is a training set of samples D, where each row of data represents a training
tuple and the tuple class{attribute value label,
} {each tuple with an } n-dimensional attribute
vector represents X = a1 , a2 , … an−1 , an , , a1 , a2 , … an−1 , an , represents the meas‐
ured value of A1 , A2 , … An−1 , An corresponding to the n characteristic attributes. Also
assume there are m class variables C1 , C2 , … Cm−1 , Cm.
Assuming tuple vector X is known, classifier needs to predict class Ci of X belongs
to, in fact, obtain the maximum posteriori probability of P(Ci | X) and classification
prediction result for test set X can be obtained from Eq. (2). It is the value of Ci when
P(Ci | X)P(Ci ) get the maximum value
{ }
{ } ∏
n
class(X) = arg max P(X | Ci )P(Ci ) = arg max P(Ci ) P(xk | Ci ) (2)
k=1
Hadoop distributed software framework is designed to solve the problem of large data
computing, through distributed data storage and distributed computing. Therefore,
Hadoop’s basic platform is the core structure of HDFS distributed storage system and
the distributed data processing model and execution environment of MapReduce [4].
HDFS can automatically divide large files into many parts, and then upload these
parts to the computer node of the same Hadoop cluster. Users only need to log in to the
HDFS root directory to view all the shared files in the system, rather than knowing which
494 Q. Zu and J. Wu
computer is the part of the file belong to. At the same time large-scale file block storage
is the base of achieve mass data parallel computing. HDFS distributed file system has
the advantages of detecting data node failure, setting file copy, high availability and low
operating costs and so on. MapReduce parallel computing model needs to implement
Map and Reduce two functions, with the way of “divide and rule” to achieve the idea
of distributed parallel computing [5]. Map function part of the MapReduce calculation
program is mapped to the part of the data stored on the machine to calculate the Reduce
function phase based on the actual data situation. Reduce function phase based on the
actual data situation, generate a number of Reduce tasks on the Map phase of the data
protocol, processing.
In this study, a analysis system of e-commerce review texts is designed, under distributed
storage of Hadoop platform and framework of parallel computing, which is based on
the needs of businessmen’s automatic acquisition of product reviews, secure storage and
emotional analysis. Based on the B/S architecture WEB management system, the system
is divided into Hadoop distributed cluster product review data analysis module, data
storage module and data display module.
Data analysis module is the core module of the whole system. Its main job is divided
into three parts: first part is preprocessing and distributed storage the date of the product
sales and review. Second part is the implementation of Mapreduce parallel computing
framework analysis mining commodity review data information, the processed data
transferred to the database server. Third part is the data migration work. Soop is chosen
to be the cross-platform data migration tool, it is good at dealing with transporting the
data of HDFS distributed storage system to the database server.
Start Start
E-commerce Ttext
review training preprocessing
Feature
set lexicon
Characteristics
and weight
NB classification selection
training
NB classify
NB classifier Classification
result
End End
according to analytic hierarchy process to determine the weight of the Review elements,
constructing review elements and weight table, the results are shown in Table 1.
At the same time, it has been observed that there are many elements of reviews are
not directly described, but are described by other words, taking into account the lack of
elements. Therefore, it is necessary to establish a classification table for the review
elements, which can be classified into the corresponding review elements when
matching the words of the classification table. Table 2 shows the review elements.
496 Q. Zu and J. Wu
According to the review element words contained in each review text, the weights
of the review elements are calculated according to the weight table of the review factor
words. The calculation is shown in Eq. 3:
Ni
Pi =
∑
n
(3)
Ni
i=1
In Eq. 3, Ni represents the weight setting of the i-th review element in the review,
∑
n
Ni represents the sum of the weight values of all the review elements analyzed in the
i=1
review, and Pi is the weight proportion that represents the i-th review element in the
reviewary of this process.
The effective factors in the review mainly include emotional words, negative words
and turning words. Combined with Chinese Emotional Vocabulary Ontology Library
[6], according to the review to improve the emotional word annotation, it is shown in
Table 3.
In a commodity review sentence, although all the review elements are not evaluated,
there are often include two or more concerns, positive or negative. If all for the same
emotional tendencies, the emotional polarity of the sentence is easy to judge. But if there
are two kinds of emotional tendencies, the general approach is to take a simple weighting
approach to evaluate the weight of the elements, but this approach tends to have an
erroneous effect on the results [9]. Therefore, during the emotional analysis of weighted
calculation, a turning word before the review element, the weight of the previous clause
can be weaken or strengthen the weight of themselves. When there are three or more
review elements, the weakening of the previous term will increase the number of calcu‐
lations, combined with Table 1 to evaluate the elements of the word and the weight table,
using Eq. 4 to calculate the influence of the turning point on the weighting elements of
the review element.
⎧ Ni
⎪ Pi = n (1 + 𝜃)
⎪ ∑
Ni
⎪
⎨
i=1
N 𝜃 ∈ (0, 1) (4)
⎪𝜃 = 1− n i
⎪ ∑
⎪ Ni
⎩ i=1
P(X | Ci ) are unknown, the prior probabilities of class Ci, which can be obtained from
training set data, Naive Bayesian algorithm is to calculate the prior probability and
conditional probability value, Fig. 2 shows the algorithm flow.
Start
End
After calculating the priori probability, calculating probability P(X | Ci ) of the tuple
data X in class Ci. It can be known that the attributes of X are independent of each other,
from the Naive Bayesian assumption, and P(X | Ci ) is obtained by multiplying the
conditional probability of each attribute value under class Ci.
According to different types of reviews, using the algorithm to deal with, various
categories of P(X | Ci ) probability value can be obtained. By comparing the size of
P(X | Ci )P(Ci ), it can predicted the label of current data tuple X class. When a review
text has multiple review elements, the emotional tendency of whole sentence can be
determined by comparing P(X | Ci )P(Ci ) ∗ wi (wi represents the weight value) with the
weighted value.
MapReduce parallel computing module is that the calculation of the classification stage
is integrated into the Hadoop platform, to achieve the emotional classification of parallel
processing. In general among the Hadoop clusters in MapReduce parallel computing,
Map phase is responsible for the main data processing work, Reduce phase is primarily
responsible for data protocol operations, such as statistical operations, maximum fetched
Big Data Analysis of Reviews on E-commerce Based on Hadoop 499
operation. And training set of naive Bayesian emotional classifier construction stage is
artificially annotated, and the amount of data is not large. The massive training set of
data may also cause over-fitting problems. So the structure of classifier and classification
stage are processed in the Map phase, classified statistic results is processed by the
Reduce phase. The MapReduce parallel computing flow is shown in Fig. 3.
E-commerce
comment
text
Reduce stage
Statistics, Statistics,
combined combined
classification classification
results results
Map stage
Output Output
result result
The precision ratio P is the proportion of the correct value of the classification a in
the classification result a and b. Then the precision P calculation equation as shown in
Eq. 5:
a
P= (5)
a+b
The recall rate R is the correct value of the classification a in the test set attributable
to “praise” the proportion of a and c. The recall rate R is calculated as shown in Eq. 6:
a
R= (6)
a+c
Figure 4, from left to right are the product brand code, electricity business platform
coding, date, review category code, commodity category code, and the statistics of
commodity, medium and bad reviews. The recall rate P of the experiment is zero point
eight six two and the recall rate R is zero point eight eight. The experimental results
Big Data Analysis of Reviews on E-commerce Based on Hadoop 501
show that the accuracy rate of the classifier is more than eighty-five percent, which means
the system is good at predicting the emotional tendency of the review texts.
After analyzing the output result, it is found that the classification result is more
accurate when the review text has only one review factor or only include emotional
word. When the review text has multiple review factors, the classification result is prone
to appear classification error, which leads to the decrease of classification accuracy.
5 Conclusion
Under the background of online shopping normalization and big data age, the e-
commerce reviews are choosen to be the object of study, to analyze the review texts
features, and designs the system of information retrieval based on Hadoop data
processing technology. Extraction, analysis of additional business information from a
large number of unstructured reviews text data, providing the important reference basis
for merchants to adjust production and sales strategy. Although the work of the corre‐
sponding research is completed, getting a better analysis of the results, but there are
some deficiencies, following aspects need to be improved:
(1) Only the naive Bayesian classification algorithm is adopted. Although it is better
to classify the commodity in the text, it can not get all the attribute characteristics
of the text when it is large.
(2) Hadoop platform is good at dealing with massive data, due to limited hardware,
only in the virtual machine to achieve a real cluster structures, its efficiency in large-
scale data can not be effectively tested.
Acknowledgment. This paper was supported by the project in the Hubei Science and Technology
Pillar Program (No. 2015BKA222).
References
1. CNNIC. Chinese Online Shopping Market Research Report of 2015 [EB/OL] (2016). http://
www.cnnic.net.cn/hlwfzyj/hlwxzbg/dzswbg/201606/t20160622_54248.htm
2. Liu, B.: Sentiment Analysis and Opinion Mining. Synthesis Lectures on Human Language
Technologies, vol. 5, no. 1, pp. 1–167 (2012)
3. Mao, C., Hu, B., Wang, M., Moore, P.: Learning from neighborhood for classification with
local distribution characteristics. In: 2015 International Joint Conference on Neural Networks
(IJCNN), pp. 1–8. IEEE, July 2015
4. Welcome to Apache Hadoop! [EB/OL] (2015). www.hadoop.apache.org
5. Dittrich, J., Quiané-Ruiz, J.A.: Efficient big data processing in Hadoop MapReduce. Proc.
VLDB Endowment 5(12), 2014–2015 (2012)
6. Xu, L., Lin, F., Yu, P., et al.: The structure of emotional vocabulary ontology. J. Intell. 27(2),
180–185 (2008)
7. Zhao, P., Zhao, H., Tao, X., et al.: Based on the semantic model of TriPos Chinese subjective
and objective analysis method. Res. Comput. Appl. 29(9), 3285–3288 (2012)
502 Q. Zu and J. Wu
8. Chen, J., Hu, B., Moore, P., Zhang, X., Ma, X.: Electroencephalogram-based emotion
assessment system using ontology and data mining techniques. Appl. Soft Comput. 30, 663–
674 (2015)
9. Xie, S., Liu, B., Wang, T.: Apply the semantic relation to construct the emotional dictionary
automatically. J. Natl. Univ. Defense Technol. 36(3), 111–115 (2014)
Analysis on Structural Vulnerability Under
the Asymmetric Information
1
Electronic Engineering Institute, Beijing University of Posts and Telecommunications,
Beijing 100876, China
hms_bupt@163.com
2
Electronic Engineering Institute of PLA, Heifei 230000, China
1 Introduction
With the rapid development of the information society, some information systems such
as the distributed network equipment, computers, databases and the application software
has attracted particular interests, they have higher resources sharing speed and stronger
coordinated ability but meanwhile the structure may be complex. To analyze the influ‐
ence of resources topology to communication, we need to detect the basic network
topology. We are capable of sending data packets to any place through a computer
terminal. By changing the survival time of the data packet we can find IP of routers.
This process can be realized by the computer tool “Traceroute” [1]. Then we can use
this tool to combine up a great deal of node’s routing tracking paths to detect the whole
network topology [2]. In addition, we can refer to the routing list information stored in
routers to complete the whole network topology. The so-called routing list information
is the connected relationship with other routing in the network. Rather than studying the
routing of the network, we would like to pay more attention to the connectivity and other
properties of the network. The present researches on the complex information system
mostly concentrated on the connectivity of the network.
In an ideal world, network topology is connected. Meanwhile the number of network
attack rises, such as Trojan, botnets, computer viruses, worms, denial of service attacks,
web page tampering, domain name hijacking and so on. These can lead to nodes or links
failed. Single node or link which is removed from the network can cause damage to the
original connected network and result in communication failure. The robust of the network
is that the network still has strong connectivity after deleting several nodes or links. On the
contrary, if the network is no longer connected due to the emergence of a large number of
isolated nodes, we consider the network to be vulnerable. The vulnerability caused by the
change of network topology is called the structural vulnerability. Considering that the
influence of the failure of nodes and links to the network is equal, in this paper, we just talk
about the failure of nodes. There are two main types of existing way of attack: Random
attack [3, 4] and Targeted attack [5, 6]. Random attack randomly selects the nodes in the
network to attack and this attack way has strong randomness. Targeted attack is ranging the
nodes in network by the nodes centrality and select the most important nodes to delete.
Many researches [7–9] compared two attack methods but they did not point out the rela‐
tionship between them. We think that these two attack forms can use a common attack
model to combine together. How to put forward a common attack model to combine two
attack methods is our research focus.
We also studied the researches about the structural vulnerability. Albert [10] found
that, in the random network, when removed node number exceeds a threshold, the network
will become fragmented and lose the network connectivity. And in scale-free networks
when the threshold phenomenon disappears, deleted node number and largest number of
nodes in the connected subgraph synchronously reduce in proportion. Li [11] proposed an
attack method based on the Maximal Vertex Coverage and conducted a lot of experiment
simulations to prove that different networks have different network structural vulnerability
in MVC attack forms. Li [12] used Percolation Theory to analyze network reliability. The
failure of network can be regarded as a percolation process. Effective nodes are corre‐
sponded to occupied nodes in percolation process and failure nodes are corresponded to
blank nodes in percolation process. Through the network simulation analysis he got the
lifetime of the network nodes, and when the scale of the nodes is larger, the network has a
longer life. Ye [13] reconstructed the network to relatively small scale-free networks in
order to increase the robustness of the network. Tanizawa [14] described random attack and
targeted attack as a series of waves, so as to put forward a series of anti-destroying ability
optimization schemes. But most researches focus on comparing network structural vulner‐
ability under a specific attack mode, and assume that the network is known, ignoring the
fact that the attacker may not get the whole information of the network, and the differ‐
ences between the defenders and attackers. In the process of actual attack, the detected
network and real network are different due to some factors such as technology and
resources. This reflects the information asymmetry between the attackers and defenders.
Analysis on Structural Vulnerability 505
2 Attack Mode
The targets of the network attack are generally the topology of network resouces such as
routes and switches, which can be abstracted as the graph G(N, E). Here N represents the
nodes in the network while E represents the edges. Let adjacency matrix A stand for the
connectivity of the network. A is an N*N matrix. If node i and node j are connected, aij = 1,
else aij = 0. When the nodes are under the attack caused by system vulnerability or other
reasons and fail in the topology of controlled resources, a disturbance occurs and the adja‐
cency matrix A will change into AN−k = A−Ak. Here, k is the proportion of removed
nodes. The analysis of structural vulnerability is equivalent to analyze the connectivity
difference between A and AN−k. We introduce R to represent the index of evaluating the
attack performance and R has many different definitions. In this paper, we choose the
largest connected branch as the major measuring object and use the local redundancy of
specific node as a supplement. Concrete expressions will be shown in Sect. 3.
506 M. He et al.
1 2
R∗ [k] = {min[R∗ ( )], min[R∗ ( )], … , min[R∗ (1)]} (1)
min N N
1 2
R∗max [k] = {max[R∗ ( )], max[R∗ ( )], … , max[R∗ (1)]}, (2)
N N
where k in R*[k] represents the proportion of removed nodes. We can get mean values
denoted as,
1 2
R∗avg [k] = {E[R∗ ( )], E[R∗ ( )], … , E[R∗ (1)]} (3)
N N
It can be confirmed that finding a node set to minimize the R*[k] is a NP hard problem
[18], and the random attack fits the mean value of R*[k]. In the case of a specific network
topology, we can get the probability of each node being selected when the nodes attacked
are randomly chosen, which can be denoted as,
{ }
1 1 1 1
PN = , , ,…, (4)
N N N N
This kind of attack doesn’t need too much algorithm complexity, and can finish the
choosing process quickly. But the attack effect is not ideal in some networks, this will be
reflected in the following simulations.
Targeted attack is to delete the nodes according to node centrality. That means the
attacker has to detect the whole network first. This attack method has high algorithm
complexity and needs to compare each node in the network. The node centrality [19, 20]
includes degree centrality, closeness centrality and betweenness centrality, etc. In this
paper, we apply the degree centrality. That is to say, the node with a higher degree value
has a greater priority to be deleted. Targeted attack has different effects in different
networks. The first step is to sort the nodes according to the degree value in the detected
network.
Centrality{n1 , n2 , ⋯ , nN } n 1 ≥ n2 ≥ ⋯ ≥ nN (5)
At this time, to all nodes in the network, the probability distribution of being chosen is
P{1,0,0,…,0}. Targeted attacks only select one node as the target at a time which ensures
the accuracy of the attack. We can assume that the targeted attack fits the minimum value
of R. We will do simulations and further explanations in Sect. 3.
Analysis on Structural Vulnerability 507
From the definition of the asymmetric information attack, we can find that it includes both
the random attack and the targeted attack. As shown in Fig. 1, light red nodes represent the
undetected nodes and red nodes are the detected nodes with high degree under the certain
508 M. He et al.
F. We can see that when F is not big enough, the attacker cannot find all nodes with big
degree. The lack of the amount of the information attackers acquires leads to inaccurate
attacks.
From Table 1, we can further understand the relationship between parameter and attack
mode. When F = 0, the attack mode is random attack. At this time H becomes meaning‐
less. When F = 1, the mode turns into the targeted attack. At this time, H is actually the
number of the deleted nodes. When F ∈ (0, 1), it is the asymmetric information attack. The
probability of selecting important nodes depends on F. Attack range H corresponds to the
number of selected nodes. It can be observed that bigger H can make up for F. This
phenomenon will be showed and analyzed in later simulation.
According to attacker’s resource we can get H and F, and the number of deleted nodes
x. To prevent nodes with big degree from repeatedly showing up in the sample, we module
the whole simulation process as the non-return unequal probability sampling. The whole
attack process is defined as follow:
Step 1: Based on H, we calculate the number of sample nodes i, which meets i = H×N
Step 2: Calculate all nodes’ degree in the network and sort them.
Step 3: Respectively give all the nodes in the network the sample rate. For the node
with the biggest degree, F is the sample rate. Other nodes’ sample rates are
calculated as follow,
Analysis on Structural Vulnerability 509
Nj = (1 − F)j−1 F (8)
Step 4: Select one node according to the sample rate. Then compare the number of the
sample nodes with i. If smaller, go to Step 2, else go next.
Step 5: When x ≤ i just delete x nodes of the sample nodes. When x > i first delete the
sample nodes, then delete x−i nodes randomly.
In the real world, when the attacker launches an attack against a network, the attacker
cannot get the whole information of the network and the defender cannot know the
choice of the attacker or what he knows about the network. Supposed that the attacker
is rational and his resource is limited, his best choice is to attack the most important
nodes in the network. The importance of the node depends on two aspect, the attacker’s
knowledge of the network and the index of the importance. For the first aspect, we
proposed two parameters, the attack arrange H and the detection degree of node F, which
represents the proportion of the detected nodes and the ratio of the detected network
information to the whole network topology information respectively. Or simply, H
shows how many nodes have been detected and F shows, for the detected nodes, how
much the attacker knows. As for the second aspect, the indicator of the importance is
usually the centrality of the nodes. In this paper, we use the degree value, and it is obvious
that for a single node, more degree value means it is connected to more nodes. And that
means if the node is attacked, there will be a greater effect. Thus, we say the more the
degree value of the node is, the more important the node is.
R measures the connectivity of the network. The common measurement of the connec‐
tivity of the network is the number of the nodes that belongs to the largest connected
component. The redundancy of the network is also used to measure the network, espe‐
cially the robustness of the network. R has two definitions explained as follow. Rm is
the number of nodes in the largest connected branch
The largest connected branch is the carrier of the network traffic. The number of
nodes in the largest connected branch measures the connectivity of the whole network
after the network is attacked, and is also an important index to judge whether the network
is failed.
Rv is the local redundancy of the specific node in the network. It can show the
robustness of the node against the attack. It is also a measurement of connectivity. It is
defined as follow.
1 ∑
Rv = min{d(v), d(j)}
|S| (10)
j∈V (Γ2v )
510 M. He et al.
Node redundancy rate refers to the ratio of the number of the paths a node goes
through to arrive to its “neighbors’ neighbors” to the number of the paths in the complete
graph. Γv represents the set of neighbor nodes. Γ2v is the set of nodes’ “neighbors’ neigh‐
bors”. |S| is the complete graph consisting of the nodes, Γv and Γ2v.
In order to test different performances of the network under various attack modes, we
use BA network for simulations. BA network is a sparse network with 5000 nodes and
5000 edges. In simulations, we use Spyder as the simulation software and python as the
simulation platform. Besides we use networkx module to call the specific function to
generate BA networks. During the experiments we assume that the attacker prepares to
delete 100 nodes. Then we adjust the value of H and F and observe the change of
parameters in the network, and based on the simulation results we analyze the different
properties of the network vulnerability under different H and F. In order to guarantee
the experimental accuracy, we do each experiment for 50 times, then average them as
the final results and plot curves.
Table 2 is the algorithm for using the number of the nodes belong to the largest
connected component to measure the connectivity of the network after the network
attacked.
First we simulate the attack on the BA network (Fig. 2). The attack range H is
respectively 0.001, 0.002 and 0.005, and the detection degree of node is respectively
0.1, 0.5 and 0.9. In figures, the horizontal axis represents the number of the deleted nodes
and the vertical axis is the number of nodes in the largest connected branch. Based on
the simulation results it can be found that:
(1) As a whole, no matter what strategy the attack process is based on, after deleting
100 nodes, the curve tends to be gentle and the network shows its robustness. That
is to say, BA network will not be directly completely broken after deleting a certain
number of nodes. However, at this time, the network may not be able to commu‐
nicate. Therefore, to the attacker, it is not necessary to delete all 100 nodes to destroy
the network.
(2) Comparing 2-a, 2-b, 2-c we can find that when H and F are big enough, the curves
go down rapidly, and then go to a plain. At this time the network is failed. When
H = 0.005 F = 0.9, deleting 20 nodes is enough to make the network failed. It is
obvious that these 20 nodes are important to the whole network connectivity. The
attacker is a rational person with limited resources, as a result he will choose the
most important nodes to attack.
(3) For the supplement effect (H and F), it can be proved from the aspect of algorithm.
From 2-a, 2-b, 2-c, when H is fixed, F just need to reach a certain value to make
network failed quickly. As in figures, curves (F = 0.5 F = 0.9) are close to each
other. From 2-d, 2-e, 2-f when F is fixed, H reaches a certain value to make curves
close to each other. We use the resource cost coefficient C to evaluate the attack
performance. Obviously when C is smaller, the attack cost is lower and corre‐
spondingly, the protect cost is lower. How to balance the relationship between the
cost and the effect is a big issue. We can also find that the BA network is a multi‐
center network. After deleting several center nodes, the network tends to be broken.
Comparing Rm with Rv, we can find that the two indicators can both be used to
measure the connectivity of the network, and can both reflect the robustness of the
network. The difference between the two indicators is that they indicate the properties
of the network from different aspects. Rv is used to analyze single node’s connectivity
and Rm is used to measure the whole network’s connectivity. The connectivity of single
node and the whole network are different but relevant. When there is a need for evalu‐
ating important nodes, we will select the Rv to be the indicator. When we need to evaluate
the robustness of the network structure, we select the Rm. From the analysis of BA
network, the effect of the nonrandom attack (F > 0 H > 0) is better than random attack
(F = 0 H = 0). When F and H are bigger, the attack effect is more obvious. And in the
beginning of the attack, the connectivity of the network reduces quickly. When the
network shows robust and the curves become gentle, the effect of the random attack and
the nonrandom attack (F > 0 H > 0) are tend to be the same.
Fig. 4. Simulations on Router network. (a) the degree distribution of Router network (b) the
change of Rm during the attack of Router network
Figure 4a is the degree distribution of Router network. And x axis is degree; y axis
is the number of nodes. Degree and number of nodes meet linear relationship. So the
degree distribution of Router network and BA network obey power law distribution.
From Fig. 4b we can see after deleting 1000 nodes, the network becomes to be failed.
The curve at the top is the asymmetric information attack which has H = 0.0003 and
F = 0.9, and the random attack (F = 0 H = 0) is under the curve (H = 0.0003 F = 0.9).
So the effect of the random attack is better than the asymmetric information attack. We
can conclude that the effect of the asymmetric information attack in BA network is
opposite to that in Router network. But the degree distribution of the two networks both
514 M. He et al.
obey the power law distribution. So the structural vulnerability does not only depend on
the degree distribution. In order to fully understand the reason of this phenomenon, we
do simulations on ER network which has different structural vulnerability with BA
network.
ER network has 5000 nodes and 12500 edges. Corresponding parameters m, n are
5000 and 0.0005. Here we attack ER network (Fig. 5) we can find that.
(1) From Fig. 5a random attack (H = 0 F = 0) is at the bottom which states that the
effect of random attack on ER network is best. And the bigger F is, the worse the
effect is.
(2) In Fig. 5b three curves almost coincide. It indicates that when the H is less than a
certain value F has no influence. This is because the effect of the asymmetric infor‐
mation attack (F > 0) is tiny in ER network.
The result shows that ER network is robust against the attack (F > 0) and the random
attack has the best effect on ER network. ER network is a random network. The nodes
of the network have similar degree value, and the status of each node is the same. If we
use the asymmetric information attack (F > 0), can not improve the attack effect not
only, rise instead counteractive. Compared with BA network the curves of attack (F > 0)
are above the random attack so the targeted attack cannot fit Rmin [k].
Thus, the effect of the asymmetric information attack on ER network is opposite to
that on BA network but is same with that on Router network. Although Router network
shares the same degree distribution with BA network, its structural vulnerability is more
close to ER network. This shows that the real network has certain randomness.
4 Conclusion
In this article, through the analysis of the random attack and the targeted attack, we
proposed an asymmetric information attack model, and the random attack and the
targeted attack are special cases of the attack model. Using node degree as the indicator,
Analysis on Structural Vulnerability 515
we do simulations on the ER, BA and Routing network, and analyze the contact among
the random attack, the targeted attack and the asymmetric information attack. Based on
the results, we reveal different structural vulnerability of different networks. Simulation
shows that BA network is fragile to the asymmetric information attack (F > 0). And ER
network shows robust to the attack (F > 0). Because the scale of the real network is large
and the real network has randomness, it is robust to the asymmetric information attack
(F > 0) when the attack uses degree centrality to measure nodes’ importance.
References
Abstract. With the rapid growth of multimedia data, cross-modal retrieval has
received great attention. Generally, learning semantics correlation is the primary
solution for eliminating heterogeneous gap between modalities. Existing
approaches usually focus on modeling cross-modal correlation and category
correlation, which can’t capture semantic correlation thoroughly for social
multimedia data. In fact, the diverse link information is complementary to
provide rich hints for semantic correlation. In this paper, we propose a novel
cross-modal correlation learning approach based on subspace learning by taking
heterogeneous social link and content information into account. Both
intra-modal and inter-modal correlation are simultaneously considered through
explicitly modeling link information. Additionally, those correlations are
incorporated into final representation, which further improve the performance of
cross modal retrieval effectively. Experimental results demonstrate that the
proposed approach performs better comparing with several state-of-the-art
cross-modal correlation learning approaches.
1 Introduction
With the rapid development of multimedia technology, there has been a massive
explosion of multimedia data on social media websites, which makes the traditional
social media show the trend of multimedia [1]. In face of large amounts of complex
social multimedia data, retrieving valuable information is of great significance. Con-
sequently, cross-modal retrieval attracts considerable attention, in which users can
input any modalities data at hand to query relevant information of other modalities.
Different from traditional single-modal retrieval, it is more comprehensive and can
meet the increasing user demands. Generally, learning semantic correlation is the main
solution for eliminating heterogeneous gap between modalities for cross-modal
retrieval. Nevertheless, these data do not exist in isolation in social multimedia, which
makes correlation learning more challenging. On the one hand, different modalities of
multimedia data are usually in coexistence. For example, in image sharing website,
users usually share images accompanied by some text to illustrate. On the other hand,
© Springer International Publishing AG 2018
Q. Zu and B. Hu (Eds.): HCC 2017, LNCS 10745, pp. 516–526, 2018.
https://doi.org/10.1007/978-3-319-74521-3_54
Combining Link and Content Correlation Learning 517
these multimedia data are closely associated with the multi social factors, such as user,
group and location information, which connect different social components into
heterogeneous social multimedia networks.
Although, several efforts have been paid to correlation learning. Most of existing
methods focus on modeling cross-modal correlation and category correlation, such as
canonical correlation analysis [2], cross-modal factor analysis [3] and semantic corre-
lation matching [4]. However, those methods fail to model the correlation thoroughly for
social multimedia data. Much social correlation between two objects is not exploited
adequately. For instance, image and text may be connected by the same user, which
indicates they have some semantic correlation. In fact, the complex heterogeneous link
structure is complementary to provide the rich hints for semantic correlation and can be
exploited to bridge the heterogeneous gap to some extent. Therefore, both link and
content information are should be captured to improve the retrieval performance.
In this paper, we propose a novel correlation learning approach based on subspace
learning. It jointly considers both heterogeneous link and multimedia content, which is
ignored by previous works. In this learning framework, firstly, multiple social links are
transformed into both intra-modal and inter-modal correlation via heterogeneous net-
works. Then heterogeneous modalities are projected into a unified subspace according
to those correlation so that the similarity between different modalities can be measured
through projection matrices. The proposed approach is experimentally evaluated better
than other prevailing approaches on a cross-modal dataset.
Along this line of research, there are two main contributions of proposed approach.
On one hand, instead of treating each semantic link equally, we design a weight
learning approach keeping link structure consistence with content information. On the
other hand, we propose incorporating heterogeneous link structure and content infor-
mation into the unified feature representation which are not only helpful to bridge the
heterogeneous gap, but also robust to noise.
The remainder of this paper is organized as follows. The related work on corre-
lation learning for cross-modal retrieval is reviewed in Sect. 2. Then we present the
proposed combining link and content correlation learning approach in Sect. 3. In
Sect. 4, experiments are shown to verify the effectiveness of proposed method. Finally,
we conclude this paper in Sect. 5.
2 Related Works
Image-image network
Text-text network
Image-text network
framework, while it can easily extend to multi-modalities case. Based on this frame-
work, the details are depicted for each step as follows.
where xi and xj are the same type, P is a meta-path, Pxi !xj is the number of path
instances between xi and xj , jPxi !xi j is that between xi and xi , and Pxj !xj is that
between xj and xj .
Different meta-paths reflect different aspects of image or text similarity. Instead of
treating each meta-path equally, we design a novel algorithm to learn different weights.
The latent hypothesis is the link-based similarity sðxi ; xj Þ is consistence with
content-based similarity cðxi ; xj Þ. The set of meta-paths are denoted as p ¼
½p1 ; p2 ; . . .; pm and the weights of different meta-paths are denoted as
w ¼ ½w1 ; w2 ; . . .; wm . We minimize the following objective function to learn different
weights,
!2
n X
X n X
m
LðwÞ ¼ cðxi ; xj Þ wd s ðxi ; xj Þ
pd
þ akwk22 ð2Þ
i¼1 j¼1 d¼1
where the first component ensures that link-based similarity is consistence with
content-based similarity, and the second term is L2 regulation. It is worth paying
attention that these similarities are normalized in [0, 1]. The above objective unction
can solved by many off-the-shelf methods, such as Newton method or stochastic
gradient decent method.
After having all the weights for each meta-path, the link-based similarity can be
computed by combining all the similarities that based on different meta-paths.
Therefore, link-based correlation can be modeled by three weight matrices, of which
each element wij is expressed as below,
8 m
<P
wd spd ði; jÞ i and j are of the same type
wij ¼ ð3Þ
: d¼1
cij otherwise
where the cij denotes whether the link exist or not. In this paper, if there exist meta-path
coexist1
“image ! text” between i and j, we set cij ¼ 1, otherwise we set cij ¼ 0. Fur-
thermore, multi-modal fusion is performed to get a common wij for image and text
domain in this research.
where the first term is the distance between different modalities in the projected space
and the second term is that between same modality, k is the parameter for balancing the
inter-modal and intra-modal correlation.
However, there are no obvious semantic meanings in the above projected space. In
fact, the semantic label information can offer some useful guide information. Motived
by this, we also add the label consistent term into above loss function. So our ultimate
objective function is defined as follows,
X 2
min wij U T xi V T zj F
U;V
i2X;j2Z
X 2 2
þk wij U T xi UT xj F þ V T zi V T zj F ð5Þ
i;j2X or Z
2 2
þ b U T X Y X F þ V T Z Y Z F þ l kU k2;1 þ kV k2;1
here, Y X and Y Z are the initial label matrices on image and text domain respectively.
k:k2;1 denote l2;1 -norm, which is used to sparse feature selection on projection matrices.
b and l are the balancing parameters.
To solve the objective function in Eq. (5), we firstly simplify it. Inspired by
Laplacian regularizer, we may arrive at a corresponding compact form as in Eq. (6).
2
minU T X V T ZF þ k trðFðD WÞFT Þ
U;V
2 2 ð6Þ
þ b U T X Y X F þ V T Z Y Z F þ l kU k2;1 þ kV k2;1
P
where tr() denote the trace of matrix, D is a diagonal matrix with dði; iÞ ¼ wij , and
T j
F ¼ X U; Y T V . Note that the parameter k have been changed to 2k. While it does
not affect the result, so we still use k for convenience.
The above objective function can be optimized by an iterative algorithm, which
solves for one variable while keeping the others fixed. Hence, the solution of U and V
can be expressed in each iteration by Eqs. (7) and (8),
1
U ¼ ð1 þ bÞXX T þ lRu þ kX ðD W ÞXT bXY TX þ XZT V ð7Þ
1
V ¼ ð1 þ bÞZZT þ lRv þ kZðD W ÞZT bZY TZ þ ZX T U ð8Þ
522 L. Zhang et al.
Where Ru is a diagonal matrix [15], each element is defined as in Eq. (9), Rv has
analogous representation with Ru as in Eq. (10).
1
ru ði; iÞ ¼ rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
P 2 ffi ð9Þ
2 U ij 2 þ e
j
1
rv ði; iÞ ¼ rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
P ffi ð10Þ
2 Vij 2 þ e
2
j
in which e is a smoothing term, and we set it to be a small constant value to avoid the
denominator being zero.
Once projection matrices for heterogeneous low-level features are learnt, all the
data are readily projected into unified feature space. To perform cross-modal retrieval,
the similarity between the query and the other modalities can be computed directly.
4 Experiments
In this section, experiments are conducted to verify effectiveness of our method and
investigate the performance by comparing with several state-of-the-art methods.
4.1 Dataset
We perform the experiments for cross-modal retrieval based on the large-scale
NUS-WIDE dataset [16]. Each image is associated with some tags, which can be
regarded as text information. The dataset contains almost 270,000 images with 5,018
unique tags collected from Flickr website. We firstly find 10 largest classes and crawl
the list of users who upload or favor a given image according to the image ID. Then,
we choose the users owing more than 10 images and obtain 6487 unique users
including 1418 authors and 5069 favorite users. What’s more, we also crawl the groups
and contacts information for each author.
In the learning link-based correlation stage, in order to catch the link-based
intra-modal correlation as much as possible, we chose 5 kinds of meta-paths for images
and texts as mentioned above. While, we only consider one meta-path
coexist1
“image ! text” as inter-modal correlation. As for low-level feature, we take
500-dimensional SIFT feature vectors for images and 1000-dimensional tag feature
vectors for texts. Besides, we random select 80% of the image-text pairs used for
training and 20% for testing to conduct experiments.
(a)
(b)
Fig. 2. MAP performance of (a) image to text and (b) text to image retrieval task for each
category.
Fig. 3. Precision recall curve of (a) image to text and (b) text to image retrieval task.
Combining Link and Content Correlation Learning 525
5 Conclusion
In this paper, we examined the problem of correlation learning for cross-modal retrieval
in heterogeneous social multimedia networks. We have proposed an effective learning
approach by combining link and content information into together to improve the
performance of semantic correlation in a unified projection subspace. Different from
traditional approaches, we integrate rich link structure to obtain accurate and robust
representation. Furthermore, there are obvious semantic meanings in the unified space
through embedding semantic label information. Extensive experiments have verified
the better performance of proposed approach comparing with several prevailing
approaches. In the future, we intend to exploit nonlinear projection to obtain unified
feature representation.
References
1. Sang, J., Xu, C., Jain, R.: Social multimedia ming: from special to general. In: IEEE
International Symposium on Multimedia, pp. 481–485 (2016)
2. Hardoon, D.R., Szedmak, S., Shawe-Taylor, J.: Canonical correlation analysis: an overview
with application to learning methods. Neural Compute. 16(12), 2639–2664 (2004)
3. Li, D., Dimitrova, N., Li, M., Sethi, I.: Multimedia content processing through cross-modal
association. In: Proceedings of ACM International Conference on Multimedia, pp. 604–611
(2003)
4. Rasiwasia, N., Pereira, J.C., Coviello, E., Doyle, G., Lanckriet, G., Levy, R., Vasconcelos,
N.: A new approach to cross-modal multimedia retrieval. In: Proceedings of the ACM
International Conference on Multimedia, pp. 251–260 (2010)
5. Andrew, G., Arora, R., Bilmes, J., Livescu, K.: Deep canonical correlation analysis. In:
Proceedings of the 30th International Conference on Machine Learning, pp. 1247–1255
(2013)
6. Rosipal, R., Krämer, N.: Overview and Recent Advances in Partial Least Squares. In:
Saunders, C., Grobelnik, M., Gunn, S., Shawe-Taylor, J. (eds.) SLSFS 2005. LNCS, vol.
3940, pp. 34–51. Springer, Heidelberg (2006). https://doi.org/10.1007/11752790_2
7. Huang, L., Peng, Y.: cross-media retrieval by exploiting fine-grained correlation at entity
level. Neurocomputing 236, 123–133 (2017)
8. Peng, Y., Zhai, X., Zhao, Y., Huang, X.: Semi-supervised cross-media feature learning with
unified patch graph regularization. IEEE Trans. Circuits Syst. Video Technol. 26(3), 583–596
(2016)
9. Jia, Y., Salzmann, M., Darrell, T.: Learning cross-modality similarity for multinomial data.
In: Proceedings of the 11th International Conference on Computer Vision, pp. 2407–2414
(2011)
10. Han, X., Thomas, S.: Toward artificial synesthesia: linking images and sounds via words. In:
NIPS Workshop on Machine Learning for Next Generation Computer Vision Challenges
(2010)
11. Wang, S., Huang, Q.: Research on heterogeneous media analytics: a brief introduction.
J. Integr. Technol. 4(2), March 2015
526 L. Zhang et al.
12. Sun, Y., Han, J.: Mining heterogeneous information networks: a structural analysis
approach. SIGKDD Explor. 14(2), 20–28 (2012)
13. Jin, X., Luo, J., Yu, J., Wang, G., Joshi, D., Han, J.: Reinforced similarity integration in
image-rich information networks. IEEE Trans. Knowl. Data Eng. 25(2), 448–460 (2013)
14. Sun, Y., Han, J., Yan, X., Yu, P., Wu, T.: Pathsim: meta path-based top-k similarity search
in heterogeneous information networks. VLDB 4(11), 992–1003 (2011)
15. Wang, K., He, R., Wang, W., Wang, L., Tan, T.: Learning coupled feature spaces for
cross-modal matching. In: IEEE International Conference on Computer Vision, pp. 2088–
2095 (2013)
16. Chua, T., Tang, J., Hong, R., Li, H., Luo, Z., Zheng, Y.: NUS-WIDE: a real-world web
image database from National University of Singapore. In: Proceedings of ACM
International Conference Image Video Retrieval, pp. 1–9 (2009)
The Research on Touch Gestures Interaction
Design for Personal Portable Computer
1 Related Researches
Touch screen technology has been applied to mobile phones for 16 years and the
exploration of touch gesture interaction has always been the focus of researchers.
User-Defined Gestures for Surface Computing [1] written by Jacob O. Wobbrock
etc. defined the commonly used gestures and basic gestures of desktop large screen,
deeply discussed the influence of human cognitive behavior on hand gesture interac-
tion. Based on user defined design method in intuitive interaction domain, foreign
scholars designed Interactive behaviors on touch screen mobile phone, somatosensory
operation device. Besides, there are also many scholars study on the pain point of the
touch screen interaction [2–4], such as “click low precision”, “fat finger” and so on.
Cedric Foucault et al designed a two-handed interaction model “Spad” [5] to enhance
the productivity of tablet, this interactive model adopts non master hand to activate the
functional mode and use a tablet application “Spad” interactive control mode cooperate
with context components in “Keynote” to complete the same task. By comparing the
operations confirmed that “Spad” completes the task more efficiently without
increasing the complexity. But “Spad” requires that all tasks be divided into four
groups. Each of groups contains three buttons that brings functional limitations.
To draw a conclusion, researchers have proposed their own design methods and
solutions to the pain points of mobile phone touch screen gesture interaction. However,
there is no research on the design of large screen personal portable computer to provide
a unified guidance for personal portable computer touch screen gesture interaction.
© Springer International Publishing AG 2018
Q. Zu and B. Hu (Eds.): HCC 2017, LNCS 10745, pp. 527–537, 2018.
https://doi.org/10.1007/978-3-319-74521-3_55
528 Q. Sheng et al.
2 Experimental Study
2.1 Experimental Subject
We invited 12 college students to participate in the experiment, all of them were
graduate students, among them, 5 were female, and the others are male. Their age
ranged from 20–25 7 formal experiment, there were no long time intense hand
movements. All of them had used touch screen equipment in the past and are familiar
with the touch operation. The subjects had 3 min to know the page layout and basic
operations before task.
2.5.1 Frequency
Select the most frequent gestures as standard of experiment task and remove the
gestures in poor consistency, we achieved the target collection of gesture. The corre-
sponding gestures and their frequency are shown in below tables (Tables 2, 3 and 4).
The Research on Touch Gestures Interaction Design 529
Reform Clockwise 7
Pre
Task Gestures Operating mode Frequency
condition
Compatible with
Multiple
Rotating three kinds of 7
choice
rotation
Two fingers
Zoom in shape 8
extended
Two fingers
Reduce the shape 8
pinched
Due to the current page layout will be trimmed by the more frequent task like copy,
cut and paste, the page layout framework will have an impact on the process of user
defined gestures, therefore, the final data of these three tasks are eliminated in the data
statistics.
2.5.2 Predictability
Through gestures set and the frequency of each gesture, the Predictability of the ges-
tures set can be calculated:
P
G¼ s2S jPS j
100%0
j Pj
In the formula: G stands for Predictability, P represents the total number of actions
defined by the user in task, Ps represents the number set of “S”. s is a subset of S, the
predictable sequence of tasks is shown in the following figure (Fig. 1).
532 Q. Sheng et al.
Zoom in the
Two fingers extended 7
words
Zoom out
Two fingers pinched 7
text
2.5.3 Uniformity
After calculating the predictable nature of gesture set, the consistency of gesture set can
be calculated.
2
P P jPij
r2R PiPr jPrj
A¼ 100%
jRj
By summarizing the experimental data further, generally, the gestures can be used in
personal portable computer are divided into: basic gestures, symbol gestures, combined
gestures.
Picture Shape
Main screen In list Thumbnail Detailed view Control Text editing Form
editing editing
Open Adjust
the Adjust the corresponding item Stop the Active control Move cursor
Click Select
corresponding sliding item function position
item
Intelligent
Scale / Switch
Double click phrase Select Enter input status
view
selection
Select whole
Triple click
segment
Active
Corresponding items get into edit
Long press application Activate hidden functions
status Activate hidden functions
editing
Double finger
copy copy
drag
Switch to Horizontal
Horizontal Switch main Transfer movement Operation
another
sliding screen delete Switch to control
sibling page
another
Vertical
Open
Vertical moving to certain moving Switch Operation
Vertical slip Notification
page to another control
bar toolbar
sibling page
Four finger
Switching application View the background running APP
slip
Two finger
Return to retraction Three
Zoom Scale
previous level or four finger
switch view
Five finger
Close the application back to the main screen
zoom
Rotate the
Rotate Rotate
picture
536 Q. Sheng et al.
4 Conclusion
At present, as a “productivity tool”, large-screen personal portable computer has not yet
formed a complete and systematic touch-screen operation model in the people’s
understanding. We applied the method put forward by Wobbrock and others in the
personal portable computer gestures interactive design, summarized a set of gesture set
that for different operators of the personal portable computer. We also divided the
personal portable computer gestures are into three categories of basic gestures, symbol
gestures, and combined gestures in accordance with the gesture operation model, and
summed up the specialty of three types of gestures. Based on this gesture set, it is
convenient for the personal portable computer designers to build gestures that conform
to the user’s mental model, and giving the user a more natural and intuitive interaction
experience.
References
1. Wobbrock, O., Morris, M.R., Wilson, A.D.: User-defined gestures for surface computing. In:
Proceedings of CHI, no. 9, p. 183 (2009)
2. Wobbrock, O.: Maximizing the guessability of symbolic input. In: CHI Extended Abstracts,
no. 5, pp. 69–72 (2005)
3. Ruiz, J., Li, Y., Lank, E.: User-defined motion gestures for mobile interaction. In: Proceedings
of the SIGCHI Conference on Human Factors in Computing Systems, Vancouver. ACM
(2011)
4. Abadi, H.G.N., Peng, L.Y., Zadeh, A.M.H.: Guessability study on considering cultural values
in gesture design for different user interfaces. In: International Proceedings of Economics
Development & Research, no. 4, pp. 76–80 (2012)
5. Foucault, C.: Spad:a bimanual interaction technique for productivity applications on
multi-touch tablets. In: CHI, no. 4, pp. 1879–1884 (2014)
6. Guiard, Y.: Asymmetric division of labor in human skilled bimanual action: The kinematic
chain as a model. J. Motor Behav. 19(4), 486–517 (1987)
Research on Mobile User Dynamic Trust
Model Based on Mobile Agent System
1 Introduction
However, when user preference prediction is carried out by using the traditional col-
laborative filtering method, the prediction accuracy of the user preference is relatively
low due to the sparsity problem. At the same time, due to the limited input and output
capability of mobile phones and the characteristics of mobile user’s real-time access to
information, mobile network users have higher demand on mobile users’ forecasting
accuracy [3]. Traditional user preference forecasting methods are not suitable for
personalized mobile network service system.
The basic idea of collaborative filtering is to find the nearest neighbor of the target
user by calculating the similarity between the users and then to predict the interest
preference of the target user by the nearest interest preference of the neighbor. How-
ever, with the expansion of data size, data sparsity problems gradually appear, it leads
to algorithm efficiency and accuracy reducing. Some scholars mitigate this problem by
integrating trust relationships between users into traditional collaborative filtering. In
literature [4], the trust relationship between mobile users is calculated by the mobile
user’s explicit score. However, in the mobile communication network, because the
mobile phone’s input and output capability is limited, the explicit evaluation infor-
mation between mobile users is relatively small, Generally, we use the implicit cal-
culation method to obtain the trust relationship among mobile users. In literature [2],
for the characteristics of mobile communication networks, the trust relationship
between mobile users is introduced into the traditional collaborative filtering, and the
trust relationship between mobile users is analyzed by analyzing the communication
behavior between mobile users and gives a linear calculation method to calculate the
trust of mobile users. In literature [5], it describes an e-mail-based social network
computing method, which analyzes the communication behavior between users to build
a social network, and through experiments to prove that based on the logarithm
function to calculate the user trust relationship is more accurate than that through the
linear function [6].
In addition, some researchers solve the sparsity problem from another aspect. In
literature [7], it propose a collaborative filtering improvement method based on link
prediction, which builds a network model that connects users and project nodes and by
analyzing the relation between the user and the project can effectively reduce the
sparsity problem existing in the traditional collaborative filtering. In the traditional link
prediction method, node only represents the user/project, edge only represents the
relationship between the user/project, and the literature [7] improved the traditional link
prediction methods, nodes in the network represent users and items, and edges rep-
resent relationships between the user and the project. Although the literature [7]
reduces the sparsity problem of collaborative filtering through link prediction, it does
not consider the impact of trust on mobile user preference prediction for the charac-
teristics of mobile communication network.
In this paper, we analyze the mobile user behavior to calculate the trust degree
between mobile users and select the approximate neighbor of the target user by
combining with the similarity of the mobile user’s preference; Then, the improved link
prediction method is used to calculate the correlation information between the mobile
user and the mobile network service, and the mobile network service which the target
user is most likely to use is determined according to the calculated correlation; Finally,
540 W. Jiang and Y. Xu
based on the above research, a mobile user preference prediction method based on trust
and link prediction is proposed.
In order to improve the accuracy of mobile user preference forecasting, this paper
introduces the trust degree between mobile users and the correlation between mobile
user-mobile network services into the mobile user preference prediction model [8].
So, In Guu1 < V, E >, The weight between e < ui, uj > is T (ui, uj);
542 W. Jiang and Y. Xu
(5) Each user node in U repeat (3) to form Guu2 < V, E >;
(6) Calculate the indirect trust between mobile users. In the Mole Trust algorithm
proposed in document [8], we introduce the attenuation factor into the trust cal-
culation, and think that the trust degree is declining in the transmission process.
According to Definition 2, when calculating the indirect trust degree, because
there may be multiple co-trust neighbors between the two mobile user nodes, the
indirect trust degree can be calculated by the method of trust synthesis [12]. Let
the confidence factor in the transmission process is D, taking the mobile user
nodes ui and uk as an example, if the node uj is the common trust neighbor, the
indirect trust degree of ui through uj and node uk can be expressed as
Tj ðui ; uk Þ ¼ D Min T ui ; uj ; T uj ; uk ð3Þ
For each of the mobile users uii and uk, the trust neighbor uj calculates Tj (ui, uk), and
finally obtains the indirect trust degree of ui and uk. The formula can be expressed as
XL
T ð ui ; uk Þ ¼ j
Tj ðui ; uk Þ=L ð4Þ
Where L represents the total number of mobile users ui and uk common trust
neighbors.
(7) For each pair of mobile user nodes satisfying Definition 2 for Guu2 < V, E >
carried out (5), forming a complete mobile user trust network Guu3 < V, E >.
where C(u) represents the set of neighbor nodes of node u, C(s) represents the set of
S
neighbor nodes of node s and CðsÞ ¼ c 2 CðsÞ CðcÞ.
8 P
> ðrui ;s rui Þðruj ;s ruj Þ
< qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
P
s2Sui uj
; Sui uj t min jSui j; Suj
2 P
simðui ; uj Þ ¼ ðrui ;s rui Þ s2sui uj ðruj ;s ruj Þ
2
ð6Þ
>
:
s2sui uj
0; Sui uj \t min jSui j; Suj
Where Sui and Suj represent the scoring service set of mobile users ui and uj
respectively, Sui uj represents the service set, rui ;s and ruj ;s of the common score of
mobile user ui and uj, respectively represent the score of mobile user ui and uj to the
service, rui represents the average score of the mobile user ui for the rating service.
Where W (ui, uj) represents the harmonic weight of the mobile user ui and the
mobile user uj; sim(ui, uj) represents the similarity degree of the mobile user ui and the
mobile user uj; T(ui, uj) represents the trust of the mobile user ui and the mobile user uj,
a1, a2, b1, b2 are the harmonic factors, and a1 + b1 = 1, a2 + b2 = 1, k is the trust
threshold.
When determining the approximate neighbor set of mobile users, the approximate
neighbor of t% is selected as the approximate neighbor set of the target user. When
t = 0, the approximate neighbor set is empty set, t = 100, then the approximate
neighbor set includes all the mobile users in the data set, so we cannot get the desired
544 W. Jiang and Y. Xu
result when the value of t is too small or too large. In this paper, according to a number
of experimental results to select a reasonable value of t.
(2) Identify the target mobile network service set that needs to be predicted
According to literature [14], the higher the correlation between the mobile user and
the mobile network service, the greater the possibility that the mobile user will use the
corresponding mobile network service in the future. Therefore, when determining the
target mobile network service set that needs to be predicted, selecting the relevance of
the former h% of the mobile network services as the target user is most likely to use a
collection of mobile network services.
(3) Predict mobile user preferences
Combining the harmonic weight with the traditional preference prediction formula,
the weighted average forecast formula [15] is
X
rui ;s ¼ rui þ k uj 2 Nui W ui ; s; uj ruj ;s ruj ð8Þ
s 2 Sui
Where Nui is the nearest neighbor node set for mobile user ui, Sui is the set of
mobile network services
needed
for mobile user ui, k is a normalization factor
P
k ¼ 1= u 2 N 6W ui ; s; uj :
j ui
s 2 Su i
data. The Filmtipset community not only provides the relevant scoring informa-
tion based on user records, but also includes the social relations between users. In
this paper, we randomly selected 500 of the users and these users to make the
evaluation of the 1000 films for the experiment.
Since the MIT data set does not include the user’s evaluation of the mobile network
service, the data set is processed as follows before the experiment: By using the
behavior of the mobile service for all users in the nine months, the user’s preference for
the mobile service is sorted in the order of time, and the user’s preference information
is quantized in the form of scores (1 to 10).
During the experiment, the two data sets are processed in the same way, each data
set is divided into five groups randomly, each group is divided into two parts, one is the
training set, the other is the test set, which training set accounts for 80% of each group
of data, test set accounted for 20%.
The F indicator is a comprehensive assessment of the accuracy and recall rate, The
higher the F index is, the higher the accuracy is, In this paper, according to the
experimental results evaluation criteria for the precision and recall rate is redefined as
follows:
based on the linear function. The range of the indirect trust attenuation factor D is set to
[0.5, 1.0), the step size is 0.05, according to the experimental results, the D is set to
0.55. The threshold value of the confidence threshold is set to [0.1,0.9] and the step
length is 0.1. The experimental results are shown in Fig. 1. You can see from Fig. 1,
based on the logarithmic function of trust calculation method is superior to linear
function calculation method based on trust, this is because between user preference and
traffic with diminishing marginal utility theory, the logarithmic function can better
reflect the relationship between them.
You can see from Fig. 1, the F index is less than 0.6 in a general increasing trend,
in the case of more than 0.6 overall decreasing trend, this is because when a value is
small, too many users trust, when is large, remove too much trust users, both caused a
decline in trust the degree of accuracy. The results show that when the threshold of trust
is 0.6, the F index is larger. In this paper, the trust threshold is set to 0.6.
(2) Harmonic weighting parameter selection experiment
This paper makes use of genetic algorithm in Matlab (Genetic Algorithm GA)
toolbox respectively on the two datasets, the parameter a1 and b1 and a2 and beta b2
tuning experiments. The fitness function of genetic algorithm is defined as follows:
X
n jpi qi j
nMRE ¼ =n ð13Þ
i¼1 qi
In this paper, the range of the harmonic eights a1 and a2 in the Eq. (7) is set to a1 2
(0,1) and a2 2 (0,1), while b1 = 1–a1 and b2 = 1–a2. The optimal parameter values
obtained by genetic algorithm optimization are shown in Table 1.
From the experimental results, we can see that in determining the mobile user
approximate neighbors, if the trust is greater than a given threshold, the trust plays a
major role. If the degree of trust is less than a given threshold, the score similarity plays
a major role. This is because people are always more trusted about the more familiar
people, and their preferences are more likely to be affected by the surrounding people
(such as family, friends, etc.), and for relatively unfamiliar people, their preferences are
primarily influenced by user preferences that are similar to their ratings. When the trust
is greater than a given threshold, mobile users are more likely to be family members,
friends, etc., and trust is less than a given threshold, the mobile users are more likely to
be strangers [19, 20].
Research on Mobile User Dynamic Trust Model 547
4 Conclusion
Acknowledgments. This work was supported by the National Natural Science Foundation of
China (61472136; 61772196), the Hunan Provincial Focus Social Science Fund (2016ZBB006)
and Hunan Provincial Social Science Achievement Review Committee results appraisal identi-
fication project (Xiang social assessment 2016JD05).
References
1. Ricci, F.: Mobile recommender systems. J. Inf. Technol. Tourism 12(3), 205–231 (2011)
2. Wuhan, H., Xiangwu, M., Licai, W.: Collaborative filtering algorithm based on user
socialization relation mining in mobile communication network. J. Electron. Inf. Technol. 33
(12), 3002–3007 (2011)
3. Jiang, W., Zhang, L., Wang, P.: Research on grid resource scheduling algorithm based on
MAS cooperative bidding game. Sci. China F 52(8), 1302–1320 (2009)
4. Wang, H.M., Yin, G., Xie, B., et al.: Research on network-based large-scale collaborative
development and evolution of trustworthy software. Sci. Sin. Inf. 44, 1–19 (2014)
Research on Mobile User Dynamic Trust Model 549
5. Fengling, X., Xiangwu, M., Licai, W.: Collaborative filtering recommendation algorithm
based on mobile user context similarity. J. Electron. Inf. Technol. 33(11), 2785–2789 (2011)
6. Ding, Y., Wang, H., Shi, P., et al.: Trusted cloud service. Chin. J. Comput. 38(1), 133–149
(2015)
7. Jiang, W., Yusheng, X., Guo, H., Zhang, L.: Dynamic trust calculation model and credit
management mechanism of online transaction. Sci. China F Inf. Sci. 44(9), 1084–1101
(2014). https://doi.org/10.1360/N112013-00202
8. Zhang, S.B., Xu, C.X.: Study on the trust evaluation approach based on cloud model. Chin.
J. Comput. 36(2), 422–431 (2013)
9. Jiang, W.J., Zhong, L., Zhang, L.M., Shi, D.J.: Dynamic cooperative multi-agent model of
complex system based-on sequential action’ logic. Chin. J. Comput. 36(5), 115–1124 (2013)
10. Lim, S.L., Finkelstein, A.: StakeRare: using social networks and collaborative filtering for
large-scale requirements elicitation. IEEE Trans. Softw. Eng. 38(3), 707–735 (2012)
11. Xu, J., Si, G.N., Yang, J.F., et al.: An internetware dependable entity model and trust
measurement based on evaluation. Sci. Sin. Inform. 43, 108–125 (2013)
12. Yuxiang, W., Xiuquan, Q., Xiaofeng, L.: Research on the mechanism of mobile social
service selection based on context. J. Comput. Sci. 33(11), 2126–2135 (2010)
13. Weijin, J.: Dynamic Modeling and Quantification Trust More Research Methods Agent.
Science Press, Beijin (2014). 6
14. Wang, J., Li, S.-J., Yang, S., Jin, H., Yu, W.: A new transfer learning model for
cross-domain recommendation. Chin. J. Comput. 40(33), 1–15 (2017). Online publication
number
15. Zhang, W.-L., Guo, B., Shen, Y., et al.: Computation offloading on intelligent mobile
terminal. Chin. J. Comput. 39(5), 1021–1038 (2016)
16. Eagle, N., Pentland, A., Lazer, D.: Inferring friendship network structure by using mobile
phone data. Proc. Natl. Acad. Sci. 106(36), 15274–15278 (2009)
17. Fernando, D., Chavarriaga, J., Pedro, G., et al.: Movie recommendations based in explicit
and implicit features extracted from the Filmtipset dataset. In: Proceedings of the Workshop
on Context-Aware Movie Recommendation, Barcelona, Spain, pp. 45–52 (2010)
18. Xiong, C.Q., Ouyang, Y., Mei, Q.: Argumentation model based on certainty-factor and
algorithms of argument evaluation. J. Softw. 25(6), 1225–1238 (2014)
19. Huang, D.J., Arasan, V.: On measuring email-based social network trust. In: Proceedings of
the Global Telecommunications Conference (GLOBECOM), Miami, FL, pp. 1–5 (2010)
20. Xiuquan, Q., Chun, Y., Xiaofeng, L.: A trustworthiness method based on user context in
social network service. J. Comput. Sci. 34(12), 2403–2413 (2011)
21. Benchettara, N., Kanawati, R., Rouveirol, C.: Supervised machine learning applied to link
prediction in bipartite social networks. In: International Conference on Advances in Social
Networks Analysis and Mining, Odense, pp. 326–330 (2010)
A Group Decision-Making Method Based
on Evidence Theory in Uncertain
Dynamic Environment
1 Introduction
The above model achieves the uncertainty description of the debate model in different
ways, but it is not comprehensive enough to construct the relationship between argu-
ments in the debate system.
This paper expands the description of the relationship between the arguments by
referring to the model of the literature [11] in the description of the internal structure of
the argument. There is not only a mutual attack and support relationship between the
arguments, but also to support such attacks and support relationships. Against the
theory of evidence is used to analyze the reasoning process of uncertainty under the
condition of uncertainty, and the quantitative description of the reliability of the
argument is realized.
Group discussion is carried out through dialogue among speakers, and the arguments
(views) generated during the dialogue process constitute the content of the whole study.
The research framework is a formal description of the research arguments, including
the description of the internal structure of the argument and the interrelationship
between the arguments. Based on the literature [11], this paper proposes an extended
research model, which includes not only the attack and support relationship between
the arguments, but also the support or opposition to such attacks and support
relationships.
Definition 1. [11] The statement is a description of things, it can be objective data,
subjective judgments, factual phenomena, theoretical knowledge, etc., constitute the
basic constituent units of the argument, recorded as h.
Definition 2. [11] Reasoning argument is a tuple of At ¼ ðH; hÞ; Among them, H ¼
fh1 ; h2 ; ; hn g is a statement, and satisfy: (1) H is consistent, namely 6 9hi ; hj 2
H; hi :hj ; (2) Logically, H infers the h, denoted as H ) h. The H is called the
premise of the reasoning argument, the h is called the reasoning argument, and the set
of all arguments is denoted as At .
Definition 3. [11] The definition is provided with two theories of At ¼ ðH1 ; h1 Þ,
Bt ¼ ðH2 ; h2 Þ: (1) if 9h 2 H2 ; h1 :h, At attack Bt, denoted by At ! Bt or (At,Bt)−;
(2) if 9h 2 H2 ; h1 h; At support Bt, denoted as At ⊸ Bt or (At, Bt)+. The relationship
between attack and support relationship referred to as an argument-the argument
relation, denoted as rll ¼ ðAt ; Bt Þ; as shown in Fig. 1. All the arguments-the set of
arguments are recorded as ℛll.
Definition 4. The rule argument is a two tuple Ag ¼ ðH; rll Þ; where H is a set of
statements, rll is an argument-argument relationship, and satisfies: (1) H is compatible,
(2) Logically H inference out of rll ; denoted as H ) rll . The set of all rules argument is
denoted as At .
Definition 5. A rule Cg ¼ ðH1 ; r1 Þ; argument-argument relationship r2 ¼ ðAt ; Bt Þ:
(1) if r1 :r2 ; Cg against r2 ; denoted as Cg !r2 or ðCg ; r2 Þ ; (2) if r1 r2 ; Cg support
r2 , denoted as Cg (r2 or ðCg ; r2 Þ þ . The support and opposition of the argument to
552 W. Jiang and Y. Xu
A B A B
A B A B
C C
a b
A B A B
C C
c d
In a dynamic environment, the relationship between the argument and its arguments is
uncertain. The evidence theory is an effective method to express and deal with the
uncertainty reasoning. This paper analyzes the argument and the relationship of the
research model based on the evidence theory.
A Group Decision-Making Method Based on Evidence Theory 553
Definition 7. [12] For a verdict, the complete set of all possible answers is H ¼
fh1 ; h2 ; ; hn g; and the elements in H are mutually exclusive, then H is the recog-
nition frame. The power set of H is denoted by 2H .
Definition 8. [13] For any proposition A22H ; the definition of mapping m : 2H !½0; 1
satisfies the following conditions:
ð;Þ ¼ 0
(1) mP
(2) mð AÞ ¼ 1
AH
Then we called m for H on the basic belief assignment (basic belief assignment,
BBA).
Definition 9. [14] Assuming that the two independent evidence m1 and m2 under
frame H are identified, m ¼ m1 m2 is defined as
8 P
< m1 ðAi Þm2 ðBj Þ
Ai \ Bj ¼A
mð AÞ ¼ ; A 6¼ ; ð1Þ
: 1K
0; A¼;
P
according to the Dempster combination rule, where K ¼ m1 ðAi Þm2 Bj is the
Ai \ Bj ¼;
degree of conflict between the evidence m1 and m2 .
Definition 10. The truth of a statement h as a question to be judged, construct a
statement recognition framework, recorded as Hh ¼ fTrue; Falseg, where True rep-
resents the statement is true, False on behalf of the statement is false.
Definition 11. For the statement recognition frame Hh ¼ fTrue; Falseg, the reliability
assignment function mh : 2Hh !½0; 1 defined on 2Hh , where mh ðTrueÞ ¼ ah is the
reliability of the statement that the statement is true according to the current situation,
mh ðFalseÞ ¼ bh indicates that the statement is false, mh ðHh Þ ¼ ch ¼ 1 a b is the
confidence of the statement unknown reliability, the confidence vector of statement h is
denoted as wðhÞ ¼ ðah ; bh ; ch Þ.
Definition 12. The authenticity of a certain premise H ¼ fh1 ; h2 ; ; hn g as a pending
decision, the premise of the framework to identify the framework, denoted as
HH ¼ fTrue; Falseg.
The reliability vector of the premise H is denoted as wðH Þ ¼ ðaH ; bH ; cH Þ. When
the current H is the combination of multiple statements,
that is H ¼ h1 ^ ^hn ; then
wðH Þ ¼ wðhi Þ; where i satisfies: ahi ¼ min ahj ; that is, the premise of the relia-
j2f1;;ng
bility of the value of the vector from the lowest credibility of the statement. When the
current H is the recursion of multiple statements,
that is H ¼ h1 _ _hn ; then wðH Þ ¼
wðhi Þ; where i satisfies: ahi ¼ max ahj ; that is the prerequisite reliability vector is
j2f1;;ng
determined by the highest degree of confidence.
554 W. Jiang and Y. Xu
Definition 13. [10, 15] Evidence mapping refers to the mapping of the prerequisite
framework HH to the conclusion recognition framework Hh , which describes the
reasoning relationship between the elements in HH and the elements in Hh . Evidence
mapping from HH to Hh is expressed as C : HH !22 h ½0;1 . The evidence mapping for
H
where hHi is the prerequisite element, hhij is the element of the conclusion, and
f hHi !hhij is the rule strength of hHi supporting hhij .
S
Make HHi ¼ m j¼1 hhij a prerequisite for hHi to infer all the conclusion of the
collection.
Equation (2) satisfies the following condition:
(1) hhij 6¼ ; j ¼ 1; 2; ; m
(2) 0\f hHi !hhij 1
Pm
(3) f hHi !hhij ¼ 1
j¼1
among which,
ðr11 þ r21 Þ=2; r11 r21 6¼ 0
r31 ¼ ð3Þ
0; r11 r21 ¼ 0
A Group Decision-Making Method Based on Evidence Theory 555
ðr12 þ r22 Þ=2; r12 r22 ¼
6 0
r32 ¼ ð4Þ
0; r12 r22 ¼ 0
According to the reliability value vector wðH Þ and the mapping matrix RðAt Þ of the
premise H, the reliability value vector
where r31 and r32 are calculated by the same formula (3) and (4).
According to the reliability value vector wðH Þ and the mapping matrix R1 of the
premise H, the first behavior
ð1Þ
RðAt ÞH ¼ wðH Þ R1 ð6Þ
ð2Þ
matrix defined by the second row RðAt Þ of the
Definition 16. Similarly, the mapping
prerequisites H to RðAt Þ is R2 ¼ rij 33 . The row headings are H, :H and the com-
plete set of HH ; the column heading is the intensity of the rules deduced by :H1 ,
followed by :H1 h1 , :H1 :h1 and the complete set H:H1 . The specific form of R2 is
2 3
H ð:H1 h1 Þ H ð:H1 :h1 Þ 1 r11 r12
R2 ¼ 4 :H ð:H1 h1 Þ :H ð:H1 :h1 Þ 1 r21 r22 5
r31 r32 1 r31 r32
where r31 and r32 are calculated as follows (3) and (4). The second row of the mapping
matrix RðAt Þ can be obtained
ð2Þ
RðAt ÞH ¼ wðH Þ R2 ð7Þ
Thus, under the influence of H, the original mapping matrix RðAt Þ is updated to:
(1) Update the first row of RðAt Þ according to the Dempster combination rule:
ð1Þ
RðAt Þðnew
1Þ
¼ RðAt ÞH RðAt Þð1Þ ;
ð2Þ
(2) Similarly, the second line of RðAt Þ is updated as RðAt Þðnew
2Þ
¼ RðAt ÞH RðAt Þð2Þ ;
(3) The third line of RðAt Þ is calculated by the same formula (3) and (4).
Time node 1 produces argument A1, the 2reliability value 3w(H1)1 = w(h11) of the
0:9 0 0:1
premise H1, the rule mapping matrix R11 ¼ 4 0 0:8 0:2 5, the confidence value
0 0 1
1 1
wðhÞ ¼ wðH1 Þ R11 = (0.63,0.16,0.21) of the conclusion h.
Time node 2 produces argument A2, the reliability value of w(H2)2 = w(h23) of the
premise H2, and the confidence value wðh11 Þ2 ¼ wðH2 Þ2 R21 of the conclusion h11. The
reliability value of h11 is updated to wðh11 Þ2new ¼wðh11 Þ2 wðh11 Þ = (0.8730,0.0847,
0.0423). In this case, the reliability value of H1 is w(H1)2 = w(h12), the reliability value
of h is updated to wðhÞ2 ¼ wðH1 Þ2 R11 = (0.7650,0.0800,0.1550). It can be seen that
the reliability of A1 conclusion h is improved by support of A2 for A1.
Time node 3 produces argument A3, A3 reduces the credibility of A2 premise, and
thus reduces the credibility of h11 , the final h of the reliability value is updated to
wðhÞ3 = (0.7267,0.1027,0.1706).
The time node 4 produces the argument A4, wðh12 Þ4new ¼ wðh12 Þ4 wðh12 Þ =
(0.7466,0.2095,0.0439), the confidence degree of h12 , is reduced by A4, and the reli-
ability value of h is updated to wðhÞ4 = (0.6720,0.1676,0.1605).
The time node 5 produces the argument A5, A5 against the attack of A4 to A1, and
ð1Þ ð1Þ ð1Þ
the first row of the mapping matrix R41 of A4 is updated to R41new ¼ R41H R41 =
(0.1176 0.7059 0.1765). Thus the reliability value of h is updated to wðhÞ5 =
(0.7018,0.1442,0.1540).
Assuming that the acceptability of the statement is discriminated by rule 1, if the
threshold is a0 ¼ 0:7, then h is acceptable at the end of the study.
In the whole study, the update process of the presentation reliability value is shown
in Table 2.
Table 2. (continued)
Statement Time 1 Time 2 Time 3 Time 4 Time 5
h41 – – – (0.6,0.1,0.3) (0.6,0.1,0.3)
h42 – – – (0.8,0.2,0) (0.8,0.2,0)
h43 – – – (0.7,0.1,0.2) (0.7,0.1,0.2)
h51 – – – – (0.9,0,0.1)
h52 – – – – (0.8,0.2,0)
4 Conclusion
In this paper, the debate model based on evidence theory is constructed for the problem
of group decision-making under uncertain conditions. First, a more comprehensive
framework of the debate system is defined, which defines the relationship between the
internal structure of the argument and the argument, which not only includes attacks
and support relationships, but also allows for support or opposition to such attacks and
support relationships. Then we introduce the theory of evidence to describe the
uncertainty of the argument, apply the evidence mapping method to the uncertainty
process of the debate process, and realize the transmission and renewal of the reliability
of the argument.
Acknowledgments. This work was supported by the National Natural Science Foundation of
China (61472136; 61772196), the Hunan Provincial Focus Social Science Fund (2016ZBB006)
and Hunan Provincial Social Science Achievement Review Committee results appraisal identi-
fication project (Xiang social assessment 2016JD05).
References
1. Amgoud, L., Prade, H.: Using arguments for making and explaining decisions. Artif. Intell.
173(3–4), 413–436 (2009)
2. Chen, Y., Xiang, D., Zhao, Y.: Argumentation-based group discussion modeling and
decision approach. Syst. Eng. Theory Pract. 33(7), 1633–1639 (2013)
3. Dung, P.M.: On the acceptability of arguments and its fundamental role in nonmonotonic
reasoning, logic programming and n-person games. Artif. Intell. 77(2), 321–357 (1995)
4. Caminada, M., Amgoud, L.: On the evaluation of argumentation formalisms. Artif. Intell.
171(5-6), 286–310 (2007)
5. Cayrol, C., Lagasquie-Schiex, M.C.: Bipolarity in argumentation graphs: towards a better
understanding. Int. J. Approximate Reasoning 54(7), 876–899 (2013)
6. Chen, J.L., Wang, C.C., Chen, C.: Extended bipolar argumentation model. J. Softw. 23(6),
1444–1457 (2012)
7. Amgoud, L., Vesic, S.: A new approach for preference-based argumentation frameworks.
Ann. Math. Artif. Intell. 63(2), 149–183 (2011)
8. Tang, Y., Hang, C.W., Parsons, S., et al.: Towards argumentation with symbolic
dempster-shafer evidence. In: Proceedings of the Computational Models of Argument
(COMMA 2012), Vienna, Austria, pp. 462–469. IOS Press (2012)
560 W. Jiang and Y. Xu
9. Xiong, C.Q., Ouyang, Y., Mei, Q.: Argumentation model based on certainty-factor and
algorithms of argument evaluation. J. Softw. 25(6), 1225–1238 (2014)
10. Xiong, C.Q., Zhan, Y.F., Chen, S.B.: An argumentation model based on evidence theory. In:
The 10th International Conference on Computer Science and Education (ICCSE 2015),
Cambridge University, UK, pp. 451–454. IEEE Press (2015)
11. Xiong, C.Q., Li, D.H.: Model of argumentation. J. Softw. 20(8), 2181–2190 (2009)
12. Zhang, X., Mu, L.H.: Evidence combination rule based on local conflict elimination. Syst.
Eng. Electron. 38(7), 1594–1599 (2016)
13. Han, D.Q., Yang, Y., Han, C.Z.: Advances in DS evidence theory and related discussions.
Control Decis. 29(1), 1–11 (2014)
14. Shafer, G.: A Mathematical Theory of Evidence. Princeton University Press, Princeton
(1976)
15. Liu, W., Hughes, J.G., McTear, M.F.: Representing heuristic knowledge in DS theory. In:
Proceedings of the Eighth International Conference on Uncertainty in Artificial Intelligence,
pp. 182–190. Morgan Kaufmann Publishers Inc. (1992)
Analysis and Estimate the Effect of Knowledge on Software
Reliability Distribution
1 Introduction
growth model and other reliability growth model is used in the test phase. All dynamic
reliability models can be denoted as a function of time or its logic in the development
phase.
Knowledge can provide guidance for software testing and many studies have been
carried out. Xu [9] indicated that the cause of software errors is the errors of knowledge
used and its usage in the software system, and proposed software testing method based
on the knowledge by practice analysis, which can make up the inadequacy of testing
mothed adequacy and suitability effectively. Vijaykumar [10] proposed the concept of
Reference Ontology of Software Testing by identifying and reuse of the ontology in the
software testing process, which can gives semantics to a large number of software testing
information, so as to support the knowledge management in the software testing process.
However, the research of the effect of Software knowledge on software reliability is
very few. How to describe the relationship between knowledge and software reliability,
so as to achieve high quality of software testing with knowledge, and improve software
quality and reliability, scientific and quantitative analysis is necessary. Therefore, this
paper estimates and is focus on the relationship between the software knowledge and
software reliability.
The remainder of this paper is structured as follows. In Sect. 2, Weibull distribution
is briefly summarized. The relationship between the software knowledge and software
reliability are derived in detail, and a formalization expression of the relationship are
described briefly in Sect. 3. Engineering Experiments results and analysis are reported
in Sect. 4 and Sect. 5 concludes this paper.
Weibull distribution [11, 12] is used for reliability analysis in various engineering fields
for many years, which from bearing fatigue life of the deep groove ball to the tube fault
and river flood. Weibull distribution is one of the three famous extreme value distribu‐
tions. The significant feature of Weibull distribution is that the probability density grad‐
ually tends to zero. Figure 1 shows the Weibull probability density curves with different
shape parameter m values.
The Cumulative Distribution Function(CDF) and Probability Density Func‐
tion(PDF) of Weibull distribution can be formulated as:
m
CDF: F(t) = 1 − e−(t∕c) (1)
( )
m t m −(t∕c)m
PDF: f (t) = e (2)
t c
where m denotes the shape parameter, c denotes scale parameter, t denotes time.
Rayleigh model is special case of the Weibull distributions which shape parameter
m is equal to 2. The CDF and PDF are expressed as:
( )
2 t 2 −(t∕c)2
PDF: f (t) = e (4)
t c
Exponential model is another special case of the Weibull distributions which shape
parameter m is equal to 1. Exponential model is suitable for statistical process that
monotonous decline to progressive values. The CDF and PDF are expressed as:
1 −(t∕c)
PDF: f (t) = e = 𝜆e−𝜆t (6)
c
In engineering practice, the above formulations need to be multiplied by the total
defect number or total defect cumulative probability K, where K is an estimated param‐
eter for deriving specific model from the dataset.
This paper takes the system testing phase for example, where the shape parameter
m is equal to 1. There are three hypotheses as follow.
Hypothesis 1: The knowledge quantity of testing team A and B are SA and SB corre‐
spondingly.
Hypothesis 2: Testing team A and B test the same software product respectively, while
they can find out all defects through learning and experience accumu‐
lation.
Hypothesis 3: The fault distribution of software product is subject to Weibull distri‐
bution in software reliability field.
{ }
Assuming the fault set X = x1 , x2 , x3 , … is the sum of all kind fault modes found
at t moment.
K −(t∕c)
X = K ⋅ f (t) = e (7)
c
where K denotes the total defect number, c denotes scale parameter.
As observed from the engineering practice, the experience and knowledge of testing
team can affect the test efficiency and the found problems. Therefore, the knowledge
quantity of testing team is positively related to the amount of defects can be found.
K = K(S) ∝ S (8)
By plugging Formula (8) to Formula (7), the following formula is obtained as:
K(S) −(t∕c)
X= e (9)
c
which means that the defect number X found at moment t is related to knowledge quantity
S and scale factor c. X can be expressed by S and c as:
X = X(S, c, t) (10)
( )
Assuming the fault probability distribution of testing team A and B are f1 t, c1 and
( )
f2 t, c2 correspondingly:
K(SA ) −(t∕c1 )
X1 = K(SA ) ⋅ f1 (t) = e (11)
c1
K(SB ) −(t∕c2 )
X2 = K(SB ) ⋅ f2 (t) = e (12)
c2
Subtract Formula (11) and (12) after logarithmic, the following formula is obtained
as:
Analysis and Estimate the Effect of Knowledge on Software Reliability Distribution 565
c1 − c2
ln X1 − ln X2 = ln K(SA ) − ln K(SB ) − ln c1 + ln c2 − t ∗ (13)
c1 c2
High-level testing team can find problem faster and more, which means that if SA >
SB, c1 < c2. For any moment t as a constant, team A finds more faults and more efficient
than team B. Team A reaches the testing goal faster than team B at the end of the test
timing.
Because Δ ln K is a constant, the relationship of Δ ln X and |Δ ln c| can be shown in
Fig. 2.
c = 𝜆g(D, AD , R, AR , H, I) (15)
where D denotes the concept set of software knowledge, AD denotes the attribute set of
software knowledge, R denotes the relationship and rules of knowledge, AR denotes
attribute set of knowledge relationship, H denotes he level of knowledge concept, I
denotes instance set, 𝜆 denotes coefficient. Function g() denotes the measurement of the
knowledge quantity needed in software development or testing, which is proportional
to the scale parameter c.
566 C. Yang et al.
The knowledge quantity required by software testers has direct effect on the test. If
the tester’s understanding of the software system is more comprehensive and profound,
the less knowledge quantity required, the smaller scale parameter c is, which makes
software fault detection faster and software reliability can be improved effectively. On
the contrary, if the knowledge and experience of tester is weak, the more knowledge
quantity required and the larger c is, which makes fault detection slower.
Similarly, the effect of knowledge on the software testing phase is shown in Fig. 4.
The less of knowledge quantity tester requires, the smaller scale parameter c is, which
makes software fault found timely and effectively. On the contrary, if the knowledge
and experience of tester is weak, the more knowledge quantity required, the larger scale
parameter c is, which makes fault detection slower.
4 Experiments
In this section, software engineering testing of a certain Linux operation system is carried
out, in order to verify the proposed conclusion of the relationship between knowledge
and software reliability.
There are total 10 round testing designed, the test strategies are shown as follow:
(1) Each round is independent.
(2) For each round, the found problems and fault modes are concluded as knowledge.
(3) For each round, test cases are added based on the increased problems and fault
modes of the previous round.
(4) For each round, test case distribution is improved based on the fault mode distri‐
bution of all previous rounds.
The test results of 10 rounds are shown in Table 1, and the analysis is shown in
Fig. 5. Firstly, tester can find more faults by learning the software product more as the
testing going on. The knowledge of tester can be enriched after analyzing these defects
and faults. Secondly, the number of test problems and fault modes are increased each
round, while the increasing speed decreases and fault mode tends to 35 finally, which
means that the knowledge of tester has been improved.
K −(t∕c)
X = K ⋅ f (t) = e (16)
c
The simulation result by Matlab tool is shown in Fig. 6, which K = 27.943, c = 0.911.
By plugging the above two c into the probability density function of Formula (7),
two probability density curves are obtained as shown in Fig. 8.
With the knowledge increasing of tester, more defects are found, and smaller the
scale parameter c is, which means that the knowledge mainly affect the scale parameter
c of software reliability distribution.
5 Conclusion
On the basis of the theoretical analysis and derivation on the Weibull distribution of
defect density, this paper gives the conclusion that the knowledge qualification varies
inversely to scale parameter c of Weibull distribution via the analyses of the quantitative
relation between the tester knowledge level and defect distribution. Accordingly, the
proposed relationship can help to analysis the effect of software test team on the product
reliability curve, so as to predict the trend of finding problems along with software
reliability, and determine the end time of the test process.
570 C. Yang et al.
References
1. Lyu, M.R.: Handbook of Software Reliability Engineering. McGraw Hill and IEEE Computer
Society Press, New York (1996)
2. Bansal, A., Pundir, S.A.: Review on approaches and models proposed for software reliability
testing. Int. J. Comput. Commun. Technol. 4(2), 7–9 (2013)
3. Xavier, J., Macêdo, A., Matias, R., et al.: A survey on research in software reliability
engineering in the last decade. In: Proceedings of the 29th Annual ACM Symposium on
Applied Computing, pp. 1190–1191. ACM (2014)
4. Duran, J.W., Wiorkowski, J.J.: Capture-recapture sampling for estimating software error
content. IEEE Trans. Softw. Eng. 1, 147–148 (1981)
5. Nathan, I.: A deteministric model to predict “error-free” status of complex software
development. In: Workshop on Quantitative Software Models for Software Reliability,
Complexity and Cost: An Assessment of the State of the Art
6. Musa, J.: Operational profiles in software-reliability engineering. IEEE Softw. 10(2), 14–32
(1993)
7. Littlewood, B., Verrall, J.L.: Likelihood function of a debugging model for computer software
reliability. IEEE Trans. Reliab. 30(2), 145–148 (1981)
8. Goel, A.L., Okumotu, K.: Time-dependent error detection rate model for software reliability
and other performance measures. IEEE Trans. Reliab. 28(3), 206–211 (1979)
9. Xu, R.: The testing method based on software knowledge. J. Wuhan Univ. (Nat. Sci. Edn.)
46(1), 61–62 (2000)
10. de Santiago Jr., V.A., Vijaykumar, N.L.: Generating model-based test cases from natural
language requirements for space application software. Softw. Qual. J. 20(1), 77–143 (2012)
11. Kan, S.H.: Metrics and models in software quality engineering (2003)
12. Covert, R.P., Philip, G.C.: An EOQ model for items with Weibull distribution deterioration.
AIIE Trans. 5(4), 323–326 (1973)
Development of Virtual Reality-Based
Rock Climbing System
1 Introduction
Virtual reality (VR) typically refers to computer technologies that use VR headsets to
generate the realistic images, sounds and other sensations, through which users can interact
with virtual objects or stimuli that are modeled from the real world. VR environments have
been used extensively in a variety of fields, such as cinema and entertainment, healthcare
and clinical therapies, engineering, education and training and so on which offers users
numerous advantages and benefits such as immersive, interactive and cost-efficient expe‐
riences [1]. The study of VR technology possesses an interdisciplinary characteristic.
Through the advancement of interface technologies, VR will eventually become widely
popular, changing our lifestyle and making our work easier [2–4].
There are many commercial indoor rock climbing machines available in the market
[5–7]. The Treadwall series indoor climbing machine, released at 2012, is one of the
most widely used climbing machine, which is the first generation of the track format
climbing machine. However, its climbing path is single and repeated which lacks
immersive sense. Currently, there are more and more research and development work
2 Development
The developed VR system consists of following components, which include the user
control interface, 3D geometric model of the entire environment and the correspondence
of the virtual and real.
The overall VR system architecture is illustrated in Fig. 1. A number of software
tools as well as hardware interfaces are used to develop the VR environment.
So as to accomplish the login function, we create a database with two sheets. They
are user base information sheet: users: id, username, avatar and user authorization infor‐
mation sheet: user_auths: id, user_id, identity_type, identifier, credential. When the user
sends the mailbox/username/phone number and password request to log in, we first
determine the type the check the credential. For example, someone use username
“xiaoming” to login, part of the code is shown below:
SELECT * FROM user_auths WHERE identity_type =‘username’ and identi‐
fier=‘xiaoming’
It will return user_id when password match with credential.
When user is in the scene “SelectMount” and select a mount, the program will load
the ‘LoadingScene’ at first, then start the Coroutine ‘BeginLoading’ on its Start function.
Part of the code is shown below:
IEnumerator BeginLoading ()
asyn = SceneManager.LoadSceneAsync(MountName);
The Setting Interface, which consists of video settings and audio settings, enable
user to change the effect of mountain scene and the sound of system (Fig. 4).
The video settings allow user to set the resolution, the anti-aliasing mode, the filtering
mode or shader to tell system how to render. The audio settings are for adjusting the
volume and quality of all sounds in the entire climbing system, giving the user the most
comfortable sound experience.
Step 2: Using the Terrain tool in Unity3D to draw the rough model of the mountain,
and then build the detailed model to shape it to be similar with the real mountain
appearance.
Step 3: Map the model and improve the mapping quality. Add the sky, sunlight, trees
and other natural elements to increase the reality degree and improve the
immersive experience.
Step 4: To determine whether the virtual rock climbing scene is similar to the real peak
scene. If there is a big gap, it is necessary to continue to optimize the model to
develop the immersion sense. If it meets the requirements, the scene is supposed
to be completed.
map. Each point of the image is accurately matched to the surface of the model object,
and the gap between the two adjacent points is processed by the smooth interpolation
of the image.
In the 3ds MAX 3D modeling software, polygon modeling technology is utilized to
build the rock model, as shown in Fig. 7(a). The rock texture map is obtained by
contrasting it to the real rock appearance. Through the UV mapping technology, the rock
model map is rendered to derive the final model, which is shown as Fig. 7(b).
Fig. 7. Rock model built by polygon modeling technology and UV map method
The brush tool of the terrain editor is utilized to build the mountain model. By refer‐
ring the model to the actual mountain surface of the Roraima, the mountains, valleys
and plains in the real peaks as shown in Fig. 8(a) are carved out. By using Photoshop
software to deal with the color map and the normal map of the environment, the corre‐
sponding textures are obtained. Then the terrain editor tool is adopted to select the
Development of Virtual Reality-Based Rock Climbing System 577
different materials on the terrain mapping. The model effect is shown in Fig. 8(b). The
close-range mountain is shown in Fig. 8(c).
a LOD 0 b LOD1
c LOD2 d LOD3
Through the establishment of different level of detail complex models and drawing
corresponding levels, it can realize the optimization of computing resources on the basis
of ensuring visual effects. The system is based on Unity3d engine. By applying LOD
Group components and add different level models, optimal resource adjustment is
578 Y. Su et al.
achieved. Taking the tree model as an example, the LOD level detail rendering technique
is designed as shown in Fig. 10.
Vector3 tracker_position =
SteamVR_Controller.Input(deviceIndex).transform.pos +
GameObject.Find("LMHeadMountedRig").transform.position;
Quaternion tracker_rotation =
SteamVR_Controller.Input(deviceIndex).transform.rot *
GameObject.Find("LMHeadMountedRig").transform.rotation;
HoldsGroup plane on the virtual mountain surface so that the virtual hold can coincide
with the real hold.
Besides, we acquire the running speed of the rock climbing machine by using encoder
during the operation of it, and send it to PC via Serial Port. When the machine is running,
the rock climbing surface will rotate and the real holds will continuously move down.
We move up the Room in the VR environment at the speed received from machine so
as to match the virtual hold with the real hold through.
3 Interactive Device
Leap Motion is use as hands and fingers tracking device [16]. In order to use Leap Motion
normally, we need to install Leap Motion Orion software and download and import the
UNITY CORE ASSETS package to the Unity3D “IndoorClimbing” project. Then we
drag a LMHeadMountedRig into the scene from the path: LeapMotion/Prefabs. The
LMHeadMountedRig prefab uses the camera location provided by Unity to place the
LeapHandController at the correct position in the virtual world. The coordinates in the
tracking data are then transformed from Leap space to Unity space relative to the position
and orientation of the LeapHandController game object.
We import the UI Input module as well in order to control the standard Unity UI
widgets by naked hands. The primary component of the UI input module is the LeapE‐
ventSystem prefab. The LeapInputModule script component of this prefab implements
a Unity Input Module that uses tracking data to allow the user to manipulate standard
UI controls with their hands.
Vive Tracker have many modes to communicate with PC, the mode we used is that track
moving objects using a wireless interface in VR, with the accessory (refer to the rock
climbing machine) passing data to a PC via ‘FATFS-SDIO-USB’ [15]. The dongle is
used to transfer tracking data from the Vive Tracker to a PC, but the accessory transfers
data to/from a PC directly for a specific purpose based on our design that transfer the
path file to the rock climbing machine’s controller. The communication mode is shown
in Fig. 11.
Fig. 12. The virtual mountain surface superimposed on the rock climbing surface
The overall drive module controls the rock climbing machine to work through human
body weight as well as hydraulic resistance, to achieve the dynamic balance of people
climbing process. The path change module is used to achieve the path change of the rock
climbing process during climbing. We design the variable amplitude module to change the
slope of climbing machine through the worm gear mechanism in order to simulate the
different slope of the rock in the real outdoor rock climbing. The security module is used
to reduce the risk of the user during the climbing process and the failure rate of the machine.
4 Conclusion
climbing. Users can stay at home to enjoy the beauty of the world’s major famous climbing
mountains via this system. This project is of great significance to the development of rock
climbing and the combination of virtual reality technology and sports [17].
References
1. Chen, D.: Colorful Virtual Reality World. China Water & Power Press, Beijing (2015)
2. Yu, X.: Virtual Reality Technology Basic Tutorial. Tsinghua University Press, Beijing (2015)
3. Shen, Y.: Virtual Reality Technology. Tsinghua University Press, Beijing (2009)
4. Wang, H.: Virtual Reality: Leading the Future Human-Computer Interaction Revolution.
China Machine Press, Beijing (2016)
5. Magiera, A., Roczniok, R.: The climbing preferences of advanced rock climbers. Hum. Mov.
14(3), 254–264 (2013)
6. Hinch, T.: “It’s a place to climb”: place meanings of indoor rock climbing facilities. Leisure/
loisir 38(3–4), 271–293 (2014)
7. Udrea, P.E.: Rock climbing exercising machine. CA, CA 2794590 A1 (2014)
8. Zulj, S., Seketa, G., Dzaja, D., Celic, L., Magjarevic, R.: Virtual reality system for assisted
exercising using WBAN. In: IFMBE Proceedings of 6th European Conference of the
International Federation for Medical and Biological Engineering, vol. 45, pp. 719–722 (2014)
9. Zhang, L., Liu, Q.: Application of simulation and virtual reality to physical education and
athletic training. In: Pan, Z., Cheok, A.D., Müller, W., Chang, M., Zhang, M. (eds.)
Transactions on Edutainment VII. LNCS, vol. 7145, pp. 24–33. Springer, Heidelberg (2012).
https://doi.org/10.1007/978-3-642-29050-3_3
10. Jerald, J., Giokaris, P., Woodall, D., Hartbolt, A., Chandak, A., Kuntz, S.: Developing virtual
reality applications with unity. In: Virtual Reality, pp. 1–3. IEEE (2014)
11. Linowes, J.: Unity Virtual Reality Projects. Kybernetes, vol. 29, no. 3 (2015)
12. Jackson, S.: Unity3D UI Essentials (2015)
13. Mooney, T.: 3ds Max Speed Modeling for Games. Packt Publishing, Birmingham (2012)
14. Xudong, Q.: Research on 3D Model Codec. (Doctoral dissertation, University of Science and
Technology of China)
15. HTC Corporation: HTC Vive Tracker Developer Guideline (2017)
16. Nandy, A.: Leap Motion for Developers. Apress, New York (2016)
17. Chang, L., Chen, D.: The development of immersive rock climbing system based on virtual
reality technology. In: National Conference on Intelligent Manufacturing. Chinese
Association of Artificial Intelligence (2016)
Kinematics and Simulation Analysis of a 3-DOF Mobile
Handling Robot Based ADAMS and MATLAB
Jingbo Hu, Dingfang Chen ✉ , Lijie Li, Jie Mei, Qin Guo, and Huafeng Shi
( )
1 Introduction
The handling robot made in KUKA, ABB and so on has been used in the production
process of assembly, welding, handling and etc. [1]. With the robot technology gradually
becoming mature, handling robots have presented an important support in the field of
industrial production [2]. In microelectronics manufacturing, welding, packaging and
many other areas, handling robots have been widely used [3]. At present, the handling
robot has been widely used in the national economy in all important aspects [4]. But,
there are less robots which combines motion and handling. In fact, robots combining
motion and handling are necessary and can play important role in many fields.
In this paper, a mobile handling robot model is established. And in the end the
displacement of the joints is analyzed, which provides the basis for the debugging of
handling robot, saving the on-site debugging time and protecting the handling robot in
an extent [5].
The robot designed in this paper is a 3 DOF mobile handling robot. Three manipulators
are rotating joints and the rotation axes are parallel to each other. As shown in Fig. 1,
the front mobile platform (1) is equipped with a servo for the steering of the mobile
platform, and the servo implements the forward and backward commands of the
platform. The joints (7), (8), (9) are individually driven by a servo to achieve steering.
All servos are controlled by the controller (3). The end part of the arm (6) can be provided
with different clamping mechanisms depending on the work object. In the actual process,
through the mobile platform, the object can be held and moved from the target point 1
to the target point 2.
Fig. 1. Three - dimensional simplified model of mobile handling robot. 1-Mobile platform, 2-
Additional weight, 3-Controller, 4-Big arm, 5-Middle arm, 6-End arm, 7-Joint 3, 8-Joint 2, 9-Joint 1
As the robot actually works, the movement and handling are not carried out at the
same time. Therefore, the model can be simplified in the analysis of the robot’s handling
characteristics. The mobile part of the robot can be neglected temperately, and only
manipulator is analyzed. According to the D-H parameter method, the coordinate system
of the links in the manipulator is built. The coordinate system of each joint is established
at the end of the link. Each coordinate system draws only two axes in the Fig. 2.
⎡ nx ox ax px ⎤
⎢n oy ay py ⎥
0
T = 01 T 12 T 23 T = ⎢ y
3 n oz az pz ⎥ (1)
⎢ z ⎥
⎣0 0 0 1⎦
In the analysis, the geometric and joint variables of links are supposed to be known,
and the position and posture of the robot terminal actuator relative to the reference
coordinate system can be solved [6]. The inverse kinematics means that the position and
posture of the robot terminal actuator are provided to derive the joint parameters. In
order to control the robot arm, the inverse kinematics solution is the basis of the robot
movement [7]. It is because that in the practically work, the variables of all joints are
determined according to the position of target point.
For a given position and posture, the transformation matrix of robot is deduced as
formula (1), The values of the joint variable 𝜃1 , 𝜃2 , 𝜃3 can be solved by using corre‐
sponding inverse transformation matrix multiply Eq. (1), as shown in Table 2:
3 Kinematics Analyses
As it can be seen from Fig. 3, the displacement of the mark point 1 does not change
because the joint shaft 1 is connected to the seat (i.e., the base), so that the displacement
does not change. From Figs. 4, 5 and 6, it is found that the displacement of the marker
points in the Z direction does not change and there is no movement in the Z direction.
In the X, Y direction, the displacement curve approximates in the form of a sin function.
The results of the above ADAMS simulation are in accordance with the actual situation,
which verifies the rationality of the designed structure. The kinematics analysis and
research provide the theoretical and simulation basis for the subsequent product design.
586 H. Jingbo et al.
4 Trajectory Planning
The trajectory planning is to map each joint variable to a smooth time function [10],
which includes Cartesian space planning and joint space planning. The Cartesian space
planning is to use the Cartesian coordinate points’ sequences of the end effector position
to constitute a trajectory, which may reach the singularity. But the situation does not
appear in the joint space planning. In this paper, the trajectory planning of robot is carried
out by using the joint space planning method (Fig. 7).
As can be seen from Fig. 8, the kinematic curve in the x, y direction of the terminal
manipulator can be seen as a trigonometric function, and the kinematic curve in the z
direction can be considered basically no change, which is the same as the previous Adam
simulation results. If the image is furtherly solved, the correlation graph of velocity and
acceleration can be obtained. It can be seen that the trajectory is in accordance with the
operational requirements of the mobile manipulator robot, indicating that the path plan‐
ning method is reasonable.
Kinematics and Simulation Analysis of a 3-DOF Mobile Handling Robot 587
5 Conclusion
In this paper, the displacement curves of the three joints and the end point of the robot
are obtained, and the rationality of the structural design is verified, which provides the
theoretical and simulation basis for the next step of the robot design. The two-dimen‐
sional and three-dimensional displacement curves of the end of the manipulator further
validate the rationality of the mechanism and the correctness of the Adam simulation.
It establishes the foundation for the structure design, dynamic analysis and control
system of the robot.
References
1. Miao, D.: Design and Dynamic Performance Analysis of Heavy Load Handling Robot. Hefei
University of Technology (2014)
2. Li, R.: Development strategy of industrial robot industrialization in China. Aviat. Manuf.
Technol. 9, 32–37 (2010)
3. Li, C., Yang, Z., Cai, T.: A motion planning of moving robot based on ADAMS simulation.
Mech. Transm. 9, 28–31, 37 (2016)
4. Song, D.: Application of industrial robot in manufacturing industry. In: China Automotive
Industry and Equipment Manufacturing Industry Development Forum (2008)
5. Han, X., Li, C., Yu, X., Zhao G.: Modeling simulation of arc welding robot based on ADAMS/
view. Weld. J. 4, 69–72, 116 (2013)
6. Cai, Z., Xie, B.: Robotics, 3rd edn. Tsinghua University Press, Beijing
7. Ding, L., Li, E., Tan, M., Wang, Y.: Design and kinematics analysis of five-DOF moving
robot system. J. Huazhong Univ. Sci. Technol. S1, 19–22 (2015). (Natural Science Edition)
8. Chen, Z., Dong, Y.: Characteristics and Examples of MSC Adams Simulation of Multi-Body
Dynamics. China Water Resources and Hydropower Press, Beijing (2012)
9. Ma, R., Hao, S., Zheng, W.: Research on joint simulation of manipulator based on MATLAB
and ADAMS. J. Mech. Des. Manuf. 4, 93–95 (2010)
10. Wang, Y., Yu, X., Li, N., Zhu, W.: Kinematical analysis of ROBONOVA-1 robot. J. Xihua
Univ. 03, 6–9, 22 (2009). (Natural Science Edition)
A New Type of 3D Printing Nozzle with PET Wire
as Raw Material
Abstract. In this paper, the wasted PET bottles were analyzed. With the experi‐
ment analysis and the related literature, we determine the feasibility of the wasted
PET plastic bottles as 3D printing materials. We design and introduced the prin‐
ciple of the 3D print nozzle of using the positive and negative screw to extrude
material. The feasibility of the design is demonstrated by three - dimensional
modeling, physical experiment analysis and ANSYS simulation analysis.
1 Introduction
The “No.1” plastic bottle which is made of the PET material accounts for more than half
of the plastic bottle market. The output is huge, but its recovery rate is low [1]. The PET
material can be used as an excellent printing supplies, because of its strong adhesion
between the layers, good mobility, easy carbonation and other advantages. However,
due to the crystallization rate of the PET material is slow, the PET material can’t meet
the requirements of 3D printing technology rapid prototyping.
In recent years, FDM technology is the most used 3D printing technology, many
open source desktop 3D printers are mostly using this program [2]. FDM printing
supplies are mostly ABS and PLA [3], supplies are expensive, and ABS in the printing
process will release toxic gases. The use of ABS and PLA for FDM printing has some
shortcomings. For example, it is easy to break the wire, poor extrusion molded product
and product strength is not enough and the surface accuracy is poor. To improve the
printing performance, Kannan [4] added iron powder to the ABS, the addition of surface
active agent material made of iron powder or ABS composite. You Shu studied the effect
of 3D printing conditions on the mechanical properties of PLA plastics [5].
In this paper, we first studied the modification of materials. We have designed a
nucleating agent addition device, positive and negative threaded screw extrusion device,
heating and cooling system. The feasibility of the design is demonstrated by three-
dimensional modeling, physical experiment analysis and ANSYS simulation analysis.
The experiment two types of mainstream PET plastic bottles on the market were DSC
experiments, the experimental results shown in Figs. 1 and 2. From the DSC experi‐
mental data obtained, wasted PET material Tm is 240 °C–255 °C, the difference between
pure PET plastic bottle is not much, and it can be used for secondary processing.
Fig. 1. Nongfushanquan
Fig. 2. Yibao
It is helpful to increase the crystallization rate and improve the mechanical properties
of the finished product by adding nucleating agent to PET during the melting process
[6]. The results showed that the use of inorganic nucleating agent talc powder to modify
the PET material, when the nucleating agent to add the quality ratio of 5% [7], the
crystallization effect is better, the printing effect is the best.
590 Y. Hong et al.
The design of new nozzle structure includes nucleating agent addition device, positive
and negative threaded screw extrusion device, heating and cooling system. Figure 3 is
a graph model for explosion.
The nucleating agent adding device comprises a feeding screw rod and a material
cylinder. It can achieve the purpose of accurately and stably adding nucleating agent.
The extruding device is composed of a mixing cylinder and a positive and negative
threaded screw extrusion device.
The heating device is arranged outside the mixing cylinder and is heated by a heating
rod, and the temperature is accurately controlled by a temperature sensor. The heat
dissipation system adopts the principle of air pump blowing. The heat dissipation effect
is better, the energy consumption is lower, and the noise is smaller. A schematic diagram
of the heat sink is shown in Fig. 4.
thread feeding extrusion effect. The middle part is anti thread, and the screw thread and
the inner wall of the sleeve are buffered by a certain clearance between the threads and
the inner wall of the sleeve. PET sheet material into the above material inlet, PET mate‐
rial into the molten state, it will melt the screw thread of PET material is transported
downwards and the formation of certain anti back flow in the thread, and after the
nucleating agent is fully mixed, melted PET material is down from the nozzle at the
lower end of the screw extrusion molding.
Figure 6 is the positive and negative screw in kind. Figure 7 shows the nozzle’s
working process.
Fig. 6. The positive and negative screw Fig. 7. The nozzle’s working process
4 Conclusion
The nozzle we designed completes the modification and extrusion molding of waste
PET material. The whole process from the plastic bottle to the actual model can be
realized. PET material does not produce toxic substances during printing which is more
environmentally friendly Compared to ABS. However the melting range is too narrow.
Suitable temperature is an important indicator to measure the printing effect. How to
control the printing temperature accurately is a problem we need to solve. Due to the
complex internal mechanical structure of the nozzle, the vibration amplitude of the
nozzle has a great influence on the model printing accuracy. It is the important direction
of mechanical debugging to make the printing more smoothly. This design has solved
the current printer can’t use PET material as 3D printing raw materials. Waste PET
material is cheap, the cost of printing is reduced by about 90% Compared to ABS and
PLA materials which provides a new opportunity for the development and wide appli‐
cation of 3D printing technology [9].
References
1. Tang, G., Hu, B., Kang, Z.: Current situation and problems of waste plastics recycling. Renew.
Resour. Recycl. Econ. 6(1) (2013)
2. Guo, K.: Three-dimensional printing equipment and data processing software a number of key
technology research. Huazhong University of Science and Technology (2008)
3. Melnikova, R., Ehrmann, A., Finsterbusch, K.: 3D printing of textile-based structures by Fused
Deposition Modelling (FDM) with different polymer materials. Mater. Sci. Eng. 62(1), 12–18
(2012)
4. Kannan, S., Senthilkumaran, D., Elangovan, K.: Development of composite materials by rapid
prototyping technology using FDM method. In: International Conference on Current Trends
in Engineering and Technology, vol. 13, pp. 281–284 (2013)
5. You, S., Hu, Y., Wei, Q.: Effects of 3D printing conditions on the mechanical properties of
degradable polylactic acid. Plastics 29(3), 91–94 (2015)
594 Y. Hong et al.
6. Legras, R., Dekoninck, J.M., Vanzieleghem, A., Mercier, J.P., Nield, E.: Crystallization of
poly (ethylene terephthalate) induced by organic salts: model compound study of the
mechanism of action of the nucleating agent. Polymer 27, 109–117 (1986)
7. Ailan, Z.: Application of nucleating agents in crystallization of PET. Polyester Ind. 05, 1–5
(2011)
8. Gao, J.: FDM rapid prototyping machine temperature field and stress field numerical simulation
9. Sun, J., Tong, Z., Ying, Z.: Analysis of the market development prospect and application of
3D printing technology. Wuhan Univ. Technol. 6(1), 1672–3918 (2014)
When Partitioning Works and When It Doesn’t:
An Empirical Study on Cache Way Partitioning
Hanfeng Qin ✉
( )
School of Computer Science and Technology, Huazhong University of Science and Technology,
Wuhan 430074, China
hanfengtsin@hust.edu.cn
1 Introduction
The poor resource utilization in data centers increases the total cost of ownership (TCO)
of IT service. For example, the average utilization achieves only around 6% to 12% in
Google’s data center [10]. Recent efforts in virtualization and cloud computing tech‐
nologies are promising to improve resource utilization by consolidating many services
on the same server in data centers. However, co-locating more applications has the risk
of degraded quality-of-service (QoS) due to uncontrolled contention for shared
resources, primarily the last-level caches (LLC) [18, 19]. Researchers leverage cache
partitioning techniques [17] to address the arbitrary access to the shared LLC. As an
important performance isolation mechanism, cache partitioning has been proposed to
restrict the available amount of shared cache lines that an applications can access when
it co-runs with other workloads. Depending on the optimization target, generally, a cache
performance when cache partitioning is applied to programs that have less similarity
in memory access intensity of high memory intensity. Co-scheduling programs of
heterogeneous cache way sensitivity can make effective use of cache partitioning.
(iii) The overhead of dynamic partitioning adjustment impacts the performance gains
of cache partitioning. Design a partitioning algorithm with a steady partitioning size
from less frequently adjustment can improve performance.
2 Motivation
As the capacity of last-level cache (LLC) in modern x86 processors is ever growing.
It is important to understand how to make effective use of cache partitioning techniques
in data center. For example, Intel releases Xeon E7-8893 v4 processor equipped with
an LLC of 60 MB, and IBM Power 8 associates a larger LLC of 96 MB. It is important
to understand how to make effective use of cache partitioning techniques in data center.
Although has been proposed for a decade around, way-partitioning has still been an
active baseline in cache partitioning research [1, 12, 19], even in complex workload in
context of cloud computing [3] and warehouse computer [7]. In practical production
system and market, Intel releases a cache allocation techniques (CAT) [4, 6] in Haswell
SKUs processor, which is also based on way-partitioning.
Although cache partitioning techniques have been extensively explored, how to use
cache partitioning effectively is still a problem and not well understood. We carry an
experiment and find that cache partitioning does not always work as an expected winner.
We consolidate various quad-core workload mixes to co-run on the same multicore
server that supports CAT, and measure their performance under the baseline LRU and
a CAT-based partitioning policy, respectively. Figure 1 reports the results under the
baseline system and CAT-based partitioning. We find that cache partitioning has diverse
impact on performance for different workload mixes. For some workloads cache parti‐
tioning outperforms LRU indeed as expected. But for some workloads, the expected
benefits of partitioning vanish. Even worse, it result in unexpected performance
598 H. Qin
degradation for some workloads under partitioning. We are very interested in seeking
the answer to this unexpected behavior which has not ever been reported in prior work.
In this paper, we are motivated to explore when partitioning works and when it does not,
and how to make effective use of it.
Fig. 1. Performance comparison for various workload mixes under the baseline LRU and the
Intel CAT-based partitioning replacement policy
3 Experimental Methodology
3.1 Simulator
We use an event-driven ZSim [16] to model our baseline multicore system. ZSim is
a fast x86 micro-architecture simulator based on Intel Pin [11]. We closely model an
Intel Sandy-bridge processor, of which configuration parameters are detailed
presented in Table 2. Each core has its private split 32 KB 4-way associative L1
instruction and data cache, and a unified 256 KB 8-way associative L2 cache. A
unified 4 MB 16-way associative L3 cache (LLC) is shared by all cores. The default
LLC replacement policy in our baseline system is LRU. We choose Utility Cache
Partitioning (UCP) [14] as way-partitioning candidate, and implement an UCP-based
replacement policy, labeled as WayPart in Sect. 4. We configure UCP with a per-
core 256-line 16-way associative UMON circuits, and enforce the Lookahead algo‐
rithm to dynamically resize the partitions every 5000 cycles, similar as [14].
3.2 Workloads
The SPEC CPU2006 [5] suite is used to perform evaluation. We compile each program
using GCC 6.2.1 with an optimization flag of -O3, and feed them with a single reference
input. A representative slice of 500 million instructions for each program is identified
with PinPoint [13]. To characterize the memory behavior, we perform both way sensi‐
tivity analysis and memory intensity analysis for the total 29 programs, respectively.
increases, but would not increase any more at some point even more ways are assigned.
Way step (WT) programs improve performance as cache way increases, and they show
an increased steady phase. Figure 2 depicts the correlation curves of performance and
cache ways for typical workload within these categories.
Memory Access Intensive Analysis. We also study the memory access intensity in
terms of misses per thousand instructions (MPKI) and classify programs into four cate‐
gories: High intensive (HI) programs have MPKI larger than 30, thus generate very high
memory pressure. Medium intensive (MI) programs have MPKI larger than 10 but less
than 30. Low intensive (LI) programs have MPKI larger than 1 but less than 10. Non-
intensive (NI) programs have lower MPKI than 1 and thus they do not contend for the
memory bandwidth.
Combining both memory intensity and way sensitivity, we group the total 29
programs into 11 subsets as shown in Table 3. We perform evaluation with a total of 77
quad-workloads mixed with programs from these categories by varying memory inten‐
sity and way sensitivity.
∑ IPCishared
Weighted Speedup = ,
IPCialone
When Partitioning Works and When It Doesn’t 601
where both the IPCishared and IPCialone are performances of the i -th program when it co-
runs with other programs, and that when it runs alone, respectively.
4 Empirical Studies
In this section, we report the results of our empirical studies. We are restricted to present
the evaluation to a subset of these workloads due to page limit instead of the total 77
workload mixes, nevertheless, the results are also applied to the remaining workloads.
Specially, we perform evaluations on workload mixes of medium memory intensity and
varying way sensitivity as shown in Table 4 in studying the impact of cache configura‐
tion, cache way sensitivity, and partitioning variations to performance. We use workload
mixes of way step sensitivity and varying memory intensity as shown in Table 5 to
investigate the impact of memory intensity to performance.
Table 4. Workload mixes of medium memory intensity and varying way sensitivity
Category Workload mixes Category Workload mixes
4WS (450,471,483,471) 4WT (437,482,459,437)
4WI (462,433,462,433) 3WS + 1WI (450,471,483,462)
3WS + 1WT (450,471,483,459) 2WS + 2WI (471,483,462,433)
2WS + 1WI + 1WT (483,450,433,459) 2WS + 2WT (471,483,437,482)
1WS + 1WI + 2WT (450,462,437,482) 1WS + 3WI (450,462,433,462)
1WS + 2WI + 1WT (471,462,433,459) 1WS + 3WT (471,437,482,459)
Table 5. Workload mixes of way step sensitivity and varying memory intensity
Category Workload mixes Category Workload mixes
1HI + 3LI (473,434,445,436) 1HI + 1MI + 1LI + 1NI (473,459,436,435)
1HI + 3MI (429,437,482,459) 1HI + 1MI + 2LI (473,437,434,445)
2HI + 2MI (429,473,437,482) 1HI + 1MI + 2NI (429,482,435,400)
2HI + 2LI (429,473,445,436) 1HI + 2MI + 1LI (429,482,459,434)
2HI + 2NI (429,473,435,400) 1HI + 2MI + 1NI (473,437,482,400)
3HI + 1MI (429,473,429,482) 2HI + 1MI + 1LI (429,473,459,434)
3HI + 1LI (429,473,429,436) 2HI + 1MI + 1NI (429,473,437,435)
3HI + 1NI (429,473,429,435) 2HI + 1LI + 1NI (429,473,445,400)
1HI + 3NI (429,400,435,400)
Cache Set Number. Figure 3 reports the normalized performances across a group of
cache size from 4 MB to 16 MB by increasing the number of cache sets while keeping
a fixed cache way number. We can see that the set number has a direct impact of
performance gains of way-partitioning. To our surprised, the baseline LRU outperforms
way-partitioning for most workloads (7 out of 12) with a small cache of 4 MB. As we
602 H. Qin
increase the set number, way-partitioning improves these workloads gradually. When
the cache size is increase to 16 MB, besides 2 workloads which have the similar
performance under both partitioning and LRU, way-partitioning wins LRU for 10 out
of the total 12 workloads, including the 7 workloads which has poor performance in
small caches of 4 MB. Way-partitioning achieves a normalized performance improve‐
ment to LRU at a geometric mean of 11.5% in the large cache of 16 MB.
Fig. 3. Normalized performance of way-partitioning to LRU in LLC of the same way numbers
but different set numbers
Fig. 4. Normalized performance of way-partitioning to LRU in LLC of the same set numbers
but different way numbers
Cache Way Number. We set a fixed cache size of 4 MB but varying the associativity
to 16-way and 32-way, respectively. Figure 4 reports the normalized performances of
way-partitioning to LRU with an associativity of 16-way and 32-way. Considering that
the workloads used in this evaluation are picked in terms of cache way sensitivity, we
expect to see performance improvement for workload mixes containing more cache way
sensitive programs under a high associative cache. However, we hardly observe any
performance improvement for all workload mixes, only by a geometric mean of 1%, as
we increase the cache associativity. Compared with the impact of increasing the cache
When Partitioning Works and When It Doesn’t 603
set number, the performance improvement from increasing cache way number is tiny
(1% v.s. 11.5%).
The different impact of cache set number and cache way number attributes to the
amount of conflict misses. As the cache set number increases, the address stream is
scattered among more cache sets, consequently, decreases the conflict misses within a
cache set. However, increasing the cache way number decreases capacity misses but
does not prevent conflict misses. Excessive conflict misses diminish the performance
gains form cache partitioning.
Fig. 5. Normalized performance of way-partitioning to LRU for workload mixes of cache way
step sensitivity programs under different memory intensities
604 H. Qin
Table 6. Cache misses comparison of workload mixes with high diverse memory access intensity
under LRU and way-partitioning
Cache misses 473.astar 429.mcf 473.astar 436.cacutsADM
LRU 45.38 68.27 45.38 5.04
WayPart 48.29 64.60 48.28 102.74
Fig. 6. Cache blocks received by each program in a medium memory intensive workload mixes
consisting of one way sensitive and three way step sensitive programs under the Lookahead
partitioning algorithm
4.4 Implications
We summarize our findings and their implications to future cache system design or
effective cache resource exploration for cloud data centers as follows.
• Cache partitioning does not work in caches of small capacity. Increasing capacity by
increasing the set number instead of the way number can improve the performance
gains of cache partitioning. Considering the design complication, high overhead and
energy of associative lookup, it is not beneficial to design high-associative caches in
data center because they do not make cache partitioning work better.
• Consolidating services in data centers should take the memory characteristics of
programs into account since they have direct impact on performance gains of cache
partitioning. It degrades performance when cache partitioning is applied to programs
that have less similarity in memory access intensity of high memory intensity. Co-
scheduling programs of heterogeneous cache way sensitivity can make effective use
of cache partitioning.
• The overhead of dynamic partitioning adjustment impacts the performance gains of
cache partitioning. Design a partitioning algorithm with a steady partitioning size
from less frequently adjustment can improve performance.
5 Conclusion
Service consolidation is promising to improve the poor resource utilization in cloud data
centers but at a risk of suffering performance due to uncontrolled access to shared last-
level cache. Although cache partitioning schemes have been exploited extensively, how
to make effect use of cache partitioning is still not well understood. In this paper, we are
seeking to explore when partitioning works and when it does not with an empirical study
on a commonly used way-partitioning policy for a variety of workloads. We investigate
the impact of cache configuration, memory characteristic of program, and partitioning
variation to the performance gain under partitioning. We identify several interesting find‐
ings which help us in future cache system design and optimization for cloud data centers.
606 H. Qin
References
1. Cook, H., Moreto, M., Bird, S., Dao, K., Patterson, D.A., Asanovic, K.: A hardware evaluation
of cache partitioning to improve utilization and energy-efficiency while preserving
responsiveness. In: Proceedings of ISCA, pp. 308–319 (2013)
2. Eyerman, S., Eeckhout, L.: System-level performance metrics for multiprogram workloads.
IEEE Micro 28(3), 42–53 (2008)
3. Ferdman, M., Adileh, A., Kocberber, O., Volos, S., Alisafaee, M., Jevdjic, D., Kaynak, C.,
Popescu, A.D., Ailamaki, A., Falsafi, B.: Clearing the clouds: a study of emerging scale-out
workloads on modern hardware. In: Proceedings of ASPLOS, pp. 37–48 (2012)
4. Funaro, L., Ben-Yehuda, O.A., Schuster, A.: Ginseng: market-driven LLC allocation. In:
Proceedings of ATC, pp. 295–308 (2016)
5. Henning, J.L.: SPEC CPU2006 benchmark descriptions. ACM SIGARCH Comput. Archit.
News 34(4), 1–17 (2006)
6. Herdrich, A., Verplanke, E., Autee, P., Illikkal, R., Gianos, C., Singhal, R., Iyer, R.: Cache
QoS: from concept to reality in the Intel® Xeon® processor e5-2600 v3 product family. In:
Proceedings of HPCA, pp. 657–668 (2016)
7. Kanev, S., Darago, J.P., Hazelwood, K., Ranganathan, P., Moseley, T., Wei, G.Y., Brooks,
D.: Profiling a warehouse-scale computer. In: Proceedings of ISCA, pp. 158–169 (2015)
8. Kasture, H., Sanchez, D.: Ubik: efficient cache sharing with strict QoS for latency-critical
workloads. In: Proceedings of ASPLOS, pp. 729–742 (2014)
9. Lin, J., Lu, Q., Ding, X., Zhang, Z., Zhang, X., Sadayappan, P.: Gaining insights into multicore
cache partitioning: bridging the gap between simulation and real systems. In: Proceedings of
HPCA, pp. 367–378 (2008)
10. Lo, D., Cheng, L., Govindaraju, R., Barroso, L.A., Kozyrakis, C.: Towards energy propor-
tionality for large-scale latency-critical workloads. In: Proceedings of ISCA, pp. 301–312
(2014)
11. Luk, C., Cohn, R., Muth, R., Patil, H., Klauser, A., Lowney, G., Wallace, S., Reddi, V.J.,
Hazelwood, K.: Pin: building customized program analysis tools with dynamic
instrumentation. In: Proceedings of PLDI, pp. 190–200 (2005)
12. Pan, A., Pai, V.S.: Imbalanced cache partitioning for balanced data-parallel programs. In:
Proceedings of MICRO, pp. 297–309 (2013)
13. Patil, H., Cohn, R., Charney, M., Kapoor, R., Sun, A., Karunanidhi, A.: PinPointing
representative portions of large Intel® Itanium® programs with dynamic instrumentation. In:
Proceedings of MICRO, pp. 81–92 (2004)
14. Qureshi, M.K., Patt, Y.N.: Utility-based cache partitioning: a low-overhead, high
performance, runtime mechanism to partition shared caches. In: Proceedings of MICRO, pp.
423–432 (2006)
15. Sanchez, D., Kozyrakis, C.: Vantage: scalable and efficient fine-grain cache partitioning. In:
Proceedings of ISCA, pp. 57–68 (2011)
16. Sanchez, D., Kozyrakis, C.: ZSim: fast and accurate microarchitectural simulation of
thousand-core systems. In: Proceedings of ISCA, pp. 475–486 (2013)
17. Suh, G.E., Rudolph, L., Devadas, S.: Dynamic partitioning of shared cache memory. J.
Supercomput. 28(1), 7–26 (2004)
18. Tang, L., Mars, J., Vachharajani, N., Hundt, R., Soffa, M.L.: The impact of memory subsystem
resource sharing on datacenter applications. In: Proceedings of ISCA, pp. 283–294. ACM
(2011)
When Partitioning Works and When It Doesn’t 607
19. Yang, H., Breslow, A., Mars, J., Tang, L.: Bubble-Flux: precise online QoS management for
increased utilization in warehouse scale computers. In: Proceedings of ISCA, pp. 607–618
(2013)
20. Ye, Y., West, R., Cheng, Z., Li, Y.: COLORIS: a dynamic cache partitioning system using
page coloring. In: Proceedings of PACT, pp. 381–392 (2014)
Track Maintenance Feedback Mechanism
Based on Hadoop Big Data Analysis
Yong Zhu ✉ , Jiawei Fan ✉ , Guangyue Liu ✉ , Mou Wang ✉ , and Qian Wang ✉
( ) ( ) ( ) ( ) ( )
Intelligent Manufacturing and Control Institute, Wuhan University of Technology, Wuhan, China
zhuyong2016@gmail.com, 1228915795@qq.com, 491534654@qq.com,
137078388@qq.com, 775316545@qq.com
Abstract. With the rapid development of economy and the people’s growing
material needs, increased frequency and intensity of railway transportation, the
requirement of increasing the railway maintenance, security is becoming more
and more attention. The current routine of daily maintenance is done mainly by
manual and large rail inspection vehicles. The maintenance method is of high
strength, low efficiency, high risk and low maintenance accuracy. Based on the
above background, the project team has designed an efficient track inspection
machine based on the collaborative working method of the mother-machine. The
railway maintenance and data collection is achieved through the collaborative
work of the mother-machine. In this case, the mother machine detects and collects
the data, the sub-machine repairs and collects the data, the upper machine imple‐
ments the coordination, the big data processing and the feedback system. Data
collected by a railway big data, to take advantage of these data, the team set up
big data processing system based on hadoop, adopting clustering analysis, inte‐
grated analysis and time prediction analysis method, experience about defect
distribution map, so as to optimize the workings of a composite aircraft, constantly
improve the maintenance system based on composite aircraft performance. The
design of the project team is based on the system of the railway maintenance
system, which is intelligent and timely. Can be automated and dehumanized,
realize railway maintenance, and can improve the efficiency of railway mainte‐
nance system and reduce cost.
irregularity, orbital defect, bolt looseness and other issues detection and maintenance.
However, because of the wide distribution of China’s railway and large the number of
long tracks, China’s railway sector mainly rely on artificial or semi-automated machi‐
nery inspection method to detect and repair rail damage. Under this background, the
project team designed an efficient rail maintenance machine which based on the coherent
work mode of the parent machine. At the same time, the machine can also achieve the
rail detection and maintenance of automation and intelligence. The most important is,
people could establish a big data processing system on the hadoop based on the process
of detection and maintenance of this machine.
Although the current equipment has data collection and data feedback, but it did not
achieve real-time interaction, the two part are distributed with a certain lag. The data
can not be timely fed back to the data processing system through the host computer after
the collection. And the data which has been dealt by the data processing terminal could
not be sent back to the work of equipment in time, so it could not met the requirement
of immediate maintenance.
Through the collection of big data on the railway, the Hadoop for data processing
and integration, and by the hadoop system for information feedback, the equipment
could examines the defects of the relevant data more accurately according the relevant
data. Besides, people could start a directional work through the integration of informa‐
tion and the types of defects and multiple locations. So the fixed section of the railway
can be detected, and the efficiency can be improved.
From the background of railway’s development, we can see that there are two kinds of
information on the existing railway tracks. Among them, irregularity refers to the orbital
geometry, size and spatial position of the deviation. In the broadest sense, the position
of the center line of the linear track, the height of the track, deviates; the curve center
curve deviation; curvature, high, gauge value, slope changes in the size of the deviation,
collectively referred to as the track is not smooth. Track surface defects include cracks
of rail surface, abrasions, blocks, and the surface appears dark or black lines and so on.
This information, needs to work through the way composite machine to work to collect.
Track maintenance machine is as shown in Fig. 1 [1].
The acquisition process is divided into three parts. The first is the image, instrument
and other information of data acquisition. Then these directly collected information,
through the processing of data conversion, into the corresponding digital information.
Finally, the data finished, and through digital-analog conversion, the digital information
combined with positioning information into a digital group, and through the FPI bus-4G
communication networks to the host computer. The concrete data of detection and
acquisition methods are as follows:
In order to obtain the above non-smooth information, we use the parent machine in
the detection of the parent machine for data collection. For different types of irregular‐
ities, we use different data collection. At work, each testing hardware works simultane‐
ously. Using displacement sensors to measure displacement variables, the actual gauge
values are added to the standard gauge. Using gyroscope to get the machine’s side rolling
Angle, the high value is calculated through trigonometric relation. The lateral acceler‐
ation of the vehicle is measured by the transverse acceleration sensor, and the integral
is obtained by the horizontal displacement deviation. By using the vertical acceleration
sensor, the vertical acceleration of the body is measured, and the integral obtains the
vertical displacement offset, and the calculation is high and low [2].
categories: One is the limitation of the manufacturing process, the defects of the steel
rail during the forging process; The other is generated by the high intensity and high
density fatigue wear of the locomotive and rail.
As a result, we use machine vision technology to detect defects caused by this insta‐
bility. By CCD camera track surface image acquisition in the first place, again by FPGA
combined with single chip microcomputer to feature extraction, image through image
processing the final defect type and location. Then, according to the image pixel infor‐
mation, you can coordinate the defect [3].
After the data collection is the corresponding data processing, in this project, the most
important is the analysis of these data analysis and data feedback after this, you can
simplify the construction process of maintenance machinery in the railway track,
improve the efficiency of rail maintenance and management.
For the railway track data can be divided into structured data and unstructured data
two categories. Unstructured data refer to the data tracked by the rail maintenance
machine in this project, the most important of which is the generating data after the
image processing, and the processing of this part of the data is the key to the project.
In the sub-machine work together for some time, it will inevitably produce a large
number of railway maintenance data, the analysis and optimization of the massive data
have certain influence on the distribution and proportion of the work on the railway.
Therefore, we are for the detection and collection of data, data processing, the analysis,
to obtain big data based on the railway maintenance profile, that is, “track spectrum.”
Through the data analysis of the track spectrum, we find the frequency distribution rule
of the daily maintenance work, so as to optimize the working scheduling of the coop‐
erating sub-machine, realize the principle of the twenty-eight in the course of railway
daily maintenance, improve the efficiency of overhaul and reduce the cost of mainte‐
nance.
Data monitoring and analysis platform as shown in Fig. 3.
Track Maintenance Feedback Mechanism 613
In order to analyse the data, we use two methods of clustering analysis and integration
analysis on the basis of Hadoop, and carry out the comprehensive evaluation and analysis
of the data by the two types of data and the state of track maintenance. Maintenance of
the real state and its forecast analysis, and targeted adjustment of the proportional
number of sub-machine maintenance and distribution of the scope of work. In the
following, two analytical methods will be briefly introduced:
The clustering analysis is divided into two parts: the test set and the training set. The
detailed analysis is as follows, Fig. 4.
Through the prediction of strength, we can get predictive state distribution of railway
track defects and irregularities and bolts loosening, based on Hadoop, so as to make
corresponding repair and distribution rules for the occurrence rate of defects of different
frequency in different regions, and further to achieve the principle of twenty-eight, to
improve the efficiency of railway railways daily maintenance, reduce the cost of railway
railways daily maintenance.
Based on the results of the above two analyzes, we can obtain the comprehensive distri‐
bution of railway track defects, irregularities, bolts and other information, and predict
the occurrence rate and time of occurrence of railway rails. For different regions,
different intensity and the frequency of maintenance planning, improve the efficiency
of maintenance sub-machine, reduce the corresponding cost, in order to achieve the
railway track maintenance operations really “twenty-eighty principle.” Maintenance
machine feedback workflow is shown as Fig. 5.
5 Summary
Today, the main way of railway maintenance is still labour, it exists high duration,
difficult and other defects. The current machinery can only be automated, can not be
intelligent, can not meet today’s needs.
This project designed a highly efficient intelligent rail maintenance machine, and
based on this design based on the hadoop for big data processing system, and its collec‐
tion, processing, analysis and feedback maintenance machine to detect the massive data,
the project has the following innovations:
(1) Mechanical structure design adaptive strong, can be automatically detected main‐
tenance operations
(2) Modular, integrated, systematic design can be adapted to a variety of complex
operating requirements
(3) Design of the data and instructions two-wire transmission function, to achieve the
machine’s online detection and status monitoring, and real-time access to the state
of the rail changes
(4) Designed data on-line analysis and big data processing platform to achieve the
analysis of the state of the railway maintenance and sub-machine work mechanism
feedback, shorten costs and improve efficiency
The project team uses cloud computing as a platform for heterogeneous and diver‐
sified data storage and analysis and the platform after the operation of the massive data
based on the maintenance of state maintenance, system feedback optimization, isolated
information system interoperability Support, and become a candidate after integration.
This work has a low cost, good system scalability (unlimited storage capacity), high
reliability, parallel analysis and so on, will become one important system of intelligent
way of railway maintenance in the future.
References
1. Chen, C., Kong, J., et al.: Modern Mechanical Designer Manual. Mechanical Industry Press,
Beijing (2014)
2. Wang, Y., Yu, Z., Bia, B., Xu, X., Zhu, L.: Study on crack identification algorithm of metro
tunnel based on image processing. J. Instrum. 07, 1489–1496 (2014)
3. Wang, Y.: Study on key technology of big data processing flow based on Hadoop. Inf. Technol.
09, 143–146, 151 (2014)
4. Ma, S., Wang, X., Fang, K.: Integration analysis of big data. J. Stat. Res. 11, 3–11 (2015)
5. Cao, Y.: Hadoop Performance Optimization in Big Data Environment. Dalian Maritime
University (2013)
6. Tang, D.: Hadoop-based affine propagation big data clustering analysis method. Comput. Eng.
Appl. 04, 29–34 (2015)
Crowdsourcing and Stigmergic Approaches
for (Swarm) Intelligent Transportation Systems
With an ever growing availability of embedded, mostly personal and mobile com-
puting devices for everyday tasks, there is an almost limitless potential for tap-
ping onboard resources, especially sensing-related ones, as well as corresponding
compute nodes to be exploited for locally executable tasks. Mobile CrowdSens-
ing (MCS) comprises by definition a category of applications where individuals
carrying sensor-hosting embedded computers (e.g. smartphones) get collectively
c Springer International Publishing AG 2018
Q. Zu and B. Hu (Eds.): HCC 2017, LNCS 10745, pp. 616–626, 2018.
https://doi.org/10.1007/978-3-319-74521-3_64
Crowdsourcing and Stigmergic Approaches 617
traffic accordingly by e.g. network coding strategies, otherwise leaving the back-
end (and the network itself) prone to scalability issues over huge population
scenarios.
Given the paradigm, i.e. MCS, and forthcoming use cases, with specific regard
to ITS where mobility is really going to match crowds at scale, we conceived a
design pattern. Such a scheme may lend resilience when faced with big swarms,
while at the same time helping DevOps with their coding and deployment strate-
gies. More specifically, collaborative approaches, collective and swarm intelligent
ones could be good metaheuristics than can provide effective solutions to ITS
problems such as vehicle routing, usually NP-hard. The collaboration of multiple
agents is the best solution available in the ITS scenario to reduce latency and
be effective in addressing the routing problem also when traffic conditions must
be considered. This perfectly matches with crowdsourcing based paradigms as
MCS, mainly aiming at supporting applications able to exploit the collective
intelligence and their emerging properties in problem solving. In this paper, we
propose to address the routing problem adopting a swarm intelligence approach
able to take into account the road traffic conditions. It mainly consists of an ant
colony optimization (ACO) algorithm implemented and deployed on mobiles
constituting an MCS-contributed infrastructure, able to interact each other fol-
lowing an opportunistic patterns. We therefore framed our approach in the class
of Swarm ITS (SITS).
In the following, we are going to first lay out an overview of ITS systems, then
discussing MCS solutions for them. After that, we focus on swarm intelligence,
stigmergy and ACO, coming up with SITS. This way, we define our traffic-
aware vehicle routing solution based on a modified version of an ACO. This
SITS solution is thus evaluated by a simulator which preliminary results are
discussed. Finally some remarks and hints for future work close the paper.
of food are found. As other colony members step into pheromone trails, they
tend to stick to the beaten path accordingly. Moreover, the trace gets reinforced
as individuals follow the same trail, leaving pheromone of their own, in turn
resulting increasingly attractive for other ants. For any complex problem which
can be reduced to a search for optimal paths, ACO may work as a probabilistic
solver, by emulating such naturally occurring behavior. The probability pkij for
an artificial ant k, placed in vertex i, to move toward node j is defined as follows:
β
α
τij · ηij
pkij = (1)
α
l∈Nik (τil · ηilβ )
where τij corresponds to the quantity of pheromones laid over arc aij , ηij to
a-priori attractiveness of the move, computed by some heuristic embedding the
cost of choosing arc aij along the path that leads to the destination, and Nik is
the set of neighbours in node i for ant k, i.e., the nodes available for the ant to
move in. Coefficients α and β are global parameters for the algorithm. According
to typical ACO variants, ants bring food back home (i.e. nest) after being done
with their movement. Denoting Tk as the tour of ant k, Ck is defined as the
length of Tk , and used to specify the amount of pheromones to be placed by ant
k on each arc on the trail that led to the food source:
1 k
k C k if arc (i, j) belongs to T
Δτij = (2)
0 otherwise
m
k
τij = τij + Δτij . (3)
k=1
At the end of a round, after each ant has completed a move, the extent of
pheromones laid over each arc gets reduced (e.g. evaporates), according to:
the distance (i.e., cost) cdij from node i towards destination d along node j as
the sum of that between i and j, cij , and that from j to destination, cjd . Then
we specify the cost over a certain arc aij :
as the ratio between a certain property we want to use as metric, hij , and its
weight, wij .
Given all the above, the value of the a-priori gain, η j , for a certain choice
i→d
leading to destination d is computed according to the following formula:
η j = δ/c j (8)
i→d i→d
Even in this case, far for the solution from losing generality, on the contrary
the modification expands the scope of applicability, by relaxing constraints over
the scenario.
5 Preliminary Evaluation
To evaluate our proposed technique, we examined the traffic of an urban area,
near down-town the city of Messina. Arcs weights represent the length lij of each
road segment such that, using Eqs. 6 and 7 we can compute for each destination d
the cost to take the arc (i, j) in the path towards node d without considering the
road traffic. Providing these values as a metric to define the heuristic function
of the A* search algorithm, we can find the shortest path length, which can be
very different from the optimal path when the traffic is considered. We then
implemented a simulator of the overall system. The evaluation of the model will
provide a different metric to the heuristic function of the A* search algorithm
that takes in account also the traffic.
In the evaluation through the simulator we considered two scenarios where
either we take into account the pheromone value or not. According to Eq. 9 this
can be obtained by setting either α = −1 or α = 0. According to the results
thus obtained, in both cases the flows of traffic concentrate on the neighbors
of the specific destination and are directed towards it, thus confirming that
the messages go in the right directions. In the case not taking into account
pheromones we can observe a strong flow from node 1 to node 3 which is justified
by the presence of a source in node 1. Moreover, the preferred paths used to
reach node 3 are clearly visible. However, as expected, the pheromone effect
is to redistribute the traffic. In such case the flow on congested arcs decreases
indeed, while underexploited arc one increase.
To evaluate the effectiveness of the proposed approach in distributing the
overall traffic on the graph, we investigated the distribution of the aggregated
traffic (i.e. related to all destinations) for all the arcs. Since the traffic is evaluated
as a rate of an exponential process, we can compute the aggregated one just as
the sum of the traffics of each destination.
The resulting frequency or probability mass functions of the traffic with and
without pheromone are shown in Fig. 4. It can be observed that the pheromone
improves the distribution of the overall traffic intensities. Indeed, without
pheromone, several arcs present a high-intensity traffic, especially around the
values 0.2 and 0.55. The distribution thus obtained appears to be bimodal. When
the pheromone is considered, the traffic results more evenly distributed, with a
peak around the value 0.6. From these values, we can argue that the pheromone
allows the overall traffic to be more evenly distributed than when the pheromone
is not considered.
6 Conclusions
People, crowds and critical masses are becoming more and more powerful, not
only from an abstract point of view related to the opinion they express, but
also in more physical terms, due to their work potential. Indeed, volunteer and
crowd-based approaches are spreading like wild-fire across different disciplines
and sciences. Example are crowdsourcing, crowdfunding, geocomputing and vol-
unteer geographic information, just to name a few in different areas. A very fertile
ground for new approaches and technologies is computer science and engineering,
where volunteer and crowd-based ones have gained solid roots as in the cases of
crowdsourcing, crowdsearching, crowdsensing. In particular, Mobile Crowdsens-
ing is a very promising approach for involving people, citizen, crowds into sensing
campaigns according to participatory and/or opportunistic schemes. Although
the Mobile Crowdsensing paradigm is quickly rising interests and funds, there
is still untapped potential, as well as unexplored solutions this paradigm may
empower.
This paper tries to partially fill this gap by first defining a scenario and identi-
fying some specific features for a novel opportunistic, distributed, self-organizing
approach, applied to a specific class of MCS application, tackling local optimiza-
tion problems, in the ITS domain. The solution proposed adapts and extends
an ant colony optimization algorithm to a problem of pathfinding and graph
traversal according to a given distance metric. This way a new class of intelli-
gent transportation systems is identified: the swarm intelligent transportation
systems - SITS - one.
626 S. Distefano et al.
References
1. European Parliament: Directive 2010/40/EU, Brussels (2010)
2. Da Lio, M., Biral, F., Bertolazzi, E., Galvani, M., Bosetti, P., Windridge, D., Saroldi,
A., Tango, F.: Artificial co-drivers as a universal enabling technology for future intel-
ligent vehicles and transportation systems. IEEE Trans. Int. Transp. Syst. 16(1),
244–263 (2015)
3. Groot, N., De Schutter, B., Hellendoorn, H.: Toward system-optimal routing in
traffic networks: a reverse stackelberg game approach. IEEE Trans. Int. Transp.
Syst. 16(1), 29–40 (2015)
4. Du, L., Dao, H.: Information dissemination delay in vehicle-to-vehicle communi-
cation networks in a traffic stream. IEEE Trans. Int. Transp. Syst. 16(1), 66–80
(2015)
5. Lee, E., Lee, E.-K., Gerla, M., Oh, S.: Vehicular cloud networking: architecture and
design principles. IEEE Commun. Magaz. 52(2), 148–155 (2014)
6. Lv, Y., Duan, Y., Kang, W., Li, Z., Wang, F.-Y.: Traffic flow prediction with big
data: a deep learning approach. IEEE Trans. Int. Transp. Syst. 16, 1–9 (2014)
7. Hamdar, S., Talebpour, A., Dong, J.: Travel time reliability versus safety: a stochas-
tic hazard-based modeling approach. IEEE Trans. Int. Transp. Syst. 16(1), 264–273
(2015)
8. Ganti, R., Ye, F., Lei, H.: Mobile crowdsensing: current state and future challenges.
IEEE Commun. Magaz. 49(11), 32–39 (2011)
9. Dorigo, M., Stützle, T.: Ant Colony Optimization. Bradford Company, Scituate
(2004)
Research on Visual Feedback Based on Natural Gesture
1
School of Digital Media and Design Art, Beijing University of Posts and Telecommunications,
Beijing 100876, China
hwj1505@bupt.edu.cn, super_bupt@126.com
2
Network System and Network Culture Key Laboratory of Beijing, Beijing 100876, China
1 Introduction
Virtual and real interaction is a new and emerging form of human-computer interaction,
highlighting the virtual reality and augmented reality. With the help of powerful
computing and graphics capabilities of computers, it achieves more intelligent under‐
standing of human order and enables more input modes [1].
Various input modes expand the interaction between virtual and real worlds. In the
virtual and real interaction, the most reasonable and efficient interaction should be
natural gestures. It allows users to abandon the external devices and interact with the
virtual scene.
There are plenty forms of feedback during virtual and real interaction, such as vision,
hearing, touch, space conversion etc. In virtual and real interaction, the content is mainly
based on the three-dimensional scene and model objects, and the input and output of
information are different.
The input of visual information accounts for more than 80% of the total intake of
information, and can provide interactive feedback immediately, which has a positive
effect on reducing the user’s cognitive load. Since the feedback comes from the percep‐
tion of vision [2], environment and input gestures, the visual feedback plays an important
role in virtual and real interaction.
2 Related Work
In view of the current research, the study of visual feedback among the virtual and real
interaction was divided into the following three directions: hardware development and
extensions, hardware algorithm based on recognition accuracy and natural gestures
design.
Mores Prachuabrued et al. explored the visual feedback when fingers penetrating
virtual objects during virtual grasping and evaluated the performance (penetration, finger
release time, accuracy) of several common visual feedbacks. Results showed that
recommend visual feedback is color change [3]. And another experiment showed the
combination of touch and visual feedback is optimal [4].
Faieza Abdul Aziz et al. studied visual feedback mechanism, they found out that
users achieved higher efficiency in finishing assigned tasks with visual feedback. In
addition, the results showed that color changes are more effective than text prompts [5].
Based on previous studies, YingKai designed visual feedback for virtual hand and
grasping objects. She finally conclued:1. overall performance was better than the local
performance. 2. visual change of objects was more effective than the change of virtual
hand, 3. red color obtained the best feedback [6].
Zhang Wei et al. presented that static color recognition efficiency in virtual envi‐
ronment is far higher than in the real world, and suggested using dynamic [7] visual
feedback in the virtual scene.
Previous research had laid a certain foundation for related fields, but it also exposed
some shortcomings. In the virtual and real interaction visual feedback, it is not enough
only to study the changes of static visual feedback, and in the course of experiment, the
level of the user’s operation is not accurately regulated, so it is necessary to use the
auxiliary information of other channels to authenticate of conclusions under single visual
feedback.
In order to improve the research content in this field, on the basis of previous research,
we conducted an experimental study on natural grasping gestures in the virtual reality
interaction with respect to visual feedback. This paper mainly studied the difference and
validity of visual feedback form of the target object. The feedback form variables include
object color, transparency, flashing, highlight, rotation, shaking, scale, and deformation.
Besides, auxiliary electromyography and heart rate test were applied for evaluation.
Experimental results were analyzed.
3 Experiment Design
First, we define natural gestures. Natural gestures refer to gestures that do not add any
markings, do not wear any auxiliary equipment, and use bare hand to interact with
Research on Visual Feedback Based on Natural Gesture 629
objects directly. According to Wu and his co-author [8], a gesture includes Three stages,
concluding the following Fig. 1:
Fig. 1. Three stages of the gesture input, OOR means the device “out of range”
We defined the natural grasping gesture; sub-action can be divided into “encounter
object” and “pick up the object”. We introduced the way of visual feedback.
So how do you judge the validity of different visual feedback types in virtual reality
interactions? In fact, it is the in process of the grab gesture from “encounter object” to
“pick up the object”. From the “encounter object”, to see the visual feedback, and then
quickly “pick up the object”, the moment of seeing the visual feedback will differ
because of the different kinds of visual feedback; this difference can represent the
difference of feedback ways, and then we can analyze its effectiveness. On this basis,
we construct the experimental platform, and do the experimental study.
The experiment used Unity 5.4 as the platform, C# as the programming language, and
Leap Motion camera to input the gesture coordinates and use grasping gestures to
perform experimental operations and record data.
At the same time, the Ergo LAB physiological detection equipment was used to
detect the EMG and heart rate in the experiment. The physiological indexes of the
subjects were recorded under different visual feedback conditions (Fig. 2). The EMG
detector (sEMG) electrode was pasted at the extensor carpi radials brevis muscle of the
right arm [9], which was the main muscle of the activity in the operation of the grasping
gesture.
630 W. Hou et al.
Fig. 2. Based on leap motion system and Ergo LAB, visual feedback - cognitive experiment
Object color, transparency change - For the feedback form in this study, the independent
variable of object color change was red, the self-variable of object transparency change
was alpha equal to 0.4 concluding from the literature [6].
Object flashing - For the feedback form in this study, the independent variable of
object blink control was the red and transparent alternation of the object in this paper [7].
Object highlighting – The paper [10] suggested that in the three-dimensional virtual
scene, the use of highlighted form to interact with object was desirable. So, we designed
highlighting color selection red, display on the edge of the small ball.
Rotation, shaking, and scaling of objects – The author [11] used scaling, translation
and rotation operations in interactive operations of 3D models in virtual reality, so we
chose these three ways in the experiment.
Object deformation - When the virtual finger is grasping, the places where fingers
and small ball collide will have deformation [12], which control the deformation of the
radius and the intensity in the appropriate range, in order to achieve the deformation of
the feedback form. The interface was shown in Fig. 3.
Research on Visual Feedback Based on Natural Gesture 631
After pre-tests and screening, 20 participants were selected: 10 boys and 10 girls, aged
from 20 to 25 (mean age 23.4), were all right handed and familiar with computer oper‐
ation. All participants were in good health and no symptoms of physical fatigue such as
massive exercise before the test. All participants were willing to accept the experiment
voluntarily.
Before the start of experiment, the Ergo LAB device will be fixed, and the heart
index detector (HRV) was clamped on the left index finger (HRV) of the subject, wipe
the right side of the arm extensor parts with alcohol and fixed the electromyography
(sEMG).
The experiment will begin after the detection of heart rate and myoelectric physio‐
logical signal remaining stable. The experiment was divided into 2 groups and 11
experiments. Each process requires 3 s of rest after the participant completes each
grasping, each procedure include 5 times of this process. After each procedure, let the
participant take a short break while saving Ergo LAB physiological data. And so on,
followed by the implementation of 11 experimental process, the data will automatically
be saved in the local file. After the experiment was completed, the subjects were graded
on the satisfaction of several visual feedback methods, and they were graded from -2,
-1, 1, and 2 by Likert 5 scale.
Reaction time ΔT = T1−T2, which T1 was the time when the subject encounters the
ball to the release of the ball, T2 was the time when the user begin performing the
grasping gesture to the release. Therefore, ΔT could be expressed as the user to see the
feedback to the beginning of the implementation of the time difference, that is, the reac‐
tion time, as the form of feedback effectiveness of the important indicators. Satisfaction
S was the satisfaction of the user after the completion of the experiment, the evaluation
of the eight variables of the satisfaction score, respectively, -2, -1, 0, 1, 2, post-data
analysis was standardized, all returned to 0 to -1 score.
The physiological signal data were sorted out in the Ergo LAB analysis software,
and the fragments from the signal equalization change to the signal equalization change
were selected and processed for analysis. Analysis of sEMG signal according to the
literature [13] selected these three data indicators commonly used in time domain and
frequency domain analysis. Of this, YAverage represented the average level of the
physiological signal amplitude of the segment, YVariance represented the amplitude
variance of the physiological signal of the fragment. They all obey to the zero mean
Gaussian distribution, proportional to the degree of muscle contraction; AVHR repre‐
sents the mean heart rate of the segment.
4 Experiment Design
It can be seen from the chart that the first ΔT without feedback was much larger than
that of the other eight feedback modes, indicating that the operational efficiency value
in the form of visual feedback was improved and the visual feedback mode is reasonable
and necessary;
The ΔT of the highlight-feedback was the smallest, and the user satisfactions were
also the highest. The ΔT of these four modes of feedback such as rotation-feedback,
shake-feedback, scale-feedback, and deformation-feedback were too large, and
customer satisfaction were relatively low, the initial description of these the form of
feedback was inappropriate.
In order to test the rationality of the ΔT and S data results, ΔT and S were tested for
paired samples T, and the results showed that, in addition to the transparent and deformed
feedback methods, the other six feedback patterns had a significant correlation in the
relevance of the sample sig <0.05, showing a high degree of correlation, which was also
consistent with the description of the statistical feedback time is small and high satis‐
faction Feedback time and satisfaction with low consistency.
In the process of data analysis, we found that the results of different variables showed
a certain consistency, such as color, transparency, highlight the feedback method are
better. And rotation, scaling, shaking, deformation of these feedback methods was rela‐
tively poor, but the results were also close to the guess may also be a certain degree of
relevance. Therefore, cluster analysis was performed and the clustering results were
shown in Fig. 4.
The data of the physiological indexes after the errata was sorted out, and the mean
value of the case variables was output. At the same time, the satisfaction score was
normalized and the overall data results were presented in Table 2.
From the overall analysis, the physiological index data of different feedback forms
was different, we could guess that there existence a certain psychological model between
the visual and tactile perception.
Research on Visual Feedback Based on Natural Gesture 635
The average EMG and reaction time of visual feedback was positively related to the
reaction time, the greater the average EMG value is higher, but the effectiveness of the
feedback form is small;
The heart rate associated with visual feedback, visual feedback was more intense,
the heart rate is higher, but the effectiveness is not high, the use of context information,
such as presentation, warning effect, interaction effects etc.;
The muscle fatigue and cognition, more familiar and more acceptable form of feed‐
back of the lower degree of muscle fatigue, exercise performance was better than the
intuitive, people were more likely to perceive, but the effect was not good.
5 Conclusion
Through inner-group analysis and inter-group analysis, the data were analyzed by hori‐
zontal and vertical, respectively, summarized as follows:
(1) In general, the feedback form of visual expression was superior to the performance
of movement;
(2) The more obvious the physiological index, the strongest the stimulus. The color
feedback and highlight feedback performance were the best, and also had the
highest subjective satisfaction;
(3) The flashing feedback was the most intense stimulus, but it was not applicable in
interactive feedback, and subjective satisfaction was low;
(4) The deformation feedback was the most consistent cognitive approach to grasp
gestures, and was recommended in the virtual scene collision detection;
(5) The scale feedback was not the best in cognition, but it could reduce the user’s
operation fatigue and reduce the user’s learning process in interactive operation;
In summary, the form of visual feedback was suggested as follows:
In the virtual and real interaction, it was best to use the visual expression of visual
feedback form, mainly color, local performance and overall performance should be
considered;
636 W. Hou et al.
In scene content design, the dynamic visual feedback was more easily perceived than
the static;
In the multimodal input and output model, visual feedback of motor form was better
and more in line with human cognition;
In the multi-channel information input and output mode, the visual feedback of motor
expression form is better and more consistent with human cognition;
When there were a lot of visual feedback forms in the interaction of virtual and real
interaction, priority should be given to using a consistent and consistent feedback
method to select the correct form of visual expression and the performance of movement;
In this paper, we studied the effects of different visual feedback forms on cognitive and
user subjective satisfaction under the natural gesture. At the same time, we used phys‐
iological detection equipment to detect the physiological characteristics of the experi‐
ment process and concluded that visual feedback affects user reaction and user experi‐
ence to some extent, and found that there was a relationship between physical charac‐
teristics and visual feedback.
In this paper, the study of the type of visual feedback was still at the initial stage.
The variable level of scaling and deformation feedback was derived from the actual
project, the selected deformation parameters needed further study. In addition, visual
feedback had more dimensions and angles of variable types that need further study.
References
1. Zhao, Q.: Virtual reality review. Sci. China Ser. F Inf. Sci. 39(1), 2–46 (2009)
2. Zhao, Z.: Virtual reality based on human-computer interaction. J. Wuhan Univ. Sci. Technol.
(2007)
3. Prachyabrued, M., Borst, C.W.: Visual feedback for virtual grasping. In: IEEE Symposium
on 3D User Interfaces (3DUI) (2014)
4. Prachyabrued, M., Borst, C.W.: Virtual grasp release method and evaluation. Int. J. Hum
Comput Stud. 70(11), 828–848 (2012)
5. Prachyabrued, M., Borst, C.W.: Visual interpenetration tradeoffs in whole-hand virtual
grasping. In: Proceedings of IEEE 3DUI, pp. 39–42 (2012)
6. Ying, K.: Visual feedback for natural interaction study. Beijing University of Posts and
Telecommunications (2016)
7. Meng, F.-X., Zhang, W.: Way-finding during a fire emergency: an experimental study in a
virtual environment, vol. 57(6), June (2014). Taylor & Francis, Oxon. ISSN:0014-0139
8. Wu, M., Shen, C., Ryall, C., Balakrishnan, R.: Gesture registration, relaxation, and reuse for
multipoint direct-touch surfaces. In: Proceedings of the First IEEE International Workshop
on Horizontal Interactive Human-Computer Systems (TABLETOP 06), pp. 183–190 (2006)
9. Lin, Y., Zhang, F.: Carpi brevis tendon twin countershaft one case. Chin. J. Clin. Anat. 23(6),
582 (2005)
10. Chen, D.: Interactive real-time rendering of three-dimensional scene style of lead color.
Nanjing University (2016)
Research on Visual Feedback Based on Natural Gesture 637
11. Li, S.-S.: Interactive 3D model of operational studies. Dalian University of Technology
(2009)
12. Hou, W., Guo, Y., Li, T.-M., et al.: Research on three - dimensional object deformation
technology for natural interaction
13. Wang, J.: Research progress on signal analysis and its application of sEMG. J. Phys. Educ.
20(4), 56–60 (2000)
Information Security Technology and Application
in Logistics Traceability System of Aquatic Products
Based on QR Code
Abstract. In view of the security hidden danger of the business data in the
logistics traceability system of aquatic product, the design of the information
security strategy is carried out from the two angles, which are data collection and
transmission. Firstly, the aquatic traceability code is designed. Then the data
encryption algorithm is studied, and the data security transmission of information
security protection system is built. In the intelligent collection of traceability data,
QR Code is used to express the logistics trace data, then uses the improved data
encryption algorithm to encryption storage about QR Code. In the transmission
of traceability data, the HTTPS protocol is used to construct the client-server
transmission encryption channel to ensure the integrity of the data. Finally those
information security technologies about Logistics traceability data are used in a
logistics traceability system of aquatic product, and the effectiveness of RC4-RSA
hybrid encryption algorithm is verified. And the integrated applications enhance
the system’s information security.
1 Introduction
Symmetric encryption algorithms suit for a large amount of data encryption because
of its low complexity of decryption and fast encryption speed, but they also have key
shortcomings. Asymmetric algorithms have the feature for the high security, simple
public key public key management, but relatively slow encryption speed.
QR Code using in the aquatic products logistics traceability, information read a large
amount of encrypted data more, speed and security of encryption to protect data, using
symmetric encryption algorithm and asymmetric encryption algorithm combination.
Through the comparison of common algorithms in 3–1, this paper uses RC4 algorithm
and RSA algorithm for QR Code hybrid, the original plaintext data is encrypted by RC4
algorithm, improve the speed of encryption and decryption of the RC4 algorithm, key
is encrypted using the RSA algorithm, to ensure the security of the key, solve the problem
of key distribution and management, further to improve the security of encryption.
(3) Encryption scheme for QR Code
Use RC4 and RSA algorithms in the encoding link of QR Code, to encrypt the data.
The QR Code encryption process is shown in Fig. 2.
Information Security Technology and Application in Logistics Traceability System 641
Through the mixed encryption with RC4 and RSA algorithm, the speed and security
are improved while the security requirements of QR Code encryption and decryption
are ensured. When QR Code is used in the aquatic traceability system, the aquatic enter‐
prise uses the RSA algorithm to generate the public key and the private key, in which
the public key can be made public, and the private key is reserved inside the enterprise
for encryption. In the QR Code generating process, according to the encrypted data,
enterprise distribute RC4 key or randomly generate, guarantee different QR Code using
different encryption keys, dynamically generated QR Code encryption, and improve
system security. The logistics staff read or write information by professional equipment
with QR Code decryption algorithm, input the information to aquatic product traceability
system, and data acquisition and encryption to achieve traceability to logistics supply
chain, enhance the security of information.
is a low-diffusion state. Therefore, the RC4 algorithm has some weak keys and is
vulnerable to attack.
In order to improve the security of RC4 algorithm in QR code encryption, the
Logistic chaotic map is used to improve the randomness of pseudo-random sub-code
generation, which make a better diffusion effect and reduce the occurrence of weak key.
The so-called chaos is the disorder and random phenomenon in a deterministic system,
the chaotic sequence generated by the chaotic system has unpredictability [7].
Logistic one-dimensional mapping which is widely used is a chaos mapping of
mathematical form, the mathematical expression is:
( )
Xi+1 = μXi 1 − Xi (2.1)
② Generate the random key sequence by substituting the chaotic value and iteration in
the RC4 key scheduling algorithm.
Information Security Technology and Application in Logistics Traceability System 643
y=u*x(1-x);
x=y;
j=(j+S[i]+T[i]+ y*256)mod256;
swap(s[i],s[j]);
③ Recursive execution after adding chaotic maps during the process of generation and
replacement of Pseudo-random sub-code.
y=u*x(1-x);
x=y;
j=(j+S[i]+y*256)mod 256;
swap(S[i],S[j]);
t=(S[i]+S[j])mod 256;
key[r]=S[t];
Do the XOR operation for the sequence generated by the third step with the plaintext,
until the algorithm encryption step is completed. And the randomness of the random
code generation is greatly improved after using the Logistic mapping.
The encryption security of QR Code information is improved by RC4 based on
chaotic mapping and RSA algorithm, which are used to encrypt and decrypt the QR
Code. It can effectively prevent the forgery and tampering of logistics business data and
ensure the security of data in data acquisition and generation rooting.
In this example, the aquatic logistics traceability system is based on the background of
an aquatic product processing enterprises in Hubei Province, and it is constructed
according to the actual needs of the enterprises. The two-dimensional bar code is used
to collect the information intelligently and realize the resource sharing of aquatic prod‐
ucts. In the event of aquatic products quality problems, a quick inquiry to trace the
relevant breeding distribution information can found where the problem is through the
information contained in the two-dimensional code.
644 Q. Zu and R. Chen
From the logistics business data security considerations, use QR Code data encryp‐
tion to protect the security of data in the supply chain information transmission, and the
HTTPS protocol to ensure the safety of data transmission channel.
intermediate key formed in the encryption process, and according to the decrypted RC4
key to obtain the original of the processing information. Using of ordinary scanning
software without decryption procedures and RSA public key, can just access to a mean‐
ingless garbled.
The effect before and after the QR Code encryption is shown in Fig. 6.
(3.2)
π ∫x
erfc(x) is the complementary error function, and Sn is the result of the addition of the
random sequence to −1 and 1. By running the RC4 algorithm in MATLAB to obtain
the random sequence, and then conversion and addition to get Sn, the value of Sn into
646 Q. Zu and R. Chen
the formula, get P − Value, as shown in Fig. 7. It can be seen from the figure
P − Value > 0.01, in line with the requirements of the ideal random sequence.
Vn (obs) represents the sum of all 0 and 1 values in the random sequence of the RC4
algorithm, and n is its length. Replace the relevant parameters in MATLAB, to calculate
P − Value. As can be seen from Figs. 3 and 4, P − Value > 0.01.
Through the frequency test and run test, the P − Value obtained is greater than the
NIST specified 0.01, in line with the requirements of the ideal random sequence.
5 Conclusion
Under the background of supply chain logistics system, this paper analyzes the hidden
dangers of aquatic logistics traceability system in business data, and constructs data
acquisition and coding encryption. For the RC4 algorithm based on chaotic mapping is
used in QR code encryption process, the author made the randomness test, and the test
results show that the randomness is in accordance with NIST.
Through the QR Code encryption technology, using different RC4 key, the dynamic
encryption of aquatic supply chain traceability can be achieved, fully guarantee the
uniqueness of aquatic logistics data in the whole supply chain system, and avoid the
forgery of others, improve the safety of aquatic system information from the root.
Information Security Technology and Application in Logistics Traceability System 647
But the study only QR Code as a representative of the logistics information carrier
research, Future research can also expand the data acquisition information carrier object,
from the grid intrusion detection, trusted computing point of view, more adequate
protection of the logistics system information security.
Acknowledgment. This paper was supported by the project in the Hubei Science and Technology
Pillar Program (No. 2015BKA222).
References
1. Li, J.: The security challenge of network information reflected by “Prism” and its strategic
thinking. Inf. Theory Practice 37(4), 48–52 (2014)
2. Cheng, D.S., Wu, K.K., Liu, Z.Y.: Image encryption algorithm based on chaotic pseudorandom
match shift. Inf. Commun. 8, 7–9 (2016)
3. Prasad, R.P., Sudha, K.R., Sanyasi Naidu, P.: Information system with interlacing, key
dependent permutation and rotation. Int. J. Comput. Netw. Secur. 2(5), 86 (2010)
4. Soufiene, B.O., Abdullah, A.B., Abdelbasset, T.: Confidentiality and integrity for data
aggregation in WSN using homomorphic encryption. Wirel. Pers. Commun. 80(2), 867–889
(2015)
5. Liao, C.Z., Xu, J.W.: Research on AES and ECC mixed encryption algorithm. Softw. Guide
15(11), 63–64 (2016)
6. Zhao, Q., Peng, H., Hu, B., Liu, Q., Liu, L., Qi, Y., Li, L.: Improving individual identification
in security check with an EEG based biometric solution. In: Yao, Y., Sun, R., Poggio, T., Liu,
J., Zhong, N., Huang, J. (eds.) BI 2010. LNCS (LNAI), vol. 6334, pp. 145–155. Springer,
Heidelberg (2010). https://doi.org/10.1007/978-3-642-15314-3_14
7. Niu, C., Yang, Y.J., Mao, X.Q.: Protection mechanism of continuous data based on virtual
storage technology. Comput. Eng. Des. 34(4), 1207–1211 (2013)
Quantified Self: Big Data Analysis Platform for Human
Factor Efficacy Evaluation of Complex System
1
China Electronic Science Research Institute, Beijing, China
2
School of Digital Media and Design Arts, Beijing University of Post and Telecommunications,
Beijing 100876, China
hou1505@163.com
3
Beijing Key Laboratory of Network and Network Culture, Beijing University of Post
and Telecommunications, Beijing 100876, China
Abstract. By analyzing the factors affecting the airborne mission system, this
paper applied the method of Quantified Self to the evaluation of human effec‐
tiveness in the military airborne mission system. According to the depth of inter‐
action between people and information, we divide the information circumstance
into four aspects including individual, equipment, network and environment.
Then we construct a complete individual Quantified Self information interaction
system by collecting physiological data, cognitive data, behavioral data and envi‐
ronmental data. Finally, the functional architecture and composition of the ergo‐
nomic evaluation platform are given in combination with the airborne mission
system.
1 Introduction
With the development of information technology such as the internet of things, cloud
computing, mobile Internet, “Natural Interaction” and “Intelligent Decision” have
become important concepts in various information system. Developing of the informa‐
tion technology also lead to the explosive growth of data, which has had a profound
effect and even gradually change the original knowledge production mode and cognitive
framework. Big data analysis has become an integral method in many fields, and one of
the big data analysis trend is Quantified Self (Fig. 1).
The airborne cabin is an important campaign environment in high-tech war. The
working space of the cabin is airtight and narrow, and the operation environment is
complex. Noise, vibration, electromagnetic filed will affect the working condition and
operation ability of operators to varying degrees, which will affect the combat effec‐
tiveness. So, the main problems we have to face are:
(1) Large amount of data: With the development of sensors, the amount of data will
become larger, the types of data will become more diverse and the update speed of
data will become faster.
(2) High complexity of tasks: There are many kinds of tasks, such as searching,
tracking, and monitoring. The information capacity of these tasks interface is enor‐
mous, and the structural relationship of these tasks is complicated, and these tasks
have the characteristic of uncertain and difficult to predict, which is easy to cause
the disorder of cognitive and affect the judgement of the operator.
(3) Change of the working mode: The working mode of the operators is mainly
converted from operation to surveillance and decision making, which results in a
sharp increase in imbalances between the human cognitive and the interface infor‐
mation encoding [2].
The information warfare requires both commanders and operators to perform tasks
accurately and in real-time. It is a key link for improving the individual performance
and the team decision-making ability to establish a real-time and high-precision state
awareness model for combatants, which can detect the cognitive state of the commanders
and soldier combatant states in real time and adjust them in a timely manner. At the
same time, the information warfare also requires more dominant, autonomy and intel‐
ligent. Building a decision support system that provides operators with inferences about
current and upcoming behavior and assigning tasks autonomously between operators
and system.
2 Related Work
Quantifying self, as one of the big data analysis methods, can discover valuable infor‐
mation and implicit relationships in human-computer-environment system by analyzing
various data associated with human activities, which can help to improve current states.
The emergence and development of the concept of “quantifying self” was only a few
years ago and mainly in the field of health. It is used to track and record individual
behaviors such as sports, sleep, diet, mood by using multiple device such as computers,
portable sensors and smart phone. Then the collected data were used to discover valuable
information in fitness, daily physiology and diseases treatment which can be used to
improve the people’s living condition. With the development of micro sensors and
650 C. Li et al.
intelligent mobile technology, quantifying self will become an important method for
ergonomics research, information visualization and human-computer interaction
research.
At present, western countries led by the United States are in absolute leading position
in the research field of complex systems. The key technology program of United States
Department of Defense not only lists human-computer interaction interfaces as an
important part of software technology development, but also specially adds human-
computer system interface which is juxtaposed with software technology. A new project
of the DARPA hopes to improve the cognitive ability by changing the information
display mode, which can improve operators’ efficiency, and the cognitive ability is
reflected by the cognitive enhancement. The US Navy Research institute used eye
tracking technology to solve problems in the ship pilot train [6]. The following work
had been done:
(1) The researchers organized a large number of experienced crew and ordinary
students to simulate operations such as navigation, berthing and emergency avoid‐
ance. And the researchers collected eye movements, behavioral data and some
electrophysiological data from them throughout the missions.
(2) The researchers identified different eye movements patterns between experienced
crew and ordinary students by detailed analyzing above data, and established a
quantitative model of eye movements pattern.
(3) The model is applied to training students. By monitoring the eye movement data
of students, the system can automatically calculate whether the students deviate
from the correct mode of operation and give timely feedback according to the
model.
At present, the system has been applied to the simulation training of American Navy
drivers, and has achieved good results. The study of human efficacy has entered a new
stage in China, too. Quantitative research, as an important method to study human
efficiency and user experience, has been carried out in many fields such as ships, tanks,
aircraft driving and civil aviation control. User experience is also attached great impor‐
tance in the development of civil information systems such as the Internet. A large
number of quantitative research has been carried out. For example, Wen-jun Hou and
Xiao-yu Gao of the Beijing University of Posts and Telecommunications studied a
method based on measurement of pupil satisfaction. [4] This paper analyzed the rela‐
tionship between the pupil size and the user satisfaction in usability tests. And they
discovered that there is a linear significant negative correlation between the user satis‐
faction and user’s pupil diameter standard deviation rate, and build a model to describe
this relationship.
According to the depth of interaction between people and information, we divide the
information circumstance into four aspects including individual, equipment, network
and environment [3]. Then we construct a complete individual Quantified Self
Quantified Self 651
The system combines eye movement with motion capture and physiological synchro‐
nization technology. It uses wearable data acquisition technology and wireless data
transmission technology to let operators can completely free from any interference when
collecting data. Human-environment synchronization platform is capable of: Recording,
tracking and analyzing real-time psychological and physiological changes of users, such
as ECG, skin electricity and other indicators; analyzing people’s attention, cognitive
load, fatigue; Synchronizing real-time records involved human behaviors, equipment
running information and the environmental information, such as temperature and
humidity.
The system collects and analyzes the causal relationship between the changes of
human-computer-environment interaction, and realizes real-time simultaneous
recording and analysis of human-environment data. Therefore, it accomplishes multi‐
dimensional data visualization analysis. At the same time, for the effective data extrac‐
tion and visualization presentation, we improve the human-computer interaction inter‐
face and system process to build an effective monitoring and evaluation methods.
According to the different design for the information display interface of the aircraft
system, the indicators such as the physiology and eye movement of the operator are
analyzed. Meanwhile, the time, the number of occasions and the path of the operation
task are explored and analyzed. We found out the good and bad of the operation interface
and tried to improve the interface to become a standard of intelligent human-computer
interface.
Quantified Self 653
5 Conclusion
Quantified Self combining with big data technology can be applied to the evaluation of
complex information system, which greatly promotes the development of China’s mili‐
tary strategy. It not only can improve the airborne information system interface and the
operation efficiency but also provides more possibilities in real-time monitoring, combat
multi-distributed data acquisition and other aspects. Furthermore, sufficient and compre‐
hensive data contribute to display interface design, flexible layout and adaptive recon‐
figuration. The system makes an effort on establishing a real-time and high-precision
model of combat personnel state perception based on physiological data and neural
calculation real-time analysis. Moreover, it increases the accuracy of automatic online
analysis based on deep learning training.
References
1. What will Royal Navy warships look like in 2050? Jonathan Beale Defense correspondent,
BBC News, 31 August 2015
2. Jing, L.: Encoding information of human-computer interface for equilibrium of time pressure.
J. Comput. Aided Des. Comput. Graphics 25(7), 1022 (2013)
3. Wang, H.: In the information age, human-computer interaction design of complex system are
very important. China Aviation News (2016)
4. Hou, W.: Customer satisfaction evaluation model based on pupil size changes. Space Med.
Med. Eng. 05, 001 (2013)
Application of Speech Recognition Technology in Logistics
Selection System
1
Wuhan Second High School, Wuhan 400010, China
22836099@qq.com
2
Zhen Kun Hang Industrial Supermarket, Shanghai 201716, China
1 Introduction
Order picking based on product barcode is widely adopted in automatic logistics and
warehouse management systems. However, barcode order picking presents several
disadvantages when the order number increases significantly. Firstly, as the number of
order increases, system capacity does not scale-up easily, leading to delay in order
delivery. Secondly, currently barcode order picking systems may still rely heavily on
human labors. As a result, the entire system becomes error-prone and inconsistent in
system efficiency. In the present paper, we propose a voice recognition based order
picking system. It is our contention that the introduction of voice recognition and voice
based control system can significantly reduce the physical activity at the order picking
production line. This proposed approach, therefore, is applicable in labor intensive and
high throughput retail warehouse/distribution and logistics centers, in particular, those
that are not equipped with automatic and semiautomatic systems.
Automatic Speech Recognition (ASR) refers to a whole raft of technologies that aim to
collect and process human speech and digitise such data into computer understandable
and actable data and/or instructions. Based on retrained models, computers should be
able to performed the entire process without or with very limited human intervention.
Upon receiving human speech signal, ASR systems process the data based on
features such as ambient/background signals, timbre, pitch, etc. Figure 1 depicts the key
components of an ASR system. An ASR system normally consists of 5 modules, prepro‐
cessing, feature extraction, model training, model scoring and selection, and post-
processing. The workflow can be largely divided into model training and model appli‐
cation.
(1) Preprocessing module: Speech signal includes filtering, A/D conversion, filtering,
demonising, speech enhancement, signal smoothing, end-point detection, etc.
(2) Feature extraction: After preprocessing, speech signals are subject to time-
frequency analysis, Cepstral analysis and wavelet transformation. This is to extract
from the signal data such features as: timbre, language, and voice contents.
(3) Model training: Features extracted by the previous module are then fed into training
algorithms to obtain language models. Model training normally is an iterative
process wherein features are evaluated, re-processed and optimised so as to
construct models that can reflect all features of the input data. Trained models are
stored in a model library.
(4) Model scoring and selection: When ASR system is in application mode, features
extracted from speech signal are utilised to identify and retrieve the best model
from the model library. A predefined scoring scheme can guide the selection of best
matching models.
(5) Post-processing: Natural languages are ambiguous. In order to improve the system
performance, it may be necessary to introduce linguistic and semantic analysis to
the output of selected speech signal model.
The envisaged use case of the proposed system is that the speech recognition system
should be a portable device that the order picking staff can easily carry along without
hindering their normal work routines. Such a scenario derives a plethora of requirements
on speech processing speed and accuracy, device portability, device energy consumption
profile. In the present paper, we focused on local dialects and opt for non-specific speech
recognition chip. We also tuned the model library based on the core user population,
ambient and environmental noise, and typical use cases of the distribution centre where
the system will be evaluated and put into production.
656 T. Fu and B. Sun
Currently, widely used non-specific speed recognition chips include ASR M08,
LD3320 and LingYang 61A. Such chips do not require pre-recording of speech signals
for initialisation and validation and are suitable for our envisaged use case [1].
LD3320 speech recognition module is a specialised processor with external circuits
including AD, DA converters, microphone interface, audio output interface, and parallel
and serial interface. It benefits from small physical size and low power consumption,
lending itself to mobile applications. LD3320 embeds speech recognition module trained
based on a large amount of speech data. It provides an off-the-shelf solution for many
use cases. LD3320 provides a versatile of external control and interfaces, including
dynamic editing of the recognised keywords. LD3320, therefore, facilitates further
development of speech recognition functionalities and customisation against specific
application context [2].
End users of the proposed speech recognition system are order picking staff in labour
intensive manual warehouse/distribution centres. The speech recognition system is
expected to be worn by the users while they are moving between aisles, staging areas,
order picking lines and storage spaces. It, therefore, should be light weight, low energy
consumption, easy to carry and highly responsive [3]. Meanwhile, as the staff turnover
rate is very high in such a working environment, the speech recognition module should
be non-specific [4].
LD3320 presents the following key features:
(1) Integrated Flash and RAM. This eliminates the needs for external storage and thus
reduce the complexity and cost of the system and overall energy consumption
profile.
(2) Embedded speech recognition models. The on chip models are already suitable for
many generic application scenarios. It therefore presents a low learning curve for
adoption.
(3) Parameter tuning on distance and sensitivity.
(4) No requirement on prior training and recording.
(5) Dynamic keyword editing allows easy extension and adaption to new scenarios.
(6) High accuracy rate (95%) against a list of as many as 50 keywords.
(7) Integrated A/D, D/A convertors, Integrated amplifier circuit and a 550 mW speaker
interface for playback; parallel and serial interface for further development.
(8) Can easily switch between sleep and activate states to reduce energy consumption.
(9) Working power supply of 3.3 V. LD3320 can be powered by three AA batteries to
meet the power supply needs for portable systems.
LD3320 ICRoute is based on keyword recognition. Figure 2 illustrates key compo‐
nents of LD3320.
LD3320 collects speech signal through the integrated microphone (MIC). The signal
is then subject to spectrum analysis, feature extraction and feature engineering so as to
be ready for keyword extraction. A trained speech recognition model will then be applied
to the processed signal, outcome of which are candidate words which will be compared
against a predefined list provided to the system. During the final step, the system will
compute a score for each candidate keywords and output the one(s) with the highest
Application of Speech Recognition Technology in Logistics 657
number, aisle number, and shelf number of the first order. It also instruct the order
picking staff to read out bar code of the picked items.
(2) When the item is located, order picking staff reads the barcode. The portable
terminal compares recognised barcode with the one in the picking order. If matches,
portable terminal will pronounce quantity of item for the order; otherwise, portable
terminal will read out location information again.
(3) When the current order is finished, order picking staff instruct the device with “task
complete”. The portable device changes the status of the current order and moves
to the next one in the order queue. If the order queue is empty, the portable terminal
pronounces “operation complete”.
Barcode comparison in step two acts as a key checking point. This is to avoid poten‐
tial cost introduced by human errors. If the barcode does not match, order picking staff
will be remaindered the correct location information or provided means to manually
double-check the order details.
The envisaged order picking scenario limits the keyword search space to reduce
system complexity. As shown above, interactions between human order picking staff
and portable terminal are restricted to the following keywords: “start operation”,
“repeat”, “task complete” and digits based item barcodes. Apart from barcodes, the other
instruction speech patterns can be preloaded, tuned and stored locally in the portable
terminal to enable offline non-specific speech signal recognition. Terminal and server
communication is only needed for handshaking, system initialisation, and downloading
order details. Downlink and uplink bandwidth can be kept to minimum and thus improve
system performance by reducing network latency.
A preliminary evaluation and comparative study has been carried out. Figure 3 demon‐
strates the difference between conventional and speech recognition based order picking
workflows.
Fig. 3. The difference between conventional and speech recognition based order picking
workflows
It is evident that our proposed solution can significantly reduce physical activities
that are inevitable in conventional approach based on barcode or RF scanning. Reduced
physical activity leads to improved work efficiency and reduced human errors due to
fatigue and negligence. Speech recognition also simplified the overall workflow from
Application of Speech Recognition Technology in Logistics 659
original six steps to three by integrating and removing such activities as operating scan‐
ners. The reduced steps also contributed to cost reduction through paperless work space,
per order time reduction, and reduced error-rate.
The speech recognition based terminal optimises human system interface. The
underlying system efficiency needs to be based on route optimisation and shelving and
stock optimisation which are beyond the scope of this paper [5].
6 Conclusion
References
1. Qingxiu, M., Zhengyun, R.: Voice picking system based on indoor positioning technology of
RFID. Artif. Intell. 18, 50–53 (2015)
2. Qin, Y., Yong, L.V.: Speech Signal Processing and Recognition, pp. 7–9. National Defend
Industry Press, Beijing (2015)
3. Jinping, L.: Simulation analysis of picking path and storage allocation strategy in distribution
center. Simul. Intell. 04, 310–315 (2015)
4. Zhao, Q., Hu, B., Shi, Y., Li, Y., Moore, P., Sun, M., Peng, H.: Automatic identification and
removal of ocular artifacts in EEG—improved adaptive predictor filtering for portable
applications. IEEE Trans. Nanobiosci. 13(2), 109–117 (2014)
5. Qingguo, Z., Lihua, H., Rongrong, W., et al.: The application of speech picking technology in
drug storage work. Chin. Pharm. 27, 501–503 (2014)
Author Index