Distributed Algorithem
Distributed Algorithem
Abstract
Nowadays, we are living in the midst of a data explosion and seeing a massive growth in
databases so with the wide availability of huge amounts of data; necessarily we are
become in need for turning this data into useful information and knowledge, where Data
mining uncovers interesting patterns and relationships hidden in a large volume of raw
data and big data is a new term used to identify the datasets that are of large size and have
grater complexity. The knowledge gained from data can be used for applications such as
market analysis, customer retention and production control. Data mining is a massive
computing task that deals with huge amount of stored data in a centralized or distributed
system to extract useful information or knowledge. In this paper, we will discuss
Distributed Data Mining systems, approaches, Techniques and algorithms to deal with
distributed data to discover knowledge from distributed data in an effective and efficient
way.
Distributed Algorithms
Distributed algorithms have two advantages. First, for some applications, no central
processor is available to handle the calculations. Second, when a large network must
forward all measurement data to a single central processor, there is a communication
bottleneck and higher energy drain at and near the central processor. Distributed
algorithms for cooperative localization generally fall into one of two schemes.
1- Classical multilateration: Each sensor estimates its multihop range to the nearest
land–reference nodes. These ranges can be estimated via the shortest path between
the node and reference nodes, that is, proportional to the number of hops or the
sum of measured ranges along the shortest path. Note that the shortest path
algorithm is already executed in a distributed manner across the network. When
each node has multiple range estimates to known positions, its coordinates are
calculated locally via multilateration.
2- Successive refinement: These algorithms try to find the optimum of a global cost
function (e.g., LS, WLS, or ML). Each node estimates its location and then
transmits that assertion to its neighbors . Neighbors must then recalculate their
location and transmit again, until convergence. A device starting without any
coordinates can begin with its own local coordinate system and later merge it with
neighboring coordinate systems . Typically, better statistical performance is
achieved by successive refinement compared to network multilateration, but
convergence issues must be addressed.
References :
2019.