Cellular Neural Networks

Download as pdf or txt
Download as pdf or txt
You are on page 1of 8

Chapter 2

Cellular Neural Network

2.1 Introduction

The cellular neural network (abbreviated as CNN) is proposed by Chua and Yang
in 1988 [1]-[2]. It is more general than Hopfield neural network. The state value of
one node (cell) at the next time is influenced by inputs and outputs of nodes near this
node. Cellular neural networks have the characteristic of parallel processing. The next
states of all nodes can be evaluated at the same time. So the operation speed is very
fast. And because CNN has the characteristic of local connection of nodes, so CNN is
suited for realizing with VLSI. The structure of cellular neural networks is regular,
parallel array and local connection etc., not only suitable for the integrated circuit, but
also can deal with some special calculations that have the characteristic of local
regular connection. Cellular neural networks mostly were applied to the image
processing application at first, for example noise removal, edge detection etc.

2.2 The Structure of Cellular Neural Networks

The primary element of CNN is a cell. Cells are arranged in a two-dimensional


array usually, as Fig. 2.1 shows. Every cell is only influenced directly by its
neighboring cells in a cellular neural network, it is not influenced directly by other
cells, and the cells not neighboring on each other are influenced each other indirectly
by the characteristic of continuous time dynamical propagation of cellular neural
networks.

Cell

Fig. 2.1 A two-dimensional 6 × 10 cell array.

3
Let us consider one M × N two-dimensional cell array. There are total M × N
cells arranged in M rows and N columns. The cell in ith row and jth column is labeled
Cij . The range of neighboring cells is called neighborhood. The number of
neighboring cells of every cell is the same, it is determined by neighborhood radius,
and this radius is different from the general round radius. The neighborhood of a cell
is defined as the following.

N ij (r ) = { C kl | max ( | k – i |, | l – j | ) ≤ r, 1 ≤ k ≤ M ; 1 ≤ l ≤ N } (2-1)

N ij ( r ) represents the set of neighboring cells C kl of cell Cij , r is the radius of


N ij ( r ) , r is a positive integer. N ij ( r ) expresses one (2r +1) × (2r +1) cell array. For
being simple and convenient, we are used to omitting r and express N ij ( r ) with N ij .
Such as r = 1, the range of cell Cij and its neighboring cells is size of 3 × 3, as Fig.
2.2 shows [3]. The cell set contained by the grey square in Fig. 2.2 is N ij (1) . N ij (1)
represents a 3 × 3 cell array.

3x3 sphere
Cell Cij
of influence

Fig. 2.2 Radius r = 1, the range of cell Cij and its neighboring cells.

The relevant parameters of a cell are defined as follows and shown in Fig. 2.3 [3]:
u ij : Input of cell Cij ;
x ij : State of cell Cij ;
y ij : Output of cell Cij ;
I ij : Threshold value of cell Cij .

4
Cell Cij
uij
Input Output
State
Threshold xij yij
Iij

Fig. 2.3 The primary element of a cellular neural network - cell Cij .

uij xij yij


Iyx Ry
Eij -
Iij C Rx ․․․
Ixu(i,j;k,l) Ixy(i,j;k,l)

Fig. 2.4 The circuit structure of cell Cij .

The basic circuit structure of a cell is shown in Fig. 2.4 [1]. According to this
circuit structure, the motion equation of a cell Cij can be written by

dxij (t ) 1
C =− xij (t ) + ∑ A(i, j; k , l ) y kl (t ) + ∑ B(i, j; k , l )u kl + I ij
dt Rx C kl ∈N ij C kl ∈N ij (2-2)
1 ≤ i ≤ M ;1 ≤ j ≤ N

For being simple and convenient and without losing generality, we can suppose C =
R x = 1, then Equation (2-2) can be rewritten to Equation (2-3):

dxij (t )
dt
= − xij (t ) + ∑ A(i, j; k , l ) y
C kl ∈N ij
kl (t ) + ∑ B(i, j; k , l )u
C kl ∈N ij
kl + I ij
(2-3)
1≤ i ≤ M; 1≤ j ≤ N

5
According to view of neural networks, the meaning of each parameter of Equation
(2-3) is listed as follows:
x ij : State of cell Cij ;
y kl : Output of neighboring cell C kl of cell Cij ;
u kl : Input of neighboring cell C kl of cell Cij ;
I ij : Threshold value of cell Cij ;
A(i, j; k, l): The weighting of output of neighboring cell C kl of cell Cij ;
B(i, j; k, l): The weighting of input of neighboring cell C kl of cell Cij .

A(i, j; k, l), B(i, j; k, l) and I ij are total 2 × (2r + 1) 2 + 1 values. They determine the
behavior of a two-dimensional cellular neural network. We call that A = [A(i, j; k, l)]
is the feedback template and B = [B(i, j; k, l)] is the control template. The output y ij
is a piecewise linear function of x ij . It is given by the following output equation and
shown in Fig. 2.5 [1]:

yij (t ) = f (xij (t ) ) =
1
2
(
xij (t ) + 1 − xij (t ) − 1 ) (2-4)

y ij

-1 1 xij
-1

Fig. 2.5 Output function.

We can receive the following results from output function (2-4):


1. y ij (t ) ≤ 1

⎧ 1, xij < 1

dy ij
2. =⎨
dxij ⎪ 0, xij ≥ 1

3. If xij < 1 , then xij = y ij

Each cell is influenced by neighboring cells’ inputs and outputs. Cells’ inputs multiply
template B and cell’s outputs multiply template A, as Fig. 2.6 shows [3].

6
Multiply Multiply
Standard CNN: ( A, B, I )
template A template B

A B

The center cell be


influenced

Output Y State X Input U

Fig. 2.6 The basic model of cellular neural networks.

The motion equation can be expressed by Fig. 2.7 [3]:

Control Bias
B Iij
Template

Local x& ij x ij Output


Input + Integrator
f(.) y ij
u ij
-1
Feedback
A
Template

Input from
Neighborhood
ukl

Feedback from
Neighborhood
y kl

Fig. 2.7 Motion equation.

If A(i, j; i, j ) > 1 , then after the transient decayed to zero, each cell of a cellular
neural network must stop in a stable equilibrium point. And the magnitude of each
stable equilibrium point is greater than 1. In other words, we have the following

7
characteristics:
lim xij (t ) > 1, 1≤ i ≤ M; 1≤ j ≤ N (2-5)
t →∞

and
lim y ij (t ) = ±1 , 1≤ i ≤ M; 1≤ j ≤ N (2-6)
t →∞

Namely if the center element A(i, j; i, j ) of template A satisfies A(i, j; i, j ) > 1 , then
the absolute value of the output of each cell in stable equilibrium point is equal to 1.
This guarantees that cellular neural networks have binary value output; this property
is essential for solving classification problems in image processing applications.

2.3 Comparison of Cellular Neural Network and Hopfield Neural Network

It can be said that cellular neural networks are evolved from Hopfield neural
network. Cellular neural networks are more general than Hopfield neural network.
Because in cellular neural networks each cell is influenced by only near cells, it is not
like Hopfield neural network, in which each cell is influenced by all other cells. So
cellular neural networks can be easily realized by VLSI technique. The followings are
some differences between cellular neural networks and Hopfield neural network:
(1) The weight matrix of Hopfield neural network must be symmetric, but the weight
matrices A and B of a cellular neural network are not necessary to be symmetric.
(2) Hopfield neural network is allowed to operate asynchronously, but cellular neural
networks must operate synchronously.
(3) The connections in the cellular neural network are local. Hopfield neural network
is a fully connected neural network. In general, the number of interconnections in
the cellular neural network is less than the number of interconnections in a
Hopfield neural network.
(4) The self-feedback coefficient A(i, j; i, j) of cell Cij in a cellular neural network is
greater than 1 in order to guarantee that the steady-state output of each cell is
either +1 or -1. This condition is always violated in a Hopfield neural network
since its diagonal coupling coefficients are all assumed to be zero [6].

Design of CNN System

The design of a CNN system is finding one or more templates that realize a
certain input-output behavior. The design methods of templates found in literature can
be divided into two classes:
Design by synthesis: given an explicit problem specification, a set of parameters is
found that satisfies the specified requirements.
8
Design by learning: given a vague description of the task through a large number of
input-target pairs, a learning process is applied that minimizes some kind of cost
function.
Rather than being alternatives, these two types of design can be considered as
supplementing methods. The choice between them is based on the type of information
that is available about the problem at hand. In case an explicit description of the
desired functional behavior is available, a synthesis method is used. In case only
implicit knowledge is available, a learning method is the obvious choice.

2.4 Discrete-Time Cellular Neural Networks

Harrer and Nossek recommend that discrete time cellular neural networks
(Discrete Time Cellular Neural Network, is abbreviated as DT-CNN) [8] is as the
discrete time version of CNN that Chua and Yang introduced (is abbreviated as
Chua−Yang-CNN here). The network dynamics of Chua−Yang-CNN is described by
a set of differential equations:

dxij (t )
dt
= − xij (t ) + ∑ A(i, j; k , l ) y
Ckl ∈N ij
kl (t ) + ∑ B(i, j; k , l )u
Ckl ∈N ij
kl (t ) + I ij

1≤ i ≤ M; 1≤ j ≤ N

t in the previous equation is a discrete time variable, then we can rewrite the previous
equation:

xij (h + Δt ) − xij (h)


Δt
= − xij (h) + ∑ A(i, j; k , l ) y
C kl ∈N ij
kl ( h) + ∑ B(i, j; k , l )u
Ckl ∈N ij
kl (h) + I ij
(2-7)
1 ≤ i ≤ M; 1 ≤ j ≤ N

If Δt = 1, the Equation (2-7) can be rewritten as the Equation (2-8):

xij (h + 1) − xij (h) = − xij (h) + ∑ A(i, j; k , l ) y


C kl ∈N ij
kl ( h) + ∑ B(i, j; k , l )u
C kl ∈N ij
kl (h) + I ij

xij (h + 1) = ∑ A(i, j; k , l ) y
C kl ∈N ij
kl ( h) + ∑ B(i, j; k , l )u
C kl ∈N ij
kl (h) + I ij (2-8)

1≤ i ≤ M; 1≤ j ≤ N

DT-CNN is a timing system. A set of discrete motion equations describes its dynamic

9
behavior. At discrete time h, the state xij (h) of a cell Cij depends on time-invariant
input u kl and time-variant output y kl (h − 1) of its neighboring cell C kl . The output
function of DT-CNN is the hardlimiter activation function.

⎧+ 1 for x ij (h + 1) ≥ 0
y ij (h + 1) = f ( x ij (h + 1)) = ⎨
⎩− 1 for x ij (h + 1) < 0 (2-9)
1≤ i ≤ M; 1≤ j ≤ N

References

[1] L. O. Chua and Lin Yang, “Cellular Neural Networks: Theory,” IEEE Trans. on
CAS, vol. 35 no. 10, pp.1257-1272, 1988.
[2] L. O. Chua and Lin Yang, “Cellular Neural Networks: Applications,” IEEE Trans.
on CAS, vol.35, no.10, pp. 1273-1290, 1988.
[3] Leon O. Chua, CNN:A PARADIGM FOR COMPLEXITY. World Scientific,
1998.
[4] Special Issue on Chaotic Systems, Proc. IEEE, Aug. 1987.
[5] L. O. Chua and R. N. Madan, “The sights and sounds of chaos,” IEEE Circuits
Devices Mag., pp. 3-13, Jan., 1988.
[6] J. J. Hopfield, “Neural networks and physical systems with emergent
computational abilities,” Proc. Natl. Acad. Sci. USA., vol. 79, pp. 2554-2558,
1982.
[7] J. J. Hopfield, “Neurons with graded response have collective computational
properties like those of two-state neurons,” in Proc. Natl. Acad. Sci. USA, vol. 81,
pp. 3088-3092, 1984.
[8] H. Harrer and J. A. Nossek, “Discrete-time Cellular Neural Networks,”
International Journal of Circuit Theory and Applications, Vol. 20, pp. 453-468,
1992.

10

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy