Advances in Digital Image Compression by Adaptive Thinning
Advances in Digital Image Compression by Adaptive Thinning
Advances in Digital Image Compression by Adaptive Thinning
the target function f is approximated by the unique continuous function L(f, Y ), whose restriction on any triangle
in the Delaunay triangulation DY is a linear function and
which satisfies the interpolation conditions
L(f, Y )(y) = f (y),
artefacts, such as in the case of wavelets. In contrast, irregular adaptive triangulations offer much more flexibility, and they support the appropriate concept of adaptivity
for representing natural features of the image. In view of
the required compression, however, this enhanced flexibility may lead to high coding costs required for coding the
node coordinates, and the topology coding (connectivity
between nodes) of the corresponding mesh.
In order to entirely avoid the required costs for the connectivity coding, adaptive thinning algorithms work with
Delaunay triangulations. Recall that a Delaunay triangulation of a discrete planar point set X R2 is a triangulation, such that the circumcircle of each of its triangles does
not contain any point from X in its interior. An example
of such a triangulation is shown in Figure 2. For further
details concerning triangulation methods, we refer to the
textbook [5].
Now we associate, with any finite set X of points its
unique Delaunay triangulation DX (in the case of co-circular points in X there may be ambiguities which we exclude in the following discussion for the sake of simplicity). Thus, at the decoder, the set of points X can directly
be used in order to uniquely reconstruct the triangulation
DX .
The adaptive thinning algorithm, first developed in [4],
is concerned with the approximation of a bivariate function
f from a finite set of scattered data points X R2 and
sample values {f (x)}xX . To this end, a data hierarchy
X = XN XN 1 . . . Xn
(1)
of nested subsets of X = {x1 , . . . , xN } R2 is constructed. This is done by recursively removing points from
X. At any removal step, one point is removed from the current subset Xp X in (1), so that Xp is of size |Xp | = p,
n p N.
The multiresolution method, associated with adaptive
thinning, works with decremental Delaunay triangulations
over the subsets in (1). To this end, for any subset Y X,
for all y Y .
(a)
(a)
(b)
(c)
(d)
(b)
Figure 4: Barbara. Reconstruction from n = 7336 most
significant pixels (a) with L -norm and without LSA; (b)
with L2 -norm and with LSA.
3. CODING SCHEME
This section briefly explains the coding of the most significant pixels, which is also subject of the previous paper [3]. First note that any coding scheme requires a nonambiguous decoding rule which enables the receiver to uniquely reconstruct the image. As already discussed in the
previous section, this can be accomplished by using any set
Xn of n most significant pixels output by adaptive thinning.
The subset Xn can be considered as a set of tridimen(1)
(2)
(1)
sional points (xi , xi , zi ), 1 i n, where xi and
(2)
xi are the integer coordinates of the point xi and where
(1)
(2)
(1)
(2)
zi = f (xi , xi ) = Q(f (xi , xi )) is a quantized
(1)
(2)
value, where the luminances f (xi , xi ) are output by
least squares approximation. We use a uniform quantization step q, so that Q(z) = dz/qe. As shown in Section 2,
the use of Delaunay triangulations avoids the coding of
any connectivity information. Only the tridimensional locations of the points in the subset Xn are required at the decoder. Furthermore, the ordering of the nodes is not needed
for the reconstruction.
W
NW NE
SW SE
+NW +NE
SW SE
+
2552
PSNR = 10 log10
,
MSE
given in dB, where MSE denotes the Mean Square Error
X
1
j)|2 .
MSE =
|I(i, j) I(i,
N M i,j
In our first example, we decided to use a test image
called Peppers, shown in Figure 6 (a), whose size is 256by-256 pixels. Note that this image contains very few textured areas. Our method provides a PSNR value of 31.13 dB
(corresponding to the image in Figure 6 (b)), whereas SPIHT
yields a better PSNR value of 31.65 dB (Figure 6 (c)).
A second example is shown in Figure 7. The size of the
test image, called Fruits (shown in Figure 7 (a)), is also
256-by-256 pixels. In this test case, our method provides
a PSNR value of 32.13 dB (Figure 7 (b)), whereas SPIHT
yields a PSNR value of 32.77 dB (Figure 7 (c)).
Both examples show that our algorithm achieves, in
contrast to SPIHT, accurate localization of sharp edges, and
so it avoids spurious ringing artefacts. Although our method
is slightly inferior to SPIHT in terms of its PSNR, we believe that it is quite competitive. This is supported by the
good visual quality of the image reconstructions by our
compression method (see Figure 6 (b) and Figure 7 (b)).
SW SE
5. ACKNOWLEDGEMENT
Figure 5: First three splits of the cubic domain .
We code the pixel points by performing a recursive splitting of the domain = [0..N ] [0..M ] [0..P ], where N
and M are the dimensions of the image and P is the num
ber of possible values for fi (typically, P = 28 1 = 255
for unquantized data, but P = 255/q when the quantization step is q).
At each step, we split a non-empty domain , initially
= , into two subdomains 1 and 2 of equal size. If
m denotes the number of most-significant pixels in the
domain , then we have m = m1 + m2 . Thus only one
of the two numbers, say m1 , is added to the bitstream.
At the decoder, the number m2 will be deduced from m
and m1 . Each number is coded by the minimal number
of required bits. For instance, since 0 m1 m , the
number m1 is coded by dlog2 (m +1)e bits. The splitting
of the subdomains is performed recursively until the points
are exactly localized. For the purpose of illustration, the
first three splits of the cubic domain are shown in Figure 5. For further details on this particular coding scheme,
we refer to [3].
4. NUMERICAL RESULTS
We have implemented the compression scheme proposed
in this paper. In this section, numerical examples are used
in order to evaluate the performance of our method. To
this end, we compare our compression scheme with the
wavelet-based compression scheme SPIHT [6]. We evaluate the reconstruction quality of the decoded image I by
(a)
(a)
(b)
(b)
(c)
(c)