Digital Image Processing Basics
Digital Image Processing Basics
Digital Image Processing means processing digital image by means of a digital computer. We
can also say that it is a use of computer algorithms, in order to get enhanced image either to
extract some useful information. Digital image processing is the use of algorithms and
mathematical models to process and analyze digital images. The goal of digital image processing
is to enhance the quality of images, extract meaningful information from images, and automate
image-based tasks.
According to block 1, if input is an image and we get out image as an output, then it is termed
as Digital Image Processing.
According to block 2, if input is an image and we get some kind of information or description
as a output, then it is termed as Computer Vision.
According to block 3, if input is some description or code and we get image as an output, then
it is termed as Computer Graphics.
According to block 4, if input is description or some keywords or some code and we get
description or some keywords as a output, then it is termed as Artificial Intelligence.
Advantages of Digital Image Processing:
1. Improved image quality: Digital image processing algorithms can improve the visual
quality of images, making them clearer, sharper, and more informative.
2. Automated image-based tasks: Digital image processing can automate many image-
based tasks, such as object recognition, pattern detection, and measurement.
3. Increased efficiency: Digital image processing algorithms can process images much
faster than humans, making it possible to analyze large amounts of data in a short amount
of time.
4. Increased accuracy: Digital image processing algorithms can provide more accurate
results than humans, especially for tasks that require precise measurements or quantitative
analysis.
Disadvantages of Digital Image Processing:
1. High computational cost: Some digital image processing algorithms are
computationally intensive and require significant computational resources.
2. Limited interpretability: Some digital image processing algorithms may produce
results that are difficult for humans to interpret, especially for complex or sophisticated
algorithms.
3. Dependence on quality of input: The quality of the output of digital image processing
algorithms is highly dependent on the quality of the input images. Poor quality input
images can result in poor quality output.
4. Limitations of algorithms: Digital image processing algorithms have limitations, such
as the difficulty of recognizing objects in cluttered or poorly lit scenes, or the inability to
recognize objects with significant deformations or occlusions.
5. Dependence on good training data: The performance of many digital image
processing algorithms is dependent on the quality of the training data used to develop the
algorithms. Poor quality training data can result in poor performance of the algorithm.
Components of an Image Processing System
The components of an image processing system differ depending on the type of image, for
example, satellite images and x-ray images cannot be processed by the same type of image
processing system. However, the following components are the minimum requirement of an
image processing system as shown in Fig. 1.
Processing Machine (Computer): It may be a general-purpose computer according to task to be
performed. The basic use of this device is that it will perform all digital image processing tasks
off-line. Storage Device: Storage devices are used for storing of original and processed images
for different purposes and use.
Display Device: It is used for displaying data. Example of a display device is generally a colour
monitor.
Image Processing Software: Image processing software such as IGIS, ERDAS, ENVI and
Geomatica are specially designed programming modules that perform specific tasks on remotely
sensed images.
Printing Device: It is used for representing and storing image data in hard copy format. It could
be laser, inkjet or any other printer.
Fig: 1. Block diagram showing different components of a digital image processing system.
It is important to note that following factors should be considered while selecting a digital image
processing system:
Number of Analysts and Mode of Operation: You should consider a software which could be
accessed and interactively used for data processing by at least the number of people involved in
the study.
Memory and Processing Specifications of Computer: Processing of different types of digital
remote sensing data require different processing capabilities and memory. You should choose
software that is compatible with the specifications of computer. You can use a computer having
at least or better than the minimum specification that is required to process the data of your
interest.
Operating System: The operating system must be powerful and easy to use. MS-DOS, UNIX
and WINDOW’s are the most universally used operating systems, though MS-DOS is no longer
relevant. The chosen image processing software and hardware should be compatible with the
operating system in your computer.
Storage: Digital images are stored usually in matrix form with various multispectral bands and
in different formats. Image processing software which is capable of storing and processing the
concerned image formats should be considered.
Display Resolution: Different image types require different display resolution hence you should
consider software which is capable of handling highest resolution.
Steps in Digital Image Processing: It is now apparent to you that the subject of digital image
processing is very broad. For your understanding, we can generalize the image processing into
following four basic steps (Fig. 2):
Image Pre-processing
Image Enhancement
Image Transformation
Image Classification
Accuracy Assessment
The DIP aids in generating thematic maps and decision-making process (Fig. 2).
Steps of Digital Image Processing
IMAGE PRE-PROCESSING
Remotely sensed image data are obtained from the imaging sensors mounted on satellite
platforms. This data contains raw images, which are susceptible to a variety of distortions that
have to be corrected before making them suitable for any application. Distortions can affect the
values recorded by the sensors in different wavelengths. As a result, they affect the geometry of
the pixel, alignment of an image with other images and reference maps. Thus, it becomes
important for us to know types of these distortions and their effects of on the images.
Image Distortions: Any kind of errors present in remote sensing images are described as image
distortions. These distortions occur due to data recording procedure, shape and rotation of the
Earth and environmental conditions prevailing at the time of data acquisition. Distortions
occurring in remote sensing images can be categorized into two types:
• Radiometric distortions
• Geometric distortions.
Radiometric distortions: These distortions refer to the incorrect values in the pixels (i.e.,
variation in the pixel intensities) recorded in an image irrespective of the object or scene being
scanned and resultant distortions are known as radiometric distortions. The recorded values get
distorted due to one or more of the following factors:
• Sensor ageing
• Random malfunctioning of the sensor elements
• Atmospheric interference at the time of image acquisition and
• Topographic effects.
Geometric distortions: These distortions usually occur when actual geometry of an
object/spatial feature is not properly captured in an image. As a consequence, distance, area,
direction and shape properties of objects vary across an image and these errors are known as
geometric errors/distortions. These distortions are inherent in images because we attempt to
represent three-dimensional Earth surface as a two-dimensional image. Further, the images are
captured of a moving target (Earth) with a camera mounted on a moving platform (satellite). The
relative motion between the camera and target also results in geometric distortion. Geometric
errors originate during the process of data collection and vary in type and magnitude. There are
several factors causing geometric distortions such as:
• Earth’s rotation
• Earth’s curvature
• Satellite platform instability
• Instrument error
Image Corrections: Raw remote sensing images always contain significant amount of
distortions, they cannot be used directly for image analysis. The image correction involves image
processing operations which normally precede manipulation and analysis of image data to extract
specific information.
The primary aim of image correction operations is to correct distorted image data to
create a more accurate representation of the original scene. Image corrections are also
known as preprocessing of remotely sensed images.
It is a preparatory phase that improves quality of images and serves as a basis for further
image analysis. Depending upon the kinds of errors which are present in images, the
image correction functions are comprised radiometric and geometric corrections. Both of
these corrections are made prior to actual use of remote sensing data in resource
management, environmental monitoring and change detection applications by application
scientists.
Image correction generates a corrected image, which is both radiometrically and
geometrically as close as to the radiant energy characteristics of the original scene.
Complete chain of processing of remote sensing images is shown in Fig. 3. It becomes
clear from this figure that image correction forms an integral part of processing of
images.
Fig. 4: Landsat 7 FCC images of the Kangra region, Himachal Pradesh, India (A) Before and (B)
After geometric correction.
Geometric errors present in remote sensing images can be categorised into internal or systematic
and external or non-systematic geometric errors. Internal geometric errors are introduced by the
sensor system itself and/or by the effects of Earth’s rotation and curvature. These errors are
predictable or computable and often referred to as systematic that can be identified and corrected
using pre-launch or platform ephemeris. Reasons of geometric distortions causing internal
geometric errors in remote sensing images include the following:
Skew caused by the Earth’s rotation effects.
Scanning system induced variation in ground resolution cell size and dimensional relief
displacement.
External geometric errors are usually introduced by phenomena that vary in nature through
space and time. The most important external variables that can cause geometric error in remote
sensor data are random movements by the spacecraft at exact time of data collection, which
usually involve:
Altitude changes
Attitude changes (yaw, roll and pitch).
There are several methods used to remove the geometric errors. It may be noted that systematic
geometric error is generally easier to identify and correct than non-systematic or random
geometric error.
IMAGE ENHANCEMENT: Different errors/distortions affect the quality of remotely sensed
data and are removed by using image correction techniques. After that you can extract usable
information from an image that has been corrected for radiometric and geometric errors. Visual
interpretation of remote sensing images is one of the most powerful ways to retrieve information.
Depending on the quality of image and human expertise in image interpretation, many of the
attributes are readily identified by human eyes. However, at times it is required to highlight
(enhance) certain features of interest in images to make the image better interpretable. This is
achieved by a technique known as image enhancement.
Image enhancement simply refers to the techniques used to increase the human interpretability
and ability to derive meaningful information from remotely sensed images. It improves the
quality and clarity of images.
The enhancement operations include enhancing contrast and suppressing noise, for
revealing details to a human interpreter. For example, image enhancement of a blurred
and low contrast satellite image of the Himalayan Mountains could be made clear by
increasing contrast range in the original image.
The original image might have areas of very high and very low intensity, which mask
details. The enhanced image will have improved contrast providing details of the masked
image regions.
It is important to know that the removal of blur is known as digital image restoration. It
requires knowledge of the point spread function of the source of blur. It is a highly formal
and mathematical model based process while image enhancement is often subjective, to
suit the liking of the interpreter.
The primary goal of image enhancement is to improve the visual interpretability of the
images. This is done by increasing the apparent distinction between features in the image.
Image enhancement involves use of a number of statistical and image manipulation
functions available in an image processing software. Image features are enhanced by the
following two operations:
1. Point Operations: In this operation, value of pixel is enhanced independent of
characteristics of neighbourhood pixels within each band. Image enhancement techniques
based on point operations are used for contrast enhancement.
2. Local (Neighbourhood) Operations: In this operation, value of pixel is modified based
on neighbouring brightness values. The image processing techniques based on
neighbourhood operations are also known as spatial enhancement techniques.
IMAGE TRANSFORMATION: Image transformation involves manipulation of multiple
bands of data, from a single multispectral image or from two or more images of the same area
acquired at different times. In this process, a raw image or several sets of images undergo some
mathematical treatment to achieve a new imagery (Fig. 5). As a consequence, new resultant
transformed image generated from two
or more sources highlights particular features or properties of interest better than original input
images. Hence, the transformed image may have properties that make it more suited to a
particular purpose than original input images.
There are several transforms namely, Fourier transform, wavelet transform, or textural transform
that are applied to a single band or the panchromatic image. Therefore, it is essential to quality
the operations involving more than one band as multiband transformations or multiband
operations. Image transformation by itself does not imply that it requires multiple bands.
Fig.5. Landsat 7 FCC images of Kangra region, Himachal Pradesh, India: (A) Before; and
(B) After Normalised Difference Vegetation Index transformation.
Multiband transformation techniques are useful for compressing bands having similar kind of
information into fewer bands and also to extract new bands of data that are more interpretable to
human eye. Such operation is also known as spectral enhancement method. It is important to
note that these transformation techniques require more than one band of data. There are a number
of multiband transformation techniques available, in which arithmetic operations and image
fusion are the two common techniques used for image transformation.
IMAGE CLASSIFICATION: Different image correction, enhancement and transformation
techniques, which help us to better visualize and interpret remotely sensed images. It may be
noted that all these techniques aid in providing only better visual information and, therefore,
have limited utility. Now, we will move a step further and will learn how to make more sense of
the landscape/image by dividing it into separate classes based on surface characteristics. This
process is known as image classification. It involves conversion of image data into finite set of
classes that represent surface types in the imagery. It may be used to identify vegetation types,
anthropogenic structures, mineral resources, etc. or transient changes in any of these features.
Additionally, classified remotely sensed image can be converted to vector features (e.g.,
polygons) in order to compare with other data sets or to calculate spatial attributes (e.g., area,
perimeter, etc.). Image classification is a very active field of study broadly related to the field of
pattern recognition
Image classification is the process of assigning spectral classes into information classes. Spectral
classes are groups of pixels that are uniform with respect to their brightness values in different
spectral channels of data. Information classes are categories of interest that an analyst attempts to
identify in the image on the basis of his knowledge and experience about the area. For example,
a remote sensing image contains spectral signatures of several features present on the ground
covered by pixels of DN values in different spectral bands. An interpreter or analyst identifies
homogeneous groups of pixels having similar values and labels the groups as information classes
such as water, agriculture, forest, etc. while generating a thematic map. When this thematic
information is extracted with the help of software, it is known as digital image classification.
It is important to note that there could be many spectral classes within an information class
depending upon the nature of features the image represents or the purpose of the classification. In
other words, different spectral classes may be grouped under one information class. In short, we
can define image classification as a process of assigning all pixels in the image to particular
classes or themes based on spectral information represented by the digital numbers. The
classified image comprises a mosaic of pixels, each of which belongs to a particular theme and is
a thematic map of the original image.
There are two general approaches to image classification:
• Supervised Classification: It is the process of identification of classes within a remote sensing
data with inputs from and as directed by the user in the form of training data.
• Unsupervised Classification: It is the process of automatic identification of natural groups or
structures within a remote sensing data, without considering their thematic class labels.
Both the classification approaches differ in the way the classification is performed. In the case of
supervised classification, specific land cover types are delineated based on statistical
characterization of sample data drawn from known examples in the image (known as training
sites). In unsupervised classification, however, clustering algorithms are used to uncover the
commonly occurring land cover types, with the analyst providing interpretations of those cover
types at a later stage. The underlying assumption for this approach is that spectrally similar
pixels are also commonly found in physical proximity of each other. Merits and demerits of both
the supervised and unsupervised classification methods are summarized in Table 14.1
Merits and demerits of supervised and unsupervised classification methods.
ACCURACY ASSESSMENT: Once you have a prepared a thematic map using image
classification, the next step is that you would want to know how much accurate those outputs are
because inaccuracies in outputs will have their bearing on the map’s utility and users would have
greater confidence in utilising data if its accuracy is high
Accuracy assessment is the final step in the analysis of remote sensing data which helps us to
verify how accurate our results are. It is carried out once the interpretation/classification has been
completed. Here, we are interested in assessing accuracy of thematic maps or classified images
which is known as thematic or classification accuracy. The accuracy is concerned with the
correspondence between class label and ‘true’ class. A ‘true’ class is defined as what is observed
on the ground during field surveys. For example, a class labeled as water on a classified
image/map is actually water on the ground.
In order to perform accuracy assessment correctly, we need to compare two sources of
information which include:
Interpreted map/classified image derived from the remote sensing data and
Reference map, high resolution images or ground truth data.
Numerical reporting of the classification accuracy requires a separate ground truth data set
(labeled pixel sets belonging to different classes) that is different from the training data set.
Relationship between these two sets of information is commonly expressed in two forms,
namely:
Error matrix that describes the comparison of these two sources of information.
Kappa coefficient which consists of a multivariate measure of agreement between rows and
columns of error matrix.
Without performing the essential measures to evaluate the accuracy or validity of the map, a map
created utilizing remotely sensed data or other spatial data cannot be considered the finished
product. Visual inspection, non-site-specific analysis, creating difference images, error budget
analysis, and quantitative accuracy assessment are just a few of the techniques available to
analyze accuracy/error in spatial data
Digital Image Interpretation
Digital image interpretation involves computer-based manipulation of the original digital
images to enhance their quality so that more information can be extracted from them.
Digital image interpretation only enhances the visualization of the image for extracting more
accurate information but it cannot add more information than what the original image already
has.
Processing machine (computer), storage device, display device, image processing software and
printing device are the main components of an image processing system.
Image pre-processing, image enhancement, image transformation, image classification and
accuracy assessment are the main steps in digital image processing.
Any kind of errors present in remote sensing images are described as image distortions and
radiometric and geometric distortions are two main types image distortions.
Image correction involves the removal of distortions in the image data to create a more
accurate representation of the original scene.
Image enhancement is the technique used to increase the human interpretability and ability to
derive meaningful information from remotely sensed images.
Multiband image transformation involves manipulation of multiple bands of data, from a single
multispectral image or from two or more images of the same area acquired at different times.
Image classification is the technique is used to divide into separate thematic classes based on
surface characteristics.
Accuracy assessment is the final step in the analysis of remote sensing data which help us to
verify how accurate our results or thematic maps are.
Image interpretation & analysis
In order to take advantage of and make good use of remote sensing data, we must be able to
extract meaningful information from the imagery. This brings us to the topic of discussion in this
chapter - interpretation and analysis - the sixth element of the remote sensing process which
we defined in Chapter 1. Interpretation and analysis of remote sensing imagery involves the
identification and/or measurement of various targets in an image in order to extract useful
information about them. Targets in remote sensing images may be any feature or object which
can be observed in an image, and have the following characteristics:
Targets may be a point, line, or area feature. This means that they can have any form,
from a bus in a parking lot or plane on a runway, to a bridge or roadway, to a large
expanse of water or a field.
The target must be distinguishable; it must contrast with other features around it in the
image.
Much interpretation and identification of targets in remote sensing imagery is performed
manually or visually, i.e. by a human interpreter. In many cases this is done using
imagery displayed in a pictorial or photograph-type format, independent of what type of
sensor was used to collect the data and how the data were collected. In this case we refer
Satellite image of area in (a) grey scale and in (b) standard FCC.
B. Second order — Geometric Arrangements of Objects
i) Texture: Texture refers to the frequency of tonal variation in an image. Texture is produced
by an aggregate unit of features which may be too small to be clearly discerned individually on
the image. It depends on shape, size, pattern and shadow of terrain features. Texture is always
scale or resolution dependent. Same reflected objects may have difference in texture helps in
their identification. As an example in a high resolution image grassland and tree crowns have
similar tone, but grassland will have smooth texture compared to tree. Smooth texture refers to
less tonal variation and rough texture refers to abrupt tonal variation in an imagery or
photograph.
High resolution image showing different textures.
The similar kinds of reflectance/colour (green) appear throughout the image, three distinct land
cover types can be seen from image texture. Triangular patch at the bottom left corner is the
plantation which has rough texture where individual trees can be seen. Right side on top of
image, the trees are closer together and tree canopies merge together, forming medium textural
pattern and at the right bottom corner with smooth texture indicating that it is probably an open
field with short grass.
ii) Pattern: Pattern refers to the spatial arrangement of the objects. Objects both natural and
manmade have a pattern which aids in their recognition. The repetition of certain general form or
relationship in tones and texture creates a pattern, which is characteristic of this element in image
interpretation. In the Fig. 9.3 it could be easily understood that at the left bottom corner of the
image, it is plantation, where the tress are nearly equally spaced. Whereas at the upper right and
bottom right corners show natural vegetation.
iii) Size: Size of objects on images must be considered in the context of the image scale or
resolution. It is important to assess the size of a target relative to other objects in the scene, as
well as the absolute size, to aid in the interpretation of that target. A quick approximation of
target size can make direct interpretation to an appropriate result more quickly. The most
measured parameters are length, width, perimeter, area, and occasionally volume. For example,
if an interpreter had to distinguish zones of land use, and had identified an area with a number of
buildings in it, large buildings such as factories or warehouses would suggest commercial
property, whereas small buildings would indicate residential use.
Edge of
Species Crown shape Tone Pattern Texture
Crown
Dark but
Conical with Circular but
Cypress lighter than Spotted Hard and fine
round crown not sharp
cedar