Dip Notes Module-1 Part-1 17ec72
Dip Notes Module-1 Part-1 17ec72
MODULE-1
When x, y, and the amplitude values of f are all finite, discrete quantities, we call the image
a Digital image.
A digital image is composed of a finite number of elements, each of which has a particular location
and value. These elements are referred to as picture elements, image elements, pels, and pixels.
2. Write a note on origins(History) of Digital Image Processing
a) Early 1920s: One of the first applications of digital imaging was in the newspaper industry
– The Bartlane cable picture transmission service
– Images were transferred by submarine cable between London and New York
– Pictures were coded for cable transfer and reconstructed at the receiving end on a telegraph printer
Mid to late 1920s: Improvements to the Bartlane system resulted in higher quality images
– New reproduction processes based on photographic techniques
– Increased number of tones in reproduced images
Page 11
DIGITAL IMAGE PROCESSING[17EC72]
b) 1960s: Improvements in computing technology and the onset of the space race led to a surge of
work in digital image processing
– 1964: Computers used to improve the quality of images of the moon taken by the Ranger 7 probe
– Such techniques were used in other space missions including the Apollo landings
d) 1980s - Today: The use of digital image processing techniques has exploded and they are now used
for all kinds of tasks in all kinds of areas – Image enhancement/restoration – Artistic
effects – Medical visualisation – Industrial inspection – Law enforcement – Human
computer interfaces
X-rays for medical and industrial imaging are generated using an x-ray tube, which is a vacuum tube with
a cathode and anode. The cathode is heated, causing free electrons to be released. These electrons flow at
high speed to the positively charged anode. W en the electron strike a nucleus, energy is released in the
form x-ray radiation.
h
Angiography is another major application in an area called contrast enhancement radiography. The
procedure is used to obtain images of blodd vessels.
Page 12
DIGITAL IMAGE PROCESSING[17EC72]
iii) Imaging in the microwave band
Dominant application in microwave band is radar. The unique feature of imaging radar is its ability to
collect data over virtually any region at any time, regardless of weather or ambient lighting conditions. Some
radar waves can penetrate clouds and under certain conditions can also see through vegetation, ice and
extremely dry sand.
Major applications of imaging in the radio band are in medicine and astronomy. In medicine radio waves
are used in magnetic resonance imaging (MRI). This techniques places a patient in a powerful magnet and passes
radio waves through his or her body in short pulses. Each pulse causes a responding pulse of radio waves to be
emitted by patient’s tissues. The location from which theses signals orginate and their strength are
determined by a computer which produces a two-dimensional picture of a section of the patient.
v) Other Imaging Modalities Acoustic images, electron microscopy and synthetic (computer –
generated images)
Imaging using sound finds application in geological exploration, industry and medicine. The most
important commercial applications of image processing in geology are in mineral and oil exploration.
Ultrasound imaging is used routinely in manufacturing; the best known applications of this technique
are in medicine, especially in obsterics, where unborn babies are imaged to determine the health of
their development.
4. FUNDAMENTAL STEPS IN DIGITAL IMAGE PROCESSING
3. Image Restoration
Image restoration is an area that also deals with improving the appearance of an image. However, unlike
enhancement, which is subjective, image restoration is objective, in the sense that restoration techniques tend to
be based on mathematical or probabilistic models of image degradation.
6. Compression
Compression deals with techniques for reducing the storage required to save an image or the bandwidth to
transmit it. Particularly in the uses of internet it is very much necessary to compress data.
7. Morphological Processing
Morphological processing deals with tools for extracting image components that are useful in the
representation and description of shape.
8. Segmentation
Segmentation procedures partition an image into its constituent parts or objects. In general, autonomous
segmentation is one of the most difficult tasks in digital image processing. A rugged segmentation
procedure brings the process a long way toward successful solution of imaging problems that require objects to
be identified individually.
3. Computer
Page 15
DIGITAL IMAGE PROCESSING[17EC72]
The computer in an image processing system is a general-purpose computer and can range from a PC to a
supercomputer. In dedicated applications, sometimes specially designed computers are used to achieve a
required level of performance
Page 16
DIGITAL IMAGE PROCESSING[17EC72]
c)Retina – it is innermost membrane of the eye. When the eye is properly focused, light from an object outside the
eye is imaged on the retina. There are various light receptors over the surface of the retina
The two major classes of the receptors are-
1) cones- it is in the number about 6 to 7 million. These are located in the central portion of the retina
called the fovea. These are highly sensitive to color. Human can resolve fine details with these cones because
each one is connected to its own nerve end. Cone vision is called photopic or bright light vision
2) Rods – these are very much in number from 75 to 150 million and are distributed over the entire
retinal surface. The large area of distribution and the fact that several roads are connected to a single
nerve give a general overall picture of the field of view. They are not involved in the color vision and are
sensitive to low level of illumination. Rod vision is called is scotopic or dim light vision.
The absent of reciprocators is called blind spot
ii. Distribution of rods and cones in Retina
Page 17
The absence of receptors in this area results in the so-called blindspot. Fig.shows that cones are most dense in
the center of the retina(in the center area of the fovea) receptors Density is measured in degrees from fovea.
Cones are more dense in the centre of the retina (fovea) & distributed lightly in the remaining part of eye. Rods
are distributed equally over the surface of eye.
iii. Image Formation in the Eye
The major difference between the lens of the eye and an ordinary optical lens in that the former is flexible. The
shape of the lens of the eye is controlled by tension in the fiber of the ciliary body. To focus on the distant object
the controlling muscles allow the lens to become thicker in order to focus on object near the eye it becomes relatively
flattened.
The distance between the center of the lens and the retina is called the focal length and it
varies from 17mm to 14mm as the refractive power of the lens increases from its minimum to its maximum.
When the eye focuses on an object farther away than about 3m.the lens exhibits its lowest
refractive power. When the eye focuses on a nearly object. The lens is most strongly refractive.
The retinal image is reflected primarily in the area of the fovea. Perception then takes place by the relative
excitation of light receptors, which transform radiant energy into electrical impulses that are ultimately
decoded by the brain.
DIGITAL IMAGE PROCESSING[17EC72]
iv. Brightness Adaption and Discrimination
Digital image are displayed as a discrete set of intensities. The range of light intensity levels to which the
human visual system can adopt is enormous- on the order of 1010 from scotopic threshold to the glare limit.
Experimental evidences indicate that subjective brightness is a logarithmic function of the light intensity
incident on the eye.
The curve represents the range of intensities to which the visual system can adopt. But the visual system cannot
operate over such a dynamic range simultaneously. Rather, it is accomplished by change in its overcall sensitivity
called brightness adaptation.
For any given set of conditions, the current sensitivity level to which of the visual system
is called brightness adoption level , Ba in the curve. The small intersecting curve represents the range of
subjective brightness that the eye can perceive when adapted to this level. It is restricted at level Bb , at and
below which all stimuli are perceived as
indistinguishable blacks. The upper portion of the curve is not actually restricted. whole simply raise the adaptation
level higher than Ba .
The ability of the eye to discriminate between change in light intensity at any specific adaptation level is also
of considerable interest.
Faculty: Prof.Sunitha.R
Page 13
DIGITAL IMAGE PROCESSING[17EC72]
Take a flat, uniformly illuminated area large enough to occupy the entire field of view of the subject. It may
be a diffuser such as an opaque glass, that is illuminated from behind by a light source whose intensity, I can be
varied. To this field is added an increment of illumination ΔI in the form of a short duration flash that appears
as circle in the center of the uniformly illuminated field. If ΔI is not bright enough, the subject cannot
see any perceivable changes.
As ΔI gets stronger the subject may indicate of a perceived change. ΔIc is the increment of illumination
discernible 50% of the time with background illumination I. Now, ΔIc /I is called the Weber ratio.
Small value means that small percentage change in intensity is discernible representing “good” brightness
discrimination.
Large value of Weber ratio means large percentage change in intensity is required representing “poor
brightness discrimination”.
Optical illusion
a) Single image
sensor
b) Line sensor
c) Array sensor
*The incoming energy is transformed into a voltage by the combination of input electrical power
and sensor material.
*Output voltage waveform is response of the sensor(s)
*A digital quantity is obtained from each sensor by digitizing its response
a. Image acquisition using a single sensor
Page 16
DIGITAL IMAGE PROCESSING[17EC72]
A geometry that is used much more frequently than single sensors consists of an in-line arrangement of sensors
in the form of a sensor strip. The strip provides imaging elements in one direction. Motion perpendicular to the
strip provides imaging in the other direction This is the type of arrangement used in most flat bed scanners. Sensing
devices with 4000 or more in-line sensors are possible. In-line sensors are used routinely in airborne imaging
applications, in which the imaging system is mounted on an aircraft that flies at a constant altitude and
speed over the geographical area to be imaged.
One-dimensional imaging sensor strips that respond to various bands of the electromagnetic spectrum
are mounted perpendicular to the direction of flight. The imaging strip gives one line of an image at a time, and
the motion of the strip completes the other dimension of a two-dimensional image. Lenses or other focusing
schemes are used to project the area to be scanned onto the sensors. Sensor strips mounted in a ring
configuration are used in medical and industrial imaging
c. Image Acquisition Using Sensor Arrays
Individual sensors can be arranged in the form of a 2-D array. Numerous electromagnetic and
some ultrasonic sensing devices are arranged frequently in an array format. This is also the predominant
arrangement found in digital cameras. A typical sensor for these cameras is a CCD array, which can be
manufactured with a broad range of sensing properties and can be packaged in rugged arrays of 4000 *
4000 elements or more.
CCD sensors are used widely in digital cameras and other light sensing instruments. The response
of each sensor is proportional to the integral of the light energy projected onto the surface of the sensor, a property
that is used in astronomical and other applications requiring low noise images.
8. A Simple Image Formation Model
• To create a digital image, we need to convert the continuous sensed data into digital form. This involves
two processes: sampling and quantization.
• Figure 2.16(a) shows a continuous image, f(x, y) , that we want to convert to digital form.
• An image may be continuous with respect to the x - and y – coordinates, and also in amplitude.
• To convert it to digital form, we have to sample the function in both coordinates and in amplitude.
• Digitizing the coordinate values is called sampling • Digitizing the amplitude values is called
quantization.
• In order to form a digital function, the gray-level values must be converted (quantized) into
discrete quantities.
• Sampling means that the values of the continuous function f(x,y) are retained only in specific
positions (i,j) where 0≤i≤Nx and 0≤j≤Ny, where Nx and Ny are integer values. The sampling
topology depends on the spatial arrangement and size of the sensors that are used to acquire the image.
• Clearly, the quality of a digital image is determined to a large degree by the number of samples and
discrete gray levels used in sampling and quantization.
f (1,1)
f (1, 2) ... f (1, N)
f f (2, 2) ... f (2, N)
(2,1)
f (x, y)
... ... ... ...
f (M ,1) f (M , 2) ... f (M , N)
• The number of intensity levels is an integer power of 2.
DIGITAL IMAGE PROCESSING[17EC72]
i.e L = 2k
256 = 28
k=8
• The range of values spanned by the gray scale is called dynamic range of an image. The upper limit is
determined by saturation and the lower limit by noise.
• An image with high dynamic range is said to be a high contrast image.
• An image with low dynamic range leads to a dull, washed out look.
• The number of bits required to store a digitized image is
bMNk
when M = N, b =N2k
For example, a 256 X 256 image represented with 8 bits takes 5,24,288 bits.
11. RELATIONSHIPS BETWEEN PIXELS
In this topic, several important relationships between pixels in a digital image are considered.
• Neighbors
• Adjacency
• Connectivity
• Paths
• Regions and boundaries
• Distance measures
i.Neighbors of a Pixel
• A pixel p at coordinates (x,y) has four horizontal and vertical neighbors whose coordinates are given
by: (x+1,y), (x-1, y), (x, y+1), (x,y-1)
(x, y-1)
(x-1, y) P (x,y) (x+1, y)
DIGITAL IMAGE PROCESSING[17EC72]
(x, y+1)
This set of pixels, called the 4-neighbors or p, is denoted by N4(p). Each pixel is one unit distance from (x,y) and
some of the neighbors of p lie outside the digital image if (x,y) is on the border of the image.
P (x,y)
(x-1, y-1) (x+1, y+1)
These points, together with the 4-neighbors, are called the 8-neighbors of p, denoted by N8 (p). As before, some of
the points in ND (p) and N8 (p) fall outside the image if (x,y) is on the border of the image.
ii. Adjacency
Image boundary and regions are defined by set of connected pixels. To determine if the two pixels are
connected/adjacent or not, there are two conditions
a) Two pixels should be neighbors
b) Their gray levels should be similar or should be in the set
Ex-1: V = {1, 2}
4- Adjacency – two pixel P and Q with value from V are 4–adjacency if A is in the set n4(P)
Ex.1
DIGITAL IMAGE PROCESSING[17EC72]
0 q1 1 q2 1
0 2P 0
0 0 1q3
p and q1 are not 4-adjacent because q1 is not N4(p), q1 is diagonal to p
p and q2 are neighbors and are 4-adjacent because q1 is not N4(p) and p=2 & q2 = 1 are in set V
p and q3 are not 4-adjacent because q3 is not N4(p), q3 is diagonal to p
8-Adjacency – two pixel P and Q with value from V are 8–adjacency if A is in the set N8(P)
Ex.1
0 q1 1 q2 1
0 2P 0
0 0 1q3
p and q1 are not 8- adjacent, q1 is ND(p) i.e q1 is diagonal to p but q1=0 is not in the set V
p and q2 are 8- adjacent because p & q2 are 8- neighbors and because p=2 & q2 = 1 are in the set V
p and q3 are 8- adjacent because p & q2 are 8- neighbors and because p=2 & q2 = 1 are in the set V
M-adjacency –two pixel P and Q with value from V are m– adjacency if
· Q is in n4 (p) or
· Q is in nd (q) and the set N4(p) ∩ N4(p) has no pixel whose values are from set V
0 q1 0 1 q2
0 2P 0
0 0 1q3
1 0 1
p and q1 are not m- adjacent, but q1=0 is not in the set V
q2 is diagonal to ‘p’ and
N4(p) ∩ N4(q2)
{0,0,0,0}∩{0,0}
{0} here 0 is not belongs to set V. therefore p and q2 are m-adjacent.
q3 is diagonal to ‘p’ and
DIGITAL IMAGE PROCESSING[17EC72]
N4(p) ∩ N4(q3)
{0,0,0,0}∩{0,0,1}
{0,1} here 0 is not belongs to set V and 1 is belongs to set V therefore p and q3
are not m-adjacent.
NOTE: In 8-adjacency sometimes ambiguity exists between pixels , to avoid this m-adjacency is used.
iii. Connectivity
iv. Path
A (digital) path (or curve) from pixel p with coordinates (x0, y0) to pixel q with coordinates (xn, yn) is a
sequence of distinct pixels with coordinates
(x0, y0), (x1, y1), …, (xn, yn)
Where (xi, yi) and (xi-1, yi-1) are adjacent for 1 ≤ i ≤ n. Here n is the length of the path.
If (x0, y0) = (xn, yn), the path is closed path. We can define 4-, 8-, and m-paths
De(p, q) = [(x-s)2 + (y-t)2]1/2 All the pixels that have Euclidean distance less than or equal to a
value r from pixel p(x,y) are contained in a disk of radius r centered at (x,y).
D4(p, q) = |x-s| + |y-t| .All thepixels having D4 distance less than or equal to a value r are contained in a
diamond centered at (x,y). Also, all the pixels with D4 =1 are the 4- neighbors of (x,y).
D8(p, q) = max(|x-s|, |y-t|). All the pixels with D8 distance less than or equal to some value r from (x,y) are
contained in a square centered at (x,y). All pixels with D8 =1 are the 8-neigbors of (x,y).
12. Linear and Non-linear operations