Sei sulla pagina 1di 19

CP7004 Image Processing and Analysis Two Marks Question Bank

UNIT 1
SPATIAL DOMAIN FILTERING
1. Define Image.
An Image may be defined as a two dimensional function f(x, y) where x & y are spatial
(plane) coordinates, and the amplitude of f at any pair of coordinates (x, y) is called intensity or
gray level of the image at that point. When x, y and the amplitude values of f are all finite,
discrete quantities we call the image as Digital Image.
2. Define Image Sampling.
Digitization of spatial coordinates (x, y) is called Image Sampling. To be suitable for
computer processing, an image function f(x,y) must be digitized both spatially and in magnitude.
3. Define Quantization.
Digitizing the amplitu-de values is called Quantization. Quality of digital image is
determined to a large degree by the number of samples and discrete gray levels used in sampling
and quantization.
4. What is Dynamic Range?
The range of values spanned by the gray scale is called dynamic range of an image.
Image will have high contrast, if the dynamic range is high, and image will have dull washed out
gray look if the dynamic range is low.
5. Define Mach band effect.
The spatial interaction of Luminance from an object and its surround creates a
phenomenon called the mach band effect.
6. Define Brightness.
Brightness of an object is the perceived luminance of the surround. Two objects with
different surroundings would have identical luminance but different brightness.
7. What is meant by Tapered Quantization?
If gray levels in a certain range occur frequently while others occurs rarely, the
quantization levels are finely spaced in this range and coarsely spaced outside of it. This method
is sometimes called Tapered Quantization.
8. What do you meant by Gray level?
Gray level refers to a scalar measure of intensity, that ranges from black to grays and
finally to white.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

9. What do you meant by Color model?


A Color model is a specification of 3D-coordinates system and a subspace within that
system where each color is represented by a single point.
10. List the hardware oriented color models.
The hardware oriented color models are as follows,
 RGB model
 CMY model
 YIQ model
 HSI model
11. What is Hue of saturation?
Hue is a color attribute that describes a pure color where saturation gives a measure of the
degree to which a pure color is diluted by white light.
12. List the applications of color models.
The applications of color models are,
 RGB model--- used for color monitor & color video camera
 CMY model---used for color printing
 HIS model----used for color image processing
 YIQ model---used for color picture transmission
13. What is Chromatic Adoption?
The hue of a perceived color depends on the adoption of the viewer. For example, the
American Flag will not immediately appear red, white, and blue, the viewer has been subjected
to high intensity red light before viewing the flag. The color of the flag will appear to shift in hue
toward the red component cyan.
14. Define Resolutions.
Resolution is defined as the smallest number of discernible detail in an image. Spatial
resolution is the smallest discernible detail in an image and gray level resolution refers to the
smallest discernible change is gray level.
15. What is meant by pixel?
A digital image is composed of a finite number of elements, each of which has a
particular location of value. These elements are referred to as pixels or image elements or picture
elements or pels elements.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

16. Define digital image.


A digital image is an image f(x, y), that has been discretized both in spatial coordinates
and brightness.
17. List the categories of digital storage.
The categories of digital storage are,
 Short term storage for use during processing.
 Online storage for relatively fast recall.
 Archival storage for frequent access.
18. List the steps involved in digital image processing.
The steps involved in digital image processing are,
 Image Acquisition.
 Preprocessing.
 Segmentation.
 Representation and description
 Recognition and interpretation.
19. What is recognition and interpretation?
Recognition is a process that assigns a label to an object based on the information
provided by its descriptors. Interpretation means assigning to a recognized object.
20. Specify the elements of DIP system.
The elements of DIP system are,
 Image acquisition.
 Storage.
 Processing.
 Communication.
 Display.
21. Write the two types of light receptors.
The two types of light receptors are,
 Cones.
 Rods.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

22. Differentiate photopic and scotopic vision.


Sl. Photopic Vision Scotopic Vision
No
1 The human being can resolve the fine Several rods are connected to one nerve
details with these cones because each end. So it gives the overall picture of the
cone is connected to its own nerve image.
end.
2 This is also known as bright light This is also known as dim light vision.
vision.

23. How ones and rods are distributed in retina?


In each eye, cones are in the range 6-7 million, and rods are in the range 75-150
million.
24. Define subjective brightness and adaption.
Subjective brightness means, intensity as preserved by the human visual system.
Brightness adaptation means the human visual system can operate only from scotopic to glare
limit. It cannot operate over the range simultaneously. It accomplishes this large variation by
changes in its overall intensity.
25. Define weber ratio.
The ratio of increment of illumination to background if illumination is known as weber
ratio i.e. (ΔIc /I).
 If the ratio (ΔIc /I) is small, then small percentage of change in intensity is needed ie
good brightness adaptation.
 If the ratio (ΔIc /I) is large, then large percentage of change in intensity is needed ie
poor brightness adaptation
26. What is mach band effect?
Mach band effect means the intensity of the stripes is constant. Therefore it preserves the
brightness pattern near the boundaries; these bands are called as mach band effect.
27. What is simultaneous contrast?
The region’s perceived brightness does not depend on its intensity but also on its
background. All centre square have exactly the same intensity. However they appear to the eye
to become darker as the background becomes lighter.
28. What do you meant by Zooming of digital images?
Zooming may be viewed as over sampling. It involves the creation of new pixel locations
and the assignment of gray levels to those new locations.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

29. What do you meant by shrinking of digital images?


Shrinking may be viewed as under sampling. To shrink an image by one half, we delete
every row and column. To reduce possible aliasing effect, it is a good idea to blue an image
slightly before shrinking it.
30. Define the term Radiance.
Radiance is the total amount of energy that flows from the light source, and it is usually
measured in watts (w).
31. Define the term Luminance.
Luminance measured in lumens (lm), gives a measure of the amount of energy an
observer perceiver from a light source.
32. Write any four applications of DIP.
The applications of DIP are,
 Remote sensing
 Image transmission and storage for business application
 Medical imaging
 Astronomy
33. What is the effect of Mach band pattern?
The intensity or the brightness pattern perceive a darker stribe in region D, and brighter
stribe in region B. This effect is called Mach band pattern or effect.

UNIT 2

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

FREQUENCY DOMAIN PROCESSING


1. What is Image Transform?
An image can be expanded in terms of a discrete set of basis arrays called basis images.
These basis images can be generated by unitary matrices. Alternatively, a given NXN image can
be viewed as an N^2X1 vectors. An image transform provides aset of coordinates or basis
vectors for vector space.
2. List the applications of transform.
The applications of transforms are,
 To reduce band width
 To reduce redundancy
 To extract feature.
3. Give the conditions for perfect transform.
The conditions of perfect transform are,
 Transpose of matrix = Inverse of a matrix.
 Orthoganality.
4. What are the properties of unitary transform?
The properties of unitary transform are,
 Determinant and the Eigen values of a unitary matrix have unity magnitude.
 The entropy of a random vector is preserved under a unitary Transformation.
 Entropy is a measure of average information.
5. Write the expression of one-dimensional discrete Fourier transforms.
The expression of one-dimensional discrete Fourier transforms in
Forward transform. The sequence of x(n) is given by x(n) = { x0,x1,x2,… xN-1}.

Reverse transforms

6. List the Properties of twiddle factor.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

The properties of twiddle factor are,


 Periodicity : WN^(K+N)= WN^K
 Symmetry : WN^(K+N/2)= -WN^K
7. Give the Properties of one-dimensional DFT.
The properties of one-dimensional DFT are,
 The DFT and unitary DFT matrices are symmetric.
 The extensions of the DFT and unitary DFT of a sequence and their inverse
transforms are periodic with period N.
 The DFT or unitary DFT of a real sequence is conjugate symmetric about N/2.
8. Give the Properties of two-dimensional DFT.
The properties of two-dimensional DFT are,
 Symmetric
 Periodic extensions
 Sampled Fourier transform
 Conjugate symmetry.
9. Write the properties of cosine transform.
The properties of cosine transform are as follows,
 Real & orthogonal.
 Fast transform.
 Has excellent energy compaction for highly correlated data
10. Write the properties of Hadamard transform.
The properties of hadamard transform are as follows,
 Hadamard transform contains any one value.
 No multiplications are required in the transform calculations.
 The no. of additions or subtractions required can be reduced from N2 to about
Nlog2N
 Very good energy compaction for highly correlated images.
11. Define Haar transform.
The Haar functions are defined on a continuous interval Xe [0, 1] and for K=0,1,
N1,where N=2^n..The integer k can be uniquely decomposed as K=2^P+Q-1.

12. Write the properties of Haar transform.


The properties of Haar transform are,

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

 Haar transform is real and orthogonal.


 Haar transform is a very fast transform
 Haar transform has very poor energy compaction for images
 The basic vectors of Haar matrix sequencely ordered.
13. Write the Properties of Slant transform.
The properties of Slant transform are,
 Slant transform is real and orthogonal.
 Slant transform is a fast transform
 Slant transform has very good energy compaction for images
 The basic vectors of Slant matrix are not sequencely ordered.
14. Define of KL Transform
KL Transform is an optimal in the sense that it minimizes the mean square error between
the vectors X and their approximations X^. Due to this idea of using the Eigen vectors
corresponding to largest Eigen values. It is also known as principal component transform.
15. Justify that KLT is an optimal transform.
Since mean square error of reconstructed image and original image is minimum and the
mean value of transformed image is zero so that uncorrelated.
16. What is Image Enhancement?
Image enhancement is to process an image so that the output is more suitable for
specific application.
17. Name the categories of Image Enhancement and explain.
The categories of Image Enhancement are,
 Spatial domain
 Frequency domain
Spatial domain refers to the image plane and it is based on direct manipulation of pixels
of an image. Frequency domain techniques are based on modifying the Fourier transform of an
image.
18. What do you mean by Point processing?
Image enhancement at any Point in an image depends only on the gray level at that point
is often referred to as Point processing.

19. Define Histogram.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

The histogram of a digital image with gray levels in the range [0, L-1] is a discrete
function h (rk) = nk, where rk is the kth gray level and nk is the number of pixels in the image
having gray level rk.
20. What is histogram equalization?
It is a technique used to obtain linear histogram. It is also known as histogram
linearization. Condition for uniform histogram is Ps(s) = 1
21. What is contrast stretching?
Contrast stretching reduces an image of higher contrast than the original, by darkening
the levels below m and brightening the levels above m in the image.
22. What is spatial filtering?
Spatial filtering is the process of moving the filter mask from point to point in an image.
For linear spatial filter, the response is given by a sum of products of the filter coefficients, and
the corresponding image pixels in the area spanned by the filter mask.
23. Define averaging filters.
The output of a smoothing, linear spatial filter is the average of the pixels contained in
the neighborhood of the filter mask. These filters are called averaging filters.
24. What is a Median filter?
The median filter replaces the value of a pixel by the median of the gray levels in the
neighborhood of that pixel.
25. What is maximum filter and minimum filter?
The 100th percentile is maximum filter is used in finding brightest points in an image.
The 0th percentile filter is minimum filter used for finding darkest points in an image.
26. Define high boost filter.
High boost filtered image is defined as
HBF = A (original image) – LPF
= (A-1) original image + original image – LPF
HBF = (A-1) original image + HPF
27. State the condition of transformation function s=T(r).
 T(r) is single-valued and monotonically increasing in the interval 0_r_1
 0_T(r) _1 for 0_r_1.

28. Write the application of sharpening filters.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

The applications of sharpening filters are as follows,


 Electronic printing and medical imaging to industrial application
 Autonomous target detection in smart weapons.

UNIT 3
SEGMENTATION AND EDGE DETECTION

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

1. What is segmentation?
The first step in image analysis is to segment the image. Segmentation subdivides an
image into its constituent parts or objects.
2. Write the applications of segmentation.
The applications of segmentation are,
 Detection of isolated points.
 Detection of lines and edges in an image.
3. What are the three types of discontinuity in digital image?
Three types of discontinuity in digital image are points, lines and edges.
4. How the discontinuity is detected in an image using segmentation?
The steps used to detect the discontinuity in an image using segmentation are
 Compute the sum of the products of the coefficient with the gray levels contained
in the region encompassed by the mask.
 The response of the mask at any point in the image is R = w1z1+ w2z2 + w3z3 +
………..+ w9z9
 Where zi = gray level of pixels associated with mass coefficient wi.
 The response of the mask is defined with respect to its center location.
5. Why edge detection is most common approach for detecting discontinuities?
The isolated points and thin lines are not frequent occurrences in most practical
applications, so edge detection is mostly preferred in detection of discontinuities.
6. How the derivatives are obtained in edge detection during formulation?
The first derivative at any point in an image is obtained by using the magnitude of the
gradient at that point. Similarly the second derivatives are obtained by using the Laplacian.
7. Write about linking edge points.
The approach for linking edge points is to analyse the characteristics of pixels in a small
neighborhood (3x3 or 5x5) about every point (x,y)in an image that has undergone edge detection.
All points that are similar are linked, forming a boundary of pixels that share some common
properties.

8. What is the advantage of using sobel operator?

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

Sobel operators have the advantage of providing both the differencing and a smoothing
effect. Because derivatives enhance noise, the smoothing effect is particularly attractive feature
of the sobel operators.
9. What is pattern?
Pattern is a quantitative or structural description of an object or some other entity of
interest in an image. It is formed by one or more descriptors.
10. What is pattern class?
It is a family of patterns that share some common properties. Pattern classes are denoted
as w1 w2 w3 ……… wM , where M is the number of classes.
11. What is pattern recognition?
It involves the techniques for arranging pattern to their respective classes by
automatically and with a little human intervention.
12. What are the three principle pattern arrangements?
The three principal pattern arrangements are vectors, Strings and trees. Pattern vectors are
represented by old lowercase letters such as x ,y, z and it is represented in the form x=[x1,
x2, ……….., xn ]. Each component x represents I th descriptor and n is the number of such
descriptor.
13. What is edge?
An edge is a set of connected pixels that lie on the boundary between two regions. Edges
are more closely modeled as having a ramp like profile. The slope of the ramp is inversely
proportional to the degree of blurring in the edge.
14. What is meant by object point and background point?
To execute the objects from the background is to select a threshold T that separate these
modes. Then any point(x,y) for which f(x,y)>T is called an object point. Otherwise the point is
called background point.
15. Define region growing.
Region growing is a procedure that groups pixels or sub regions into layer regions based
on predefined criteria. The basic approach is to start with a set of seed points and from the grow
regions by appending to each seed these neighboring pixels that have properties similar to the
seed.

16. Define chain code.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

Chain codes are used to represent a boundary by a connected sequence of straight line
segment of specified length and direction. Typically this representation is based on 4 or 8
connectivity of segments. The direction of each segment is coded by using a numbering scheme.
17. What are the demerits of chain code?
The demerits of chain code are,
 The resulting chain code tends to be quite long.
 Any small disturbance along the boundary due to noise cause changes in the code
that may not be related to the shape of the boundary.
18. What is polygonal approximation method?
Polygonal approximation is an image representation approach in which a digital
boundary can be approximated with arbitrary accuracy by a polygon. For a closed curve the
approximation is exact when the number of segments in polygon is equal to the number of points
in the boundary so that each pair of adjacent points defines a segment in the polygon.
19. Specify the various polygonal approximation methods.
The various polygonal approximation methods are
 Minimum perimeter polygons.
 Merging techniques.
 Splitting techniques.
20. Name few boundary descriptors.
 Simple descriptors.
 Shape descriptors.
 Fourier descriptors.
21. Define length of a boundary.
The length of a boundary is the number of pixels along a boundary. Example, for a chain
coded curve with unit spacing in both directions, the number of vertical and horizontal
components plus √2 times the number of diagonal components gives its exact length.
22. Define shape numbers.
Shape number is defined as the first difference of smallest magnitude. The order n of a
shape number is the number of digits in its representation.
23. Name few measures used as simple descriptors in region descriptors.
 Area.
 Perimeter.
 Mean and median gray levels.
 Minimum and maximum of gray levels.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

 Number of pixels with values above and below mean.

UNIT 4

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

INTEREST POINTS, MORPHOLOGY AND TEXTURE


1. Define texture.
Texture is one of the regional descriptors. It provides measure measures of properties
such as smoothness, coarseness and regularity.
2. Define compactness.
Compactness of a region is defined as (perimeter)2 / area. It is a dimensionless quantity
and is insensitive to uniform scale changes.
3. List the approaches to describe texture of a region.
The approaches to describe the texture of a region are,
 Statistical approach.
 Structural approach.
 Spectural approach.
4. What is meant by markers?
An approach used to control over segmentation is based on markers. Marker is a
connected component belonging to an image. We have internal markers, associated with objects
of interest and external markers associated with background.
5. What are the two principle steps involved in marker selection?
The two steps are
 Preprocessing.
 Definition of a set of criteria that markers must satisfy.

UNIT 5
COLOR IMAGES AND IMAGE COMPRESSION

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

1. What is Data Compression?


Data compression requires the identification and extraction of source redundancy. In
other words, data compression seeks to reduce the number of bits used to store or transmit
information.
2. What are two main types of Data compression?
The two main types of data compression are lossless compression and lossy compression.
3. What is Lossless compression?
Lossless compression can recover the exact original data after compression. It is used
mainly for compressing database records, spreadsheets or word processing files, where exact
replication of the original is essential.
4. What is lossy compression?
Lossy compression will result in a certain loss of accuracy in exchange for a substantial
increase in compression. Lossy compression is more effective when used to compress graphic
images and digitized voice where losses outside visual or aural perception can be tolerated.
5. What is the need for compression?
In terms of storage, the capacity of a storage device can be effectively increased with
methods that compress a body of data on its way to a storage device and decompresses it when it
is retrieved. In terms of communications, the bandwidth of a digital communication link can be
effectively increased by compressing data at the sending end and decompressing data at the
receiving end.
At any given time, the ability of the Internet to transfer data is fixed. Thus, if data can
effectively be compressed wherever possible, significant improvements of data throughput can
be achieved. Many files can be combined into one compressed document making sending easier.
6. What are different compression methods?
The different compression methods are,
 Run Length Encoding (RLE)
 Arithmetic coding
 Huffman coding and
 Transform coding
7. What is run length coding?
Run-length Encoding, or RLE is a technique used to reduce the size of a repeating string
of characters. This repeating string is called a run; typically RLE encodes a run of symbols into
two bytes, a count and a symbol. RLE can compress any type of data regardless of its
information content, but the content of data to be compressed affects the compression ratio.
Compression is normally measured with the compression ratio:

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

8. Define compression ratio.


Compression ratio is defined as the ratio of original size of the image to compressed size
of the image .It is given as
Compression Ratio = original size / compressed size
9. Give an example for Run length Encoding.
Consider a character run of 15 'A' characters, which normally would require 15 bytes to
store: AAAAAAAAAAAAAAA coded into 15A With RLE, this would only require two bytes
to store; the count (15) is stored as the first byte and the symbol (A) as the second byte.
10. What is Huffman Coding?
Huffman compression reduces the average code length used to represent the symbols of
an alphabet. Symbols of the source alphabet, which occur frequently, are assigned with short
length codes. The general strategy is to allow the code length to vary from character to character
and to ensure that the frequently occurring characters have shorter codes.
11. What is Arithmetic Coding?
Arithmetic compression is an alternative to Huffman compression; it enables characters
to be represented as fractional bit lengths. Arithmetic coding works by representing a number by
an interval of real numbers greater or equal to zero, but less than one. As a message becomes
longer, the interval needed to represent it becomes smaller and smaller, and the number of bits
needed to specify it increases.
12. What is JPEG?
The acronym is expanded as "Joint Photographic Expert Group". It is an international standard in
1992. It perfectly Works with colour and greyscale images, Many applications e.g., satellite,
medical etc.
13. What are the basic steps in JPEG?
The Major Steps in JPEG Coding involve:
 DCT (Discrete Cosine Transformation)
 Quantization
 Zigzag Scan
 DPCM on DC component
 RLE on AC Components
 Entropy Coding

14. What is MPEG?

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

The acronym is expanded as "Moving Picture Expert Group". It is an international


standard in 1992. It perfectly Works with video and also used in teleconferencing
15. What is transform coding?
Transform coding is used to convert spatial image pixel values to transform coefficient
values. Since this is a linear process and no information is lost, the number of coefficients
produced is equal to the number of pixels transformed. The desired effect is that most of the
energy in the image will be contained in a few large transform coefficients. If it is generally the
same few coefficients that contain most of the energy in most pictures, then the coefficients may
be further coded by loss less entropy coding. In addition, it is likely that the smaller coefficients
can be coarsely quantized or deleted (lossy coding) without doing visible damage to the
reproduced image.
16. Define I-frame.
I-frame is intraframe or independent frame. An I-frame is compressed independently of
all frames. It resembles a JPEG encoded image. It is the reference point for the motion estimation
needed to generate subsequent P and P-frame.
17. Define p-frame.
P-frame is called predictive frame. A P-frame is the compressed difference between the
current frame and a prediction of it based on the previous I or P-frame.
18. Define B-frame.
B-frame is the bidirectional frame. A B-frame is the compressed difference between the
current frame and a prediction of it based on the previous I or P frame or next P-frame.
Accordingly the decoder must have access to both past and future reference frames.
19. What is zig zag sequence?
The purpose of the Zig zag Scan are,
 To group low frequency coefficients in top of vector.
 .Maps 8 x 8 to a 1 x 64 vector
20. What is block code?
Each source symbol is mapped into fixed sequence of code symbols or code words. So it
is called as block code.
21. Define instantaneous code.
A code that is not a prefix of any other code word is called instantaneous or prefix code
word.

22. Define B2 code.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department


CP7004 Image Processing and Analysis Two Marks Question Bank

Each code word is made up of continuation bit c and information bit which are binary
numbers. This is called B2 code or B code. This is called B2 code because two information bits
are used for continuation bits.
23. Define arithmetic coding.
In arithmetic coding, one to one corresponds between source symbols and code word
doesn’t exist whereas the single arithmetic code word assigned for a sequence of source symbols.
A code word defines an interval of number between 0 and 1.
24. What is bit plane decomposition?
An effective technique for reducing an image’s inter-pixel redundancies is to process the
image’s bit plane individually. This technique is based on the concept of decomposing multilevel
images into a series of binary images and compressing each binary image via one of several
well-known binary compression methods.

Prepared by L. R. Jonisha Miriam, Assistant Professor, ECE Department

Potrebbero piacerti anche