You are on page 1of 5

Computer Vision

Krishna Pareek

CSC 592, Graduate Student

Q1: How is an Image formed?


Response: White light comprises of all wavelengths of the visible light
together. Every object or part of the object absorbs a particular
wavelength(s) and reflects the rest. The reflected wavelength is the color
perceived for that particular wavelength. For Image to be formed in the
human eye, the reflected (wavelength) by the object reaching the eye
passes through the convex lens of the eye and an inverted image is formed
on the retina. The brain does the processing and inverts the image again on
the retina to perceive the correct orientation of the object. The Images below
show the different kinds of Images formed by various kind of lenses.

A Digital Image is basically a numeric representation of a two dimensional


Image. Generally the term digital image refers to Raster by type. A color
1 | Page

Computer Vision

Krishna Pareek

CSC 592, Graduate Student

digital image consists of certain number of pixels which are the smallest
entity of an image. We can say that greater the number of pixels greater is
the detail content of the image.
Image Formation Phenomenon Using CMOS (Complementary Metal
Oxide Semiconductor): The CMOS sensor consists of pixels of which the
main component is the photodiode. The Photodiode converts the incoming
energy of the photons into electrical charge. Since photodiodes are
monochromatic and cannot differentiate between the colors therefore each
pixel is covered by a micro-lens which only measures the one primary color
or we can say it filters out the other colors.

Out of the total number of pixels on the CMOS sensor, 25% of the pixels
measure the intensity for Red component of the light, 25% of them for Blue
and 50% for Green. Green constitutes for the 50% as our eyes are most
sensitive to this part of the spectrum.
Missing color information in each pixel are estimated by the software and
these approximations reduce the sharpness in the image and introduces
inaccuracies called Demosaic Artifacts.
After the photons hit the pixels, the A to D converter classifies the analogue
voltage of each pixel into one level on 256 level scale and assigns a digital
value to it. In binary 00000000 for 0 level and 11111111 for 255 level (0255), hence representing an 8 bit digital image.

2 | Page

Computer Vision

Krishna Pareek

CSC 592, Graduate Student

Color bit depth is calculated by 256 * 256 * 256 = 16777216 color values.
Example RGB value (200,150, 50) = (11001000, 10010110, 00110010)
digital.

Q2. Difference between human vision and machine vision.


Response:
S.
No
1
2

Human Vision

Machine Vision

Humans can only recognize


roughly about 30 levels of gray.
Very
small
variations
in
Luminosity cannot be resolved
by Human Vision.
Human brain processes visual
information in semantic space
mainly, that is, extracting the
semantically
meaningful
features such as line-segments,
boundaries, shape and so on.

Machine Vision can identify up to


256 different levels of gray.
Machine Vision can easily identify
even the slightest variations in
these levels.
In Machine Vision by recent
information
processing
techniques,
these
kinds
of
features cannot be detected by
computers robustly so that in
computer vision it's still difficult to
process visual information as
humans do.
The human eye is diffraction Machine vision can be improvised
limited
(resolution,
depth and perfected by the algorithms
perception et cetera)
and
associated
optics/electronics/systems.
Human Vision is susceptible to Machine vision is not susceptible
optical illusions.
to optical illusions.

3 | Page

Computer Vision

Krishna Pareek

CSC 592, Graduate Student

Q3. Enumerate major fields in Computer Vision and explain


them.
Response: The major fields in computer vision are:
1. Artificial Intelligence
Computer Vision can be applied to Artificial Intelligence for recognizing
patterns or objects and performing tasks on them or related to them.
We can obtain control information through visual data. One can
recognize handwritings, texts and drawings. Another important part
where computer vision can be applied in artificial intelligence is
passive observation and analysis to analyze certain objects over time.
2. Machine Learning
Machine learning in computer vision is the study of pattern recognition
and computational learning theory. Machine learning explores the
study and construction of algorithms that can learn from and make
predictions on data. It is the science of getting computers to act
without being explicitly programmed.
3. Image Processing
Image processing is the processing of images using mathematical
operations by using any form of signal processing for which the input is
an image. Most image processing techniques treat the image as a 2-D
signal and applying signal processing techniques to it. Image
processing includes a wide array of operations applied to any kind of
image, let it be contrast setting, background isolation et cetera.
4. Pattern Recognition
Pattern recognition focuses on the recognition of patterns and
regularities in data. Pattern recognition systems are in many cases
trained from labeled training data. If no data is available then other
algorithms can be used to discover previously unknown patterns.
Pattern recognition generally aims at providing a reasonable answer
from all matching inputs and to perform most likely / exact matching of
inputs.
5. Biology
Computer Vision can be used in various areas in biology like disease
detection, cell segmentation analysis. The algorithms can be used to
4 | Page

Computer Vision

Krishna Pareek

CSC 592, Graduate Student

analyze biological and medical image data, in particular, cell


microscopy images and medical tomographic images. For example a
focus on automated quantification of cell microscopy images from
large scale high-throughput screens for establishing models in systems
biology.
6. Robotics
Computer Vision provides artificial intelligence to deal with
autonomous planning or deliberation for robotical systems to navigate
through an environment. For example, the robots in RoboCup football
use Pattern recognition to recognize football, goal post, other players
et cetera. These things imply that image sensor technologies and
control theory often are integrated with the processing of image data
to control a robot and that real time processing is emphasized by
means of efficient implementations in hardware and software.
7. Physics
In solid state physics, image sensors rely on computer vision to detect
electromagnetic radiation which is typically in the form of visible or
infrared light. Sophisticated image sensors require quantum mechanics
to understand complete image formation process. Motion in fluids can
be visualized and analyzed using computer vision.
8. Cognitive Science
Cognitive Science which is an interdisciplinary scientific study of mind
and its processes makes use of computer vision to visualize or map the
neural signals generated in the brain. The artificial intelligence
9. Maths
Mathematics is an integral part of computer vision. All the processing is
done using various calculus principles and by applying mathematical
operations. For example, in image processing, to get desired outputs
we can either want to multiply, add, subtract, apply Gaussian kernel to
obtain the edges et cetera. All the processing is mostly done using
mathematical principles.

5 | Page

You might also like