Introduction To Machine Vision Systems
Introduction To Machine Vision Systems
1
Machine Vision
• To become familiar with technologies used
for machine vision as a sensor for robots.
– Camera and lighting technology (obtaining a
digital representation of an image)
– Software (computational techniques to process
or modify the image data)
– Analysis/decisions: using the results of the
processing in robot control
• Additional material in CS766, ECE 533, ME
739
1
Machine Vision in Automation
• Use a camera to inspect parts to:
– Guide a robot or control automated equipment
– Support statistical analysis in a computer-assisted-
manufacturing (CAM) system
– Ensure quality in manufacturing process:
• dimensions/alignment
• Determine if all components are present
• Other quality issues: color, placement, …
1
Why use Vision?
• Dynamic Range
• Can be remotely situated
• Passive
– emits no energy (cf. Laser,
sonar, IR)
– no contact required
• Flexibility
• Affordable
1
Why avoid Vision?
• Computation
– must process images
– data = information
/ Because the lighting is different, these 3 images
appear substantially different to a computer – to
• Calibration a human we easily adapt our perception for
variations in illumination and recognize that all
• Sensitivity to lighting conditions three images are of the same object.
Microscope Table
1
A machine vision system often includes
the following elements:
Image Acquisition (generally from a camera
placed above the production line),
Image Pre-Processing (e.g. increasing the
contrast, motion de-blur, etc),
Feature Extraction (e.g. measuring a distance,
checking a screw is in place etc),
Decisions (i.e. is the part OK to a tolerance, is a
label in the correct position), and,
Control (e.g. give the result to a Programmable
Logic Controller (PLC) or robot controller).
1
Image Acquisition
• Transforms the visual image of a physical
objects into a set of digitized data
– Illumination
– Image formation (including focusing)
– Image detection or sensing
– Formatting camera output signal
1
Image Formation and Detection
Vision systems have an optical-electro device that
converts electromagnetic radiation from the image of
the physical object into an electric signal used by the
vision processing unit
Image is formed by:
– Illumination flux
from object
– Optics (lens)
– Photosensitive
detectors
(photodiodes on
solid state
cameras)
1
Vision – Image Formation
•Shape
•Lighting
•Relative Positions
•Sensor sensitivity
1
Lighting
• Polarized lighting • Direct front lighting
1
Lig
htin
• Diffuse front lighting g • Fibre optic near in-
lighting
1
Image Formation and Detection
Light source
1
Digitization of Camera Signal
• Analog image data (voltage) is sampled and
quantized (often to 8 bits greyscale or 24 bits of
color)
1
Software: Processing the
Data
• The software allows the image to be
processed, analyzed, and stored.
– Different types of software packages are available, ranging from
easy-to-use packages with pre-defined tools, to SDKs (software
development kits) that allow programmers to build custom imaging
applications.
– Matlab™ has an image processing tool box
• Image Pre-processing
• Feature Extraction
1
Image Pre-
processing
• What to do with the image?
– May need to preprocess the image in order to
analyze it
• Remove motion blur (ECE 533/738)
• Enhance contrast
1
I Can See It – Why can’t the Computer?
• Minimize possible problems – The human eye and brain are elaborate and
versatile systems, capable of identifying objects in a wide variety of conditions.
For example, we are able to identify familiar people even when they are
wearing different clothes, and recognize familiar landmarks when driving on a
foggy day. A PC-based imaging system is not as versatile; it can only perform
what it has been programmed to perform. Knowing what the system can and
cannot "see" are important points to keep in mind to obtain the results you
want, and reduce errors and incorrect measurements. Common variables
include:
Changes in object’s color
Changes in surrounding lighting
Changes in camera focus or position
Improperly mounted camera
Environmental vibration
1
Find the man….
1
Distractors
Engineer the
environment!
Great examples include
commercial motion capture
systems
1
Feature
Extraction/Analysis
• 2D Geometric Analysis:
– Must have high contrast to separate (“segment”)
part from background
• In practice back lighting is often used
• The silhouette is used to determine:
– part dimensions: Width, height, orientation, etc
– Part features (e.g. number of holes)
– Relationships between parts
1
Controlled Environment
1
Goals for ME 439 and ME 739
• Modeling Cameras • Modeling Cameras
– Pinhole model
– Basic of pinhole
– Projective mapping
• Kinematics of Vision – Calibration Procedures
– Coordinate • Kinematics of Vision
transformations
– Coordinate transformations
• Processing Images – Motion field equations