Multiple cameras help perform intricate handling

Then interpreted as an intricate robotics system, the human body is a versatile, mobile, detection system with all-encompassing vision, processing power, memory, sensory feedback, and movement. Comparatively, many industrial robotic inspection systems are still fairly primitive. Although robots used in mechanical and electronic assembly lines implement several axes of movement, they often have to be instructed by human operators to perform specific motion patterns.

Oct 1st, 1997
Th Vsd51289 26

Multiple cameras help perform intricate handling

By Dave Wilson

Then interpreted as an intricate robotics system, the human body is a versatile, mobile, detection system with all-encompassing vision, processing power, memory, sensory feedback, and movement. Comparatively, many industrial robotic inspection systems are still fairly primitive. Although robots used in mechanical and electronic assembly lines implement several axes of movement, they often have to be instructed by human operators to perform specific motion patterns.

To improve robot effectiveness, machine-vision designers are adding more sensory functions and devices. These include noncontact sensors, such as vision systems, and contact sensors, such as transducer force sensors. Whereas vision systems are used for visual input, contact sensors measure the components of force and torque from robot arms and grippers.

Mimicking the eye

Many industrial robots use a single charge-coupled-device (CCD) camera to detect the parts being inspected (see Fig. 1). Generally, if more viewing information is required, more than one camera is used. In some applications, three cameras are attached to gather information in the x, y, and¥planes.

For example, a three-camera robotic system is being used in the inspection of fragile, fine-pitch surface-mounted devices (SMDs), according to Kevin Keras, marketing manager at Seiko Instruments Robotics Division (Torrance, CA). During robotic inspection, the SMDs are examined in a matrix tray feeder prior to selection and placement. A downward-facing camera attached to a robotic arm is directed over each part and determines whether the part is present. A second, fixed, upward-looking camera inspects the leads of the SMD for conformance to specifications. A third camera looks down onto the printed-circuit board for fiducial reference marks and then instructs the robot where to place the SMDs.

To build more reliable inspection systems, additional sensors such as vacuum sensors could determine whether the part was indeed picked up and placed. Whereas a vision system could be used for this function, a moving camera looking for parts would reduce overall system speed. For this reason, many robots have multitasking controllers that allow other sensors to check for parts presence while the robot arm is moving toward the placement area.

Integrated systems

A number of vendors are now offering completely integrated robot/vision systems. For instance, Seiko Instruments has produced a Microsoft Windows-based product called Vision Guide that uses an object-oriented approach as well as commercially available image-processing boards from Cognex Corp. (Natick, MA) and Matrox Electronic Systems (Dorval, Quebec, Canada). Calibration of the robot is automated, and operators need to learn only three commands to use the system.

For parts under inspection that are difficult to detect, such as parts stored in a bin with only one opening, stereoscopic viewing might solve the problem. In such a system, predominantly where range is to be measured, a pair of cameras or sonar transmitters and receivers can be used to acquire an image. A typical configuration, such as the one developed at the University of Rochester Vision and Robotics Laboratory (Rochester, NY), includes two movable color CCD cameras providing input to a MaxVideo image processor from Datacube Inc. (Danvers, MA). Camera motors are used to control both the tilt angle of the dual-eye platform and each camera`s pan angle. According to university research ers, camera movements of 400°/s can be achieved, which are somewhat close to human-eye movements of 700/s.

Three-dimensional (3-D) vision systems are preferred to manage more detailed tasks. At the Purdue University Robot Vision Laboratory (West Lafayette, IN), a 3-D vision system is recognizing tubular objects that are mixed together (see Fig. 2). The system, called the Tubular Objects Bin-Picking System, was originally developed with funding from Nippondenso (Japan); it is currently undergoing refinement for installation in a high-volume auto mobile-parts production line in Japan. In this system, an overhead light scanner illuminates the objects.

Following parts scanning, triangulation methods are usually implemented to generate a depth map of the scene. A sequence of processing steps, which include thresholding, edge detection, and noise removal, are then performed on the depth map to recognize specific patterns in the image. On a given scan of a random pile of tubes, the 3-D vision system generally recognizes between one and four complete tubes, in addition to between five and ten partial tube fragments.

However, experiments performed with the robotic system have indicated that structured light scanning is not a promising technique for image acquisition. Illuminating the many tiny holes and crevices of the parts results in extremely noisy images; therefore, robust depth-map generation is not possible. Consequently, system developers plan to use stereo vision to acquire the raw data.

Reducing scanning errors is important in binocular vision systems. For example, at the Robotics Institute of Carnegie Mellon University (Pittsburgh, PA), vision-system developers are using three cameras on their ranging platform, rather than two. The three cameras help to compensate for computational inaccuracy, introduce redundancy, reduce the chance of false matches, and increase overall system accuracy. Claimed to be one of the fastest depth-ranging systems ever implemented, the three-camera system is based on a 20 million-floating-point-operation-per-second synchronous computer.

Internal sensors

Whereas vision systems can guide a robot arm, tactile sensors perform delicate gripping and assembly operations. These sensors provide positional detail more accurately than vision systems can do alone. However, research performed on touch sensors and robots has been limited, despite the fact that multifingered robotic hands with touch sensors can move, probe, and change the vision-system environment.

As a robot approaches an object, the view from a conventional vision system might be obscured. Consequently, the combination of errors in the positioning of the gripper device with respect to the object proves difficult in applying the correct forces and torques. Having an imaging and/or a ranging sensor in the gripper can overcome this problem. Close-up imaging and ranging also provides information about an object`s surface shape and texture, a useful feature in an application such as following a weld line.

Jeremy Wilson, engineering manager of Kinetic Sciences (Vancouver, BC, Canada), says that in the Vision Skin Project his company explored for the Canadian Space Agency (CSA), a close-up imaging and ranging system was built into the fingertips of a robot gripper. Teaming with a silicon-micromachining group at Simon Fraser University (Burnaby, Canada), Kinetic Sciences researchers developed a close-up imaging and laser ranging sensor. The project was funded under the CSA STEAR (Strategic TEchnologies in Automation and Robotics) program, a long-term research initiative that supports Canada`s role in the International Space Station program.

If the objects in a particular scene are positioned in an unspecified location or orientation, an active robotic reasoning system can be used to seek and gather selective information by exploring the entire visual field based upon information that is relevant to the performance of a particular task. In the department of mechanical engineering and robotics research at the University of Surrey (Guildford, Surrey, England), John Pretlove and his team have built a mechatronic, lightweight, eye-in-hand sensor for this purpose. The function of the sensor is to provide enough information so that the robot can be guided to a known object at an unspecified location and orientation within a robot workcell. Once there, the robot can perform functions such as assembly, pick and place, bin-picking, visual tracking, and object interception in a conveyor system.

The ultimate machine

The Cog robot, which has been developed at the Massachusetts Institute of Technology (Cambridge, MA), nearly represents the ultimate robot. It contains a set of sensors and actuators that approximate human sensory and motor dynamics (see Fig. 3). The robot is directed by a scalable, multiple-instruction, multiple-data computer with 239 processor nodes. Each node has a Motorola 68332 microprocessor that communicates with front-end processor, motor controller, frame grabber, and display boards.

For Cog viewing, two monochrome CCD cameras are used to achieve disparity and approximate depth--important factors for object discrimination. This setup is able to closely duplicate human-eye speed and range of motion. Mounted on the robot`s head-like structure, the camera system moves with reasonable speed. It has a wide field of view for detection of objects and motion, a high-resolution capability, and a wide peripheral view.

Although machine-vision systems constitute a reliable means of acquiring imaging information about a scene, in applications such as robotics, sensors that provide tactile feedback are equally as important. New sensory devices are expected to empower robotic systems with increasingly more-effective detection and mobility capabilities.

DAVE WILSON is a science writer in London, England.

Click here to enlarge image

FIGURE 1. For pick-and-place parts-inspection applications, the D-TRAN system robot from Seiko Instruments can be fitted with a downward-facing camera to determine the presence of parts.

Click here to enlarge image

Click here to enlarge image

FIGURE 2. At Purdue University, a three-dimensional vision system called the Tubular Objects Bin-Picking System recognizes and localizes tubular objects. The intensity image shows a typical scene on which the system operates (top), while the resulting range map has been segmented into a set of tube fragments (bottom).

Click here to enlarge image

FIGURE 3. Under development at the Massachusetts Institute of Technology, the Cog robot can perform many human functions--except walking. Each robot eye consists of two monochrome charge- coupled-device cameras that approximate the wide peripheral view and high resolution of human eyes.

Company Information

Carnegie Mellon University

Pittsburgh, PA 15213-3890

(412) 268-2446

Fax: (412) 268-6944


Cognex Corp.

Natick, MA 01760

(508) 650-3000


Dorval, QC, Canada H9P 2T4

(514) 969-6061

Fax: (514) 969-6273

Massachusetts Institute of Technology

Cambridge, MA 02139-4307

(617) 253-1000


University of Rochester

Vision and Robotics Laboratory

Rochester, NY 14627

(716) 275-2121

Fax: (716) 275-2190


Datacube Inc.

Danvers, MA 01923-4505

(508) 777-4200

Fax: (508) 750-0938



Kinetic Sciences

Vancouver, BC, Canada V6T 1W5

(604) 822-2144

Fax: (604) 822-6188



Purdue University

School of Electrical and Computer Engineering

West Lafayette, IN 47907

(765) 494-0620

Fax: (765) 494-6440

Seiko Instruments

Factory Automation Division

Torrance, CA 90505

(310) 517-7850

Fax: (310) 517-8158


Simon Fraser University

Burnaby, Canada

(604) 291-3111

University of Surrey

Guildford, Surrey GU2 5XH England

1483 259681

Fax: 1483 306039


More in Robotics