Academic Commons


Alignment using an uncalibrated camera system

Yoshimi, Billibon; Allen, Peter K.

We describe a method for the visual control of a robotic system which does not require the formulation of an explicit calibration between image coordinates and the world coordinates. By extracting control information directly from the image, we free our technique from the errors normally associated with a fixed calibration. We attach a camera system to a robot such that the camera system and the robot's gripper rotate simultaneously. As the camera system rotates about the gripper's rotational axis, the circular path traced out by a point-like feature projects to an elliptical path in image space. We gather the projected feature points over part of a rotation and fit the gathered data to an ellipse. The distance from the rotational axis to the feature point in world space is proportional to the size of the generated ellipse. As the rotational axis gets closer to the feature, the feature's projected path will form smaller and smaller ellipses. When the rotational axis is directly above the object, the trajectory degenerates from an ellipse to a single point. We demonstrate the efficacy of the algorithm on the peg-in-hole problem.



Also Published In

IEEE Transactions on Robotics and Automation

More About This Work

Academic Units
Computer Science
Published Here
October 25, 2012