Hand-eye coordination for grasping moving objects

Allen, Peter K.; Timcenko, Aleksandar; Yoshimi, Billibon; Michelman, Paul

Most robotic grasping tasks assume a stationary or fixed object. In this paper, we explore the requirements for grasping a moving object. This task requires proper coordination between at least 3 separate subsystems: dynamic vision sensing, real-time arm control, and grasp control. As with humans, our system first visually tracks the object's 3-D position. Because the object is in motion, this must be done in a dynamic manner to coordinate the motion of the robotic arm as it tracks the object. The dynamic vision system is used to feed a real-time arm control algorithm that plans a trajectory. The arm control algorithm is implemented in two steps: 1) filtering and prediction, and 2) kinematic transformation computation. Once the trajectory of the object is tracked, the hand must intercept the object to actually grasp it. We present 3 different strategies for intercepting the object and results from the tracking algorithm.



Also Published In

Sensor Fusion III: 3D Perception and Recognition
Society of Photo-optical Instrumentation Engineers

More About This Work

Academic Units
Computer Science
Proceedings of SPIE--The International Society for Optical Engineering, 1383
Published Here
November 12, 2012