Published May 2011
| public
Book Section - Chapter
Fusion of stereo vision, force-torque, and joint sensors for estimation of in-hand object location
Chicago
Abstract
This paper develops a method to fuse stereo vision, force-torque sensor, and joint angle encoder measurements to estimate and track the location of a grasped object within the hand. We pose the problem as a hybrid systems estimation problem, where the continuous states are the object 6D pose, finger contact location, wrist-to-camera transform and the discrete states are the finger contact modes with the object. This paper develops the key measurement equations that govern the fusion process. Experiments with a Barrett Hand, Bumblebee 2 stereo camera, and an ATI omega force-torque sensor validate and demonstrate the method.
Additional Information
© 2011 IEEE. The author gratefully acknowledges the support from the National Science and Engineering Research Council of Canada (NSERC).Additional details
- Eprint ID
- 74879
- Resolver ID
- CaltechAUTHORS:20170307-170440938
- Natural Sciences and Engineering Research Council of Canada (NSERC)
- Created
-
2017-03-08Created from EPrint's datestamp field
- Updated
-
2021-11-15Created from EPrint's last_modified field