Skip to main content Accessibility help
×
Home
Hostname: page-component-5cfd469876-tkzrn Total loading time: 0.19 Render date: 2021-06-25T08:15:59.551Z Has data issue: true Feature Flags: { "shouldUseShareProductTool": true, "shouldUseHypothesis": true, "isUnsiloEnabled": true, "metricsAbstractViews": false, "figures": true, "newCiteModal": false, "newCitedByModal": true, "newEcommerce": true }

Article contents

Multi-cameras visual servoing for dual-arm coordinated manipulation

Published online by Cambridge University Press:  12 January 2017

Jiadi Qu
Affiliation:
State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, P. R. China. E-mails: qujiadi@hotmail.com, 18745007416@163.com
Fuhai Zhang
Affiliation:
State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, P. R. China. E-mails: qujiadi@hotmail.com, 18745007416@163.com
Yili Fu
Affiliation:
State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, P. R. China. E-mails: qujiadi@hotmail.com, 18745007416@163.com
Shuxiang Guo
Affiliation:
State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, P. R. China. E-mails: qujiadi@hotmail.com, 18745007416@163.com

Summary

Although image-based visual servoing (IBVS) provides good performance in many dual-arm manipulation applications, it reveals some fatal limitations when dealing with a large position and orientation uncertainty. The object features may leave the camera's field of view, and the dual-arm robot may not converge to their goal configurations. In this paper, a novel vision-based control strategy is presented to resolve these limitations. A visual path planning method for dual-arm end-effector features is proposed to regulate the large initial poses to the pre-alignment poses. Then, the visual constraints between the position and orientation of two objects are established, and the sequenced subtasks are performed to attain the pose alignment of two objects by using a multi-tasks IBVS method. The proposed strategy has been implemented on a MOTOMAN robot to perform the alignment tasks of plug–socket and cup–lid, and results indicate that the plug and socket with the large initial pose errors 145.4 mm, 43.8○ (the average errors of three axes) are successfully aligned with the allowed pose alignment errors 3.1 mm, 1.1○, and the cup and lid with the large initial pose errors 131.7 mm, 20.4○ are aligned with the allowed pose alignment errors −2.7 mm, −0.8○.

Type
Articles
Information
Robotica , Volume 35 , Issue 11 , November 2017 , pp. 2218 - 2237
Copyright
Copyright © Cambridge University Press 2017 

Access options

Get access to the full version of this content by using one of the access options below.

References

1. Stuckler, J. and Behnke, S. S., “Following Human Guidance to Cooperatively Carry a Large Object,” Proceedings of the IEEE International Conference on Humanoid Robots, Piscataway, NJ: IEEE, Bled, Slovenia (Oct. 2011) pp. 218223.Google Scholar
2. Shin, S. Y. and Kim, C. H., “Human-like motion generation and control for humanoid's dual arm object manipulation,” IEEE Trans. Ind. Electron. 62 (4), 22652276 (Apr. 2015).CrossRefGoogle Scholar
3. Ren, Y., Liu, Y. C., Jin, M. H. and Liu, H., “Biomimetic object impedance control for dual-arm cooperative 7-DOF manipulators,” Robot. Auton. Syst. 75, 273287 (Jan. 2016).CrossRefGoogle Scholar
4. Dauchez, P., Fraisse, P. and Pierrot, F., “A Vision/Position/Force Control Approach for Performing Assembly Tasks with a Humanoid Robot,” Proceedings of the IEEE/RSJ International Conference on Humanoid Robots, Piscataway, NJ: IEEE, Tsukuba, Japan (Dec. 2005) pp. 277282.Google Scholar
5. Yamazaki, K., Oya, R., Nagahama, K., Okada, K. and Inaba, M., “Bottom dressing by a dual-arm robot using a clothing state estimation based on dynamic shape changes,” Int. J. Adv. Robot. Syst. 13 (5), 113 (Jan. 2016).CrossRefGoogle Scholar
6. Beetz, M., Klank, U., Maldonado, A., Pangercic, D. and Ruhr, T., “Robotic Roommates Making Pancakes,” Proceedings of the IEEE International Conference on Humanoid Robots, Piscataway, NJ: IEEE, Bled, Slovenia (Oct. 2011) pp. 529536.Google Scholar
7. Smith, C., Karayiannidis, Y., Nalpantidis, L., Gratal, X., Qi, P., Dimarogonas, D. V. and Kragic, D., “Dual arm manipulation: A survey,” Robot. Auton. Syst. 60, 13401353 (Oct. 2012).CrossRefGoogle Scholar
8. Kragic, D. and Christensen, H. I., “Survey on visual servoing for manipulation,” Royal Inst. Technol., Stockholm, Sweden, Tech. Rep. ISRN KTH/NA/P-02/01-SE, (2002).Google Scholar
9. Hutchinson, S., Hager, G. D. and Corke, P. I., “A tutorial on visual servo control,” IEEE Trans. Robot. Automat. 12 (5), 651670 (Feb. 1996).CrossRefGoogle Scholar
10. Kermorgant, O. and Chaumette, F., “Multi-sensor Data Fusion in Sensor-based Control: Application to Multi-camera Visual Servoing,” Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Piscataway, NJ: IEEE, ShangHai, China (May 2011) pp. 45184523.Google Scholar
11. Malis, E., Chaumette, F. and Boudet, S., “Multi-cameras Visual Servoing,” Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Piscataway, NJ: IEEE, San Francisco, CA (Apr. 2000) pp. 31833188.Google Scholar
12. Chaumette, F. and Hutchinson, S., “Visual servo control, part i: Basic approaches,” IEEE Robot, Automat. Mag. 13 (4), 8290 (Dec. 2006).CrossRefGoogle Scholar
13. Chaumette, F. and Hutchinson, S., “Visual servo control, part ii: Advanced approaches,” IEEE Robot, Automat. Mag. 14 (1), 109118 (Mar. 2007).CrossRefGoogle Scholar
14. Lippiello, V., Siciliano, B. and Villani, L., “Position-based Visual Servoing in Industrial Multi-robot Cells Using a Hybrid Camera Configuration,” IEEE Trans. Robot. 23 (1), 7386 (Feb. 2007).CrossRefGoogle Scholar
15. Shauri, R. L. A. and Nonami, K., “Assembly manipulation of small objects by dual-arm manipulator,” Assem. Autom. 31 (3), 263274 (2011).CrossRefGoogle Scholar
16. Vahrenkamp, N., Boge, C., Welke, K., Asfour, T., Walter, J. and Dillmann, R., “Visual Servoing for Dual Arm Motions on a Humanoid Robot,” Proceedings of the IEEE/RSJ International Conference on Humanoid Robots, Paris, France (Dec. 2009) pp. 208214.Google Scholar
17. Huebner, K., Welke, K., Przybylski, M., Vahrenkamp, N., Asfour, T., Kragic, D. and Dillmann, R., “Grasping Known Objects with Humanoid Robots; a Box-based Approach,” Proceedings of the International Conference on Advanced Robitics (ICAR), Munich, Germany (2009) pp. 179184.Google Scholar
18. Shen, Y., Xu, D., Tan, M. and Yu, J. Z., “Mixed visual control methods for robots with self-calibrated stereo rig,” IEEE Trans. Instrum. Meas. 59 (2), 470479 (Feb. 2010).CrossRefGoogle Scholar
19. Han, S., See, W., Lee, J., Lee, M. and Hashimoto, H., “Image-based Visual Servoing Control of a Scara Type Dual-arm Robot,” Proceedings of the IEEE International Symposium Industrial Electronics (ISIE), Cholula, Mexico (2000) vol. 2, pp. 517522.Google Scholar
20. Hynes, P., Dodds, G. and Wilkinson, A. J., “Uncalibrated Visual Servoing of a Dual Arm Robot for Mis Suturing,” Proceedings of the IEEE/RAS-EMBS International Conference on Biomedical Robotics and Biomechatronics, Pisa, Italy (Feb. 2006) pp. 420425.Google Scholar
21. Kruse, D., Wen, J. T. and Radke, R. J., “A sensor-based dual-arm tele-robotic system,” IEEE Trans. Autom. Sci. Eng. 12 (1), 418 (Jan. 2015).CrossRefGoogle Scholar
22. Zereik, E., Sorbara, A., Casalino, G. and Didot, F., “Autonomous Dual-arm Mobile Manipulator Crew Assistant for Surface Operations: Force/vision –Guided Grasping,” Proceedings of the International Conference on Recent advances in Space Technologies, Istanbul, Turkey (Jun. 2009) pp. 710715.Google Scholar
23. Miyabe, T., Konno, A. and Uchiyama, M., “Automated Object Capturing with a Two-arm Flexible Manipulator,” Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Taipei, Taiwan China (May 2003) vol. 2, pp. 25292534.Google Scholar
24. Fleurmond, R. and Cadenat, V., “Multi-cameras Visual Servoing to Perform a Coordinated Task Using a Dual Arm Robot,” Proceedings of the International Conference on Informatics in control, automation and Robotics (ICINCO), Vienna, Austria (2014) vol. 2, pp. 3744.Google Scholar
25. Huang, S., Murakami, K., Yamakawa, Y., Senoo, T. and Ishikawa, M., “Fast Peg-and-hole Alignment Using Visual Compliance,” Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Tokyo, Japan (Nov. 2013) pp. 286292.Google Scholar
26. Kim, H. W., Cho, J. S. and Kweon, I. S., “A Novel Image-Based Control Law for the Visual Servoing System Under Large Pose Error,” Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Takamatsu, Japan (Oct. 2000) vol. 1, pp. 263268.Google Scholar
27. Kermorgant, O. and Chaumette, F., “Dealing with constraints in sensor-based robot control,” IEEE Trans. Robot. 30 (1), 244257 (Feb. 2014).CrossRefGoogle Scholar
28. Schramm, F., Geffard, F., Morel, G. and Micaelli, A., “Calibration Free Image Point Path Planning Simultaneously Ensuring Visibility and Controlling Camera Path,” Proceedings of the 2007 IEEE International Conference on Robotics and Automation, Roma, Italy (2007) pp. 20742079.CrossRefGoogle Scholar
29. Schramm, F., Micaelli, A. and Morel, G., “Calibration Free Path Planning for Visual Servoing Yielding Straight Line Behavior Both in Image and Work Space,” Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Edmonton, Canada (Oct. 2005) vol. 1, pp. 22162221.Google Scholar
30. Mansard, N. and Chaumette, F., “Tasking sequencing for high-level sensor-based control,” IEEE Trans. Robot. 23 (1), 6072 (Feb. 2007).CrossRefGoogle Scholar
2
Cited by

Send article to Kindle

To send this article to your Kindle, first ensure no-reply@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about sending to your Kindle. Find out more about sending to your Kindle.

Note you can select to send to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be sent to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

Multi-cameras visual servoing for dual-arm coordinated manipulation
Available formats
×

Send article to Dropbox

To send this article to your Dropbox account, please select one or more formats and confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your <service> account. Find out more about sending content to Dropbox.

Multi-cameras visual servoing for dual-arm coordinated manipulation
Available formats
×

Send article to Google Drive

To send this article to your Google Drive account, please select one or more formats and confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your <service> account. Find out more about sending content to Google Drive.

Multi-cameras visual servoing for dual-arm coordinated manipulation
Available formats
×
×

Reply to: Submit a response

Please enter your response.

Your details

Please enter a valid email address.

Conflicting interests

Do you have any conflicting interests? *