Deep learning-based method for vision-guided robotic grasping of unknown objects |
| |
Affiliation: | 1. The BioRobotics Institute, Scuola Superiore Sant''Anna, Viale R. Piaggio 34, 56025 Pontedera (PI), Italy;2. Department of Excellence in Robotics & A.I., Sant''Anna School of Advanced Studies, 56127 Pisa (PI), Italy;3. Nuovo Pignone Tecnologie S.r.l., Baker-Hughes Company, Via Felice Matteucci 2, 50127 Firenze (FI), Italy;4. Healthcare Engineering Innovation Center (HEIC), Khalifa University, Abu Dhabi, PO Box, 127788, UAE |
| |
Abstract: | Nowadays, robots are heavily used in factories for different tasks, most of them including grasping and manipulation of generic objects in unstructured scenarios. In order to better mimic a human operator involved in a grasping action, where he/she needs to identify the object and detect an optimal grasp by means of visual information, a widely adopted sensing solution is Artificial Vision. Nonetheless, state-of-art applications need long training and fine-tuning for manually build the object’s model that is used at run-time during the normal operations, which reduce the overall operational throughput of the robotic system. To overcome such limits, the paper presents a framework based on Deep Convolutional Neural Networks (DCNN) to predict both single and multiple grasp poses for multiple objects all at once, using a single RGB image as input. Thanks to a novel loss function, our framework is trained in an end-to-end fashion and matches state-of-art accuracy with a substantially smaller architecture, which gives unprecedented real-time performances during experimental tests, and makes the application reliable for working on real robots. The system has been implemented using the ROS framework and tested on a Baxter collaborative robot. |
| |
Keywords: | Collaborative robotics Deep learning Vision-guided robotic grasping Industry 4.0 |
本文献已被 ScienceDirect 等数据库收录! |
|