Deep learning-based smart task assistance in wearable augmented reality |
| |
Affiliation: | 1. School of mechanical engineering and automation, Beihang university, 37 xueyuan road, haidian district, Beijing, 100191, China;2. Beijing Spacecrafts, China Academy of Space Technology, 102 dengzhuang south road, haidian district, Beijing, 100094, China;1. Cranfield Manufacturing, School of Aerospace, Transport and Manufacturing, Cranfield University, UK;2. HSSMI a Manufacturing Innovation Institute, London, UK;1. School of Modern Post (School of Automation), Beijing University of Posts and Telecommunications, Beijing, China;2. School of Mechanical Engineering and Automation, Beihang University, Beijing, China;3. Department of Digital Machining, Sandvik Coromant, Stockholm, Sweden;4. COMAC Beijing Aircraft Technology Research Institute, Beijing, China;5. Department of Production Engineering, KTH Royal Institute of Technology, Stockholm, Sweden |
| |
Abstract: | Wearable augmented reality (AR) smart glasses have been utilized in various applications such as training, maintenance, and collaboration. However, most previous research on wearable AR technology did not effectively supported situation-aware task assistance because of AR marker-based static visualization and registration. In this study, a smart and user-centric task assistance method is proposed, which combines deep learning-based object detection and instance segmentation with wearable AR technology to provide more effective visual guidance with less cognitive load. In particular, instance segmentation using the Mask R-CNN and markerless AR are combined to overlay the 3D spatial mapping of an actual object onto its surrounding real environment. In addition, 3D spatial information with instance segmentation is used to provide 3D task guidance and navigation, which helps the user to more easily identify and understand physical objects while moving around in the physical environment. Furthermore, 2.5D or 3D replicas support the 3D annotation and collaboration between different workers without predefined 3D models. Therefore, the user can perform more realistic manufacturing tasks in dynamic environments. To verify the usability and usefulness of the proposed method, we performed quantitative and qualitative analyses by conducting two user studies: 1) matching a virtual object to a real object in a real environment, and 2) performing a realistic task, that is, the maintenance and inspection of a 3D printer. We also implemented several viable applications supporting task assistance using the proposed deep learning-based task assistance in wearable AR. |
| |
Keywords: | |
本文献已被 ScienceDirect 等数据库收录! |
|