首页 | 本学科首页   官方微博 | 高级检索  
     


Robotics‐based telepresence using multi‐modal interaction for individuals with visual impairments
Authors:Chung Hyuk Park  Ayanna M Howard
Affiliation:1. School of Engineering and Computing Sciences, New York Institute of Technology, New York, NY, USA;2. School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA, USA
Abstract:Although many assistive devices have been developed and utilized to aid in daily living, the most general assistive means for individuals with visual impairments are the walking cane and guide dogs. These assistive means are effective in assisting the user in navigating within an environment; however, the navigation space is limited to the proximal environment of the user. Thus, in this paper, we discuss a method to increase the range of accessibility to a remote environment through robotic embodiment that enables teleoperation and teleperception through multi‐modal feedback. In order to transform remote spatial information into a non‐visual modality, we present a framework for utilizing an RGB‐D‐based depth camera, a mobile robot, and a haptic interface for 3D haptic rendering to accomplish the goal of haptic exploration of a remote environment. Experiments with three different control methods for robot interaction are designed for users with and without visual impairments. Several hypothesis are built to study the correlation between control/feedback modality and performance in telerobotic operations. Results show that users performed best when combining semi‐autonomous navigation with 3D haptic exploration and also rated their experience with our system as fairly good. Copyright © 2014 John Wiley & Sons, Ltd.
Keywords:assistive robotics  3D haptic rendering  telepresence  individuals with visual impairments  telerobotics
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号