Learning object,grasping and manipulation activities using hierarchical HMMs |
| |
Authors: | Mitesh Patel Jaime Valls Miro Danica Kragic Carl Henrik Ek Gamini Dissanayake |
| |
Affiliation: | 1. Faculty of Engineering and IT, University of Technology Sydney (UTS), 15 Broadway, Ultimo, NSW, 2007, Australia 2. Members of the Computer Vision and Active Perception Laboratory, Centre for Autonomous Systems, School of Computer Science and Communication, The Royal Institute of Technology (KTH), Stockholm, Sweden
|
| |
Abstract: | This article presents a probabilistic algorithm for representing and learning complex manipulation activities performed by humans in everyday life. The work builds on the multi-level Hierarchical Hidden Markov Model (HHMM) framework which allows decomposition of longer-term complex manipulation activities into layers of abstraction whereby the building blocks can be represented by simpler action modules called action primitives. This way, human task knowledge can be synthesised in a compact, effective representation suitable, for instance, to be subsequently transferred to a robot for imitation. The main contribution is the use of a robust framework capable of dealing with the uncertainty or incomplete data inherent to these activities, and the ability to represent behaviours at multiple levels of abstraction for enhanced task generalisation. Activity data from 3D video sequencing of human manipulation of different objects handled in everyday life is used for evaluation. A comparison with a mixed generative-discriminative hybrid model HHMM/SVM (support vector machine) is also presented to add rigour in highlighting the benefit of the proposed approach against comparable state of the art techniques. |
| |
Keywords: | |
本文献已被 SpringerLink 等数据库收录! |
|