首页 | 本学科首页   官方微博 | 高级检索  
     

结合轻量Openpose和注意力引导图卷积的动作识别
引用本文:张富凯,贺天成.结合轻量Openpose和注意力引导图卷积的动作识别[J].计算机工程与应用,2022,58(18):180-187.
作者姓名:张富凯  贺天成
作者单位:河南理工大学 计算机科学与技术学院,河南 焦作 454000
基金项目:河南省高等学校重点科研项目(22A520028);
摘    要:现有人体姿态动作识别方法忽视前期姿态估计算法的作用,没有充分提取动作特征,提出一种结合轻量级Openpose和注意力引导图卷积网络的动作识别方法。该方法包含基于shufflenet的Openpose算法和基于不同尺度邻接矩阵注意力的图卷积算法。输入视频由轻量Openpose处理得到18个人体关键点信息,表达为基础时空图数据形式。节点的不同尺度邻居信息对应的邻接矩阵通过自注意力机制计算影响力,将各尺度邻接矩阵加权合并输入图卷积网络提取特征。提取到的鉴别特征通过全局平均池化和softmax分类器输出动作类别。在Le2i Fall Detection数据集和自定义的UR-KTH数据集上的实验表明,动作识别的准确率分别为95.52%和95.07%,达到了预期效果。

关 键 词:动作识别  姿态估计  注意力  图卷积网络

Action Recognition Combined with Lightweight Openpose and Attention-Guided Graph Convolution
ZHANG Fukai,HE Tiancheng.Action Recognition Combined with Lightweight Openpose and Attention-Guided Graph Convolution[J].Computer Engineering and Applications,2022,58(18):180-187.
Authors:ZHANG Fukai  HE Tiancheng
Affiliation:School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan 454000, China
Abstract:Existing human pose action recognition methods ignore the role of the previous pose estimation algorithms and do not fully extract action features. This paper proposes an action recognition method that combines lightweight Openpose and attention-guided graph convolutional network. The method includes an Openpose algorithm based on shufflenet and a graph convolution algorithm based on attention of different scales of adjacency matrices. The input video is processed by lightweight Openpose to obtain 18 human body keypoint information, which is expressed as the basic spatiotemporal graph data form. The adjacency matrix corresponding to the neighbor information of different scales of the node calculates the influence through the self-attention mechanism, and weights the adjacency matrix of each scale. The input graph convolutional network is merged to extract features. The extracted discriminative features are output action categories through global average pooling and softmax classifier. Experiments on the Le2i Fall Detection dataset and the custom UR-KTH dataset show that the accuracy of action recognition is 95.52% and 95.07%, respectively, achieving the expected results.
Keywords:action recognition  pose estimation  attention  graph convolutional network  
点击此处可从《计算机工程与应用》浏览原始摘要信息
点击此处可从《计算机工程与应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号