首页 | 官方网站   微博 | 高级检索  
     

多视角数据融合的特征平衡YOLOv3行人检测研究
引用本文:陈丽,马楠,,逄桂林,高跃,李佳洪,,张国平,吴祉璇,姚永强.多视角数据融合的特征平衡YOLOv3行人检测研究[J].智能系统学报,2021,16(1):57-65.
作者姓名:陈丽  马楠    逄桂林  高跃  李佳洪    张国平  吴祉璇  姚永强
作者单位:1. 北京联合大学 北京市信息服务工程重点实验室,北京 100101;2. 北京联合大学 机器人学院,北京 100101;3. 北京交通大学 计算机与信息技术学院,北京 100044;4. 清华大学 软件学院,北京 100085
摘    要:针对复杂场景下行人发生遮挡检测困难以及远距离行人检测精确度低的问题,本文提出一种多视角数据融合的特征平衡YOLOv3行人检测模型(MVBYOLO),包括2部分:自监督学习的多视角特征点融合模型(Self-MVFM)和特征平衡YOLOv3网络(BYOLO)。Self-MVFM对输入的2个及以上的视角数据进行自监督学习特征,通过特征点的匹配实现多视角信息融合,在融合时使用加权平滑算法解决产生的色差问题;BYOLO使用相同分辨率融合高层语义特征和低层细节特征,得到平衡的语义增强多层级特征,提高复杂场景下车辆前方行人检测的精确度。为了验证所提出方法的有效性,在VOC数据集上进行对比实验,最终AP值达到80.14%。与原YOLOv3网络相比,本文提出的MVBYOLO模型精度提高了2.89%。

关 键 词:多视数据  自监督学习  特征点匹配  特征融合  YOLOv3网络  平衡特征  复杂场景  行人检测

Research on multi-view data fusion and balanced YOLOv3 for pedestrian detection
CHEN Li,MA Nan,,PANG Guilin,GAO Yue,LI Jiahong,,ZHANG Guoping,WU Zhixuan,YAO Yongqiang.Research on multi-view data fusion and balanced YOLOv3 for pedestrian detection[J].CAAL Transactions on Intelligent Systems,2021,16(1):57-65.
Authors:CHEN Li  MA Nan    PANG Guilin  GAO Yue  LI Jiahong    ZHANG Guoping  WU Zhixuan  YAO Yongqiang
Affiliation:1. Beijing Key Laboratory of Information Service Engineering, Beijing Union University, Beijing 100101, China;2. College of Robotics, Beijing Union University, Beijing 100101, China;3. School of Computer and Information Technology, Beijing Jiaoton
Abstract:Because of the occlusion and low accuracy of long-distance detection, pedestrian detection in complex scenes is difficult. Therefore, a pedestrian detection method based on multi-view data fusion and balanced YOLOv3 (MVBYOLO) is proposed, including the self-supervised network for multi-view fusion model (Self-MVFM) and balanced YOLOv3 network (BYOLO). Self-MVFM fuses two or more input perspective data through a self-supervised network and incorporates a weighted smoothing algorithm to solve the color difference problem during the fusion; BYOLO uses the same resolution to fuse high- and low-level semantic features to obtain balanced semantic information, thereby enhancing multi-level features and improving the accuracy of pedestrian detection in front of vehicles in complex scenes. A comparative experiment is conducted on the VOC dataset to verify the effectiveness of the proposed method. The final AP value reaches 80.14%. The experimental results indicate that compared with the original YOLOv3 network, the accuracy of the MVBYOLO is increased by 2.89%.
Keywords:multi-view data  self- supervised learning  feature point matching  feature fusion  YOLOv3 network  balanced feature  complex scene  pedestrian detection
点击此处可从《智能系统学报》浏览原始摘要信息
点击此处可从《智能系统学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号