首页 | 本学科首页   官方微博 | 高级检索  
     

基于对象特征的深度哈希跨模态检索
引用本文:朱杰,白弘煜,张仲羽,谢博鋆,张俊三.基于对象特征的深度哈希跨模态检索[J].计算机科学与探索,2021,15(5):922-930.
作者姓名:朱杰  白弘煜  张仲羽  谢博鋆  张俊三
作者单位:中央司法警官学院 信息管理系,河北 保定 071000;河北大学 数学与信息科学学院,河北 保定 071002;中国石油大学(华东)计算机科学与技术学院,山东 青岛 266580
基金项目:2019年中央司法警官学院省级大学生创新创业训练计划项目;河北大学高层次创新人才科研启动经费项目;河北省高等学校科学技术研究项目;国家自然科学基金;河北省自然科学基金青年基金项目
摘    要:随着不同模态的数据在互联网中的飞速增长,跨模态检索逐渐成为了当今的一个热点研究问题。哈希检索因其快速、有效的特点,成为了大规模数据跨模态检索的主要方法之一。在众多图像-文本的深度跨模态检索算法中,设计的准则多为尽量使得图像的深度特征与对应文本的深度特征相似。但是此类方法将图像中的背景信息融入到特征学习中,降低了检索性能。为了解决此问题,提出了一种基于对象特征的深度哈希(OFBDH)跨模态检索方法。此方法从特征映射中学习到优化的、有判别力的极大激活特征作为对象特征,并将其融入到图像与文本的跨模态网络学习中。实验结果表明,OFBDH能够在MIRFLICKR-25K、IAPR TC-12和NUS-WIDE三个数据集上获得良好的跨模态检索结果。

关 键 词:对象特征  跨模态损失  网络参数学习  检索

Object Feature Based Deep Hashing for Cross-Modal Retrieval
ZHU Jie,BAI Hongyu,ZHANG Zhongyu,XIE Bojun,ZHANG Junsan.Object Feature Based Deep Hashing for Cross-Modal Retrieval[J].Journal of Frontier of Computer Science and Technology,2021,15(5):922-930.
Authors:ZHU Jie  BAI Hongyu  ZHANG Zhongyu  XIE Bojun  ZHANG Junsan
Affiliation:(Department of Information Management,The National Police University for Criminal Justice,Baoding,Hebei 071000,China;College of Mathematics and Information Science,Hebei University,Baoding,Hebei 071002,China;College of Computer Science and Technology,China University of Petroleum,Qingdao,Shandong 266580,China)
Abstract:With the rapid growth of data with different modalities on the Internet,cross-modal retrieval has gradually become a hot research topic.Due to its efficiency and effectiveness,Hashing based methods have become one of the most popular large-scale cross-modal retrieval strategies.In most of the image-text cross-modal retrieval methods,the goal is to make the deep features of the images similar to the corresponding deep text features.However,these methods incorporate background information of the images into the feature learning,as a result,the retrieval performance is decreased.To solve this problem,OFBDH(object feature based deep Hashing)is proposed to learn optimal discriminative maximum activations of convolutions from the feature maps to represent the object features,and then the learned object features are integrated into the image-text cross-modal network learning.Experimental results show that OFBDH can obtain satisfactory cross-modal retrieval results on MIRFLICKR-25K,IAPR TC-12 and NUS-WIDE.
Keywords:object feature  cross-modal loss  network parameters learning  retrieval
本文献已被 维普 万方数据 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号