首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到19条相似文献,搜索用时 171 毫秒
1.
本文提出的方法基于深度卷积神经网络和哈希算法,该方法对VGG16网络模型进行了改进:模型中全连接层保留FC6和FC7,去掉的FC8用哈希层替换,构建哈希函数获得哈希编码,在损失层对损失函数做了优化计算,微调模型初始化参数。图像检索过程主要包括模型训练和检索图像两个阶段,实验数据集采用CIFAR-10和NUS-WIDE。和其他几种典型的传统哈希算法和深度哈希算法进行对比分析,实验结果表明,本文所提方法能有效提高图像检索性能。  相似文献   

2.
随着图像数据的迅猛增长,当前主流的图像检索方法采用的视觉特征编码步骤固定,缺少学习能力,导致其图像表达能力不强,而且视觉特征维数较高,严重制约了其图像检索性能。针对这些问题,该文提出一种基于深度卷积神径网络学习二进制哈希编码的方法,用于大规模的图像检索。该文的基本思想是在深度学习框架中增加一个哈希层,同时学习图像特征和哈希函数,且哈希函数满足独立性和量化误差最小的约束。首先,利用卷积神经网络强大的学习能力挖掘训练图像的内在隐含关系,提取图像深层特征,增强图像特征的区分性和表达能力。然后,将图像特征输入到哈希层,学习哈希函数使得哈希层输出的二进制哈希码分类误差和量化误差最小,且满足独立性约束。最后,给定输入图像通过该框架的哈希层得到相应的哈希码,从而可以在低维汉明空间中完成对大规模图像数据的有效检索。在3个常用数据集上的实验结果表明,利用所提方法得到哈希码,其图像检索性能优于当前主流方法。  相似文献   

3.
为从海量的图像资源中既准确又快速地检索出目 标图像,在传统的 图像检索模型中,图像的特征通常是从固定尺度的图像上提取出的,这将不可避免地降低整 个系统 实际应用能力。为解决这一问题,本文引入分层稀疏编码模型, 提出一种基于分层匹配追踪(HMP)的快速图像检索技术,实现多尺度情况下的图像检索。本 文方法从图像中提取的低层稀疏编码特征传递 到高层,并将提取的高维稀疏编码特征转换为改进后的PCAH特征,利用哈希特征的汉明距 离度量, 实现图像的快速检索。在公共数据集Caltech256和Corel5K上的实验 结果可以看出,本文方法的查 准率和查全率较其他哈希法分别提高了5%和10%以上,而且所用时间 也最短,表明本文方法不仅具有较高的准确率,还能保持较高的时间效率。  相似文献   

4.
当前主流图像检索技术所采用的传统视觉特征编码缺少足够的学习能力,影响学习得到的特征表达能力。此外,由于视觉特征维数高,会消耗大量的内存,因此降低了图像检索的性能。文中基于深度卷积神经网络与改进的哈希算法,提出并设计了一种端到端训练方式的图像检索方法。该方法将卷积神经网络提取的高层特征和哈希函数相结合,学习到具有足够表达能力的哈希特征,从而在低维汉明空间中完成对图像数据的大规模检索。在两个常用数据集上的实验结果表明,所提出的哈希图像检索方法的检索性能优于当前的一些主流方法。  相似文献   

5.
为了应对手工视觉特征与哈希编码过程不能最佳地兼容以及现有哈希方法无法区分图像语义信息的问题,提出一种基于深度卷积神经网络学习二进制哈希编码的方法.该方法基本思想是在深度残差网络中增加一个哈希层,同时学习图像特征和哈希函数;以此同时提出一种更加紧凑的分级哈希结构,用来提取更加接近图像语义的特征.经MNIST、CIFAR-10、NUS-WIDE数据集的实验,结果表明该方法优于现有的哈希方法.该方法不仅统一了特征学习和哈希编码的过程,同时深层残差网络也能得到更接近图像语义的特征,进而提高了检索准确度.  相似文献   

6.
基于卷积神经网络和监督核哈希的图像检索方法   总被引:1,自引:0,他引:1       下载免费PDF全文
当前主流的图像检索方法采用的视觉特征,缺乏自主学习能力,导致其图像表达能力不强,此外,传统的特征索引方法检索效率较低,难以适用于大规模图像数据.针对这些问题,本文提出了一种基于卷积神经网络和监督核哈希的图像检索方法.首先,利用卷积神经网络的学习能力挖掘训练图像内容的内在隐含关系,提取图像深层特征,增强特征的视觉表达能力和区分性;然后,利用监督核哈希方法对高维图像深层特征进行监督学习,并将高维特征映射到低维汉明空间中,生成紧致的哈希码;最后,在低维汉明空间中完成对大规模图像数据的有效检索.在ImageNet-1000和Caltech-256数据集上的实验结果表明,本文方法能够有效地增强图像特征的表达能力,提高图像检索效率,优于当前主流方法.  相似文献   

7.
论文针对视觉词袋(BOVW)模型放弃图像空间结构的缺点,提出一种基于Hesse稀疏编码的图像检索算法。首先,建立n-words模型,获得图像局部特征表示。n-words模型由一系列连续视觉词获得,是图像特征的一种高级描述。该文从n=1到n=5进行试验,寻找最恰当的n值;其次,将二阶Hesse能量函数融入标准稀疏编码的目标函数,得到Hesse稀疏编码公式;最后,以获得的n-words序列作为编码特征,利用特征符号搜索算法求解最优Hesse系数,计算相似度,返回检索结果。实验在两类数据集上进行,与BOVW模型和已有的算法相比,新算法极大地提高了图像检索的准确率。  相似文献   

8.
由于最近邻查询算法一般需要较高时间和空间代价,往往不能满足大数据查询的需要.哈希技术可以大幅度减少查询时间和存储空间,其主要思想是将原始空间中的高维数据映射成为一组编码,且满足保相似性原则.现有的大部分哈希方法一般认为哈希编码的各维度权重相同.然而在实际情况中,不同的维度往往携带有不同的信息.为此,本文提出了新的算法,为编码的每个维度分配权重,并提出了对应的量化编码方式.理论证明了算法的可行性,在真实数据集下与其他哈希算法对比实验也验证了该算法的有效性.  相似文献   

9.
基于内容的图像检索的关键在于对图像进行特征提取和对特征进行多比特量化编码 。近年来,基于内容的图像检索使用低级可视化特征对图像进行描述,存在“语义鸿沟”问题;其次,传统量化编码使用随机生成的投影矩阵,该矩阵与特征数据无关,因此不能保证量化的精确度。针对目前存在的这些问题,本文结合深度学习思想与迭代量化思想,提出基于卷积神经网络VGG16和迭代量化(Iterative Quantization, ITQ)的图像检索方法。使用在公开数据集上预训练VGG16网络模型,提取基于深度学习的图像特征;使用ITQ方法对哈希哈函数进行训练,不断逼近特征与设定比特数的哈希码之间的量化误差最小值,实现量化误差的最小化;最后使用获得的哈希码进行图像检索。本文使用查全率、查准率和平均精度均值作为检索效果的评价指标,在Caltech256图像库上进行测试。实验结果表明,本文提出的算法在检索优于其他主流图像检索算法。   相似文献   

10.
由于人脸图像数据的维数都较高,将稀疏表示分类用于人脸识别时计算量很大,为了提高人脸识别系统的效率,提出了一种融合半监督降维和稀疏表示的人脸识别方法。首先利用半监督降维算法对图像进行降维处理,在较低的维数空间快速取得较高的识别率,然后利用稀疏表示分类进行人脸识别,取得比传统的最近邻分类器更高的识别率,最后在ORL人脸库上进行实验验证。结果表明,利用该融合算法可快速有效地提高人脸图像的识别效果。  相似文献   

11.
Hashing is one of the popular solutions for approximate nearest neighbor search because of its low storage cost and fast retrieval speed, and many machine learning algorithms are adapted to learn effective hash function. As hash codes of the same cluster are similar to each other while the hash codes in different clusters are dissimilar, we propose an unsupervised discriminative hashing learning method (UDH) to improve discrimination among hash codes in different clusters. UDH shares a similar objective function with spectral hashing algorithm, and uses a modified graph Laplacian matrix to exploit local discriminant information. In addition, UDH is designed to enable efficient out-of-sample extension. Experiments on real world image datasets demonstrate the effectiveness of our novel approach for image retrieval.  相似文献   

12.
To overcome the barrier of storage and computation, the hashing technique has been widely used for nearest neighbor search in multimedia retrieval applications recently. Particularly, cross-modal retrieval that searches across different modalities becomes an active but challenging problem. Although numerous of cross-modal hashing algorithms are proposed to yield compact binary codes, exhaustive search is impractical for large-scale datasets, and Hamming distance computation suffers inaccurate results. In this paper, we propose a novel search method that utilizes a probability-based index scheme over binary hash codes in cross-modal retrieval. The proposed indexing scheme employs a few binary bits from the hash code as the index code. We construct an inverted index table based on the index codes, and train a neural network for ranking and indexing to improve the retrieval accuracy. Experiments are performed on two benchmark datasets for retrieval across image and text modalities, where hash codes are generated and compared with several state-of-the-art cross-modal hashing methods. Results show the proposed method effectively boosts the performance on search accuracy, computation cost, and memory consumption in these datasets and hashing methods. The source code is available on https://github.com/msarawut/HCI.  相似文献   

13.
Techniques for fast image retrieval over large databases have attracted considerable attention due to the rapid growth of web images. One promising way to accelerate image search is to use hashing technologies, which represent images by compact binary codewords. In this way, the similarity between images can be efficiently measured in terms of the Hamming distance between their corresponding binary codes. Although plenty of methods on generating hash codes have been proposed in recent years, there are still two key points that needed to be improved: 1) how to precisely preserve the similarity structure of the original data and 2) how to obtain the hash codes of the previously unseen data. In this paper, we propose our spline regression hashing method, in which both the local and global data similarity structures are exploited. To better capture the local manifold structure, we introduce splines developed in Sobolev space to find the local data mapping function. Furthermore, our framework simultaneously learns the hash codes of the training data and the hash function for the unseen data, which solves the out-of-sample problem. Extensive experiments conducted on real image datasets consisting of over one million images show that our proposed method outperforms the state-of-the-art techniques.  相似文献   

14.
There exist few studies investigating the multi-query image retrieval problem. Existing methods are not based on hash codes. As a result, they are not efficient and fast. In this study, we develop an efficient and fast multi-query image retrieval method when the queries are related to more than one semantic. Image hash codes are generated by a deep hashing method. Consequently, the method requires lower storage space, and it is faster compared to the existing methods. The retrieval is based on the Pareto front method. Reranking performed on the retrieved images by using non-binary deep-convolutional features increase retrieval accuracy considerably. Unlike previous studies, the method supports an arbitrary number of queries. It outperforms similar multi-query image retrieval studies in terms of retrieval time and retrieval accuracy.  相似文献   

15.
Perceptual hashing is conventionally used for content identification and authentication. It has applications in database content search, watermarking and image retrieval. Most countermeasures proposed in the literature generally focus on the feature extraction stage to get robust features to authenticate the image, but few studies address the perceptual hashing security achieved by a cryptographic module. When a cryptographic module is employed [1], additional information must be sent to adjust the quantization step. In the perceptual hashing field, we believe that a perceptual hashing system must be robust, secure and generate a final perceptual hash of fixed length. This kind of system should send only the final perceptual hash to the receiver via a secure channel without sending any additional information that would increase the storage space cost and decrease the security. For all of these reasons, in this paper, we propose a theoretical analysis of full perceptual hashing systems that use a quantization module followed by a crypto-compression module. The proposed theoretical analysis is based on a study of the behavior of the extracted features in response to content-preserving/content-changing manipulations that are modeled by Gaussian noise. We then introduce a proposed perceptual hashing scheme based on this theoretical analysis. Finally, several experiments are conducted to validate our approach, by applying Gaussian noise, JPEG compression and low-pass filtering.  相似文献   

16.
提出一种采用经典-量子ε-universal哈希类的簇态量子模糊哈希构造方法.传统哈希与模糊哈希算法不能有效抵抗量子攻击.通过采用diamond范数方法,构建了一种哈希函数类最优子集并且提供信息论意义上的更优安全性.基于量子簇态独特的物理级单向计算属性,相应算法更接近于物理可实现.进一步,构造了一种在信息安全与生物特征识别方面的隐蔽信息搜索策略.该生物识别搜索算法基于簇态量子ε-universal模糊哈希构建.该策略能有效抵抗量子算法攻击,确保数据存储安全,并降低了计算复杂度.相比于其他类似策略,此算法具有更精简的结构,理论分析表明此算法具有较高的识别效率与更好的数据安全性.  相似文献   

17.
基于相似图像的肺结节CT图像检索辅助诊断对肺结节的发现有着重要的作用。肺结节的诊断难度较大,通常需要充分利用图像的边缘、分叶、毛刺、纹理等各类信息。文中针对目前基于哈希方法的肺结节检索中存在的不能充分利用图像分割信息从而导致部分信息丢失问题做出了改进,提出了一种基于图像分割的肺结节图像哈希检索方法。实验结果表明,在72位哈希码长度时,达到了85.3%的平均准确率。并且,将文中图像分割模块应用于其他哈希检索方法时,平均准确率皆有一定的提升。  相似文献   

18.
Video hashing is a useful technique of many multimedia systems, such as video copy detection, video authentication, tampering localization, video retrieval, and anti-privacy search. In this paper, we propose a novel video hashing with secondary frames and invariant moments. An important contribution is the secondary frame construction with 3D discrete wavelet transform, which can reach initial data compression and robustness against noise and compression. In addition, since invariant moments are robust and discriminative features, hash generation based on invariant moments extracted from secondary frames can ensure good classification of the proposed video hashing. Extensive experiments on 8300 videos are conducted to validate efficiency of the proposed video hashing. The results show that the proposed video hashing can resist many digital operations and has good discrimination. Performance comparisons with some state-of-the-art algorithms illustrate that the proposed video hashing outperforms the compared algorithms in classification in terms of receiver operating characteristic results.  相似文献   

19.
为了有效地实现图像Hash函数在图像认证检索中的应用,提出了结合Harris角点检测和非负矩阵分解(NMF)的图像Hash算法,首先提取图像中的角点,对角点周围图像块信息进行非负矩阵分解得到表征图像局部特征的系数矩阵,进一步量化编码产生图像Hash。实验结果表明,得到的图像Hash对视觉可接受的操作如图像缩放、高斯低通滤波和JPEG压缩具有良好的稳健性,同时能区分出对图像大幅度扰动或修改的操作。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号