首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 62 毫秒
1.
该文提出一种基于原型理论的极化SAR图像表达方法。该方法首先利用原型理论构建原型集,然后以正则化逻辑回归函数计算测试样本与每个原型集的相似度,最后通过集成投影获得图像的特征表达。在极化SAR数据上的非监督分类实验结果表明,该方法能够准确表达图像中各类地物的极化特性,达到较好的分类效果。   相似文献   

2.
胡正平  刘立真 《信号处理》2018,34(4):448-456
针对样本集距离分类算法忽略样本集内部变化的不足,利用图像多重描述提供的互补信息,提出图像集闭包建模的协同表示人脸识别算法。首先,扩展具有多重描述能力的图像集,图像的中等强度像素携带鉴别信息利用原始图像生成中等像素图像,镜像图像可增添图像细节信息利用原始图像产生镜像图像,将此两种源域图像与原始图像联合构成扩展的图像集。然后,以无参建模构建扩展的图像集为字典闭包,同类异源域的测试图像构成图像集且构建为测试闭包,借鉴协同表示思想利用字典学习迭代求解闭包系数。最后,采用残差判别函数进行模式分类。本文方法不仅构建具有多重描述能力的图像集,而且充分利用样本集内部关联性从而获得较好的分类结果。本文分别在ORL、GT(Georgia Tech Face Database)、CMU PIE人脸数据库上进行实验。   相似文献   

3.
4.
Recent studies have shown that sparse representation (SR) can deal well with many computer vision problems, and its kernel version has powerful classification capability. In this paper, we address the application of a cooperative SR in semi-supervised image annotation which can increase the amount of labeled images for further use in training image classifiers. Given a set of labeled (training) images and a set of unlabeled (test) images, the usual SR method, which we call forward SR, is used to represent each unlabeled image with several labeled ones, and then to annotate the unlabeled image according to the annotations of these labeled ones. However, to the best of our knowledge, the SR method in an opposite direction, that we call backward SR to represent each labeled image with several unlabeled images and then to annotate any unlabeled image according to the annotations of the labeled images which the unlabeled image is selected by the backward SR to represent, has not been addressed so far. In this paper, we explore how much the backward SR can contribute to image annotation, and be complementary to the forward SR. The co-training, which has been proved to be a semi-supervised method improving each other only if two classifiers are relatively independent, is then adopted to testify this complementary nature between two SRs in opposite directions. Finally, the co-training of two SRs in kernel space builds a cooperative kernel sparse representation (Co-KSR) method for image annotation. Experimental results and analyses show that two KSRs in opposite directions are complementary, and Co-KSR improves considerably over either of them with an image annotation performance better than other state-of-the-art semi-supervised classifiers such as transductive support vector machine, local and global consistency, and Gaussian fields and harmonic functions. Comparative experiments with a nonsparse solution are also performed to show that the sparsity plays an important role in the cooperation of image representations in two opposite directions. This paper extends the application of SR in image annotation and retrieval.  相似文献   

5.
场景识别是计算机视觉研究中的一项基本任务.与图像分类不同,场景识别需要综合考虑场景的背景信息、局部场景特征以及物体特征等因素,导致经典卷积神经网络在场景识别上性能欠佳.为解决此问题,文中提出了一种基于深度卷积特征的场景全局与局部表示方法.此方法对场景图片的卷积特征进行变换从而为每张图片生成一个综合的特征表示.使用CAM...  相似文献   

6.
In the tensor representation, a two-dimensional (2-D) image is represented uniquely by a set of one-dimensional (1-D) signals, so-called splitting-signals, that carry the spectral information of the image at frequency-points of specific sets that cover the whole domain of frequencies. The image enhancement is thus reduced to processing splitting-signals and such process requires a modification of only a few spectral components of the image, for each signal. For instance, the alpha-rooting method of image enhancement can be fulfilled through processing separately a maximum of 3N/2 splitting-signals of an image (N x N), where N is a power of two. In this paper, we propose a fast implementation of the a-rooting method by using one splitting-signal of the tensor representation with respect to the discrete Fourier transform (DFT). The implementation is described in the frequency and spatial domains. As a result, the proposed algorithms for image enhancement use two 1-D N-point DFTs instead of two 2-D N x N-point DFTs in the traditional method of alpha-rooting.  相似文献   

7.
In this paper, we address the problem of classifying image sets for face recognition, where each set contains images belonging to the same subject and typically covering large variations. By modeling each image set as a manifold, we formulate the problem as the computation of the distance between two manifolds, called manifold-manifold distance (MMD). Since an image set can come in three pattern levels, point, subspace, and manifold, we systematically study the distance among the three levels and formulate them in a general multilevel MMD framework. Specifically, we express a manifold by a collection of local linear models, each depicted by a subspace. MMD is then converted to integrate the distances between pairs of subspaces from one of the involved manifolds. We theoretically and experimentally study several configurations of the ingredients of MMD. The proposed method is applied to the task of face recognition with image sets, where identification is achieved by seeking the minimum MMD from the probe to the gallery of image sets. Our experiments demonstrate that, as a general set similarity measure, MMD consistently outperforms other competing nondiscriminative methods and is also promisingly comparable to the state-of-the-art discriminative methods.  相似文献   

8.
Sparse representation is a new approach that has received significant attention for image classification and recognition. This paper presents a PCA-based dictionary building for sparse representation and classification of universal facial expressions. In our method, expressive facials images of each subject are subtracted from a neutral facial image of the same subject. Then the PCA is applied to these difference images to model the variations within each class of facial expressions. The learned principal components are used as the atoms of the dictionary. In the classification step, a given test image is sparsely represented as a linear combination of the principal components of six basic facial expressions. Our extensive experiments on several publicly available face datasets (CK+, MMI, and Bosphorus datasets) show that our framework outperforms the recognition rate of the state-of-the-art techniques by about 6%. This approach is promising and can further be applied to visual object recognition.  相似文献   

9.
10.
In the tensor representation, a two-dimensional (2-D) image is represented uniquely by a set of one-dimensional (1-D) signals, so-called splitting-signals, that carry the spectral information of the image at frequency-points of specific sets that cover the whole domain of frequencies. The image enhancement is thus reduced to processing splitting-signals and such process requires a modification of only a few spectral components of the image, for each signal. For instance, the$alpha$-rooting method of image enhancement can be fulfilled through processing separately a maximum of$3N/2$splitting-signals of an image$(Ntimes N)$, where$N$is a power of two. In this paper, we propose a fast implementation of the$alpha$-rooting method by using one splitting-signal of the tensor representation with respect to the discrete Fourier transform (DFT). The implementation is described in the frequency and spatial domains. As a result, the proposed algorithms for image enhancement use two 1-D$N$-point DFTs instead of two 2-D$Ntimes N$-point DFTs in the traditional method of$alpha$-rooting.  相似文献   

11.
We present a new method for compact representation of large image datasets. Our method is based on treating small patches from a 2-D image as matrices as opposed to the conventional vectorial representation, and encoding these patches as sparse projections onto a set of exemplar orthonormal bases, which are learned a priori from a training set. The end result is a low-error, highly compact image/patch representation that has significant theoretical merits and compares favorably with existing techniques (including JPEG) on experiments involving the compression of ORL and Yale face databases, as well as a database of miscellaneous natural images. In the context of learning multiple orthonormal bases, we show the easy tunability of our method to efficiently represent patches of different complexities. Furthermore, we show that our method is extensible in a theoretically sound manner to higher-order matrices (“tensors”). We demonstrate applications of this theory to compression of well-known color image datasets such as the GaTech and CMU-PIE face databases and show performance competitive with JPEG. Lastly, we also analyze the effect of image noise on the performance of our compression schemes.   相似文献   

12.
通过互联网易获得同一对象的多个无约束的观测样本,针对如何解决无约束观测样本带来的识别困难及充分利用多观测样本数据信息提高其分类性能问题,提出基于低秩分解的联合动态稀疏表示多观测样本分类算法.该算法首先寻找到一组最佳的图像变换域,使得变换图像可以分解成一个低秩矩阵和一个相关的稀疏误差矩阵;然后对低秩矩阵和稀疏误差矩阵分别进行联合动态稀疏表示,以便充分利用类级的相关性和原子级的差异性,即使多观测样本的稀疏表示向量在类级别上分享相同的稀疏模型,而在原子级上采用不同的稀疏模型;最后利用总的稀疏重建误差进行类别判决.在CMU-PIE人脸数据库、ETH-80物体识别数据库、USPS手写体数字数据库和UMIST人脸数据库上进行对比实验,实验结果表明本方法的优越性.  相似文献   

13.
In this paper, a group-sensitive multiple kernel learning (GS-MKL) method is proposed for object recognition to accommodate the intraclass diversity and the interclass correlation. By introducing the "group" between the object category and individual images as an intermediate representation, GS-MKL attempts to learn group-sensitive multikernel combinations together with the associated classifier. For each object category, the image corpus from the same category is partitioned into groups. Images with similar appearance are partitioned into the same group, which corresponds to the subcategory of the object category. Accordingly, intraclass diversity can be represented by the set of groups from the same category but with diverse appearances; interclass correlation can be represented by the correlation between groups from different categories. GS-MKL provides a tractable solution to adapt multikernel combination to local data distribution and to seek a tradeoff between capturing the diversity and keeping the invariance for each object category. Different from the simple hybrid grouping strategy that solves sample grouping and GS-MKL training independently, two sample grouping strategies are proposed to integrate sample grouping and GS-MKL training. The first one is a looping hybrid grouping method, where a global kernel clustering method and GS-MKL interact with each other by sharing group-sensitive multikernel combination. The second one is a dynamic divisive grouping method, where a hierarchical kernel-based grouping process interacts with GS-MKL. Experimental results show that performance of GS-MKL does not significantly vary with different grouping strategies, but the looping hybrid grouping method produces slightly better results. On four challenging data sets, our proposed method has achieved encouraging performance comparable to the state-of-the-art and outperformed several existing MKL methods.  相似文献   

14.
杨晓玲  冯山  袁钟 《电子学报》2020,48(5):937-945
针对分布复杂且离群类型多样的数据集进行离群检测困难的问题,提出基于相对距离的反k近邻树离群检测方法RKNMOD(Reversed K-Nearest Neighborhood).首先,将经典欧氏距离、对象局部密度和对象邻域结合,定义了对象的相对距离,能同时有效检出全局和局部离群点.其次,以最小生成树结构为基础,采取最大边切割法以快速分割离群点和离群簇.最后,人工合成数据集和UCI数据集试验均表明,新算法的检测准确率更高,为分布异常且离群类型多样的数据集的离群检测提供了一条有效的新途径.  相似文献   

15.
朱继洪  裴继红  赵阳 《信号处理》2019,35(4):640-648
本文提出了一种基于样本图像局部模式聚类的卷积核初始化方法,该方法可用于卷积神经网络(Convolutional neural network, CNN)训练中卷积核的初始化。在卷积神经网络中,卷积核的主要作用可看成是利用匹配滤波提取图像中的局部模式,并将其作为后续图像目标识别的特征。为此本文在图像训练集中选取一部分典型的样本图像,在这些图像中抽取与卷积核相同大小的子图作为图像局部模式矢量集合。首先对局部模式子图集合应用拓扑特性进行粗分类,然后对粗分类后的每一子类采用势函数聚类的方法获取样本图像中的典型局部模式子图,构成候选子图模式集,用它们作为CNN的初始卷积核进行训练。实验结果表明,本文方法可以明显加速CNN网络训练初期的收敛速度,同时对最终训练后的网络识别精度也有一定程度的提高。   相似文献   

16.
李宁  王军敏  司文杰  耿则勋 《红外与激光工程》2021,50(12):20210233-1-20210233-7
针对合成孔径雷达(Synthetic aperture radar,SAR)目标分类问题,提出基于最大熵准则的多视角方法。采用经典的图像相似度测度构建不同视角SAR图像之间的相关性矩阵,在此基础上分别计算不同视角组合条件下的非线性相关信息熵值。非线性相关信息熵值可分析多个变量之间的统计特性,熵值的大小即可反映不同变量之间的内在关联。根据最大熵的原则选择最优的视角子集,其中SAR图像具有最大的内在相关性。分类过程以联合稀疏表示为基础,对具有最大熵值的多个视角进行联合表示。联合稀疏表示模型同时处理若干稀疏表示问题,在它们具有关联的条件下具有提升重构精度的优势。根据不同视角求解得到的表示系数,按照类别分别计算对于选取多视角的重构误差,并根据误差最小的准则进行最终决策。文中方法可有效对多视角SAR图像样本进行相关性分析,并利用联合稀疏表示利用这种相关性,能够更好提高分类精度。采用MSTAR数据集对方法进行分析测试,通过与几类其他方法在多种测试条件下进行对比,结果显示了最大熵准则在多视角选取中的有效性和文中方法对SAR目标分类性能的优越性。  相似文献   

17.
We address the problem of visual classification with multiple features and/or multiple instances. Motivated by the recent success of multitask joint covariate selection, we formulate this problem as a multitask joint sparse representation model to combine the strength of multiple features and/or instances for recognition. A joint sparsity-inducing norm is utilized to enforce class-level joint sparsity patterns among the multiple representation vectors. The proposed model can be efficiently optimized by a proximal gradient method. Furthermore, we extend our method to the setup where features are described in kernel matrices. We then investigate into two applications of our method to visual classification: 1) fusing multiple kernel features for object categorization and 2) robust face recognition in video with an ensemble of query images. Extensive experiments on challenging real-world data sets demonstrate that the proposed method is competitive to the state-of-the-art methods in respective applications.  相似文献   

18.
Distance sets for shape filters and shape recognition   总被引:2,自引:0,他引:2  
  相似文献   

19.
基于稀疏表示及光谱信息的高光谱遥感图像分类   总被引:11,自引:1,他引:10  
该文结合稀疏表示及光谱信息提出了一种新的高光谱遥感图像分类算法。首先提出利用高光谱遥感图像数据集构造学习字典,然后根据学习字典计算每个像元的稀疏系数,从而获得像元的稀疏表示特征,最后根据稀疏表示特征和光谱信息分别构造随机森林,通过投票机制得到最终的分类结果。在AVIRIS高光谱遥感图像上的实验结果表明:该文所提方法能够提高分类效果,且其分类总精度和Kappa系数要高于光谱信息和稀疏表示特征方法。  相似文献   

20.
Transferring visual prior for online object tracking   总被引:1,自引:0,他引:1  
Visual prior from generic real-world images can be learned and transferred for representing objects in a scene. Motivated by this, we propose an algorithm that transfers visual prior learned offline for online object tracking. From a collection of real-world images, we learn an overcomplete dictionary to represent visual prior. The prior knowledge of objects is generic, and the training image set does not necessarily contain any observation of the target object. During the tracking process, the learned visual prior is transferred to construct an object representation by sparse coding and multiscale max pooling. With this representation, a linear classifier is learned online to distinguish the target from the background and to account for the target and background appearance variations over time. Tracking is then carried out within a Bayesian inference framework, in which the learned classifier is used to construct the observation model and a particle filter is used to estimate the tracking result sequentially. Experiments on a variety of challenging sequences with comparisons to several state-of-the-art methods demonstrate that more robust object tracking can be achieved by transferring visual prior.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号