首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
针对Zhai和Shah提出的原始时空显著性检测模型在空间显著性方面仅仅使用了图像的亮度信息,忽略彩色图像中的色彩信息的不足,提出了一种基于HSV颜色模型的空间显著性计算方法。该方法充分利用图像中的亮度信息和彩色信息,从像素级和区域级两个层次上进行显著性的计算。将改进的空间显著性计算与Zhai和Shah提出的时间显著性计算以及时空融合框架进行整合,检测视频中的显著目标。实验证明改进方法在光照不均和背景较复杂的情况下获取的空间显著区域和显著目标比原始方法更准确。  相似文献   

2.
针对先前的立体图像显著性检测模型未充分考虑立体视觉舒适度和视差图分布特征对显著区域检测的影响,提出了一种结合立体视觉舒适度因子的显著性计算模型.该模型在彩色图像显著性提取中,首先利用SLIC算法对输入图像进行超像素分割,随后进行颜色相似区域合并后再进行二维图像显著性计算;在深度显著性计算中,首先对视差图进行预处理;然后基于区域对比度进行显著性计算;最后,结合立体视觉舒适度因子对二维显著图和深度显著图进行融合,得到立体图像显著图.在不同类型立体图像上的实验结果表明,该模型获得了85%的准确率和78%的召回率,优于现有常用的显著性检测模型,并与人眼立体视觉注意力机制保持良好的一致性.  相似文献   

3.
A biologically inspired object-based visual attention model is proposed in this paper. This model includes a training phase and an attention phase. In the training phase, all training targets are fused into a target class and all training backgrounds are fused into a background class. Weight vector is computed as the ratio of the mean target class saliency and the mean background class saliency for each feature. In the attention phase, for an attended scene, all feature maps are combined into a top-down salience map with the weight vector by a hierarchy method. Then, top-down and bottom-up salience map are fused into a global salience map which guides the visual attention. At last, the size of each salient region is obtained by maximizing entropy. The merit of our model is that it can attend a class target object which can appear in the corresponding background class. Experimental results indicate that: when the attended target object doesn’t always appear in the background corresponding to that in the training images, our proposed model is excellent to Navalpakkam’s model and the top-down approach of VOCUS.  相似文献   

4.
Bottom-up spatiotemporal visual attention model for video analysis   总被引:3,自引:0,他引:3  
The human visual system (HVS) has the ability to fixate quickly on the most informative (salient) regions of a scene and therefore reducing the inherent visual uncertainty. Computational visual attention (VA) schemes have been proposed to account for this important characteristic of the HVS. A video analysis framework based on a spatiotemporal VA model is presented. A novel scheme has been proposed for generating saliency in video sequences by taking into account both the spatial extent and dynamic evolution of regions. To achieve this goal, a common, image-oriented computational model of saliency-based visual attention is extended to handle spatiotemporal analysis of video in a volumetric framework. The main claim is that attention acts as an efficient preprocessing step to obtain a compact representation of the visual content in the form of salient events/objects. The model has been implemented, and qualitative as well as quantitative examples illustrating its performance are shown.  相似文献   

5.
针对视觉选择性注意模型化计算过程中不同特征在整合阶段的权值判定,提出一种基于特征图分布的权值估计方法,并在静态图像显著性区域提取中取得了令人满意的应用效果。首先提取原始图像的颜色、方向和强度特征图像,然后计算各个特征图的广义高斯分布参数与方差,进而给出一种特征图权值估计算法,最后通过对特征图的加权整合与归一化实现对原始图像的显著性区域提取。实验结果表明,通过此方法计算的权值对特征进行加权调制所提取的显著性区域的效果更加符合人眼的观测结果。  相似文献   

6.
A dynamic saliency attention model based on local complexity is proposed in this paper. Low-level visual features are extracted from current and some previous frames. Every feature map is resized into some different sizes. The feature maps in same size and same feature for all the frames are used to calculate a local complexity map. All the local complexity maps are normalized and are fused into a dynamic saliency map. In the same time, a static saliency map is acquired by the current frame. Then dynamic and static saliency maps are fused into a final saliency map. Experimental results indicate that: when there is noise among the frames or there is change of illumination among the frames, our model is excellent to Marat?s model and Shi?s model; when the moving objects do not belong to the static salient regions, our model is better than Ban?s model.  相似文献   

7.
目的 立体视频能提供身临其境的逼真感而越来越受到人们的喜爱,而视觉显著性检测可以自动预测、定位和挖掘重要视觉信息,可以帮助机器对海量多媒体信息进行有效筛选。为了提高立体视频中的显著区域检测性能,提出了一种融合双目多维感知特性的立体视频显著性检测模型。方法 从立体视频的空域、深度以及时域3个不同维度出发进行显著性计算。首先,基于图像的空间特征利用贝叶斯模型计算2D图像显著图;接着,根据双目感知特征获取立体视频图像的深度显著图;然后,利用Lucas-Kanade光流法计算帧间局部区域的运动特征,获取时域显著图;最后,将3种不同维度的显著图采用一种基于全局-区域差异度大小的融合方法进行相互融合,获得最终的立体视频显著区域分布模型。结果 在不同类型的立体视频序列中的实验结果表明,本文模型获得了80%的准确率和72%的召回率,且保持了相对较低的计算复杂度,优于现有的显著性检测模型。结论 本文的显著性检测模型能有效地获取立体视频中的显著区域,可应用于立体视频/图像编码、立体视频/图像质量评价等领域。  相似文献   

8.
There is need to detect regions of small defects in a large background, when product surface quality in line is inspected by machine vision systems. A computational model of visual attention was developed for solving the problem, inspired by the behavior and the neuronal architecture of human visual attention. Firstly, the global feature was extracted from input image by law’s rules, then local features were extracted and evaluated with an improved saliency map model of Itti. The local features were fused into a single topographical saliency map by a multi-feature fusion operator differenced from Itti model, in which the better feature has the higher weighting coefficient and more contribution to fusion of the feature’s images. Finally, the regions were “popped out” in the map. Experimental results show that the model can locate regions of interest and exclude the most background regions.  相似文献   

9.
Geospatial datasets from satellite observations and model simulations are becoming more accessible. These spatiotemporal datasets are relatively massive for visualization to support advanced analysis and decision making. A challenge to visualizing massive geospatial datasets is identifying critical spatial and temporal changes reflected in the data while maintaining high interactive rendering speed, even when data are accessed remotely. We propose a view-dependent spatiotemporal saliency-driven approach that facilitates the discovery of regions showing high levels of spatiotemporal variability and reduces the rendering intensity of interactive visualization. Our method is based on a novel definition of data saliency, a spatiotemporal tree structure to store visual saliency values, as well as a saliency-driven view-dependent level-of-detail (LOD) control. To demonstrate its applicability, we have implemented the approach with an open-source remote visualization package and conducted experiments with spatiotemporal datasets produced by a regional dust storm simulation model. The results show that the proposed method may not be outstanding in some specific situations, but it consistently performs very well across different settings according to different criteria.  相似文献   

10.
提出一种基于视觉注意机制的彩色图像分割方法。受生物学启发,该方法模仿人类自下而上的视觉选择性注意过程,提取图像的底层特征,构造相应的显著图。根据显著图,检测出图像中的显著区域;将显著区域和背景分离,即得到图像分割结果。在多幅自然图像上进行实验,结果表明,该方法能够取得与人类视觉系统一致的分割结果。  相似文献   

11.
提出一种基于视觉注意机制的交通路标检测方法,该方法在图像灰度变换的基础上,引入自下而上的视觉注意模型,提取图像的初级特征,构造相应的显著图,根据总显著图,检测出图像中的显著区域,定位路标在图像中的位置。对多幅实时路况图像的实验结果表明,该方法能够适应户外自然环境检测,具有较高的准确性和实时性。  相似文献   

12.
Visual attention tends to avoid locations where previous visual attention has once focused. This phenomenon is called inhibition of return (IOR), and is known as one of the important dynamic properties of visual attention. Recently, several studies have reported that IOR occurs not only on locations, but also on visual features. In this study, we propose a visual attention model that involves a featurebased IOR by extending a recent model of the “saliency map.” Our model is demonstrated by a computer simulation, and its neuronal basis is also discussed.  相似文献   

13.
视觉注意机制在大视场目标快速定位中的应用   总被引:1,自引:0,他引:1       下载免费PDF全文
视觉心理学研究表明人类在看一个场景时,往往会在很短时间内找到几个显著区,然后再细看显著区域的内容,这样可以使得人类可以快速分析复杂图像。算法首先模拟人类视觉系统特点,根据图像的底层信息如对比度、方向、亮度等提取图像中几个最需要关注的显著区域,然后按照显著性由强到弱的顺序分别在每个显著区域利用具有尺度旋转不变性的对数极坐标变换方法进行目标的匹配定位。该方法在没有牺牲定位准确度的前提下,大幅减小了运算复杂度。实验表明该算法定位速度快而且准确。  相似文献   

14.
Xiao  Feng  Liu  Baotong  Li  Runa 《Multimedia Tools and Applications》2020,79(21-22):14593-14607

In response to the problem that the primary visual features are difficult to effectively address pedestrian detection in complex scenes, we present a method to improve pedestrian detection using a visual attention mechanism with semantic computation. After determining a saliency map with a visual attention mechanism, we can calculate saliency maps for human skin and the human head-shoulders. Using a Laplacian pyramid, the static visual attention model is established to obtain a total saliency map and then complete pedestrian detection. Experimental results demonstrate that the proposed method achieves state-of-the-art performance on the INRIA dataset with 92.78% pedestrian detection accuracy at a very competitive time cost.

  相似文献   

15.
Visual saliency is an important research topic in the field of computer vision due to its numerous possible applications. It helps to focus on regions of interest instead of processing the whole image or video data. Detecting visual saliency in still images has been widely addressed in literature with several formulations. However, visual saliency detection in videos has attracted little attention, and is a more challenging task due to additional temporal information. A common approach for obtaining a spatio-temporal saliency map is to combine a static saliency map and a dynamic saliency map. In our work, we model the dynamic textures in a dynamic scene with local binary patterns to compute the dynamic saliency map, and we use color features to compute the static saliency map. Both saliency maps are computed using a bio-inspired mechanism of human visual system with a discriminant formulation known as center surround saliency, and are fused in a proper way. The proposed model has been extensively evaluated with diverse publicly available datasets which contain several videos of dynamic scenes, and comparison with state-of-the art methods shows that it achieves competitive results.  相似文献   

16.
针对运动目标跟踪问题,提出一种利用视觉显著性和粒子滤波的目标跟踪算法.借鉴人类视觉注意机制的研究成果,根据目标的颜色、亮度和运动等特征形成目标的视觉显著性特征,与目标的颜色分布模型一起作为目标的特征表示模型,利用粒子滤波进行目标跟踪.该算法能够克服利用单一颜色特征所带来的跟踪不稳定问题,并能有效解决由于目标形变、光照变化以及目标和背景颜色分布相似而产生的跟踪困难问题,具有较强的鲁棒性.在多个视频序列中进行实验,并给出相应的实验结果和分析.实验结果表明,该算法用于实现运动目标跟踪是正确有效的.  相似文献   

17.
A spatiotemporal saliency algorithm based on a center-surround framework is proposed. The algorithm is inspired by biological mechanisms of motion-based perceptual grouping and extends a discriminant formulation of center-surround saliency previously proposed for static imagery. Under this formulation, the saliency of a location is equated to the power of a predefined set of features to discriminate between the visual stimuli in a center and a surround window, centered at that location. The features are spatiotemporal video patches and are modeled as dynamic textures, to achieve a principled joint characterization of the spatial and temporal components of saliency. The combination of discriminant center-surround saliency with the modeling power of dynamic textures yields a robust, versatile, and fully unsupervised spatiotemporal saliency algorithm, applicable to scenes with highly dynamic backgrounds and moving cameras. The related problem of background subtraction is treated as the complement of saliency detection, by classifying nonsalient (with respect to appearance and motion dynamics) points in the visual field as background. The algorithm is tested for background subtraction on challenging sequences, and shown to substantially outperform various state-of-the-art techniques. Quantitatively, its average error rate is almost half that of the closest competitor.  相似文献   

18.
This paper presents a computational method of feature evaluation for modeling saliency in visual scenes. This is highly relevant in visual search studies since visual saliency is at the basis of visual attention deployment. Visual saliency can also become important in computer vision applications as it can be used to reduce the computational requirements by permitting processing only in those regions of the scenes containing relevant information. The method is based on Bayesian theory to describe the interaction between top-down and bottom-up information. Unlike other approaches, it evaluates and selects visual features before saliency estimation. This can reduce the complexity and, potentially, improve the accuracy of the saliency computation. To this end, we present an algorithm for feature evaluation and selection. A two-color conjunction search experiment has been applied to illustrate the theoretical framework of the proposed model. The practical value of the method is demonstrated with video segmentation of instruments in a laparoscopic cholecystectomy operation.  相似文献   

19.
目的 为研究多场景下的行人检测,提出一种视觉注意机制下基于语义特征的行人检测方法。方法 首先,在初级视觉特征基础上,结合行人肤色的语义特征,通过将自下而上的数据驱动型视觉注意与自上而下的任务驱动型视觉注意有机结合,建立空域静态视觉注意模型;然后,结合运动信息的语义特征,采用运动矢量熵值计算运动显著性,建立时域动态视觉注意模型;在此基础上,以特征权重融合的方式,构建时空域融合的视觉注意模型,由此得到视觉显著图,并通过视觉注意焦点的选择完成行人检测。结果 选用标准库和实拍视频,在Matlab R2012a平台上,进行实验验证。与其他视觉注意模型进行对比仿真,本文方法具有良好的行人检测效果,在实验视频上的行人检测正确率达93%。结论 本文方法在不同的场景下具有良好的鲁棒性能,能够用于提高现有视频监控系统的智能化性能。  相似文献   

20.
针对传统显著性模型在自然图像的显著性物体检测中存在的缺陷,提出了一种利用背景原型(background prototypes)进行对比的视觉关注模型,以实现显著性物体的检测与提取;传统显著性模型主要通过计算区域中心与四周区域差异性实现显著性检测,而自然场景中显著性区域和背景区域往往都存在较大差异,导致在复杂图像中难以获得理想检测效果;基于背景原型对比度的显著性物体检测方法在图像分割生成的超像素图基础上,选择距离图像中心较远的图像区域作为背景原型区域,通过计算图像中任意区域与这些背景原型区域的颜色对比度准确检测和提取图像中的显著性物体;实验结果表明,基于背景原型对比度的显著性模型可以更好地滤除杂乱背景,产生更稳定、准确的显著图,在准确率、召回率和F-measure等关键性能和直观视觉效果上均优于目前最先进的显著性模型,计算复杂度低,利于应用推广。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号