首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 93 毫秒
1.
This paper proposes a new system for low frequency adaptive image watermarking based on the statistical data from psychological experiments on human image perception. The new approach can lead to a reduction of degrading the subjective image quality that often occurs when watermark is embedded into low frequency area. In order to reduce the degrading of image quality, the new approach determines the strength of watermark according to local image characteristics such as brightness and contrast. By conducting a behavioral experiment on human image fidelity based on the psycho-visual image association technique, we were able to infer the relationship between the watermark strength and the different levels of image brightness and contrast information. The exact watermark is extracted according to edge characteristics by adopting a so-called edge mask that exploits the coefficients of subbands in the subsampled discrete wavelet transform images. Thus, our new approach does not require original images for watermark. We also show the new approach is practically validated against some standard images.  相似文献   

2.
针对当前各种图像清晰度评价方法在清晰度判别过程中单调性和区分度不够以及适用范围较小的问题,提出了一种基于四元数小波变换(QWT)幅值与相位的图像清晰度评价方法。该算法通过四元数小波变换将图像从空间域变换到频率域,对得到的四元数小波变换系数进一步计算之后获得低频子带与高频子带的幅值与相位信息,求得低频子带幅值各方向的梯度之后与对应方向的相位相乘求和,最终得到两个清晰度指标值。采用该算法与多种现有算法对不同内容的图像、不同程度模糊的图像以及含有不同程度噪声的图像进行清晰度评价实验:相对于现有算法,所提算法在对上述多种图像的清晰度评价中都保持着很好的单调性与区分度。实验结果表明,所提算法不但克服了现有算法在单调性与区分度上的不足,而且所提清晰度评价指标可以应用在图像处理中。  相似文献   

3.
Image morphing is often used in film and television industry to create synthetic visual effects by smooth transformation of one object into another. Based upon spatial representation of images, several image morphing techniques have been proposed. Simple spatial techniques, for example cross-dissolve, suffer from lack of smooth transformation while better quality techniques, like mesh warping or field warping, have significant computational complexity. In this paper we present a simple but good quality image morphing technique based upon frequency domain representation of images. Transformation from a source image to a target image takes place by mixing low frequencies of the source image and high frequencies of the target image in varying proportions. The proposed technique has been applied to a wide variety of images. The resulting sequence of images are better in visual quality and faster in execution time.  相似文献   

4.
多曝光图像融合技术是将一组场景相同但曝光程度不同的图像序列直接融合成为一幅含有更多场景细节信息的高质量图像。针对现有算法局部对比度差和色彩失真的问题,结合Retinex理论模型提出了一种新的多曝光图像融合算法。首先,基于Retinex理论模型,利用光照估计算法将曝光序列图像分为入射光分量序列和反射光分量序列,然后分别采用不同的融合方法对这两组序列进行处理。对于入射光分量,要保证场景的全局亮度的变化特性并且削弱过曝光和欠曝光区域的影响;而对于反射光分量,要采用适度曝光的评价参数来更好地保留场景的色彩及细节信息。分别从主观和客观两方面对所提算法进行了分析。实验结果表明,同传统基于图像域合成的算法相比,该算法在结构相似度(SSIM)上平均提升了1.7%,另外在图像色彩和局部细节上的处理效果更好。  相似文献   

5.
目的 现有大多数低照度图像增强算法会放大噪声,且用于极低照度图像时会出现亮度提升不足、色彩失真等问题。为此,提出一种基于Retinex(retina cortex)的增强与去噪方法。方法 为了增强极低照度图像,首先利用暗通道先验原理估计场景的全局光照,若光照低于0.5,对图像进行初始光照校正;其次,提出一种Retinex顺序分解模型,使低照度图像中的噪声均体现在反射分量中,基于分解结果,利用Gamma校正求取增强后的噪声图像;最后,提出一种基于内外双重互补先验约束的去噪机制,利用非局部自相似性原理为反射分量构建内部先验约束,基于深度学习,为增强后的噪声图像构建外部先验约束,使内外约束相互制约。结果 将本文算法与6种算法比较,在140幅普通低照度图像和162幅极低照度图像上(有正常曝光参考图像)进行主观视觉和客观指标评价比较,结果显示本文方法在亮度提升、色彩保真及去噪方面均有明显优势,对于普通低照度图像,BTMQI(blind tone-mapped quality index)和NIQE(natural image quality evaluator)指标均取得次优值,对于极低照度图像...  相似文献   

6.
In this paper, we show how to harness both low-rank and sparse structures in regular or near-regular textures for image completion. Our method is based on a unified formulation for both random and contiguous corruption. In addition to the low rank property of texture, the algorithm also uses the sparse assumption of the natural image: because the natural image is piecewise smooth, it is sparse in certain transformed domain (such as Fourier or wavelet transform). We combine low-rank and sparsity properties of the texture image together in the proposed algorithm. Our algorithm based on convex optimization can automatically and correctly repair the global structure of a corrupted texture, even without precise information about the regions to be completed. This algorithm integrates texture rectification and repairing into one optimization problem. Through extensive simulations, we show our method can complete and repair textures corrupted by errors with both random and contiguous supports better than existing low-rank matrix recovery methods. Our method demonstrates significant advantage over local patch based texture synthesis techniques in dealing with large corruption, non-uniform texture, and large perspective deformation.  相似文献   

7.
针对夜间图像光线暗、特征不易提取的问题,提出一种基于全局和局部特征的无参考夜间图像质量评价方法.首先,利用等高线原理将图像分为亮区域和暗区域2部分,将亮区域占整幅图的比例作为特征1;其次,提取夜间图像的全局亮度信息并将其作为特征2;再次,结合微分算子法求得图像的边缘图作为特征3;然后,将夜间图像从RGB转换到HSI,提...  相似文献   

8.
This paper introduces a new class of reconstruction algorithms that are fundamentally different from traditional approaches. We deviate from the standard practice that treats images as point samples. In this work, image values are treated as area samples generated by nonoverlapping integrators. This is consistent with the image formation process, particularly for CCD and CID cameras. We show that superior results are obtained by formulating reconstruction as a two-stage process: image restoration followed by application of the point spread function (PSF) of the imaging sensor. By coupling the PSF to the reconstruction process, we satisfy a more intuitive fidelity measure of accuracy that is based on the physical limitations of the sensor. Efficient local techniques for image restoration are derived to invert the effects of the PSF and estimate the underlying image that passed through the sensor. The reconstruction algorithms derived herein are local methods that compare favorably to cubic convolution, a well-known local technique, and they even rival global algorithms such as interpolating cubic splines. Evaluations are made by comparing their passband and stopband performances in the frequency domain, as well as by direct inspection of the resulting images in the spatial domain. A secondary advantage of the algorithms derived with this approach is that they satisfy an imaging-consistency property. This means that they exactly reconstruct the image for some function in the given class of functions. Their error can be shown to be at most twice that of the "optimal" algorithm for a wide range of optimality constraints.  相似文献   

9.
We introduce a new technique for nonrigid image registration based on the composition of local deformations. The warping model is analyzed in order to guarantee continuity, differentiability and a one-to-one transformation by constraining the parameters of the nonlinear spatial transformation. A genetic algorithm solves the model by global optimization, handling constraints, and maximizing the normalized mutual information. The composition of local transformations goes throughout several levels of resolution, from coarse to fine. The performance of our technique was tested in synthetic and real medical images. The proposed method was always able to improve the similarity criterion between image pairs, demonstrating the robustness of the method for several modalities of images.  相似文献   

10.
《Information Fusion》2007,8(2):193-207
Comparative evaluation of fused images is a critical step to evaluate the relative performance of different image fusion algorithms. Human visual inspection is often used to assess the quality of fused images. In this paper, we propose some variants of a new image quality metric based on the human vision system (HVS). The proposed measures evaluate the quality of a fused image by comparing its visual differences with the source images and require no knowledge of the ground truth. First, the images are divided into different local regions. These regional images are then transformed to the frequency domain. Second, the difference between the local regional images in frequency domain is weighted with a human contrast sensitivity function (CSF). The quality of a local regional image is obtained by computing the MSE of the weighted difference images obtained from the fused regional image and source regional images. Finally, the quality of a fused image is the weighted summation of the local regional images quality measures. Our experimental results show that these metrics are consistent with perceptually obtained results.  相似文献   

11.
Cephalometric images usually have low contrast. The existing techniques for automatic cephalometric analysis usually use histogram equalization for image enhancement. This technique has the advantage of being fully automatic and nonlinear. However, it suffers from spikes, excessive enhancement, and lack of brightness preservation. The proposed technique is an adaptive histogram equalization technique that uses wavelet based gradient histograms. This paper compares its performance with two traditional techniques, three histogram modification based techniques, and two wavelet based techniques. Forty digital and scanned cephalograms are used to conduct tests. In addition to visual histograms and intensity profiles, the proposed method is compared in terms of eight quantitative measures. The various measures are applied to analyze the results in terms of contrast enhancement (EME, CNR), brightness preservation (AMBE), edge conservation and enhancement (H, TEN), preservation of image structures and non-addition distortion (MSSIM, SVD-M). The proposed method gives good contrast enhancement, with better brightness preservation without losing edge information and with the minimum addition of distortions to the enhanced cephalometric images.  相似文献   

12.
生成对抗网络(generative adversarial network,GAN)快速发展,并在图像生成和图像编辑技术等多个方面取得成功应用。然而,若将上述技术用于伪造身份或制作虚假新闻,则会造成严重的安全隐患。多媒体取证领域的研究者面向GAN生成图像已提出了多种被动取证与反取证方法,但现阶段缺乏相关系统性综述。针对上述问题,本文首先阐述本领域的研究背景和研究意义,然后分析自然图像采集与GAN图像生成过程的区别。根据上述理论基础,详细介绍了现有GAN生成图像的被动取证技术,包括:GAN生成图像检测算法,GAN模型溯源算法和其他相关取证问题。此外,针对不同应用场景介绍基于GAN的反取证技术。最后,通过实验分析当前GAN生成图像被动取证技术所面临的挑战。本文根据对现有技术从理论和实验两方面的分析得到以下结论:现阶段,GAN生成图像的被动取证技术已在空间域和频率域形成了不同技术路线,较好地解决了简单场景下的相关取证问题。针对常见取证痕迹,基于GAN的反取证技术已能够进行有效隐藏。然而,该领域研究仍存在诸多局限:1)取证与反取证技术的可解释性不足;2)取证技术鲁棒性和泛化性较弱;3)反取证技术缺乏多特征域协同的抗分析能力等。上述问题和挑战还需要研究人员继续深入探索。  相似文献   

13.
We discuss a new deblurring problems in this paper. Focus measurements play a fundamental role in image processing techniques. Most traditional methods neglect spatial information in the frequency domain. Therefore, this study analyzed image data in the frequency domain to determine the value of spatial information. but instead misleading noise reduction results . We found that the local feature is not always a guide for noise reduction. This finding leads to a new method to measure the image edges in focus deblurring. We employed an all-in-focus measure in the frequency domain, based on the energy level of frequency components. We also used a multi-circle enhancement model to analyze this spatial information to provide a more accurate method for measuring images. We compared our results with those using other methods in similar studies. Findings demonstrate the effectiveness of our new method.  相似文献   

14.
15.
This paper makes use of finger vein to propose an accurate personal authentication system. A matched filtering at various scales is applied on the vein image to reduce the effects of noise occurred due to non-uniform illumination, low local contrast, hairs and skin texture. Vein image is also enhanced by using global characteristics. Local and global characteristics of enhanced vein images are fused to obtain accurate vein pattern. Extracted veins are used for matching such that the problem of geometric deformations can be handled efficiently. For performance evaluation, a publicly available database of 3132 unconstrained finger vein images acquired from 156 subjects is used. Performance of the proposed system is found to be better than the various existing systems.  相似文献   

16.
Approximation techniques are an important aspect of digital signal and image processing. Many lossy signal compression procedures such as the Fourier transform and discrete cosine transform are based on the idea that a signal can be represented by a small number of transformed coefficients which are an approximation of the original. Existing approximation techniques approach this problem in either a time/spatial domain or transform domain, but not both. This paper briefly reviews various existing approximation techniques. Subsequently, we present a new strategy to obtain an approximation fˆ(x) of f(x) in such a way that it is reasonably close to the original function in the domain of the variable x, and exactly preserves some properties of the transformed domain. In this particular case, the properties of the transformed values that are preserved are geometric moments of the original function. The proposed technique has been applied to single-variable functions, two-dimensional planar curves, and two-dimensional images, and the results obtained are demonstrative  相似文献   

17.
一种频率域解混叠的CCD图像超分辨重建方法*   总被引:1,自引:0,他引:1  
在不增大光学系统的焦距,也不缩小CCD 器件阵元的宽度的情况下,为了获得高分辨率的图像,对多幅具有互补信息的低分辨率图像进行研究。首先,把多幅具有互补信息的低分辨率图像通过傅里叶变换到频率域;接着,对频率域中高分辨与低分辨率图像之间关系进行建模,进而得到它们之间的关系;然后,通过对模型求解,得到高分辨率图像的频率值;最后,进行傅里叶反变换即可得到高分辨率图像。实验结果表明,该重建方法获得的高分辨率影像在视觉上比用任意一帧影像经过双线性插值获得的影像要清晰得多,与实际摄取高一倍的分辨率影像几乎达到一样。对b  相似文献   

18.
Multi-focus image fusion has emerged as a major topic in image processing to generate all-focus images with increased depth-of-field from multi-focus photographs. Different approaches have been used in spatial or transform domain for this purpose. But most of them are subject to one or more of image fusion quality degradations such as blocking artifacts, ringing effects, artificial edges, halo artifacts, contrast decrease, sharpness reduction, and misalignment of decision map with object boundaries. In this paper we present a novel multi-focus image fusion method in spatial domain that utilizes a dictionary which is learned from local patches of source images. Sparse representation of relative sharpness measure over this trained dictionary are pooled together to get the corresponding pooled features. Correlation of the pooled features with sparse representations of input images produces a pixel level score for decision map of fusion. Final regularized decision map is obtained using Markov Random Field (MRF) optimization. We also gathered a new color multi-focus image dataset which has more variety than traditional multi-focus image sets. Experimental results demonstrate that our proposed method outperforms existing state-of-the-art methods, in terms of visual and quantitative evaluations.  相似文献   

19.
Image matching is an important area of research in the field of artificial intelligence, machine vision and visual navigation. A new image matching scheme in which grey scale images are quantised to form sub-band binary images is presented. The information in the binary images is then signaturised and the signatures are sorted as per significance. These sorted signatures are then normalised to transform the represented image pictorial features in the form of a hyper-dimensional vector cluster. For the image matching, the two clusters from both the images are compared in the transformed domain. This comparison yields efficient results directly in the image spatial domain avoiding the need of image inverse transformation for the interpretation of results. As compared with the conventional techniques, this comparison avoids the wide range of square error calculations all over the image. It also directly guides the solution in an iterative fashion to converge towards the true match point. The process of signaturisation is based on image local features and is moulded in a way to support the scale and rotation-invariant template matching as well. A four-dimensional solution population scheme has also been presented with an associated matching confidence factor. This factor helps in terminating the iterations when the essential matching conditions have been achieved. The proposed scheme gives robust and fast results for normal, scaled and rotated templates. Speed comparison with older techniques shows the computational viability of this new technique and its much lesser dependence on image size. The method also shows noise immunity at 30 dB additive white Gaussian noise and impulsive noise.  相似文献   

20.
Calibration based attack is one of the most important steganalytic attacks in recent past specifically for JPEG domain steganography. In calibration attack, the attacker generally predicts the cover image statistics from the stego image. Preventing attackers from such prediction is used to resist these attacks. Domain separation (or randomization) is such a technique which is used for hiding the embedding domain from the attacker. It is observed that existing domain randomization techniques cannot provide enough randomization such that they are easily be detected by recent steganalysis techniques. In this paper, we have extended our previous work based on spatial desynchronization using statistical analysis. It is also experimentally shown that proposed algorithm is less detectable against the calibration based blind as well as targeted steganalytic attacks than the existing JPEG domain steganographic schemes.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号