首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到19条相似文献,搜索用时 299 毫秒
1.
曾茜  韩华  马媛媛 《计算机工程》2022,48(10):95-102
在具有模体特征的食物链网络、社交网络中,局部朴素贝叶斯(LNB)的链路预测方法通过准确区分每个共邻节点的贡献以提高链路预测的精确度,但忽略了每个共邻节点对所在路径的贡献不同以及网络模体结构对链接形成的作用。针对LNB链路预测方法存在的局限性问题,结合路径模体特征与朴素贝叶斯理论,提出基于模体的朴素贝叶斯链路预测方法。定义模体密度以量化路径结构上模体的聚集程度。考虑路径结构上模体密度对链接形成的影响,构建每条路径的角色贡献函数,以量化每条路径结构的模体特征对节点相似性的影响。在此基础上,根据朴素贝叶斯理论与角色贡献函数推导节点相似性指标。在Football、USAir、C.elegans、FWMW、FWEW和FWFW 6个真实网络上进行实验,结果表明,该方法能有效提高预测性能且具有较优的鲁棒性,其中在具有显著模体特征的FWMW、FWEW、FWFW网络上,相比现有相似性指标中较优的Katz指标,所提相似性指标的AUC值提升了2%~7%。  相似文献   

2.
《计算机工程与科学》2017,(10):1825-1831
复杂网络包括生物性信息网络、科学家合作网络、社交关系网络等,研究复杂网络的关系预测问题有助于预测蛋白质相互关系,发现科学家合作关系,以及挖掘潜在好友关系等。目前,绝大多数关系预测算法由复杂网络的相似度模型实现,但该类型算法基于显式的网络拓扑特征构建,忽视了影响关系生成的隐含信息。针对这一问题,在朴素贝叶斯链接预测模型(LNB)基础上提出了一种加强(Enhanced)朴素贝叶斯链接预测模型(ELNB),该模型通过定义共邻节点关系概率对共邻节点构成的局部子图特征进行建模,有效缓解了LNB中的独立性假设,实现了共邻节点关系贡献的量化计算。在人工数据集和真实复杂网络数据集上的实验表明,本文提出的模型优于基准算法和其他新近提出的模型。同时,把ELNB的思想有效地拓展到其他基于共邻节点的相似度算法中,为该类模型的研究提供一种新的方案。  相似文献   

3.
反垃圾邮件技术已成为人们关注的一个焦点。基于贝叶斯理论的垃圾邮件过滤技术有着独特的优势,而其中的朴素贝叶斯模型具有算法简单、有效,易于实现等优点而成为最常用的模型。本文系统地介绍了朴素贝叶斯及其扩展模型的核心思想,并对朴素贝叶斯模型的发展作了大胆的预测,这对贝叶斯垃圾邮件过滤技术具有理论和现实的意义。  相似文献   

4.
丁超  赵海  司帅宗  朱剑 《计算机应用》2019,39(4):963-971
为了对正常衰老的人脑功能网络(NABFN)的拓扑结构变化进行探究,提出一种基于朴素贝叶斯的网络演化模型(NBM)。首先,依据朴素贝叶斯(NB)的链路预测算法与解剖距离来定义节点间存在连边的概率;其次,利用特定的网络演化算法,在青年人的脑功能网络基础上,通过不断地增加连边来逐步得到相应中年及老年时期的模拟网络;最后,为了对模拟网络与真实网络间的相似程度进行评价,提出网络相似指标(SI)值。仿真实验结果表明,与基于共同邻居的网络演化模型(CNM)相比,NBM构建的模拟网络与真实网络间的SI值(4.479 4,3.402 1)高于CNM模拟网络对应的SI值(4.100 4,3.013 2);并且,两者模拟网络的SI值均明显高于随机网络演化算法所得模拟网络的SI值(1.892 0,1.591 2)。实验结果证实NBM能够更为准确地预测出NABFN的拓扑结构变化过程。  相似文献   

5.
惠孛  吴跃  陈佳 《计算机科学》2006,33(5):110-112
使用朴素的贝叶斯(NB)分类模型对邮件进行分类,是目前基于内容的垃圾邮件过滤方法的研究热点。朴素的贝叶斯在参数之间联系不强的时候分类效果简单而有效。但是朴素的贝叶斯分类模型中对特征参数的条件独立假设无法表达参数之间在语义上的关系,影响分类性能。在朴素的贝叶斯分类模型的基础上,我们提出了一种双级贝叶斯分类模型(DLB,Double Level Bayes),既考虑到了参数之间的影响又保留了朴素的贝叶斯分类模型的优点。同时时DLB模型与朴素的贝叶斯分类模型的性能进行比较。仿真实验表明,DLB分类模型在垃圾邮件过滤应用中的效果在大部分条件下优于朴素的贝叶斯分类模型。  相似文献   

6.
针对连锁型零售企业直接营销中的问题,基于原型理论的挖掘模型选择方法,提出基于EM聚类朴素贝叶斯模型,通过实验证明了该模型在客户购买行为的预测性能上明显优于基于K-means聚类朴素贝叶斯模型和无聚类的朴素贝叶斯模型。最后,利用该模型检验了直接营销中的对新客户进行分类预测的有效性。  相似文献   

7.
惠孛  吴跃 《计算机应用》2009,29(3):903-904
由于朴素贝叶斯分类模型的简单高效,在垃圾邮件分类时可以达到较好的效果;但朴素贝叶斯的条件独立假设割裂了属性之间的关系,影响了分类的准确性。放松朴素贝叶斯分类模型关于属性之间条件独立假设,介绍一种新的基于不完全朴素贝叶斯分类模型的垃圾邮件分类模型,N平均1 依赖邮件过滤模型。使用N个1 依赖分类模型的平均概率作为分类的预测概率。实验证明,该模型在简单、高效的同时降低了对垃圾邮件分类的错误率。  相似文献   

8.
为了改善树增强朴素贝叶斯(TAN)的分类精度,对TAN结构进行了扩展,提出了一种利用可分解的评分函数构建树形贝叶斯网络分类模型的学习方法。在构建TAN网络时允许属性没有父结点。采用低阶CI测试初步剔除无效属性,再结合改进的BIC评分函数利用贪婪搜索获得每个属性结点的父结点,从而建立分类模型。对比朴素贝叶斯(NB)和TAN,提出的分类算法在分类准确率和AUC面积两个指标上表现更好,说明本文模型拥有比TAN更好的分类效果。  相似文献   

9.
基于特征加权的朴素贝叶斯分类器   总被引:13,自引:0,他引:13  
程克非  张聪 《计算机仿真》2006,23(10):92-94,150
朴素贝叶斯分类器是一种广泛使用的分类算法,其计算效率和分类效果均十分理想。但是,由于其基础假设“朴素贝叶斯假设”与现实存在一定的差异,因此在某些数据上可能导致较差的分类结果。现在存在多种方法试图通过放松朴素贝叶斯假设来增强贝叶斯分类器的分类效果,但是通常会导致计算代价大幅提高。该文利用特征加权技术来增强朴素贝叶斯分类器。特征加权参数直接从数据导出,可以看作是计算某个类别的后验概率时,某个属性对于该计算的影响程度。数值实验表明,特征加权朴素贝叶斯分类器(FWNB)的效果与其他的一些常用分类算法,例如树扩展朴素贝叶斯(TAN)和朴素贝叶斯树(NBTree)等的分类效果相当,其平均错误率都在17%左右;在计算速度上,FWNB接近于NB,比TAN和NBTree快至少一个数量级。  相似文献   

10.
11.
一种限定性的双层贝叶斯分类模型   总被引:29,自引:1,他引:28  
朴素贝叶斯分类模型是一种简单而有效的分类方法,但它的属性独立性假设使其无法表达属性变量间存在的依赖关系,影响了它的分类性能.通过分析贝叶斯分类模型的分类原则以及贝叶斯定理的变异形式,提出了一种基于贝叶斯定理的新的分类模型DLBAN(double-level Bayesian network augmented naive Bayes).该模型通过选择关键属性建立属性之间的依赖关系.将该分类方法与朴素贝叶斯分类器和TAN(tree augmented naive Bayes)分类器进行实验比较.实验结果表明,在大多数数据集上,DLBAN分类方法具有较高的分类正确率.  相似文献   

12.
Due to being fast, easy to implement and relatively effective, some state-of-the-art naive Bayes text classifiers with the strong assumption of conditional independence among attributes, such as multinomial naive Bayes, complement naive Bayes and the one-versus-all-but-one model, have received a great deal of attention from researchers in the domain of text classification. In this article, we revisit these naive Bayes text classifiers and empirically compare their classification performance on a large number of widely used text classification benchmark datasets. Then, we propose a locally weighted learning approach to these naive Bayes text classifiers. We call our new approach locally weighted naive Bayes text classifiers (LWNBTC). LWNBTC weakens the attribute conditional independence assumption made by these naive Bayes text classifiers by applying the locally weighted learning approach. The experimental results show that our locally weighted versions significantly outperform these state-of-the-art naive Bayes text classifiers in terms of classification accuracy.  相似文献   

13.
Technical Note: Naive Bayes for Regression   总被引:1,自引:0,他引:1  
Frank  Eibe  Trigg  Leonard  Holmes  Geoffrey  Witten  Ian H. 《Machine Learning》2000,41(1):5-25
Despite its simplicity, the naive Bayes learning scheme performs well on most classification tasks, and is often significantly more accurate than more sophisticated methods. Although the probability estimates that it produces can be inaccurate, it often assigns maximum probability to the correct class. This suggests that its good performance might be restricted to situations where the output is categorical. It is therefore interesting to see how it performs in domains where the predicted value is numeric, because in this case, predictions are more sensitive to inaccurate probability estimates.This paper shows how to apply the naive Bayes methodology to numeric prediction (i.e., regression) tasks by modeling the probability distribution of the target value with kernel density estimators, and compares it to linear regression, locally weighted linear regression, and a method that produces model trees—decision trees with linear regression functions at the leaves. Although we exhibit an artificial dataset for which naive Bayes is the method of choice, on real-world datasets it is almost uniformly worse than locally weighted linear regression and model trees. The comparison with linear regression depends on the error measure: for one measure naive Bayes performs similarly, while for another it is worse. We also show that standard naive Bayes applied to regression problems by discretizing the target value performs similarly badly. We then present empirical evidence that isolates naive Bayes' independence assumption as the culprit for its poor performance in the regression setting. These results indicate that the simplistic statistical assumption that naive Bayes makes is indeed more restrictive for regression than for classification.  相似文献   

14.
Feng Zeng  Lan Yao  Baoling Wu  Wenjia Li  Lin Meng 《Software》2020,50(11):2031-2045
Human contact prediction is a challenging task in mobile social networks. The existing prediction methods are based on the static network structure, and directly applying these static prediction methods to dynamic network prediction is bound to reduce the prediction accuracy. In this paper, we extract some important features to predict human contacts and propose a novel human contact prediction method based on naive Bayes algorithm, which is suitable for dynamic networks. The proposed method takes the ever-changing structure of mobile social networks into account. First, the past time is partitioned into many periods with equal intervals, and each period has a feature matrix of all node pairs. Then, with the feature matrixes used for classifiers training based on naive Bayes algorithm, we can get a classifier for each time period. At last, the different weights are assigned to the classifiers according to their importance to contact prediction, and all classifiers are weighted combination into the final prediction classifier. The extensive experiments are conducted to verify the effectiveness and superiority of the proposed method, and the results show that the proposed method can improve the prediction accuracy and TP Rate to a large extent. Besides, we find that the size of time interval has a certain impact on the clustering coefficient of mobile social networks, which further affects the prediction accuracy.  相似文献   

15.
Abstract: This research focused on investigating and benchmarking several high performance classifiers called J48, random forests, naive Bayes, KStar and artificial immune recognition systems for software fault prediction with limited fault data. We also studied a recent semi-supervised classification algorithm called YATSI (Yet Another Two Stage Idea) and each classifier has been used in the first stage of YATSI. YATSI is a meta algorithm which allows different classifiers to be applied in the first stage. Furthermore, we proposed a semi-supervised classification algorithm which applies the artificial immune systems paradigm. Experimental results showed that YATSI does not always improve the performance of naive Bayes when unlabelled data are used together with labelled data. According to experiments we performed, the naive Bayes algorithm is the best choice to build a semi-supervised fault prediction model for small data sets and YATSI may improve the performance of naive Bayes for large data sets. In addition, the YATSI algorithm improved the performance of all the classifiers except naive Bayes on all the data sets.  相似文献   

16.
操作风险数据积累比较困难,而且往往不完整,朴素贝叶斯分类器是目前进行小样本分类最优秀的分类器之一,适合于操作风险等级预测。在对具有完整数据朴素贝叶斯分类器学习和分类的基础上,提出了基于星形结构和Gibbs sampling的具有丢失数据朴素贝叶斯分类器学习方法,能够避免目前常用的处理丢失数据方法所带来的局部最优、信息丢失和冗余等方面的问题。  相似文献   

17.
朴素Bayes分类器是一种简单有效的机器学习工具.本文用朴素Bayes分类器的原理推导出"朴素Bayes组合"公式,并构造相应的分类器.经过测试,该分类器有较好的分类性能和实用性,克服了朴素Bayes分类器精确度差的缺点,并且比其他分类器更加快速而不会显著丧失精确度.  相似文献   

18.
Traditional classification algorithms require a large number of labelled examples from all the predefined classes, which is generally difficult and time-consuming to obtain. Furthermore, data uncertainty is prevalent in many real-world applications, such as sensor network, market analysis and medical diagnosis. In this article, we explore the issue of classification on uncertain data when only positive and unlabelled examples are available. We propose an algorithm to build naive Bayes classifier from positive and unlabelled examples with uncertainty. However, the algorithm requires the prior probability of positive class, and it is generally difficult for the user to provide this parameter in practice. Two approaches are proposed to avoid this user-specified parameter. One approach is to use a validation set to search for an appropriate value for this parameter, and the other is to estimate it directly. Our extensive experiments show that the two approaches can basically achieve satisfactory classification performance on uncertain data. In addition, our algorithm exploiting uncertainty in the dataset can potentially achieve better classification performance comparing to traditional naive Bayes which ignores uncertainty when handling uncertain data.  相似文献   

19.
根据RoughSet属性重要度理论,构建了基于互信息的属性子集重要度,提出属性相关性的加权朴素贝叶斯分类算法,该算法同时放宽了朴素贝叶斯算法属性独立性、属性重要性相同的假设。通过在UCI部分数据集上进行仿真实验,与基于属性相关性分析的贝叶斯(CB)和加权朴素贝叶斯(WNB)两种算法做比较,证明了该算法的有效性。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号