首页 | 本学科首页   官方微博 | 高级检索  
     

预训练语言模型实体匹配的可解释性
引用本文:梁峥,王宏志,戴加佳,邵心玥,丁小欧,穆添愉. 预训练语言模型实体匹配的可解释性[J]. 软件学报, 2023, 34(3): 1087-1108
作者姓名:梁峥  王宏志  戴加佳  邵心玥  丁小欧  穆添愉
作者单位:哈尔滨工业大学 计算学部, 黑龙江 哈尔滨 150001
基金项目:面向云边端协同的新型数据管理理论与方法研究(62232005)
摘    要:实体匹配可以判断两个数据集中的记录是否指向同一现实世界实体,对于大数据集成、社交网络分析、网络语义数据管理等任务不可或缺.作为在自然语言处理、计算机视觉中取得大量成功的深度学习技术,预训练语言模型在实体识别任务上也取得了优于传统方法的效果,引起了大量研究人员的关注.然而,基于预训练语言模型的实体匹配技术效果不稳定、匹配结果不可解释,给这一技术在大数据集成中的应用带来了很大的不确定性.同时,现有的实体匹配模型解释方法主要面向机器学习方法进行模型无关的解释,在预训练语言模型上的适用性存在缺陷.因此,以Ditto、JointBERT等BERT类实体匹配模型为例,提出3种面向预训练语言模型实体匹配技术的模型解释方法来解决这个问题:(1)针对序列化操作中关系数据属性序的敏感性,对于错分样本,利用数据集元特征和属性相似度实现属性序反事实生成;(2)作为传统属性重要性衡量的补充,通过预训练语言模型注意力机制权重来衡量并可视化模型处理数据时的关联性;(3)基于序列化后的句子向量,使用k近邻搜索技术召回与错分样本相似的可解释性优良的样本,增强低置信度的预训练语言模型预测结果.在真实公开数据集上的实验结果...

关 键 词:实体匹配  预训练语言模型  可解释性
收稿时间:2022-05-16
修稿时间:2022-07-29

Interpretability of Entity Matching Based on Pre-trained Language Model
LIANG Zheng,WANG Hong-Zhi,DAI Jia-Ji,SHAO Xin-Yue,DING Xiao-Ou,MU Tian-Yu. Interpretability of Entity Matching Based on Pre-trained Language Model[J]. Journal of Software, 2023, 34(3): 1087-1108
Authors:LIANG Zheng  WANG Hong-Zhi  DAI Jia-Ji  SHAO Xin-Yue  DING Xiao-Ou  MU Tian-Yu
Affiliation:School of Computer Science and Engineering, Harbin Institute of Technology, Harbin 150001, China
Abstract:Entity matching can determine whether records in two datasets point to the same real-world entity,and is indispensable for tasks such as big data integration,social network analysis,and web semantic data management.As a deep learning technology that has achieved a lot of success in natural language processing and computer vision,pre-trained language models have also achieved better results than traditional methods in entity recognition tasks,which have attracted the attention of a large number of researchers.However,the performance of entity matching based on pre-trained language model is unstable and the matching results cannot be explained,which brings great uncertainty to the application of this technology in big data integration.At the same time,the existing entity matching model interpretation methods are mainly oriented to machine learning methods as model-agnostic interpretation,and there are shortcomings in their applicability on pre-trained language models.Therefore,this paper takes BERT entity matching models such as Ditto and JointBERT as examples,and proposes three model interpretation methods for pre-training language model entity matching technology to solve this problem:(1) In the serialization operation,the order of relational data attributes is sensitive.We use dataset meta-features and attribute similarity to generate attribute ranking counterfactuals for misclassified samples;(2) As a supplement to traditional attribute importance measurement,the pre-trained language model attention weights are used to measure and visualize model processing;(3) Based on the serialized sentence vector,the k-nearest neighbor search technique is used to recall the samples with good interpretability similar to the misclassified samples to enhance the low-confidence prediction results of pre-trained language model.Experiments on real public datasets show that while improving the model effect through the enhancing method,our method can reach 68.8% of the upper limit of fidelity in the attribute order search space,which provides a decision explanation for the pre-trained language entity matching model.New perspectives such as attribute order counterfactual and attribute association understanding are also introduced.
Keywords:Entity Matching  Pre-trained Language Model  Interprebility
点击此处可从《软件学报》浏览原始摘要信息
点击此处可从《软件学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号