首页 | 本学科首页   官方微博 | 高级检索  
     


A semantically enhanced text retrieval framework with abstractive summarization
Authors:Min Pan  Teng Li  Yu Liu  Quanli Pei  Ellen Anne Huang  Jimmy X. Huang
Affiliation:1. School of Computer and Information Engineering, Hubei Normal University, Huangshi, China

Information Retrieval and Knowledge Management Research Lab, School of Information Technology, York University, Toronto, Canada;2. School of Computer and Information Engineering, Hubei Normal University, Huangshi, China;3. Information Retrieval and Knowledge Management Research Lab, School of Information Technology, York University, Toronto, Canada;4. Department of Computer Science, Western University, London, Canada

Abstract:Recently, large pretrained language models (PLMs) have led a revolution in the information retrieval community. In most PLMs-based retrieval frameworks, the ranking performance broadly depends on the model structure and the semantic complexity of the input text. Sequence-to-sequence generative models for question answering or text generation have proven to be competitive, so we wonder whether these models can improve ranking effectiveness by enhancing input semantics. This article introduces SE-BERT, a semantically enhanced bidirectional encoder representation from transformers (BERT) based ranking framework that captures more semantic information by modifying the input text. SE-BERT utilizes a pretrained generative language model to summarize both sides of the candidate passage and concatenate them into a new input sequence, allowing BERT to acquire more semantic information within the constraints of the input sequence's length. Experimental results from two Text Retrieval Conference datasets demonstrate that our approach's effectiveness increasing as the length of the input text increases.
Keywords:BERT  document re-ranking  passage-level relevance  pretrained language models  semantic search
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号