首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Knowledge extraction from Chinese wiki encyclopedias   总被引:1,自引:0,他引:1  
  相似文献   

2.
《Computer Physics Reports》1990,12(5):289-381
A knowldge-based project, the GRAPE system(Group Representation and Application in Physics Environment), is described in this paper. The GRAPE system is designed to provide physicists with a group theoretical environment to help them solve problems in group theory and representation. The user can communicate with GRAPE in plain English. At the present stage, it contains the knowledge of crystallography point groups, space groups as well as magnetic space groups both in group structure and group representations. The GRAPE system consists of five modules besides the knowledge base and the data base: a natural language interface, a computation module, a tutprial module, a bibliography module, and a program library. Group theoretical analysis for the Landau theory of continuous phase transitions has been the first application of the GRAPE system. The calculation for determining directions of phase transition at the Γ point for 230 space groups, 230 grey space groups and 674 black and white magnetic space groups has been performed.  相似文献   

3.
The idea of automatic summarization dates back to 1958, when Luhn invented the “auto abstract” (Luhn, 1958). Since then, many diverse automatic summarization approaches have been proposed, but no single technique has solved the increasingly urgent need for automatic summarization. Rather than proposing one more such technique, we suggest that the best solution is likely a system able to combine multiple summarization techniques, as required by the type of documents being summarized. Thus, this paper presents HAUSS: a framework to quickly build specialized summarizers, integrating several base techniques into a single approach. To recognize relevant text fragments, rules are created that combine frequency, centrality, citation and linguistic information in a context-dependent way. An incremental knowledge acquisition framework strongly supports the creation of these rules, using a training corpus to guide rule acquisition, and produce a powerful knowledge base specific to the domain. Using HAUSS, we created a knowledge base for catchphrase extraction in legal text. The system outperforms existing state-of-the-art general-purpose summarizers and machine learning approaches. Legal experts rated the extracted summaries similar to the original catchphrases given by the court. Our investigation of knowledge acquisition methods for summarization therefore demonstrates that it is possible to quickly create effective special-purpose summarizers, which combine multiple techniques, into a single context-aware approach.  相似文献   

4.
装配知识库模型及其核心设计   总被引:1,自引:0,他引:1  
余卫东  陆玉昌  张钹 《软件学报》1995,6(5):296-304
本文分析了研究和建立基于智能机器人装配系统的知识库的必要性,并探讨了实现途径.接着,我们提出了支持装配知识库的知识模型及其核心的体系结构,并讨论了支持该知识模型的建模工具EXPRESS语言.最后,给出结论.  相似文献   

5.
The prediction of the linguistic origin of surnames is a basic functionality required in the design of high-quality multilanguage speech synthesizers. The assignment of a given string representing a surname to a specific language is typically based on a set of rules which can hardly be written in an explicit form. The approach we propose faces this problem combining a rule-based system with a module based on evidential reasoning and a module based on neural networks. The resulting hybrid system combines the different sources of information, merging both knowledge from experts on linguistics and knowledge automatically acquired using learning from examples. The system has been validated on a large database containing surnames belonging to four different languages, showing its effectiveness for real-world applications.  相似文献   

6.
Building knowledge base management systems   总被引:1,自引:0,他引:1  
Advanced applications in fields such as CAD, software engineering, real-time process control, corporate repositories and digital libraries require the construction, efficient access and management of large, shared knowledge bases. Such knowledge bases cannot be built using existing tools such as expert system shells, because these do not scale up, nor can they be built in terms of existing database technology, because such technology does not support the rich representational structure and inference mechanisms required for knowledge-based systems. This paper proposes a generic architecture for a knowledge base management system intended for such applications. The architecture assumes an object-oriented knowledge representation language with an assertional sublanguage used to express constraints and rules. It also provides for general-purpose deductive inference and special-purpose temporal reasoning. Results reported in the paper address several knowledge base management issues. For storage management, a new method is proposed for generating a logical schema for a given knowledge base. Query processing algorithms are offered for semantic and physical query optimization, along with an enhanced cost model for query cost estimation. On concurrency control, the paper describes a novel concurrency control policy which takes advantage of knowledge base structure and is shown to outperform two-phase locking for highly structured knowledge bases and update-intensive transactions. Finally, algorithms for compilation and efficient processing of constraints and rules during knowledge base operations are described. The paper describes original results, including novel data structures and algorithms, as well as preliminary performance evaluation data. Based on these results, we conclude that knowledge base management systems which can accommodate large knowledge bases are feasible. Edited by Gunter Schlageter and H.-J. Schek. Received May 19, 1994 / Revised May 26, 1995 / Accepted September 18, 1995  相似文献   

7.
The Princeton WordNet® (PWN) is a widely used lexical knowledge database for semantic information processing. There are now many wordnets under creation for languages worldwide. In this paper, we endeavor to construct a wordnet for Pre-Qin ancient Chinese (PQAC), called PQAC WordNet (PQAC-WN), to process the semantic information of PQAC. In previous work, most recently constructed wordnets have been established either manually by experts or automatically using resources from which translation pairs between English and the target language can be extracted. The former method, however, is time-consuming, and the latter method, owing to a lack of language resources, cannot be performed on PQAC. As a result, a method based on word definitions in a monolingual dictionary is proposed. Specifically, for each sense, kernel words are first extracted from its definition, and the senses of each kernel word are then determined by graph-based Word Sense Disambiguation. Finally, one optimal sense is chosen from the kernel word senses to guide the mapping between the word sense and PWN synset. In this research, we obtain 66 % PQAC senses that can be shared with English and another 14 % language-specific senses that were added to PQAC-WN as new synsets. Overall, the automatic mapping achieves a precision of over 85 %.  相似文献   

8.
Technology intelligence systems are vital components for planning of technology development and formulation of technology strategies. Although such systems provide computation supports for technology analysis, much effort and intervention of experts, who may be expensive or unavailable, is required in gathering processes of information for analysis. As a remedy, this paper proposes TrendPerceptor, a system that uses a property-function based approach. The proposed system assists experts (1) to identify trends in invention concepts from patents, and (2) to perform evolution trend analysis of patents for technology forecasting. For this purpose, a module of the system uses grammatical analysis of textual information to automatically extract properties and functions, which show innovation directions in a given technology. Using the identified properties and functions, a module for invention concept analysis based on network analysis and a module for evolution trend analysis based on TRIZ (Russian acronym of the Theory of Inventive Problem Solving) trends are suggested. This paper describes the architecture of a system composed of these three modules, and illustrates two case studies using the system.  相似文献   

9.
Expert scheduling systems, which develop the schedule automatically on a real time basis, are able to respond to the changes of product demand in Flexible Manufacturing Systems (FMS). While developing an expert scheduling system, the most time-consuming and difficult step is knowledge acquisition, the process that elicits the knowledge from experts and transfers it into the knowledge base. A trace-driven knowledge acquisition (TDKA) method is proposed to extract the expertise from the schedules produced by expert schedulers. Three phases are involved in the TDKA process: data collection, data analysis, and rule evaluation. In data collection, the expert schedulers are identified and decisions made during the scheduling process are recorded as a trace. In data analysis, a set of scheduling rules is developed based on the trace. The rules are then evaluated in the last phase. If the resulting rules do not perform as well as the expert schedulers, the process returns to phase two and refines the rules. The whole process stops whenever the resulting rules perform at least as well as the expert schedulers. A circuit board production line is used to demonstrate the feasibility of the TDKA methodology. The scheduling rules perform much better than the expert schedulers from whom the rules are extracted.  相似文献   

10.
Detection of fake news has spurred widespread interests in areas such as healthcare and Internet societies, in order to prevent propagating misleading information for commercial and political purposes. However, efforts to study a general framework for exploiting knowledge, for judging the trustworthiness of given news based on their content, have been limited. Indeed, the existing works rarely consider incorporating knowledge graphs (KGs), which could provide rich structured knowledge for better language understanding.In this work, we propose a deep triple network (DTN) that leverages knowledge graphs to facilitate fake news detection with triple-enhanced explanations. In the DTN, background knowledge graphs, such as open knowledge graphs and extracted graphs from news bases, are applied for both low-level and high-level feature extraction to classify the input news article and provide explanations for the classification.The performance of the proposed method is evaluated by demonstrating abundant convincing comparative experiments. Obtained results show that DTN outperforms conventional fake news detection methods from different aspects, including the provision of factual evidence supporting the decision of fake news detection.  相似文献   

11.
文章利用面向对象的编程技术,采用Delphi5.0语言,针对由不同数据库系统构成的知识库系统,实现了不同知识库系统的无缝连接,较好地解决了不同知识库系统的接口问题,为KDD中不一致知识的的通用处理奠定了坚实的基础。  相似文献   

12.
Expert system applications in the biomedical domain have long been hampered by the difficulty inherent in maintaining and extending large knowledge bases. We have developed a knowledge-based method for automatically augmenting such knowledge bases. The method consists of automatically integrating data contained in commercially available, external, online databases with data contained in an expert system's knowledge base. We have built a prototype system, named DBX, using this technique to augment an expert system's knowledge base as a decision support aid and as a bibliographic retrieval tool. In this paper, we describe this prototype system in detail, illustrate its use, and discuss the lessons we have learned in its implementation.  相似文献   

13.
The most fascinating advantage of the semantic web would be its capability of understanding and processing the contents of web pages automatically. Basically, the semantic web realization involves two main tasks: (1) Representation and management of a large amount of data and metadata for web contents; (2) Information extraction and annotation on web pages. On the one hand, recognition of named-entities is regarded as a basic and important problem to be solved, before deeper semantics of a web page could be extracted. On the other hand, semantic web information extraction is a language-dependent problem, which requires particular natural language processing techniques. This paper introduces VN-KIM IE, the information extraction module of the semantic web system VN-KIM that we have developed. The function of VN-KIM IE is to automatically recognize named-entities in Vietnamese web pages, by identifying their classes, and addresses if existing, in the knowledge base of discourse. That information is then annotated to those web pages, providing a basis for NE-based searching on them, as compared to the current keyword-based one. The design, implementation, and performance of VN-KIM IE are presented and discussed.  相似文献   

14.
支持智能搜索的自扩展知识库模型的研究和设计*   总被引:1,自引:1,他引:0  
利用自然语言处理和理解技术,提出并实现了一种可以对网页中的中文信息进行处理,获取并存储知识,具有自我扩展特性和支持中文智能搜索功能的知识库系统模型。该知识库模型将语义Web技术与智能搜索技术结合,支持自然语言的搜索请求,采用OWL本体描述语言来表达知识,支持知识的应用与推理,具有一定的实用和研究价值。  相似文献   

15.

This work introduces a novel approach to extract meaningful content information from video by collaborative integration of image understanding and natural language processing. We developed a person browser system that associates faces and overlaid name texts in videos. This approach takes news videos as a knowledge source, then automatically extracts face and assoicated name text as content information. The proposed framework consists of the text detection module, the face detection module, and the person indexing database module. The successful results of person extraction reveal that the proposed methodology of integrated use of image understanding techniques and natural language processing technique is headed in the right direction to achieve our goal of accessing real content of multimedia information.

  相似文献   

16.
A sememe is defined as the minimum semantic unit of languages in linguistics. Sememe knowledge bases are built by manually annotating sememes for words and phrases. HowNet is the most well-known sememe knowledge base. It has been extensively utilized in many natural language processing tasks in the era of statistical natural language processing and proven to be effective and helpful to understanding and using languages. In the era of deep learning, although data are thought to be of vital importance, there are some studies working on incorporating sememe knowledge bases like HowNet into neural network models to enhance system performance. Some successful attempts have been made in the tasks including word representation learning, language modeling, semantic composition, etc. In addition, considering the high cost of manual annotation and update for sememe knowledge bases, some work has tried to use machine learning methods to automatically predict sememes for words and phrases to expand sememe knowledge bases. Besides, some studies try to extend HowNet to other languages by automatically predicting sememes for words and phrases in a new language. In this paper, we summarize recent studies on application and expansion of sememe knowledge bases and point out some future directions of research on sememes.  相似文献   

17.
介绍了一个面向应用领域的知识发现系统开发平台KDIST。将数据挖掘技术巧妙地封装在应用领域的问题中,使开发出的知识发现系统操作傻瓜化,用户无须关心数据挖掘本身,有效地减轻了领域用户使用负担,提高了数据挖掘技术实用性。所开发出的知识发现系统将挖掘得到的知识融合到已有专家系统的知识库中。两个实例系统的应用证明知识发现系统是专家系统自动半自动知识获取和知识库精化的良好工具。  相似文献   

18.
With the advancement of scientific and engineering research, a huge number of academic literature are accumulated. Manually reviewing the existing literature is the main way to explore embedded knowledge, and the process is quite time-consuming and labor intensive. As the quantity of literature is increasing exponentially, it would be more difficult to cover all aspects of the literature using the traditional manual review approach. To overcome this drawback, bibliometric analysis is used to analyze the current situation and trend of a specific research field. In the bibliometric analysis, only a few key phrases (e.g., authors, publishers, journals, and citations) are usually used as the inputs for analysis. Information other than those phrases is not extracted for analysis, while that neglected information (e.g., abstract) might provide more detailed knowledge in the article. To tackle with this problem, this study proposed an automatic literature knowledge graph and reasoning network modeling framework based on ontology and Natural Language Processing (NLP), to facilitate the efficient knowledge exploration from literature abstract. In this framework, a representation ontology is proposed to characterize the literature abstract data into four knowledge elements (background, objectives, solutions, and findings), and NLP technology is used to extract the ontology instances from the abstract automatically. Based on the representation ontology, a four-space integrated knowledge graph is built using NLP technology. Then, reasoning network is generated according to the reasoning mechanism defined in the proposed ontology model. To validate the proposed framework, a case study is conducted to analyze the literature in the field of construction management. The case study proves that the proposed ontology model can be used to represent the knowledge embedded in the literatures’ abstracts, and the ontology elements can be automatically extracted by NLP models. The proposed framework can be an enhancement for the bibliometric analysis to explore more knowledge from the literature.  相似文献   

19.
We present a tool that combines two main trends of knowledge base refinement. The first is the construction of interactive knowledge acquisition tools and the second is the development of machine learning methods that automate this procedure. The tool presented here is interactive and gives experts the ability to evaluate an expert system and provide their own diagnoses on specific problems, when the expert system behaves erroneously. We also present a database scheme that supports the collection of specific instances. The second aspect of the tool is that knowledge base refinement and machine learning methods can be applied to the database, in order to automate the procedure refining the knowledge base. In this paper we examine the application of inductive learning algorithms within the proposed framework. Our main goal is to encourage the experts to evaluate expert systems and to introduce new knowledge, based on their experience.  相似文献   

20.
基于伪自然语言理解的CAI开发平台   总被引:1,自引:0,他引:1  
基于伪自然语言理解,提出并实现了一个种高效率的知识获取方法,并把它用一诉开发中。首先知识工程师利用来自然语言的BL语言 写书本自然描述,然后利用知识编译系统处理BL程序以高效率地实现书本知识获取,再后领域专家在书本知识库的基本语义呆引导下利用知识求精系统对书本知识库加以少许求精,接着对领域知识库动态全局规划,把领域知识分解成一个个概念,最后通过方法生成组织成一个个课文传授予学生。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号