共查询到20条相似文献,搜索用时 15 毫秒
1.
Dominique Pastre 《Annals of Mathematics and Artificial Intelligence》1993,8(3-4):425-447
TheMuscadet theorem prover is a knowledge-based system able to prove theorems in some non-trivial mathematical domains. The knowledge bases contain some general deduction strategies based onnatural deduction, mathematical knowledge and metaknowledge. Metarules build new rules, easily usable by the inference engine, from formal definitions. Mathematical knowledge may be general or specific to some particular field.Muscadet proved many theorems in set theory, mappings, relations, topology, geometry, and topological linear spaces. Some of the theorems were rather difficult.Muscadet is now intended to become an assistant for mathematicians in discrete geometry for cellular automata. In order to evaluate the difficulty of such a work, researchers were observed while proving some lemmas, andMuscadet was tested on easy ones. New methods have to be added to the knowledge base, such as reasoning by induction, but also new heuristics for splitting and reasoning by cases. It is also necessary to find good representations for some mathematical objects. 相似文献
2.
Marc Kirschenbaum Leon Sterling Ashish Jain 《Annals of Mathematics and Artificial Intelligence》1993,8(3-4):229-245
This paper presents a mathematical theory underlying a systematic method for constructingProlog programs calledstepwise enhancement. Stepwise enhancement dictates building a program starting with askeleton program which constitutes the basic control flow for the problem to be solved, and adding extra computations to the skeleton program by using well-understood programming techniques. Each extra computation can be developed independently, and the separate enhancements combined to produce the final program. While intuition and motivation have focused onProlog, the methods are applicable to logic programming languages more generally. The central concept in our mathematical theory for stepwise enhancement is that of a program map between two logic programs. Our definition of a program map from an enhancement to its skeleton guarantees the lifting of computations, the essence of the enhancement methodology. In this paper, we give definitions of program map and extensions, show that the definitions preserve the property of computations lifting, give examples of extensions and programming techniques which generate them, and point to directions for future work. 相似文献
3.
Model checking is a useful method to verify automatically the correctness of a system with respect to a desired behavior, by checking whether
a mathematical model of the system satisfies a formal specification of this behavior. Many systems of interest are open, in
the sense that their behavior depends on the interaction with their environment. The model checking problem for finite-state
open systems (called module checking) has been intensively studied in the literature. In this paper, we focus on open pushdown systems and we study the related model-checking problem (pushdown module checking, for short) with respect to properties expressed by CTL and CTL
* formulas. We show that pushdown module checking against CTL (resp., CTL
*) is 2Exptime-complete (resp., 3Exptime-complete). Moreover, we prove that for a fixed CTL or CTL
* formula, the problem is Exptime-complete. 相似文献
4.
Hypothesis-formation problems occur when the outcome of an experiment as predicted by a scientific theory does not match the outcome observed by a scientist. The problem is to modify the theory, and/or the scientist's conception of the intial conditions of the experiment, such that the prediction agrees with the observation. I treat hypothesis formation as adesign problem. A program calledHypgene designs hypotheses by reasoning backward from its goal of eliminating the difference between prediction and observation. This prediction error is eliminated bydesign operators that are applied by a planning system. The synthetic, goal-directed application of these operators should prove more efficient than past generate-and-test approaches to hypothesis generation.Hypgene uses heuristic search to guide a generator that is focused on the errors in a prediction. The advantages of the design approach to hypothesis formation over the generate-and-test approach are analogous to the advantages of dependency-directed backtracking over chronological backtracking. These hypothesis-formation methods were developed in the context of a historical study of a scientific research program in molecular biology. This article describes in detail the results of applying theHypgene program to several hypothesis-formation problems identified in this historical study.Hypgene found most of the same solutions as did the biologists, which demonstrates that it is capable of solving complex, real-world hypothesis-formation problems. 相似文献
5.
Feifei Li Ke Yi Wangchao Le 《The VLDB Journal The International Journal on Very Large Data Bases》2010,19(5):715-733
The database community has devoted extensive amount of efforts to indexing and querying temporal data in the past decades.
However, insufficient amount of attention has been paid to temporal ranking queries. More precisely, given any time instance
t, the query asks for the top-k objects at time t with respect to some score attribute. Some generic indexing structures based on R-trees do support ranking queries on temporal
data, but as they are not tailored for such queries, the performance is far from satisfactory. We present the Seb-tree, a simple indexing scheme that supports temporal ranking queries much more efficiently. The Seb-tree answers a top-k query for any time instance t in the optimal number of I/Os in expectation, namely,
O(logB \fracNB+\frackB){O\left({\rm log}_B\,\frac{N}{B}+\frac{k}{B}\right)} I/Os, where N is the size of the data set and B is the disk block size. The index has near-linear size (for constant and reasonable k
max values, where k
max is the maximum value for the possible values of the query parameter k), can be constructed in near-linear time, and also supports insertions and deletions without affecting its query performance
guarantee. Most of all, the Seb-tree is especially appealing in practice due to its simplicity as it uses the B-tree as the only building block. Extensive
experiments on a number of large data sets, show that the Seb-tree is more than an order of magnitude faster than the R-tree based indexes for temporal ranking queries. 相似文献
6.
Experience from recent years has shown that it is often advantageous not to build a single product but rather a family of
similar products that share at least one common functionality while having well-identified variabilities. Such product families are built from elementary features that reach from hardware parts to software artefacts such as requirements, architectural elements or patterns, components,
middleware, or code. We use the well established mathematical structure of idempotent semirings as the basis for a product family algebra that allows a formal treatment of the above notions. A particular application of the algebra concerns the multi-view reconciliation problem that arises when complex systems are modelled. We use algebraic integration constraints linking features in one view to features
in the same or a different view and show in several examples the suitability of this approach for a wide class of integration
constraint formulations. Our approach is illustrated with a Haskell prototype implementation of one particular model of product family algebra. 相似文献
7.
Wenfei Fan Hong Gao Xibei Jia Jianzhong Li Shuai Ma 《The VLDB Journal The International Journal on Very Large Data Bases》2011,20(4):495-520
This paper investigates constraints for matching records from unreliable data sources. (a) We introduce a class of matching dependencies (mds) for specifying the semantics of unreliable data. As opposed to static constraints for schema design, mds are developed for record matching, and are defined in terms of similarity predicates and a dynamic semantics. (b) We identify a special case of mds, referred to as relative candidate keys (rcks), to determine what attributes to compare and how to compare them when matching records across possibly different relations.
(c) We propose a mechanism for inferring mds, a departure from traditional implication analysis, such that when we cannot match records by comparing attributes that
contain errors, we may still find matches by using other, more reliable attributes. Moreover, we develop a sound and complete
system for inferring mds. (d) We provide a quadratic-time algorithm for inferring mds and an effective algorithm for deducing a set of high-quality rcks from mds. (e) We experimentally verify that the algorithms help matching tools efficiently identify keys at compile time for matching,
blocking or windowing and in addition, that the md-based techniques effectively improve the quality and efficiency of various record matching methods. 相似文献
8.
Model Checking with Strong Fairness 总被引:1,自引:0,他引:1
In this paper we present a coherent framework for symbolic model checking of linear-time temporal logic (ltl) properties over finite state reactive systems, taking full fairness constraints into consideration. We use the computational
model of a fair discrete system (fds) which takes into account both justice (weak fairness) and compassion (strong fairness). The approach presented here reduces the model-checking problem into the question of whether a given fds is feasible (i.e. has at least one computation).
The contribution of the paper is twofold: On the methodological level, it presents a direct self-contained exposition of full
ltl symbolic model checking without resorting to reductions to either μ-calculus or ctl. On the technical level, it extends previous methods by dealing with compassion at the algorithmic level instead of either
adding it to the specification, or transforming compassion to justice.
Finally, we extend ctl∗ with past operators, and show that the basic symbolic feasibility algorithm presented here, can be used to model check an
arbitrary ctl∗ formula over an fds with full fairness constraints.
This research was supported in part by an infra-structure grant from the Israeli Ministry of Science and Art, a grant from
the U.S.-Israel Binational Science Foundation, and a gift from Intel. 相似文献
9.
Providing machine tractable dictionary tools 总被引:1,自引:1,他引:0
Yorick Wilks Dan Fass Cheng-ming Guo James E. McDonald Tony Plate Brian M. Slator 《Machine Translation》1990,5(2):99-154
Machine readable dictionaries (Mrds) contain knowledge about language and the world essential for tasks in natural language processing (Nlp). However, this knowledge, collected and recorded by lexicographers for human readers, is not presented in a manner for Mrds to be used directly for Nlp tasks. What is badly needed are machine tractable dictionaries (Mtds): Mrds transformed into a format usable for Nlp. This paper discusses three different but related large-scale computational methods to transform Mrds into Mtds. The Mrd used is The Longman Dictionary of Contemporary English (Ldoce). The three methods differ in the amount of knowledge they start with and the kinds of knowledge they provide. All require some handcoding of initial information but are largely automatic. Method I, a statistical approach, uses the least handcoding. It generates relatedness networks for words in Ldoce and presents a method for doing partial word sense disambiguation. Method II employs the most handcoding because it develops and builds lexical entries for a very carefully controlled defining vocabulary of 2,000 word senses (1,000 words). The payoff is that the method will provide an Mtd containing highly structured semantic information. Method III requires the handcoding of a grammar and the semantic patterns used by its parser, but not the handcoding of any lexical material. This is because the method builds up lexical material from sources wholly within Ldoce. The information extracted is a set of sources of information, individually weak, but which can be combined to give a strong and determinate linguistic data base. 相似文献
10.
Given a graph with a source and a sink node, the NP-hard maximum k-splittable s,t-flow (M
k
SF) problem is to find a flow of maximum value from s to t with a flow decomposition using at most k paths. The multicommodity variant of this problem is a natural generalization of disjoint paths and unsplittable flow problems.
Constructing a k-splittable flow requires two interdepending decisions. One has to decide on k paths (routing) and on the flow values for the paths (packing). We give efficient algorithms for computing exact and approximate
solutions by decoupling the two decisions into a first packing step and a second routing step. Usually the routing is considered
before the packing. Our main contributions are as follows:
(i) We show that for constant k a polynomial number of packing alternatives containing at least one packing used by an optimal M
k
SF solution can be constructed in polynomial time. If k is part of the input, we obtain a slightly weaker result. In this case we can guarantee that, for any fixed ε>0, the computed set of alternatives contains a packing used by a (1−ε)-approximate solution. The latter result is based on the observation that (1−ε)-approximate flows only require constantly many different flow values. We believe that this observation is of interest in
its own right.
(ii) Based on (i), we prove that, for constant k, the M
k
SF problem can be solved in polynomial time on graphs of bounded treewidth. If k is part of the input, this problem is still NP-hard and we present a polynomial time approximation scheme for it. 相似文献
11.
In many AI fields, one must face the problem of finding a solution that is as close as possible to a given configuration.
This paper addresses this problem in a propositional framework. We introduce the decision problem distance-sat, which consists in determining whether a propositional formula admits a model that disagrees with a given partial interpretation
on at most d variables. The complexity of distance-sat and of several restrictions of it are identified. Two algorithms based on the well-known Davis/Logemann/Loveland search procedure
for the satisfiability problem sat are presented so as to solve distance-sat for CNF formulas. Their computational behaviors are compared with the ones offered by sat solvers on sat encodings of distance-sat instances. The empirical evaluation allows us to draw firm conclusions about the respective performances of the algorithms
and to relate the difficulty of distance-sat with the difficulty of sat from the practical side.
A preliminary version of this paper appeared with the title “distance-sat: Complexity and Algorithms” in the proceedings of the 16
th
National Conference on Artificial Intelligence (AAAI’99), pages 642–647, 1999. 相似文献
12.
This paper introduces the continuous consensus problem, in which a core M[k] of information is continuously maintained at all correct sites of the system. All local copies of the core must be identical
at all times k, and every interesting event should eventually enter the core. The continuous consensus problem is studied in synchronous
systems with crash and omission failures, assuming an upper bound of t on the number of failures in any given run of the system. A simple protocol for continuous consensus, called ConCon, is presented. This protocol is knowledge-based: The actions processes take depend explicitly on their knowledge, as well
as on their knowledge of what other processes know about failures and about events that occurred in the system. A close connection
between continuous consensus and knowledge is established by showing that in every continuous consensus protocol, the information
in the core at any given time must be common knowledge. Based on the characterization of common knowledge by Moses and Tuttle,
it is shown that ConCon is an optimum protocol for continuous consensus, maintaining the most up-to-date core possible at all times: For every pattern
of failures and external inputs and each point in time, the core provided by ConCon contains the cores of all correct protocols for continuous consensus. Indeed, the ConCon protocol can be viewed as a simplification of the Moses and Tuttle construction for computing the common knowledge at a given
point. Finally, a uniform version of continuous consensus is considered, in which all processes (faulty and nonfaulty) are guaranteed to maintain the
same core at any given time. An algorithm for uniform continuous consensus is presented, and is also shown to be an optimum
solution.
A preliminary version of this paper appeared in the Proceedings of the TARK X conference, Singapore 2005.
Work on this paper was performed in part during a sabbatical at the School of Computer Science and Engineering, The University
of New South Wales, Sydney, NSW 2052, Australia, where it was partially supported by ARC Discovery Grant RM02036. 相似文献
13.
Databases with uncertainty and lineage 总被引:2,自引:0,他引:2
Omar Benjelloun Anish Das Sarma Alon Halevy Martin Theobald Jennifer Widom 《The VLDB Journal The International Journal on Very Large Data Bases》2008,17(2):243-264
This paper introduces uldbs, an extension of relational databases with simple yet expressive constructs for representing and manipulating both lineage and uncertainty. Uncertain data and data lineage are two important areas of data management that have been considered extensively in isolation,
however many applications require the features in tandem. Fundamentally, lineage enables simple and consistent representation
of uncertain data, it correlates uncertainty in query results with uncertainty in the input data, and query processing with
lineage and uncertainty together presents computational benefits over treating them separately. We show that the uldb representation is complete, and that it permits straightforward implementation of many relational operations. We define two notions of uldb minimality—data-minimal and lineage-minimal—and study minimization of uldb representations under both notions. With lineage, derived relations are no longer self-contained: their uncertainty depends
on uncertainty in the base data. We provide an algorithm for the new operation of extracting a database subset in the presence
of interconnected uncertainty. We also show how uldbs enable a new approach to query processing in probabilistic databases. Finally, we describe the current state of the Trio system, our implementation of uldbs under development at Stanford.
This work was supported by the National Science Foundation under grants IIS-0324431, IIS-1098447, and IIS-9985114, by DARPA
Contract #03-000225, and by a grant from the Boeing Corporation. 相似文献
14.
Limits on the majority vote accuracy in classifier fusion 总被引:7,自引:0,他引:7
We derive upper and lower limits on the majority vote accuracy with respect to individual accuracy p, the number of classifiers in the pool (L), and the pairwise dependence between classifiers, measured by Yule’s Q statistic. Independence between individual classifiers is typically viewed as an asset in classifier fusion. We show that
the majority vote with dependent classifiers can potentially offer a dramatic improvement both over independent classifiers
and over an individual classifier with accuracy p. A functional relationship between the limits and the pairwise dependence Q is derived. Two patterns of the joint distribution for classifier outputs (correct/incorrect) are identified to derive the
limits: the pattern of success and the pattern of failure. The results support the intuition that negative pairwise dependence is beneficial although not straightforwardly related
to the accuracy. The pattern of success showed that for the highest improvement over p, all pairs of classifiers in the pool should have the same negative dependence.
ID="A1"Correspondance and offprint requests to: L. I. Kuncheva, School of Informatics, University of Wales, Bangor LL57 1UT, Gwynedd, UK. Email: l.i.kuncheva@bangor.ac.uk 相似文献
15.
This paper demonstrates the capabilities offoidl, an inductive logic programming (ILP) system whose distinguishing characteristics are the ability to produce first-order
decision lists, the use of an output completeness assumption as a substitute for negative examples, and the use originally
motivated by the problem of learning to generate the past tense of English verbs; however, this paper demonstrates its superior
performance on two different sets of benchmark ILP problems. Tests on the finite element mesh design problem show thatfoidl’s decision lists enable it to produce generally more accurate results than a range of methods previously applied to this
problem. Tests with a selection of list-processing problems from Bratko’s introductory Prolog text demonstrate that the combination
of implicit negatives and intensionality allowfoidl to learn correct programs from far fewer examples thanfoil.
This research was supported by a fellowship from AT&T awarded to the first author and by the National Science Foundation under
grant IRI-9310819.
Mary Elaine Califf: She is currently pursuing her doctorate in Computer Science at the University of Texas at Austin where she is supported
by a fellowship from AT&T. Her research interests include natural language understanding, particularly using machine learning
methods to build practical natural language understanding systems such as information extraction systems, and inductive logic
programming.
Raymond Joseph Mooney: He is an Associate Professor of Computer Sciences at the University of Texas at Austin. He recerived his Ph.D. in Computer
Science from the University of Illinois at Urbana-Champaign in 1988. His current research interests include applying machine
to natural language understanding, inductive logic programming, knowledge-base and theory refinement, learning for planning,
and learning for recommender systems. He serves on the editorial boards of the journalNew Generation Computing, theMachine Learning journal, theJournal of Artificial Intelligence Research, and the journalApplied Intelligence. 相似文献
16.
We study the bandwidth allocation problem (bap) in bounded degree trees. In this problem we are given a tree and a set of connection requests. Each request consists of
a path in the tree, a bandwidth requirement, and a weight. Our goal is to find a maximum weight subset S of requests such that, for every edge e, the total bandwidth of requests in S whose path contains e is at most 1. We also consider the storage allocation problem (sap), in which it is also required that every request in the solution is given the same contiguous portion of the resource in
every edge in its path. We present a deterministic approximation algorithm for bap in bounded degree trees with ratio
. Our algorithm is based on a novel application of the local ratio technique in which the available bandwidth is divided into narrow strips and requests with very small bandwidths are allocated in these
strips. We also present a randomized (2+ε)-approximation algorithm for bap in bounded degree trees. The best previously known ratio for bap in general trees is 5. We present a reduction from sap to bap that works for instances where the tree is a line and the bandwidths are very small. It follows that there exists a deterministic
2.582-approximation algorithm and a randomized (2+ε)-approximation algorithm for sap in the line. The best previously known ratio for this problem is 7.
A preliminary version of this paper was presented at the 14th Annual European Symposium on Algorithms (ESA), 2006. 相似文献
17.
Erez Petrank 《Computational Complexity》1994,4(2):133-157
We refine the complexity analysis of approximation problems by relating it to a new parameter calledgap location. Many of the results obtained so far for approximations yield satisfactory analysis with respect to this refined parameter, but some known results (e.g.,max-k-colorability, max 3-dimensional matching andmax not-all-equal 3sat) fall short of doing so. As a second contribution, our work fills the gap in these cases by presenting new reductions.Next, we present definitions and hardness results of new approximation versions of some NP-complete optimization problems. The problems we treat arevertex cover (for which we define a different optimization problem from the one treated in Papadimitriou & Yannakakis 1991),k-edge coloring, andset splitting. 相似文献
18.
safeDpi: a language for controlling mobile code 总被引:11,自引:0,他引:11
safeDpi is a distributed version of the Picalculus, in which processes are located at dynamically created sites. Parametrised code may be sent between sites using so-called ports, which are essentially higher-order versions of Picalculus communication channels. A host location may protect itself by only accepting code which conforms to a given type associated
to the incoming port.
We define a sophisticated static type system for these ports, which restrict the capabilities and access rights of any processes
launched by incoming code. Dependent and existential types are used to add flexibility, allowing the behaviour of these launched
processes, encoded as process types, to depend on the host's instantiation of the incoming code.
We also show that a natural contextually defined behavioural equivalence can be characterised coinductively, using bisimulations
based on typed actions. The characterisation is based on the idea of knowledge acquisition by a testing environment and makes explicit some of the
subtleties of determining equivalence in this language of highly constrained distributed code. 相似文献
19.
Wenfei Fan Jeffrey Xu Yu Jianzhong Li Bolin Ding Lu Qin 《The VLDB Journal The International Journal on Very Large Data Bases》2009,18(4):857-883
We study the problem of evaluating xpath queries over xml data that is stored in an rdbms via schema-based shredding. The interaction between recursion (descendants-axis) in xpath queries and recursion in dtds makes it challenging to answer xpath queries using rdbms. We present a new approach to translating xpath queries into sql queries based on a notion of extended
XP
ath expressions and a simple least fixpoint (lfp) operator. Extended xpath expressions are a mild extension of xpath, and the lfp operator takes a single input relation and is already supported by most commercial rdbms. We show that extended xpath expressions are capable of capturing both dtd recursion and xpath queries in a uniform framework. Furthermore, they can be translated into an equivalent sequence of sql queries with the lfp operator. We present algorithms for rewriting xpath queries over a (possibly recursive) dtd into extended xpath expressions and for translating extended xpath expressions to sql queries, as well as optimization techniques. The novelty of our approach consists in its capability to answer a large class
of xpath queries by means of only low-end rdbms features already available in most rdbms, as well as its flexibility to accommodate existing relational query optimization techniques. In addition, these translation
algorithms provide a solution to query answering for certain (possibly recursive) xml views of xml data. Our experimental results verify the effectiveness of our techniques.
An extended abstract was presented at the 31st international conference on Very Large Data Bases (VLDB), 2005. 相似文献
20.
In the setting of multi-instance learning, each object is represented by a bag composed of multiple instances instead of by a single instance in a traditional learning setting. Previous works in this
area only concern multi-instance prediction problems where each bag is associated with a binary (classification) or real-valued (regression) label. However, unsupervised multi-instance learning where bags are without labels has not been studied. In this paper, the problem of unsupervised multi-instance
learning is addressed where a multi-instance clustering algorithm named Bamic is proposed. Briefly, by regarding bags as atomic data items and using some form of distance metric to measure distances
between bags, Bamic adapts the popular k
-Medoids algorithm to partition the unlabeled training bags into k disjoint groups of bags. Furthermore, based on the clustering results, a novel multi-instance prediction algorithm named Bartmip is developed. Firstly, each bag is re-represented by a k-dimensional feature vector, where the value of the i-th feature is set to be the distance between the bag and the medoid of the i-th group. After that, bags are transformed into feature vectors so that common supervised learners are used to learn from
the transformed feature vectors each associated with the original bag’s label. Extensive experiments show that Bamic could effectively discover the underlying structure of the data set and Bartmip works quite well on various kinds of multi-instance prediction problems. 相似文献