共查询到20条相似文献,搜索用时 15 毫秒
1.
Navin Kabra David J. DeWitt 《The VLDB Journal The International Journal on Very Large Data Bases》1999,8(1):55-78
In this paper we describe the design and implementation of OPT++, a tool for extensible database query optimization that uses an object-oriented design to simplify the task of implementing, extending, and modifying an optimizer. Building an optimizer using OPT++ makes it easy to extend the query algebra (to add new query algebra operators and physical implementation algorithms to the system), easy to change the search space, and also to change the search strategy. Furthermore, OPT++ comes equipped with a number of search strategies that are available for use by an optimizer-implementor. OPT++ considerably simplifies both, the task of implementing an optimizer for a new database system, and the task of evaluating alternative optimization techniques and strategies to decide what techniques are best suited for that database system. We present the results of a series of performance studies. These results validate our design and show that, in spite of its flexibility, OPT++ can be used to build efficient optimizers. Received October 1996 / Accepted January 1998 相似文献
2.
Heuristic and randomized optimization for the join ordering problem 总被引:11,自引:0,他引:11
Michael Steinbrunn Guido Moerkotte Alfons Kemper 《The VLDB Journal The International Journal on Very Large Data Bases》1997,6(3):191-208
Recent developments in database technology, such as deductive database systems, have given rise to the demand for new, cost-effective
optimization techniques for join expressions. In this paper many different algorithms that compute approximate solutions for
optimizing join orders are studied since traditional dynamic programming techniques are not appropriate for complex problems.
Two possible solution spaces, the space of left-deep and bushy processing trees, are evaluated from a statistical point of
view. The result is that the common limitation to left-deep processing trees is only advisable for certain join graph types.
Basically, optimizers from three classes are analysed: heuristic, randomized and genetic algorithms. Each one is extensively
scrutinized with respect to its working principle and its fitness for the desired application. It turns out that randomized
and genetic algorithms are well suited for optimizing join expressions. They generate solutions of high quality within a reasonable
running time. The benefits of heuristic optimizers, namely the short running time, are often outweighed by merely moderate
optimization performance. 相似文献
3.
Peter A. Boncz Martin L. Kersten 《The VLDB Journal The International Journal on Very Large Data Bases》1999,8(2):101-119
In query-intensive database application areas, like decision support and data mining, systems that use vertical fragmentation
have a significant performance advantage. In order to support relational or object oriented applications on top of such a
fragmented data model, a flexible yet powerful intermediate language is needed. This problem has been successfully tackled
in Monet, a modern extensible database kernel developed by our group. We focus on the design choices made in the Monet interpreter
language (MIL), its algebraic query language, and outline how its concept of tactical optimization enhances and simplifies
the optimization of complex queries. Finally, we summarize the experience gained in Monet by creating a highly efficient implementation
of MIL.
Received November 10, 1998 / Accepted March 22, 1999 相似文献
4.
PicoDBMS: Scaling down database techniques for the smartcard 总被引:1,自引:0,他引:1
Philippe Pucheral Luc Bouganim Patrick Valduriez Christophe Bobineau 《The VLDB Journal The International Journal on Very Large Data Bases》2001,10(2-3):120-132
Smartcards are the most secure portable computing device today. They have been used successfully in applications involving money, and proprietary and personal data (such as banking, healthcare, insurance, etc.). As smartcards get more powerful (with 32-bit CPU and more than 1 MB of stable memory in the next versions) and become multi-application, the need for database management arises. However, smartcards have severe hardware limitations (very slow write, very little RAM, constrained stable memory, no autonomy, etc.) which make traditional database technology irrelevant. The major problem is scaling down database techniques so they perform well under these limitations. In this paper, we give an in-depth analysis of this problem and propose a PicoDBMS solution based on highly compact data structures, query execution without RAM, and specific techniques for atomicity and durability. We show the effectiveness of our techniques through performance evaluation. Received: 15 October 2000 / Accepted: 15 April 2001 Published online: 23 July 2001 相似文献
5.
Aya Soffer Hanan Samet 《The VLDB Journal The International Journal on Very Large Data Bases》1998,7(4):253-274
Symbolic images are composed of a finite set of symbols that have a semantic meaning. Examples of symbolic images include
maps (where the semantic meaning of the symbols is given in the legend), engineering drawings, and floor plans. Two approaches
for supporting queries on symbolic-image databases that are based on image content are studied. The classification approach
preprocesses all symbolic images and attaches a semantic classification and an associated certainty factor to each object
that it finds in the image. The abstraction approach describes each object in the symbolic image by using a vector consisting
of the values of some of its features (e.g., shape, genus, etc.). The approaches differ in the way in which responses to queries
are computed. In the classification approach, images are retrieved on the basis of whether or not they contain objects that
have the same classification as the objects in the query. On the other hand, in the abstraction approach, retrieval is on
the basis of similarity of feature vector values of these objects. Methods of integrating these two approaches into a relational
multimedia database management system so that symbolic images can be stored and retrieved based on their content are described.
Schema definitions and indices that support query specifications involving spatial as well as contextual constraints are presented.
Spatial constraints may be based on both locational information (e.g., distance) and relational information (e.g., north of).
Different strategies for image retrieval for a number of typical queries using these approaches are described. Estimated costs
are derived for these strategies. Results are reported of a comparative study of the two approaches in terms of image insertion
time, storage space, retrieval accuracy, and retrieval time.
Received June 12, 1998 / Accepted October 13, 1998 相似文献
6.
Active rules for XML: A new paradigm for E-services 总被引:1,自引:0,他引:1
Angela Bonifati Stefano Ceri Stefano Paraboschi 《The VLDB Journal The International Journal on Very Large Data Bases》2001,10(1):39-47
XML is rapidly becoming one of the most widely adopted technologies for information exchange and representation. As the use
of XML becomes more widespread, we foresee the development of active XML rules, i.e., rules explicitly designed for the management
of XML information. In particular, we argue that active rules for XML offer a natural paradigm for the rapid development of
innovative e-services. In the paper, we show how active rules can be specified in the context of XSLT, a pattern-based language
for publishing XML documents (promoted by the W3C) which is receiving strong commercial support, and Lorel, a query language
for XML documents that is quite popular in the research world. We demonstrate, through simple examples of active rules for
XSLT and Lorel, that active rules can be effective for the implementation of e-commerce services. We also discuss the various
issues that need to be considered in adapting the notion of relational triggers to the XML context.
Received: 30 October 2000 / Accepted: 19 December 2000 Published online: 27 April 2001 相似文献
7.
MiniCon: A scalable algorithm for answering queries using views 总被引:5,自引:0,他引:5
Rachel Pottinger Alon Halevy 《The VLDB Journal The International Journal on Very Large Data Bases》2001,10(2-3):182-198
The problem of answering queries using views is to find efficient methods of answering a query using a set of previously
materialized views over the database, rather than accessing the database relations. The problem has received significant attention
because of its relevance to a wide variety of data management problems, such as data integration, query optimization, and
the maintenance of physical data independence. To date, the performance of proposed algorithms has received very little attention,
and in particular, their scale up in the presence of a large number of views is unknown. We first analyze two previous algorithms,
the bucket algorithm and the inverse-rules, and show their deficiencies. We then describe the MiniCon, a novel algorithm for
finding the maximally-contained rewriting of a conjunctive query using a set of conjunctive views. We present the first experimental
study of algorithms for answering queries using views. The study shows that the MiniCon scales up well and significantly outperforms
the previous algorithms. We describe an extension of the MiniCon to handle comparison predicates, and show its performance
experimentally. Finally, we describe how the MiniCon can be extended to the context of query optimization.
Received: 15 October 2000 / Accepted: 15 April 2001 Published online: 28 June 2001 相似文献
8.
Query processing over object views of relational data 总被引:2,自引:0,他引:2
Gustav Fahl Tore Risch 《The VLDB Journal The International Journal on Very Large Data Bases》1997,6(4):261-281
This paper presents an approach to object view management for relational databases. Such a view mechanism makes it possible for users to transparently work with data in
a relational database as if it was stored in an object-oriented (OO) database. A query against the object view is translated
to one or several queries against the relational database. The results of these queries are then processed to form an answer
to the initial query. The approach is not restricted to a ‘pure’ object view mechanism for the relational data, since the
object view can also store its own data and methods. Therefore it must be possible to process queries that combine local data
residing in the object view with data retrieved from the relational database. We discuss the key issues when object views
of relational databases are developed, namely: how to map relational structures to sub-type/supertype hierarchies in the view,
how to represent relational database access in OO query plans, how to provide the concept of object identity in the view,
how to handle the fact that the extension of types in the view depends on the state of the relational database, and how to
process and optimize queries against the object view. The results are based on experiences from a running prototype implementation.
Edited by: M.T. ?zsu. Received April 12, 1995 / Accepted April 22, 1996 相似文献
9.
Chiang Lee Chi-Sheng Shih Yaw-Huei Chen 《The VLDB Journal The International Journal on Very Large Data Bases》2001,9(4):327-343
Traditional algorithms for optimizing the execution order of joins are no more valid when selections and projections involve
methods and become very expensive operations. Selections and projections could be even more costly than joins such that they
are pulled above joins, rather than pushed down in a query tree. In this paper, we take a fundamental look at how to approach
query optimization from a top-down design perspective, rather than trying to force one model to fit into another. We present
a graph model which is designed to characterize execution plans. Each edge and each vertex of the graph is assigned a weight
to model execution plans. We also design algorithms that use these weights to optimize the execution order of operations.
A cost model of these algorithms is developed. Experiments are conducted on the basis of this cost model. The results show
that our algorithms are superior to similar work proposed in the literature.
Received 20 April 1999 / Accepted 9 August 2000 Published online 20 April 2001 相似文献
10.
Ashish Mehta James Geller Yehoshua Perl Erich Neuhold 《The VLDB Journal The International Journal on Very Large Data Bases》1998,7(1):25-47
A path-method is used as a mechanism in object-oriented databases (OODBs) to retrieve or to update information relevant to one class that
is not stored with that class but with some other class. A path-method is a method which traverses from one class through
a chain of connections between classes and accesses information at another class. However, it is a difficult task for a casual
user or even an application programmer to write path-methods to facilitate queries. This is because it might require comprehensive
knowledge of many classes of the conceptual schema that are not directly involved in the query, and therefore may not even
be included in a user's (incomplete) view about the contents of the database. We have developed a system, called path-method generator (PMG), which generates path-methods automatically according to a user's database-manipulating requests. The PMG offers the
user one of the possible path-methods and the user verifies from his knowledge of the intended purpose of the request whether
that path-method is the desired one. If the path method is rejected, then the user can utilize his now increased knowledge
about the database to request (with additional parameters given) another offer from the PMG. The PMG is based on access weights attached to the connections between classes and precomputed access relevance between every pair of classes of the OODB. Specific rules for access weight assignment and algorithms for computing access
relevance appeared in our previous papers [MGPF92, MGPF93, MGPF96]. In this paper, we present a variety of traversal algorithms
based on access weights and precomputed access relevance. Experiments identify some of these algorithms as very successful
in generating most desired path-methods. The PMG system utilizes these successful algorithms and is thus an efficient tool
for aiding the user with the difficult task of querying and updating a large OODB.
Received July 19, 1993 / Accepted May 16, 1997 相似文献
11.
Wen-Syan Li K.Selçuk Candan Kyoji Hirata Yoshinori Hara 《The VLDB Journal The International Journal on Very Large Data Bases》2001,9(4):312-326
Due to the fuzziness of query specification and media matching, multimedia retrieval is conducted by way of exploration.
It is essential to provide feedback so that users can visualize query reformulation alternatives and database content distribution.
Since media matching is an expensive task, another issue is how to efficiently support exploration so that the system is not
overloaded by perpetual query reformulation. In this paper, we present a uniform framework to represent statistical information
of both semantics and visual metadata for images in the databases. We propose the concept of query verification, which evaluates queries using statistics, and provides users with feedback, including the strictness and reformulation alternatives
of each query condition as well as estimated numbers of matches. With query verification, the system increases the efficiency
of the multimedia database exploration for both users and the system. Such statistical information is also utilized to support
progressive query processing and query relaxation.
Received: 9 June 1998/ Accepted: 21 July 2000 Published online: 4 May 2001 相似文献
12.
Optimizing multiple dimensional queries simultaneously in multidimensional databases 总被引:1,自引:0,他引:1
Weifa Liang Maria E. Orlowska Jeffrey X. Yu 《The VLDB Journal The International Journal on Very Large Data Bases》2000,8(3-4):319-338
Some significant progress related to multidimensional data analysis has been achieved in the past few years, including the
design of fast algorithms for computing datacubes, selecting some precomputed group-bys to materialize, and designing efficient
storage structures for multidimensional data. However, little work has been carried out on multidimensional query optimization
issues. Particularly the response time (or evaluation cost) for answering several related dimensional queries simultaneously
is crucial to the OLAP applications. Recently, Zhao et al. first exploited this problem by presenting three heuristic algorithms.
In this paper we first consider in detail two cases of the problem in which all the queries are either hash-based star joins
or index-based star joins only. In the case of the hash-based star join, we devise a polynomial approximation algorithm which
delivers a plan whose evaluation cost is $ O(n^{\epsilon }$) times the optimal, where n is the number of queries and is a fixed constant with . We also present an exponential algorithm which delivers a plan with the optimal evaluation cost. In the case of the index-based
star join, we present a heuristic algorithm which delivers a plan whose evaluation cost is n times the optimal, and an exponential algorithm which delivers a plan with the optimal evaluation cost. We then consider
a general case in which both hash-based star-join and index-based star-join queries are included. For this case, we give a
possible improvement on the work of Zhao et al., based on an analysis of their solutions. We also develop another heuristic
and an exact algorithm for the problem. We finally conduct a performance study by implementing our algorithms. The experimental
results demonstrate that the solutions delivered for the restricted cases are always within two times of the optimal, which
confirms our theoretical upper bounds. Actually these experiments produce much better results than our theoretical estimates.
To the best of our knowledge, this is the only development of polynomial algorithms for the first two cases which are able
to deliver plans with deterministic performance guarantees in terms of the qualities of the plans generated. The previous
approaches including that of [ZDNS98] may generate a feasible plan for the problem in these two cases, but they do not provide
any performance guarantee, i.e., the plans generated by their algorithms can be arbitrarily far from the optimal one.
Received: July 21, 1998 / Accepted: August 26, 1999 相似文献
13.
Scene change detection techniques for video database systems 总被引:1,自引:0,他引:1
Haitao Jiang Abdelsalam Helal Ahmed K. Elmagarmid Anupam Joshi 《Multimedia Systems》1998,6(3):186-195
Scene change detection (SCD) is one of several fundamental problems in the design of a video database management system (VDBMS).
It is the first step towards the automatic segmentation, annotation, and indexing of video data. SCD is also used in other
aspects of VDBMS, e.g., hierarchical representation and efficient browsing of the video data. In this paper, we provide a
taxonomy that classifies existing SCD algorithms into three categories: full-video-image-based, compressed-video-based, and
model-based algorithms. The capabilities and limitations of the SCD algorithms are discussed in detail. The paper also proposes
a set of criteria for measuring and comparing the performance of various SCD algorithms. We conclude by discussing some important
research directions. 相似文献
14.
Algebraic query optimisation for database programming languages 总被引:1,自引:0,他引:1
Alexandra Poulovassilis Carol Small 《The VLDB Journal The International Journal on Very Large Data Bases》1996,5(2):119-132
A major challenge still facing the designers and implementors of database
programming languages (DBPLs) is that of query optimisation. We investigate
algebraic query optimisation techniques for DBPLs in the context of a purely
declarative functional language that supports sets as first-class objects.
Since the language is computationally complete issues such as
non-termination of expressions and construction of infinite data structures
can be investigated, whilst its declarative nature allows the issue of side
effects to be avoided and a richer set of equivalences to be developed.
The language has a well-defined semantics which permits us to reason
formally about the properties of expressions, such as their equivalence with
other expressions and their termination. The support of a set bulk data
type enables much prior work on the optimisation of relational languages to
be utilised.
In the paper we first give the syntax of our archetypal DBPL and briefly
discuss its semantics. We then define a small but powerful algebra of
operators over the set data type, provide some key equivalences for
expressions in these operators, and list transformation principles for
optimising expressions. Along the way, we identify some caveats to
well-known equivalences for non-deductive database languages. We next
extend our language with two higher level constructs commonly found in
functional DBPLs: set comprehensions and functions with known inverses. Some
key equivalences for these constructs are provided, as are transformation
principles for expressions in them. Finally, we investigate extending our
equivalences for the set operators to the analogous operators over bags.
Although developed and formally proved in the context of a functional
language, our findings are directly applicable to other DBPLs of similar
expressiveness.
Edited by
Matthias Jarke, Jorge Bocca, Carlo Zaniolo. Received
September 15, 1994 / Accepted September 1, 1995 相似文献
15.
Alberto Bartoli Gianluca Dini Lanfranco Lopriore 《International Journal on Software Tools for Technology Transfer (STTT)》2001,3(2):235-245
With reference to a memory management system supporting the single address space abstraction and a uniform, persistent view of storage, we present a set of mechanisms that allow applications to exert explicit control over memory management activities. These mechanisms make it possible to move the contents of a virtual page to primary memory for fast processor access, or to push these contents back to secondary memory to free primary memory space. Our memory management scheme allows programs to exploit the memory reference pattern of the underlying algorithms, thereby improving utilisation of the system storage resources. This result is illustrated by using significant examples of memory management activities implemented at the application program level. Published online: 8 February 2001 相似文献
16.
Abstract. This paper describes the design of a reconfigurable architecture for implementing image processing algorithms. This architecture
is a pipeline of small identical processing elements that contain a programmable logic device (FPGA) and double port memories.
This processing system has been adapted to accelerate the computation of differential algorithms. The log-polar vision selectively
reduces the amount of data to be processed and simplifies several vision algorithms, making possible their implementation
using few hardware resources. The reconfigurable architecture design has been devoted to implementation, and has been employed
in an autonomous platform, which has power consumption, size and weight restrictions. Two different vision algorithms have
been implemented in the reconfigurable pipeline, for which some experimental results are shown.
Received: 30 March 2001 / Accepted: 11 February 2002
RID="*"
ID="*" This work has been supported by the Ministerio de Ciencia y Tecnología and FEDER under project TIC2001-3546
Correspondence to: J.A. Boluda 相似文献
17.
Julie A. Jacko Armando B. Barreto Ingrid U. Scott Josey Y. M. Chu Holly S. Vitense Frank T. Conway W. Bradley Fain 《Universal Access in the Information Society》2002,1(3):197-206
The objective of this study was to derive empirical knowledge of the visual search strategies of computer users who suffer
from age-related macular degeneration (AMD). This was accomplished by recording eye movement during the use of feature-enhanced
software. The results from this study show that there are differences between users who have AMD and users who are fully sighted
(FS). Detailed analyses confirmed the hypotheses that there would be performance differences between the AMD and FS participants,
and that specific features of the interface, namely icon size, background color, and the number of icons on a display, would
significantly affect the search strategies of users.
Published online: 29 January 2002 相似文献
18.
Semantic integrity support in SQL:1999 and commercial (object-)relational database management systems 总被引:1,自引:0,他引:1
Can Türker Michael Gertz 《The VLDB Journal The International Journal on Very Large Data Bases》2001,10(4):241-269
The correctness of the data managed by database systems is vital to any application that utilizes data for business, research,
and decision-making purposes. To guard databases against erroneous data not reflecting real-world data or business rules,
semantic integrity constraints can be specified during database design. Current commercial database management systems provide
various means to implement mechanisms to enforce semantic integrity constraints at database run-time.
In this paper, we give an overview of the semantic integrity support in the most recent SQL-standard SQL:1999, and we show
to what extent the different concepts and language constructs proposed in this standard can be found in major commercial (object-)relational
database management systems. In addition, we discuss general design guidelines that point out how the semantic integrity features
provided by these systems should be utilized in order to implement an effective integrity enforcing subsystem for a database.
Received: 14 August 2000 / Accepted: 9 March 2001 / Published online: 7 June 2001 相似文献
19.
Recent advances in computer technologies have made it feasible to provide multimedia services, such as news distribution
and entertainment, via high-bandwidth networks. The storage and retrieval of large multimedia objects (e.g., video) becomes
a major design issue of the multimedia information system. While most other works on multimedia storage servers assume an
on-line disk storage system, we consider a two-tier storage architecture with a robotic tape library as the vast near-line
storage and an on-line disk system as the front-line storage. Magnetic tapes are cheaper, more robust, and have a larger
capacity; hence, they are more cost effective for large scale storage systems (e.g., video-on-demand (VOD) systems may
store tens of thousands of videos). We study in detail the design issues of the tape subsystem and propose some novel tape-scheduling
algorithms which give faster response and require less disk buffer space. We also study the disk-striping policy and the
data layout on the tape cartridge in order to fully utilize the throughput of the robotic tape system and to minimize the
on-line disk storage space. 相似文献
20.
Gyeonghwan Kim Venu Govindaraju Sargur N. Srihari 《International Journal on Document Analysis and Recognition》1999,2(1):37-44
This paper presents an end-to-end system for reading handwritten page images. Five functional modules included in the system
are introduced in this paper: (i) pre-processing, which concerns introducing an image representation for easy manipulation
of large page images and image handling procedures using the image representation; (ii) line separation, concerning text line
detection and extracting images of lines of text from a page image; (iii) word segmentation, which concerns locating word
gaps and isolating words from a line of text image obtained efficiently and in an intelligent manner; (iv) word recognition,
concerning handwritten word recognition algorithms; and (v) linguistic post-pro- cessing, which concerns the use of linguistic
constraints to intelligently parse and recognize text. Key ideas employed in each functional module, which have been developed
for dealing with the diversity of handwriting in its various aspects with a goal of system reliability and robustness, are
described in this paper. Preliminary experiments show promising results in terms of speed and accuracy.
Received October 30, 1998 / Revised January 15, 1999 相似文献