This paper concerns the following problem: given a set of multi-attribute records, a fixed number of buckets and a two-disk system, arrange the records into the buckets and then store the buckets between the disks in such a way that, over all possible orthogonal range queries (ORQs), the disk access concurrency is maximized. We shall adopt the multiple key hashing (MKH) method for arranging records into buckets and use the disk modulo (DM) allocation method for storing buckets onto disks. Since the DM allocation method has been shown to be superior to any other allocation methods for allocating an MKH file onto a two-disk system for answering ORQs, the real issue is knowing how to determine an optimal way for organizing the records into buckets based upon the MKH concept.
A performance formula that can be used to evaluate the average response time, over all possible ORQs, of an MKH file in a two-disk system using the DM allocation method is first presented. Based upon this formula, it is shown that our design problem is related to a notoriously difficult problem, namely the Prime Number Problem. Then a performance lower bound and an efficient algorithm for designing optimal MKH files in certain cases are presented. It is pointed out that in some cases the optimal MKH file for ORQs in a two-disk system using the DM allocation method is identical to the optimal MKH file for ORQs in a single-disk system and the optimal average response time in a two-disk system is slightly greater than one half of that in a single-disk system. 相似文献
Abstract. The performance of the Geweke-Porter-Hudak (GPH) test, the modified rescaled range (MRR) test and two Lagrange multiplier (LM) type tests for fractional integration in small samples is examined using Monte Carlo methods. Both the GPH and MRR tests are found to be robust to moderate autoregressive moving-average components, autoregressive conditional heteroskedasticity effects and shifts in the variance. However, these two tests are sensitive to large autoregressive moving-average components and shifts in the mean. It is also found that the LM tests are sensitive to deviations from the null hypothesis. As an illustration, the GPH test is applied to two economic data series. 相似文献
A scanning electron microscope (SEM) is a sophisticated equipment employed for fine imaging of a variety of surfaces. In this study, prediction models of SEM were constructed by using a generalized regression neural network (GRNN) and genetic algorithm (GA). The SEM components examined include condenser lens 1 and 2 and objective lens (coarse and fine) referred to as CL1, CL2, OL-Coarse, and OL-Fine. For a systematic modeling of SEM resolution (R), a face-centered Box–Wilson experiment was conducted. Two sets of data were collected with or without the adjustment of magnification. Root-mean-squared prediction error of optimized GRNN models are GA 0.481 and 1.96×10-12 for non-adjusted and adjusted data, respectively. The optimized models demonstrated a much improved prediction over statistical regression models. The optimized models were used to optimize parameters particularly under best tuned SEM environment. For the variations in CL2 and OL-Coarse, the highest R could be achieved at all conditions except a larger CL2 either at smaller or larger OL-Coarse. For the variations in CL1 and CL2, the highest R was obtained at all conditions but larger CL2 and smaller CL1. 相似文献
Summary In this paper, we investigate the discretization of an elliptic boundary value problem in 3D by means of the hp-version of the finite element method using a mesh of tetrahedrons. We present several bases based on integrated Jacobi polynomials
in which the element stiffness matrix has nonzero entries, where p denotes the polynomial degree. The proof of the sparsity requires the assistance of computer algebra software. Several numerical
experiments show the efficiency of the proposed bases for higher polynomial degrees p.
相似文献