In 1970 Kovarik proposed approximate orthogonalization algorithms. One of them (algorithm B) has quadratic convergence but requires at each iteration the inversion of a matrix of similar dimension to the initial one. An attempt to overcome this difficulty was made by replacing the inverse with a finite Neumann series expansion involving the original matrix and its adjoint. Unfortunately, this new algorithm loses the quadratic convergence and requires a large number of terms in the Neumann series which results in a dramatic increase in the computational effort per iteration. In this paper we propose a much simpler algorithm which, by using only the first two terms in a different series expansion, gives us the desired result with linear convergence. Systematic numerical experiments for collocation and Toeplitz matrices are also described. 相似文献
The electroencephalogram (EEG) is a representative signal containing information about the condition of the brain. The shape of the wave may contain useful information about the state of the brain. However, the human observer cannot directly monitor these subtle details. Besides, since bio-signals are highly subjective, the symptoms may appear at random in the time scale. Therefore, the EEG signal parameters, extracted and analyzed using computers, are highly useful in diagnostics. The aim of this work is to compare the different entropy estimators when applied to EEG data from normal and epileptic subjects. The results obtained indicate that entropy estimators can distinguish normal and epileptic EEG data with more than 95% confidence (using t-test). The classification ability of the entropy measures is tested using ANFIS classifier. The results are promising and a classification accuracy of about 90% is achieved. 相似文献
Two parallel block tridiagonalization algorithms and implementations for dense real symmetric matrices are presented. Block tridiagonalization is a critical pre-processing step for the block tridiagonal divide-and-conquer algorithm for computing eigensystems and is useful for many algorithms desiring the efficiencies of block structure in matrices. For an “effectively” sparse matrix, which frequently results from applications with strong locality properties, a heuristic parallel algorithm is used to transform it into a block tridiagonal matrix such that the eigenvalue errors remain bounded by some prescribed accuracy tolerance. For a dense matrix without any usable structure, orthogonal transformations are used to reduce it to block tridiagonal form using mostly level 3 BLAS operations. Numerical experiments show that block tridiagonal structure obtained from this algorithm directly affects the computational complexity of the parallel block tridiagonal divide-and-conquer eigensolver. Reduction to block tridiagonal form provides significantly lower execution times, as well as memory traffic and communication cost, over the traditional reduction to tridiagonal form for eigensystem computations. 相似文献
We describe the design, implementation and experimental evaluation of new algorithms for computing the approximate factorization of multivariate polynomials with complex coefficients that contain numerical noise. Our algorithms are based on a generalization of the differential forms introduced by W. Ruppert and S. Gao to many variables, and use singular value decomposition or structured total least squares approximation and Gauss–Newton optimization to numerically compute the approximate multivariate factors. We demonstrate on a large set of benchmark polynomials that our algorithms efficiently yield approximate factorizations within the coefficient noise even when the relative error in the input is substantial (10−3). 相似文献
This paper presents a model-based approximate λ-policy iteration approach using temporal differences for optimizing paths online for a pursuit-evasion problem, where an agent must visit several target positions within a region of interest while simultaneously avoiding one or more actively pursuing adversaries. This method is relevant to applications, such as robotic path planning, mobile-sensor applications, and path exposure. The methodology described utilizes cell decomposition to construct a decision tree and implements a temporal difference-based approximate λ-policy iteration to combine online learning with prior knowledge through modeling to achieve the objectives of minimizing the risk of being caught by an adversary and maximizing a reward associated with visiting target locations. Online learning and frequent decision tree updates allow the algorithm to quickly adapt to unexpected movements by the adversaries or dynamic environments. The approach is illustrated through a modified version of the video game Ms. Pac-Man, which is shown to be a benchmark example of the pursuit-evasion problem. The results show that the approach presented in this paper outperforms several other methods as well as most human players. 相似文献
Association rules are well recognised as a data mining tool for analysis of transactional data, currently going far beyond the early basket-based applications. A wide spectrum of methods for mining associations have been proposed up to date, including batch and incremental approaches. Most of the accurate incremental methods minimise, but do not completely eliminate reruns through processed data. In this paper we propose a new approximate algorithm RMAIN for incremental maintenance of association rules, which works repeatedly on subsequent portions of new transactions. After a portion has been analysed, the new rules are combined with the old ones, so that no reruns through the processed transactions are performed in the future. The resulting set of rules is kept similar to the one that would be achieved in a batch manner. Unlike other incremental methods, RMAIN is fully separated from a rule mining algorithm and this independence makes it highly general and flexible. Moreover, it operates on rules in their final form, ready for decision support, and not on intermediate representation (frequent itemsets), which requires further processing. These features make the RMAIN algorithm well suited for rule maintenance within knowledge bases of autonomous systems with strongly bounded resources and time for decision making. We evaluated the algorithm on synthetic and real datasets, achieving promising results with respect to either performance or quality of output rules. 相似文献
Partial information in databases can arise when information from several databases is combined. Even if each database is complete for some “world”, the combined databases will not be, and answers to queries against such combined databases can only be approximated. In this paper we describe various situations in which a precise answer cannot be obtained for a query asked against multiple databases. Based on an analysis of these situations, we propose a classification of constructs that can be used to model approximations.
The main goal of the paper is to study several formal models of approximations and their semantics. In particular, we obtain universality properties for these models of approximations. Universality properties suggest syntax for languages with approximations based on the operations which are naturally associated with them. We prove universality properties for most of the approximation constructs. Then we design languages built around datatypes given by the approximation constructs. A straightforward approach results in languages that have a number of limitations. In an attempt to overcome those limitations, we explain how all the languages can be embedded into a language for conjunctive and disjunctive sets from Libkin and Wong (1996) and demonstrate its usefulness in querying independent databases. We also discuss the semantics of approximation constructs and the relationship between them. 相似文献