首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 21 毫秒
1.
This paper presents a comparative study of some concurrency control algorithms for distributed databases of computer clusters which emphasize high availability and high performance requirements. For this purpose, we have analyzed some concurrency control algorithms which are used in commercial DBMSs, such as the pessimistic locking algorithm as it verifies transaction conflicts early in their execution phase, and the optimistic algorithm which investigates the presence of conflicts after the execution phase. A new algorithm is proposed and implemented by a simulation program. The three algorithms were tested using different configurations. Simulation results showed that the locking algorithm performed better than the optimistic method in presence of conflicts between transactions, while the optimistic algorithm provided better results in the absence of conflicts. Furthermore, in a distributed database with a certain probability of conflicts, the locking algorithm can be used to guarantee strong consistency and an acceptable level of performance. However, if this probability is negligible, the system performance can be improved by using the optimistic algorithm. The proposed algorithm offers improved performance in numerous cases. As a result, it can be used in a distributed database to guarantee a satisfactory level of performance in the presence of conflicts.  相似文献   

2.
The computing power provided by high performance and low-cost PC-based clusters with Grid platforms are attractive and they are equal or superior to supercomputers and mainframes. In this paper, we present implementation and design rationale of Visuel toolkit for MPI parallel program performance measurement and analysis in cluster and grid environments. Most of performance visualization tools available today for high-performance platforms show solely system performance data (e.g., CPU load, memory usage, network bandwidth, server average load), and thus, being suitable for computing system activity visualization. The Visuel (Visuel (in French language) = to visualize) toolkit is web-based interface designed to show performance activities of all computing nodes of a distributed environment involved in the execution of MPI parallel program, such as CPU load level and memory usage of each computing node. In addition, this toolkit is able to display comparative performance data charts of MPI parallel applications and multiple executions under investigation. The usage of this toolkit shows that it outperforms in easing the process of investigation of parallel applications.
Hsun-Chang ChangEmail:
  相似文献   

3.
High-performance storage systems are evolving towards decentralized commodity clusters that can scale in capacity, processing power, and network throughput. Building such systems requires: (a) Sharing physical resources among applications; (b) Sharing data among applications; (c) Allowing customized data views. Current solutions typically satisfy the first two requirements through a cluster file-system, resulting in monolithic, hard-to-manage systems. In this paper we present a storage system that addresses all three requirements by extending the block layer below the file-system. First, we discuss how our system provides customized (virtualized) storage views within a single node. Then, we discuss how it scales in clustered setups. To achieve efficient resource and data sharing we support block-level allocation and locking as in-band mechanisms. We implement a prototype under Linux and use it to build a shared cluster file-system. Our evaluation in a 24-node cluster setup concludes that our approach offers flexibility, scalability and reduced effort to implement new functionality.  相似文献   

4.
High electricity consumption, associated with running Internet scale server farms, not only reflects on the data center’s greenhouse gas emissions, but also increases the cost of running the data center itself. In this paper, we consider the problem of maximizing the revenues of service providers running large scale data centers subject to setup cost by reducing their electricity bill, while considering the fact that clients consuming the offered services have finite non-deterministic patience. As a solution, we present and evaluate the performance of allocation policies which, in the context of both one and two-tiered systems, dynamically switch servers on and off according to changes in user demand. The algorithms we present aim at maximizing the users’ experience while minimizing the amount of electricity required to run the IT infrastructure in spite of non-stationary traffic which cannot be predicted with the absolute accuracy. The results of several experiments are presented, showing that the proposed schemes perform well under different traffic conditions.  相似文献   

5.
In a typical distributed computing system (DCS), nodes consist of processing elements, memory units, shared resources, data files, and programs. For a distributed application, programs and data files are distributed among many processing elements that may exchange data and control information via communication link. The reliability of DCS can be expressed by the analysis of distributed program reliability (DPR) and distributed system reliability (DSR). In this paper, two reliability measures are introduced which are Markov-chain distributed program reliability (MDPR) and Markov-chain distributed system reliability (MDSR) to accurately model the reliability of DCS. A discrete time Markov chain with one absorbing state is constructed for this problem. The transition probability matrix is employed to represent the transition probability from one state to another state in a unit of time. In addition to mathematical method to evaluate the MDPR and MDSR, a simulation result is also presented to prove its correction.  相似文献   

6.
In the computer science community there is a growing interest in the field of Ambient Intelligent Systems. This systems surround their human users with computing and networking technology unobtrusively embedded in their environment. This technology is aimed to provide the users with useful information and to take action to make the environment more convenient for them. As the number of users increases the resources that make Ambient Intelligence possible can be easily saturated making the system unstable and projecting an image of poor QoS to the users. The main goal of this paper is to provide the means for the Ambient Intelligent Systems to monitor themselves and take corrective action automatically if performance starts to drop. Our approach uses a Performance Ontology that structures the knowledge about Software Performance Engineering, and a reasoning engine that acts like an expert system with the Performance Ontology as its foundation. The case study at the end shows the applicability of the developed techniques.  相似文献   

7.
Data acquisition systems are mission-critical components in high-energy physics experiments. They are embedded in an environment of custom electronics, and are frequently characterized by high performance requirements. With the advent of powerful mainstream computing platforms and affordable high-speed networking equipment, system cost and time to completion can be significantly reduced. There still exists a considerable effort in custom software developments to build these systems and make them running efficiently. Therefore we strive for a software architecture flexible and robust enough to be usable in different system configurations and deployment cases. The software should cover the largest possible application domain and provide a practical balance between efficiency and flexibility. This article pinpoints the requirements imposed on such an on-line software infrastructure and sheds light on a viable design approach. As such, this article aims at laying out the foundations for a broader understanding of the importance for fostering a homogeneous architecture for high-energy physics data acquisition.  相似文献   

8.
In this paper, we propose an architecture, design and build a prototype of a novel IoT system with intelligence, distributed at multiple tiers including the network edge. Our proposed architecture hosts a modular, three-tier IoT system including the edge, gateway (fog) and cloud tiers. The proposed system relies on data acquired by edge devices to realize a distributed machine learning model and achieve timely response at the edge using a lightweight machine learning model. In addition, it employs more sophisticated machine learning models at the higher fog and cloud tiers for wider-scope, long-term decision making. One of the prime objectives of the proposed system is reducing the volume of data transferred across tiers. This is attained through intelligent data filtering at the edge/gateway tiers to distill key events that avail the most relevant data points to higher-tier machine learning models at the gateway and cloud. This, in turn, reduces the outliers and the redundant data that may impact the gateway and cloud models and reduces the inter-tier communications overhead. To demonstrate the merits of our proposed system, we build a proof-of-concept prototype hosting the three tiers, using COTS components and supporting networking technologies. We demonstrate through extensive experiments the merits of the proposed system. A major finding is that our system is capable of achieving prediction performance comparable to the centralized machine learning baseline model, while reducing the inter-tier communications overhead by up to 80%.  相似文献   

9.
In mobile machines, as well as in manufacturing, the overall productivity is essential for business competitiveness. As the operation of a modern mobile machine is affected by various parameters, they need to be tuned to reach an optimal performance – however, due to machine complexity, parameter optimisation is difficult for a typical operator. To enable parameter optimisation locally in machines, this article presents a system architecture to generate information and knowledge from machine fleet data and to utilise them in machine operations in the field. Measurement data is collected and analysed to discover the associations between machine performance and parameter values. While some results are plain statistical distributions, any resulting more sophisticated domain knowledge is stored as rules. Rule-based reasoning enables a zone of interoperation between the information system and domain experts. Once information and knowledge have been generated, they are made available to machines that run the actual parameter assessment application. Results made with forestry data indicate that the system has a considerable potential to improve machine productivity.  相似文献   

10.
We present here a performance analysis of three of current architectures that have become commonplace in the High Performance Computing world. Blue Gene/Q is the third generation of systems from IBM that use modestly performing cores but at large-scale in order to achieve high performance. The XE6 is the latest in a long line of Cray systems that use a 3-D topology but the first to use its Gemini interconnection network. InfiniBand provides the flexibility of using compute nodes from many vendors that can be connected in many possible topologies. The performance characteristics of each vary vastly, and the way in which nodes are allocated in each type of system can significantly impact on achieved performance. In this work we compare these three systems using a combination of micro-benchmarks and a set of production applications. In addition we also examine the differences in performance variability observed on each system and quantify the lost performance using a combination of both empirical measurements and performance models. Our results show that significant performance can be lost in normal production operation of the Cray XE6 and InfiniBand Clusters in comparison to Blue Gene/Q.  相似文献   

11.
This paper describes an approach to carry out performance analysis of parallel embedded applications. The approach is based on measurement, but in addition, the idea of driving the measurement process (application instrumentation and monitoring) by a behavioral model is introduced. Using this model, highly comprehensible performance information can be collected. The whole approach is based on this behavioral model, one instrumentation method and two tools, one for monitoring and the other for visualization and analysis. Each of these is briefly described, and the steps to carry out performance analysis using them are clearly defined. They are explained by means of a case study. Finally, one method to evaluate the intrusiveness of the monitoring approach is proposed, and the intrusiveness results for the case study are presented.  相似文献   

12.
Summary Distributed Mutual Exclusion algorithms have been mainly compared using the number of messages exchanged per critical section execution. In such algorithms, no attention has been paid to the serialization order of the requests. Indeed, they adopt FCFS discipline. Conversely, the insertion of priority serialization disciplines, such as Short-Job-First, Head-Of-Line, Shortest-Remaining-Job-First etc., can be useful in many applications to optimize some performance indices. However, such priority disciplines are prone to starvation. The goal of this paper is to investigate and evaluate the impact of the insertion of a priority discipline in Maekawa-type algorithms. Priority serialization disciplines will be inserted by means of agated batch mechanism which avoids starvation. In a distributed algorithm, such a mechanism needs synchronizations among the processes. In order to highlight the usefulness of the priority based serialization discipline, we show how it can be used to improve theaverage response time compared to the FCFS discipline. The gated batch approach exhibits other advantages: algorithms are inherently deadlock-free and messages do not need to piggyback timestamps. We also show that, under heavy demand, algorithms using gated batch exchange less messages than Maekawa-type algorithms per critical section excution. Roberto Baldoni was born in Rome on February 1, 1965. He received the Laurea degree in electronic engineering in 1990 from the University of Rome La Sapienza and the Ph.D. degree in Computer Science from the University of Rome La Sapienza in 1994. Currently, he is a researcher in computer science at IRISA, Rennes (France). His research interests include operating systems, distributed algorithms, network protocols and real-time multimedia applications. Bruno Ciciani received the Laurea degree in electronic engineering in 1980 from the University of Rome La Sapienza. From 1983 to 1991 he has been a researcher at the University of Rome Tor Vergata. He is currently full professor in Computer Science at the University of Rome La Sapienza. His research activities include distributed computer systems, fault-tolerant computing, languages for parallel processing, and computer system performance and reliability evaluation. He has published in IEEE Trans. on Computers, IEEE Trans. on Knowledge and Data Engineering, IEEE Trans. on Software Engineering and IEEE Trans. on Reliability. He is the author of a book titled Manufactoring Yield Evaluation of VLSI/WSI Systems to be published by IEEE Computer Society Press.This research was supported in part by the Consiglio Nazionale delle Ricerche under grant 93.02294.CT12This author is also supported by a grant of the Human Capital and Mobility project of the European Community under contract No. 3702 CABERNET  相似文献   

13.
Sheng Wan 《Automatica》2002,38(1):33-46
The proper measure of closed-loop performance variation in the presence of model-plant mismatch is discussed in this paper. A generalized closed-loop error transfer function, which is a special representation of the dual Youla parameter and has a close relationship with the pointwise ν-gap metric, is proposed as the suitable means of representing closed-loop performance variation in case of plant perturbation, and the closed-loop performance variation measure is accordingly defined as its maximum singular value frequency by frequency. It is shown that this measure is essential and informative in characterizing closed-loop performance variation. This measure is also shown to be readily applicable to on-line closed-loop performance assessment or monitoring, even without the explicit model of the plant. Its variant, defined as the η-function, which features the relative performance variation as well as generalized stability margin variation with respect to the nominal plant, is also discussed.  相似文献   

14.
15.
This paper describes an accurate and efficient method to model and predict the performance of distributed/parallel systems. Various performance measures, such as the expected user response time, the system throughput and the average server utilization, can be easily estimated using this method. The methodology is based on known product form queueing network methods, with some additional approximations. The method is illustrated by evaluating performance of a multi-client multi-server distributed system. A system model is constructed and mapped to a probabilistic queueing network model which is used to predict its behavior. The effects of user think time and various design parameters on the performance of the system are investigated by both the analytical method and computer simulation. The accuracy of the former is verified. The methodology is applied to identify the bottleneck server and to establish proper balance between clients and servers in distributed/parallel systems.  相似文献   

16.
Three-dimensional simulation of discrete particle systems is performed by the discrete element method (DEM) software on the gLite-based BalticGrid infrastructure. The performance of a parallel algorithm for particles exchanging processors is investigated by using a number of benchmarks. Polydispersed particle systems are visualized by a novel grid e-service VizLitG designed for convenient access and interactive visualization of remote data files located on the grid. Partial dataset transfer from the storage element is implemented in the visualization e-service. The efficiency tests of VizLitG are performed on the datasets of different sizes. Two granular problems associated with triaxial compaction and hopper discharge are solved.  相似文献   

17.
Understanding the behavior of large scale distributed systems is generally extremely difficult as it requires to observe a very large number of components over very large time. Most analysis tools for distributed systems gather basic information such as individual processor or network utilization. Although scalable because of the data reduction techniques applied before the analysis, these tools are often insufficient to detect or fully understand anomalies in the dynamic behavior of resource utilization and their influence on the applications performance. In this paper, we propose a methodology for detecting resource usage anomalies in large scale distributed systems. The methodology relies on four functionalities: characterized trace collection, multi‐scale data aggregation, specifically tailored user interaction techniques, and visualization techniques. We show the efficiency of this approach through the analysis of simulations of the volunteer computing Berkeley Open Infrastructure for Network Computing architecture. Three scenarios are analyzed in this paper: analysis of the resource sharing mechanism, resource usage considering response time instead of throughput, and the evaluation of input file size on Berkeley Open Infrastructure for Network Computing architecture. The results show that our methodology enables to easily identify resource usage anomalies, such as unfair resource sharing, contention, moving network bottlenecks, and harmful short‐term resource sharing. Copyright © 2011 John Wiley & Sons, Ltd.  相似文献   

18.
With Moore’s law supplying billions of transistors on-chip, embedded systems are undergoing a transition from single-core to multi-core to exploit this high transistor density for high performance. However, the optimal layout of these multiple cores along with the memory subsystem (caches and main memory) to satisfy power, area, and stringent real-time constraints is a challenging design endeavor. The short time-to-market constraint of embedded systems exacerbates this design challenge and necessitates the architectural modeling of embedded systems to reduce the time-to-market by expediting target applications to device/architecture mapping. In this paper, we present a queueing theoretic approach for modeling multi-core embedded systems that provides a quick and inexpensive performance evaluation both in terms of time and resources as compared to the development of multi-core simulators and running benchmarks on these simulators. We verify our queueing theoretic modeling approach by running SPLASH-2 benchmarks on the SuperESCalar simulator (SESC). Results reveal that our queueing theoretic model qualitatively evaluates multi-core architectures accurately with an average difference of 5.6% as compared to the architectures’ evaluations from the SESC simulator. Our modeling approach can be used for performance per watt and performance per unit area characterizations of multi-core embedded architectures, with varying number of processor cores and cache configurations, to provide a comparative analysis.  相似文献   

19.
20.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号