首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 364 毫秒
1.
The resources’ heterogeneity and unbalanced capability, together with the diversity of resource requirements in cloud computing systems, have produced great contradictions between resources’ tight coupling characteristics and user’s multi-granularities requirements. We propose a resource virtualization model and its on-demand allocation oriented infrastructure mainly providing computing services to solve that problem. A loosely coupled resource environment centered on resource users is created to complete a mapping from physical view of resources to logic view of resources. Heuristic resource combination algorithm (HRCA) is proposed to transform physical resources to logic resources, which meets two requirements: randomness in combination and fluctuation control to the size of resources granularities. On the basis of the appraisal indexes presented for the on-demand allocation, resource matching algorithm (RMA), targeting at resource satisfaction with the highest resource utilization, is designed to reuse resources. RMA can satisfy users’ requirement in limited time and keep resource satisfaction in the highest level in the condition of logic resources granularities being less than their required size. Resource reconfiguration algorithm (RRA) is presented to implement resource matching in the condition that virtual computing resource pool cannot match granularities of resource requirements. RRA assures the lowest resource refusal rate and the greatest resource satisfaction. We verify the effectiveness, performance and accuracy of algorithms in implementing the goal of resource virtualization centered on resource users and on-demand allocation.  相似文献   

2.
郭怡  茅苏 《微机发展》2012,(2):80-84
云计算资源管理系统是用于接收来自云计算用户的资源请求,并且把特定的资源封装为服务提供给资源请求者。在云计算环境下,如何为资源请求者选择合适的资源是一个值得研究的课题。文中通过对云计算下现有的资源提供策略的分析,同时根据不同云提供者提供的计算资源的成本不同的特点,综合考虑资源的计算能力、可靠性和单位成本三点因素,提出了云计算下基于CRP算法的资源提供策略。这种资源提供策略既能提供满足用户资源请求的服务,也能降低云服务提供者的运营成本,从而获得更大收益。  相似文献   

3.
Grid computing is mainly helpful for executing high-performance computing applications. However, conventional grid resources sometimes fail to offer a dynamic application execution environment and this increases the rate at which the job requests of users are rejected. Integrating emerging virtualization technologies in grid and cloud computing facilitates the provision of dynamic virtual resources in the required execution environment. Resource brokers play a significant role in managing grid and cloud resources as well as identifying potential resources that satisfy users’ application requests. This research paper proposes a semantic-enabled CARE Resource Broker (SeCRB) that provides a common framework to describe grid and cloud resources, and to discover them in an intelligent manner by considering software, hardware and quality of service (QoS) requirements. The proposed semantic resource discovery mechanism classifies the resources into three categories viz., exact, high-similarity subsume and high-similarity plug-in regions. To achieve the necessary user QoS requirements, we have included a service level agreement (SLA) negotiation mechanism that pairs users’ QoS requirements with matching resources to guarantee the execution of applications, and to achieve the desired QoS of users. Finally, we have implemented the QoS-based resource scheduling mechanism that selects the resources from the SLA negotiation accepted list in an optimal manner. The proposed work is simulated and evaluated by submitting real-world bio-informatics and image processing application for various test cases. The result of the experiment shows that for jobs submitted to the resource broker, job rejection rate is reduced while job success and scheduling rates are increased, thus making the resource management system more efficient.  相似文献   

4.
Yi Wei  M. Brian Blake 《Computing》2016,98(5):523-538
A Cloud platform offers on-demand provisioning of virtualized resources and pay-per-use charge model to its hosted services to satisfy their fluctuating resource needs. Resource scaling in cloud is often carried out by specifying static rules or thresholds. As business processes and scientific jobs become more intricate and involve more components, traditional reactive or rule-based resource management methods are not able to meet the new requirements. In this paper, we extend our previous work on dynamically managing virtualized resources for service workflows in a cloud environment. Extensive experimental results of an adaptive resource management algorithm are reported. The algorithm makes resource management decisions based on predictive results and high level user specified thresholds. It is also able to coordinate resources among the component services of a workflow so that unnecessary resource allocations and terminations can be avoided. Based on observations from previous experiments, the algorithm is extended with a new resource merge strategy in order to prevent average resource size from shrinking. Simulation results from synthetic workload data demonstrated the effectiveness of the extension.  相似文献   

5.
Each cloud service provider provides only an interface of its own cloud infrastructure for enabling clients to use its cloud resources. However, there is a number of difficulties for cloud providers to ensure proper functioning. One of the main problems of a cloud provider is the lack of resources to support a huge number of on-demand resources provisioning. Thus, resources cannot be distributed among different cloud providers since the federation is not the basic operation of the cloud provider. The most efficient way to overcome this problem is to extend the interface's cloud provider with an automatic negotiation to dynamically form the best agreement between the different cloud providers based on the service level agreement. In this article, we propose an extension for the Open Cloud Computing Interface which is the standardized interface for the cloud computing to support the automatic negotiation between the different cloud providers. To prove the efficiency and the effectiveness of our approach, we implement a prototype to evaluate the key presented in this article.  相似文献   

6.
7.
刘晓霞  刘靖 《计算机应用》2015,35(12):3530-3535
针对如何充分利用云基础架构层资源,满足上层云应用系统租户对应用系统容错的需求多样性和高可靠性要求的问题,提出一种面向租户和云服务提供商的、基于虚拟机部署策略的云平台容错即服务方法。该方法根据租户的特定容错需求适配适合的容错方法及容错级别,据此计算并最优化云服务提供商的收益和资源使用量,在此基础上对提供容错服务的虚拟机进行优化部署,充分利用底层虚拟机资源为租户的云应用系统提供更为可靠的容错服务。实验结果表明,所提方法能够在保障云服务提供商收益的基础上,为多租户云应用系统实现更灵活且可靠性更高的容错服务。  相似文献   

8.
Cloud manufacturing is becoming an increasingly popular enterprise model in which computing resources are made available on-demand to the user as needed. Cloud manufacturing aims at providing low-cost, resource-sharing and effective coordination. In this study, we present a genetic algorithm (GA) based resource constraint project scheduling, incorporating a number of new ideas (enhancements and local search) for solving computing resources allocation problems in a cloud manufacturing system. A newly generated offspring may not be feasible due to task precedence and resource availability constraints. Conflict resolutions and enhancements are performed on newly generated offsprings after crossover or mutation. The local search can exploit the neighborhood of solutions to find better schedules. Due to its complex characteristics, computing resources allocation in a cloud manufacturing system is NP-hard. Computational results show that the proposed GA can rapidly provide a good quality schedule that can optimally allocate computing resources and satisfy users’ demands.  相似文献   

9.
Yang  Jian  Xiang  Zhen  Mou  Lisha  Liu  Shumu 《Multimedia Tools and Applications》2020,79(47-48):35353-35367

The virtualized resource allocation (mapping) algorithm is the core issue of network virtualization technology. Universal and excellent resource allocation algorithms not only provide efficient and reliable network resources sharing for systems and users, but also simplify the complexity of resource scheduling and management, improve the utilization of basic resources, balance network load and optimize network performance. Based on the application of wireless sensor network, this paper proposes a wireless sensor network architecture based on cloud computing. The WSN hardware resources are mapped into resources in cloud computing through virtualization technology, and the resource allocation strategy of the network architecture is proposed. The experiment evaluates the performance of the resource allocation strategy. The proposed heuristic algorithm is a distributed algorithm. The complexity of centralized algorithms is high, distributed algorithms can handle problems in parallel, and reduce the time required to get a good solution with limited traffic.

  相似文献   

10.
Resource reconstruction algorithms are studied in this paper to solve the problem of resource on-demand allocation and improve the efficiency of resource utilization in virtual computing resource pool. Based on the idea of resource virtualization and the analysis of the resource status transition, the resource allocation process and the necessity of resource reconstruction are presented. Resource reconstruction algorithms are designed to determine the resource reconstruction types, and it is shown that they can achieve the goal of resource on-demand allocation through three methodologies: resource combination, resource split, and resource random adjustment. The effects that the resource users have on the resource reconstruction results, the deviation between resources and requirements, and the uniformity of resource distribution are studied by three experiments. The experiments show that resource reconstruction has a close relationship with resource requirements, but it is not the same with current distribution of resources. The algorithms can complete the resource adjustment with a lower cost and form the logic resources to match the demands of resource users easily.  相似文献   

11.
A key requirement of the cloud platform is the reasonable deployment of its large-scale virtual machine infrastructure. The mapping relation between the virtual node and the physical node determines the specific resource distribution strategy and reliability of the virtual machine deployment. Resource distribution strategy has an important effect on performance, energy consumption, and guarantee of the quality of service of the computer, and serves an important role in the deployment of the virtual machine. To solve the problem of meeting the fault-tolerance requirement and guarantee high reliability of the application system based on the full use of the cloud resource under the prerequisite of various demands, the deployment framework of the feedback virtual machine in cloud platform facing the individual user’s demands of fault-tolerance level and the corresponding deployment algorithm of the virtual machine are proposed in this paper. Resource distribution strategy can deploy the virtual machine in the physical nodes where the resource is mutually complementary according to the users’ different requirements on virtual resources. The deployment framework of the virtual machine in this paper can provide a reliable computer configuration according to the specific fault-tolerance requirements of the user while considering the usage rate of the physical resources of the cloud platform. The experimental result shows that the method proposed in this paper can provide flexible and reliable select permission of fault-tolerance level to the user in the virtual machine deployment process, provide a pertinent individual fault-tolerant deployment method of the virtual machine to the user, and guarantee to meet the user service in a large probability to some extent.  相似文献   

12.
Cloud computing is an ascending technology that has introduced a new paradigm by rendering a rational computational model possible. It has changed the dynamics of IT consumption by means of a model that provides on-demand services over the Internet. Unlike the traditional hosting service, cloud computing services are paid for per usage and may expand or shrink based on demand. Such services are, in general, fully managed by cloud providers that require users nothing but a personal computer and an Internet access. In recent years, this model has attracted the attention of researchers, investors and practitioners, many of whom have proposed a number of applications, structures and fundamentals of cloud computing, resulting in various definitions, requirements and models. Despite the interest and advances in the field, issues such as security and privacy, service layer agreement, resource sharing, and billing have opened up new questions about the real gains of the model. Although cloud computing is based on a 50-year-old business model, evidence from this study indicates that cloud computing still needs to expand and overcome present limitations that prevent the full use of its potential. In this study, we critically review the state of the art in cloud computing with the aim of identifying advances, gaps and new challenges.  相似文献   

13.
Resource allocation is a complicated task in cloud computing environment because there are many alternative computers with varying capacities. The goal of this paper is to propose a model for task-oriented resource allocation in a cloud computing environment. Resource allocation task is ranked by the pairwise comparison matrix technique and the Analytic Hierarchy Process giving the available resources and user preferences. The computing resources can be allocated according to the rank of tasks. Furthermore, an induced bias matrix is further used to identify the inconsistent elements and improve the consistency ratio when conflicting weights in various tasks are assigned. Two illustrative examples are introduced to validate the proposed method.  相似文献   

14.
The development of a communication infrastructure has made possible the expansion of the popular massively multiplayer online games. In these games, players all over the world can interact with one another in a virtual environment. The arrival rate of new players to the game environment causes fluctuations and players always expect services to be available and offer an acceptable service-level agreement (SLA), especially in terms of response time and cost. Cloud computing emerged in the recent years as a scalable alternative to respond to the dynamic changes of the workload. In massively multiplayer online games applications, players are allowed to lease resources from a cloud provider in an on-demand basis model. Proactive management of cloud resources in the face of workload fluctuations and dynamism upon the arrival of players are challenging issues. This paper presents a self-learning fuzzy approach for proactive resource provisioning in cloud environment, where key is to predict parameters of the probability distribution of the incoming players in each period. In addition, we propose a self-learning fuzzy autoscaling decision-maker algorithm to compute the proper number of resources to be allocated to each tier in the massively multiplayer online games by applying the predicted workload and user SLA. We evaluate the effectiveness of the proposed approach under real and synthetic workloads. The experimental results indicate that the proposed approach is able to allocate resources more efficiently than other approaches.  相似文献   

15.
近年来,云计算的发展为数据中心带来了新的应用场景和需求.其中,虚拟化作为云服务的重要使能技术,对数据中心服务器I/O系统的性能、扩展性和设备种类多样性提出了更高的要求,沿用传统设备与服务器紧耦合的I/O架构将会导致资源冗余,数据中心服务器密度降低,布线复杂度增加等诸多问题.因此,文章围绕I/O资源池化架构的实现机制和方法展开研究,目标是解除设备与服务器之间的绑定关系,实现接入服务器对I/O资源的按需弹性化使用,从根本上解决云计算数据中心的I/O系统问题.同时,还提出了一种基于单根I/O虚拟化协议实现多根I/O资源池化的架构,该架构通过硬件的外设部件高速互连接口多根域间地址和标识符映射机制,实现了多个物理服务器对同一I/O设备的共享复用;通过虚拟I/O设备热插拔技术和多根共享管理机制,实现了虚拟I/O资源在服务器间的实时动态分配;采用现场可编程门阵列(Field-Programmable Gate Array)构建了该架构的原型系统.结果表明,该架构能够为各个共享服务器提供良好的I/O操作性能.  相似文献   

16.
云服务环境下最大特点是按需交付,通过虚拟化技术将相关资源构建统一调度池,并且按照用户需求为用户提供服务,因此,云服务具有并行计算、开放性以及按需交付特性.对于实训教学平台来说,在云计算环境下需要面对各种用户需求,如请求任务各种各样,实验任务类型不尽相同,设备资源存在较大差异,通过虚拟化技术来实现规范化管理何资源共享,对云资源进行调度来才能有效满足用户需求,为此,在本文中提出了云计算环境下实训教学平台动态迁移策略.策略设计了三层协同资源调度机制来实现对资源和任务管理,重点研究了任务分割、资源划分、资源调度策略等,在此基础上对系统进行仿真实验,验证云计算环境下实训教学平台动态迁移策略可行与有效性.  相似文献   

17.
Cloud computing has been widely adopted by enterprises because of its on-demand and elastic resource usage paradigm. Currently most cloud applications are running on one single cloud. However, more and more applications demand to run across several clouds to satisfy the requirements like best cost efficiency, avoidance of vender lock-in, and geolocation sensitive service. JointCloud computing is a new research initiated by Chinese institutes to address the computing issues concerned with multiple clouds. In JointCloud, users’ diverse and dynamic requirements on cloud resources are satisfied by providing users virtual cloud (VC) for special purposes. A virtual cloud for special purposes is in essence a user’s specific cloud working environment having the customized software stacks, configurations and computing resources readily available. This paper first introduces what is JointCloud computing and then describes the design rationales, motivation examples, mechanisms and enabling technologies of VC in JointCloud.  相似文献   

18.
The cloud architecture is usually composed of several XaaS layers—including Software as a Service (SaaS), Platform as a Service (PaaS) and Infrastructure as a Service (IaaS). The paper studies efficient resource allocation to optimize objectives of cloud users, IaaS provider and SaaS provider in cloud computing. The paper proposes the composition of different layers in the cloud, such as IaaS and SaaS, and its joint optimization for efficient resource allocation. The efficient resource allocation optimization problem is conducted by subproblems. The proposed cloud resource allocation optimization algorithm is achieved through an iterative algorithm. The experiments are conducted to compare the performance of proposed joint optimization algorithm for efficient resource allocation with other related works.  相似文献   

19.
The paper is to consider resource scheduling with conflicting objectives in the grid environment. The objectives of the grid users, the grid resources and the grid system clash with each other. Grid users want to access enough system resources to achieve the desired level of quality of service (QoS). Resource providers pay more attention to the performance of their resources. Our resource scheduling employs market strategies to determine which jobs are executed at what time on which resources and at what prices. A grid resource provider uses its utility function to maximize its profit and a grid user uses its utility function to complete tasks while minimizing its spending. The paper proposes grid system objective optimization scheduling that provides a joint optimization of objectives for both the resource provider and grid user, which combines the benefits of both resource provider objective optimization and user objective optimization. Experiments are designed to study the performances of three resource-scheduling optimization algorithms. Performance metrics are classified into efficiency metrics, utility metrics and time metrics.  相似文献   

20.
Cloud datacenters host hundreds of thousands of physical servers that offer computing resources for executing customer jobs. While the failures of these physical machines are considered normal rather than exceptional, in large-scale distributed systems and cloud datacenters evaluation of availability in a datacenter is essential for both cloud providers and customers. Although providing a highly available and reliable computing infrastructure is essential to maintaining customer confidence, cloud providers desire to have highly utilized datacenters to increase the profit level of delivered services. Cloud computing architectural solutions should thus take into consideration both high availability for customers and highly utilized resources to make delivering services more profitable for cloud providers. This paper presents a highly reliable cloud architecture by leveraging the 80/20 rule. This architecture uses the 80/20 rule (80% of cluster failures come from 20% of physical machines) to identify failure-prone physical machines by dividing each cluster into reliable and risky sub-clusters. Furthermore, customer jobs are divided into latency-sensitive and latency-insensitive types. The results showed that only about 1% of all requested jobs are extreme latency-sensitive and require availability of 99.999%. By offering services to revenue-generating jobs, which are less than 50% of all requested jobs, within the reliable subcluster of physical machines, cloud providers can make their businesses more profitable by preventing service level agreement violation penalties and improving their reputations.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号