首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Modern Web-based application infrastructures are based on clustered multitiered architectures, where request distribution occurs in two sequential stages: over a cluster of Web servers and over a cluster of application servers. Much work has focused on strategies for distributing requests across a Web server cluster in order to improve the overall throughput across the cluster. The strategies applied at the application layer are the same as those at the Web server layer because it is assumed that they transfer directly. In this paper, we argue that the problem of distributing requests across an application server cluster is fundamentally different from the Web server request distribution problem due to core differences in request processing in Web and application servers. We devise an approach for distributing requests across a cluster of application servers such that the overall system throughput is enhanced, and load across the application servers is balanced.  相似文献   

2.
State-of-the-art cluster-based data centers consisting of three tiers (Web server, application server, and database server) are being used to host complex Web services such as e-commerce applications. The application server handles dynamic and sensitive Web contents that need protection from eavesdropping, tampering, and forgery. Although the secure sockets layer (SSL) is the most popular protocol to provide a secure channel between a client and a cluster-based network server, its high overhead degrades the server performance considerably and, thus, affects the server scalability. Therefore, improving the performance of SSL-enabled network servers is critical for designing scalable and high-performance data centers. In this paper, we examine the impact of SSL offering and SSL-session-aware distribution in cluster-based network servers. We propose a back-end forwarding scheme, called ssl_with_bf, that employs a low-overhead user-level communication mechanism like virtual interface architecture (VIA) to achieve a good load balance among server nodes. We compare three distribution models for network servers, round robin (RR), ssl_with_session, and ssl_with_bf, through simulation. The experimental results with 16-node and 32-node cluster configurations show that, although the session reuse of ssl_with_session is critical to improve the performance of application servers, the proposed back-end forwarding scheme can further enhance the performance due to better load balancing. The ssl_with_bf scheme can minimize the average latency by about 40 percent and improve throughput across a variety of workloads.  相似文献   

3.
To improve response time of a Web site, one replicates the site on multiple servers. The effectiveness of a replicated server system will depend on how the incoming requests are distributed among replicas. A large number of load‐balancing strategies for Web server systems have been proposed. In this paper we describe a testbed that can be used to evaluate the performance of different load‐balancing strategies. The testbed uses a general architecture which allows different load‐balancing approaches to be supported easily. It emulates a typical World Wide Web scenario and allows variable load generation and performance measurement. We have performed some preliminary experiments to measure the performance of a few policies for load balancing using this testbed. Copyright © 2003 John Wiley & Sons, Ltd.  相似文献   

4.
David Rine   《Information Sciences》2003,150(3-4):123-151
Human exploration and development of space will involve opening the space frontier by exploring, using and enabling the development of space through information technology, while expanding the human experience into the far reaches of space. At that point in time we assert that the current primitive World Wide Web (Web) will be replaced and dramatically expanded into an Interstellar Space Wide Web (SWW). The current state-of-the-art low-orbits communications satellites constellations will be dramatically expanded to higher orbits and to orbits supporting work on other remote human colonies. This will be necessary in order to furnish in a human-friendly way the necessary software and information that will be needed in support of Interstellar Space Wide Information Technologies. Many of the problems encountered in conceiving of, modeling, designing and deploying such a facility will be different from those problems encountered in today’s Web. Future research and development work will be to identify some of these problems and to conceptually model a few of their solutions. In this work we describe research into the development of scalable tools and techniques that reduce the effort associated with component integration, both with respect to network environments and with respect to other components within the application. Our approach is also targeted at increasing the reusability of software components and software architectures. Our research is to investigate current problems in leveraging adapters as a means to configure of large-scale Next Generation distributed systems software from reusable architectures and components. Our approach to solving this problem is through the development of a novel configuration model and network-aware runtime environment called SWWACXML, for Space Wide Web Adapter Configuration eXtensible Markup Language. SWWACXML provides support for cross-layer architectural configuration at both the application level and the level of individual network connections. The language associated with this environment captures component interaction properties and network level quality-of-service (QoS) constraints. Adapters will be generated automatically from SWWACXML specifications. These adapters are part of the SWWACXML runtime system. The runtime system includes facilities for automatic configuration and runtime reconfiguration, as well as efficient management of network connections and QoS options. This facilitates reuse because components are not tied to interactions or environments. Another aspect of this work will focus on development and experimentation with a novel Web-based interaction paradigm that allows client adapters to tailor themselves to servers at runtime. One of the strengths of our approach is that clients do not have to be tied to specific servers at implementation time. Rather, a client’s adapter loads an SWWACXML configuration page from the server. The SWWACXML configuration page defines the appropriate interaction, including management of heterogeneous network QoS options. We believe our approach is amenable to facilitate a style of dynamic reconfiguration, where clients can at runtime change server, communication or interaction protocol. Future researchers and developers will design and implement the SWWACXML system in a distributed test-bed. They will develop performance analysis techniques to judge the success and efficiency of our approach.  相似文献   

5.
Replication of information across multiple servers is becoming a common approach to support popular Web sites. A distributed architecture with some mechanisms to assign client requests to Web servers is more scalable than any centralized or mirrored architecture. In this paper, we consider distributed systems in which the Authoritative Domain Name Server (ADNS) of the Web site takes the request dispatcher role by mapping the URL hostname into the IP address of a visible node, that is, a Web server or a Web cluster interface. This architecture can support local and geographical distribution of the Web servers. However, the ADNS controls only a very small fraction of the requests reaching the Web site because the address mapping is not requested for each client access. Indeed, to reduce Internet traffic, address resolution is cached at various name servers for a time-to-live (TTL) period. This opens an entirely new set of problems that traditional centralized schedulers of parallel/distributed systems do not have to face. The heterogeneity assumption on Web node capacity, which is much more likely in practice, increases the order of complexity of the request assignment problem and severely affects the applicability and performance of the existing load sharing algorithms. We propose new assignment strategies, namely adaptive TTL schemes, which tailor the TTL value for each address mapping instead of using a fixed value for all mapping requests. The adaptive TTL schemes are able to address both the nonuniformity of client requests and the heterogeneous capacity of Web server nodes. Extensive simulations show that the proposed algorithms are very effective in avoiding node overload, even for high levels of heterogeneity and limited ADNS control  相似文献   

6.
Providing World Wide Web or FTP service on a small scale is a well-solved problem, but scaling server technology for a site that accepts millions of connections per day can easily push multiple machines and networks to the bleeding edge. In this case study, we describe concrete configuration techniques that helped us get the best possible performance out of server resources at Netscape Communications. We designed, implemented, and babysat the Netscape Web site from its inception until traffic grew to about five million hits per day. The site now gets more than 120 million hits per day. We describe many iterations through the loop of watching the load increase, finding various failures, and fixing what was broken. We hope our experience will help readers configure a high-performance server from the start and then know what to look for when it becomes overloaded  相似文献   

7.
Multimedia proxy plays an important role in multimedia streaming over wireless Internet. Since wireless network exhibits different characteristics from the Internet, multimedia proxy caching over wireless Internet faces additional challenges. In this paper, we present a study of cache replacement for a single server and server selection for multiple servers across wireless Internet. By considering multiple objectives of multimedia proxy, we design a unified cost metric to measure proxy performance in wireless Internet. Based on the defined unified cost metric, we propose a novel replacement algorithm for single-server and a new server-selection policy for multiple servers to improve the end-to-end performance such as throughput, media quality, and start-up latency. To effectively handle errors occurred on wireless link, channel-adaptive unequal error protection is deployed according to distinct quality of service requirements of layered or scalable media. Simulation results demonstrate that our approaches achieve significantly better performance than the known cache-replacement algorithms and sever selection schemes, respectively.  相似文献   

8.
The phenomenal growth that the World Wide Web (WWW) service currently experiences, necessitates the adaptation of legacy information systems like RDBMSs or full-text search systems to HTTP servers. Existing standards for performing this adaptation (i.e. CGI), although well-established, prove highly inefficient throughout periods of heavy load (multiple hits per unit time). In this paper, after reviewing all the relevant mechanisms, we propose a generic architecture which adheres to the existing standards and client/server model and alleviates the performance handicap of classical database gateways. The performance evaluation which was realised as part of this research effort revealed a noteworthy superiority of the proposed architecture with respect to monolithic (non-client/server) CGI based approaches.  相似文献   

9.
Concurrency control is one of the key problems in design and implementation of collaborative systems such as hypertext/hypermedia systems, CAD/CAM systems, and software development environments. Most existing systems store data in specialized databases with built-in concurrency control policies, usually implemented via locking. It is desirable to construct such collaborative systems on top of the World Wide Web, but most Web servers do not support even conventional transactions, let alone distributed (multi-Website) transactions or flexible concurrency control mechanisms oriented toward team work-such as event notification, shared locks, and fine granularity locks. We present a transaction server that operates independently of Web servers or the collaborative systems, to fill the concurrency control gap. By default, the transaction server enforces the conventional atomic transaction model, where sets of operations are performed in an all-or-nothing fashion and isolated from concurrent users. The server can be tailored dynamically to apply more sophisticated concurrency control policies appropriate for collaboration. The transaction server also supports applications employing information resources other than Web servers, such as legacy databases, CORBA objects, and other hypermedia systems. Our implementation permits a wide range of system architecture styles  相似文献   

10.
WebClip (on-line demo at http://www.ctr.columbia.edu/webclip) is a compressed video searching and editing system operating over the World Wide Web. WebClip uses a distributed client-server model including a server engine for content analysis/editing, and clients for interactive controls of video browsing/editing. It specializes several unique features, including compressed-domain video feature extraction and manipulation, multi-resolution video access, content based video browsing/retrieval, and a distributed network architecture.  相似文献   

11.
This paper presents the embedded realization and experimental evaluation of a media stream scheduler on network interface (NI) CoProcessor boards. When using media frames as scheduling units, the scheduler is able to operate in real-time on streams traversing the CoProcessor, resulting in its ability to stream video to remote clients at real-time rates. This paper presents a detailed evaluation of the effects of placing application or kernel-level functionality, like packet scheduling on NIs, rather than the host machines to which they are attached. The main benefits of such placement are: 1) that traffic is eliminated from the host bus and memory subsystem, thereby allowing increased host CPU utilization for other tasks, and 2) that NI-based scheduling is immune to host-CPU loading, unlike host-based media schedulers that are easily affected even by transient load conditions. An outcome of this work is a proposed cluster architecture for building scalable media servers by distributing schedulers and media stream producers across the multiple NIs used by a single server and by clustering a number of such servers using commodity network hardware and software.  相似文献   

12.
Banga  Gaurav  Druschel  Peter 《World Wide Web》1999,2(1-2):69-83
The World Wide Web and its related applications place substantial performance demands on network servers. The ability to measure the effect of these demands is important for tuning and optimizing the various software components that make up a Web server. To measure these effects, it is necessary to generate realistic HTTP client requests in a test‐bed environment. Unfortunately, the state‐of‐the‐art approach for benchmarking Web servers is unable to generate client request rates that exceed the capacity of the server being tested, even for short periods of time. Moreover, it fails to model important characteristics of the wide area networks on which most servers are deployed (e.g., delay and packet loss). This paper examines pitfalls that one encounters when measuring Web server capacity using a synthetic workload. We propose and evaluate a new method for Web traffic generation that can generate bursty traffic, with peak loads that exceed the capacity of the server. Our method also models the delay and loss characteristics of WANs. We use the proposed method to measure the performance of widely used Web servers. The results show that actual server performance can be significantly lower than indicated by standard benchmarks under conditions of overload and in the presence of wide area network delays and packet losses. This revised version was published online in August 2006 with corrections to the Cover Date.  相似文献   

13.
一个虚拟Internet服务器的设计与实现   总被引:11,自引:0,他引:11  
针对已有的解决Internet服务器性能瓶颈和可靠性问题的方法所存在的不足,提出基于IP层负载平衡调度的解决方法,将一组服务器构成一个可伸缩的、高可用的虚拟Internet服务器.通过在服务机群中透明地加入和删除结点以实现系统的伸缩性;通过检测结点或服务进程故障和正确地重置系统达到高可用性.详细讨论了虚拟Internet服务器的体系结构、设计方法和实现技术,并给出了相应的性能测试结果.  相似文献   

14.
丁晓光  古志民 《计算机工程》2007,33(15):112-114
Linux virtual server是典型的服务器机群结构,通过双前端负载分配器和后端服务器池提供高可用服务。当后端服务器池发生故障时,故障服务器的工作量将通过负载分配算法重新分配到其他可用的服务器上,结果导致服务器池负担变重,进而可用服务器的故障概率增加。该文设计并实现了高可用N+1冗余结构模型系统,改进了典型的服务器池的上述缺陷。实验表明模型有着极低的资源占用率和非常短的故障切换时间。  相似文献   

15.
基于机群的可扩展计算机网络已逐渐成为高性能网络服务器架构的基础。文章对几种典型的机群构架进行了研究,包括RR-DNS—Web服务器 AFS服务器、TCP路由器 Web服务器、主-从服务器等,对它们的并行性问题,例如可扩展性、可靠性以及负载平衡策略等,作了深入的比较分析。  相似文献   

16.
基于Web内容的集群服务器请求分配系统的研究与实现   总被引:1,自引:0,他引:1  
随着基于Internet上的Web应用服务迅速发展,提供高性能、高可靠性和高扩展性的各种Web服务已成为用户的迫切需求。通过对现有的请求分配策略的分析,提出一种综合考虑后端服务器的缓存局域性的基于Web请求内容的集群服务器负载分发策略的设计与实现。  相似文献   

17.
一种Web集群系统下的QoS控制策略   总被引:2,自引:0,他引:2  
本文提出了一种在Web集群环境下的QoS控制策略。集群前端分配器以会话单位来分配访问请求,保证后端服务器间负载的合理均衡分配,同时后端服务器采用基于会话的QoS控制策略,实现对集群系统的服务质量控制。  相似文献   

18.
High-performance Web sites rely on Web server `farms', hundreds of computers serving the same content, for scalability, reliability, and low-latency access to Internet content. Deploying these scalable farms typically requires the power of distributed or clustered file systems. Building Web server farms on file systems complements hierarchical proxy caching. Proxy caching replicates Web content throughout the Internet, thereby reducing latency from network delays and off-loading traffic from the primary servers. Web server farms scale resources at a single site, reducing latency from queuing delays. Both technologies are essential when building a high-performance infrastructure for content delivery. The authors present a cache consistency model and locking protocol customized for file systems that are used as scalable infrastructure for Web server farms. The protocol takes advantage of the Web's relaxed consistency semantics to reduce latencies and network overhead. Our hybrid approach preserves strong consistency for concurrent write sharing with time-based consistency and push caching for readers (Web servers). Using simulation, we compare our approach to the Andrew file system and the sequential consistency file system protocols we propose to replace  相似文献   

19.
To reduce the environmental impact, it is essential to make data centers green, by turning off servers and tuning their speeds for the instantaneous load offered, that is, determining the dynamic configuration in web server clusters. We model the problem of selecting the servers that will be on and finding their speeds through mixed integer programming; we also show how to combine such solutions with control theory. For proof of concept, we implemented this dynamic configuration scheme in a web server cluster running Linux, with soft real-time requirements and QoS control, in order to guarantee both energy-efficiency and good user experience. In this paper, we show the performance of our scheme compared to other schemes, a comparison of a centralized and a distributed approach for QoS control, and a comparison of schemes for choosing speeds of servers.  相似文献   

20.
线程池是提高服务器程序性能的一种很好技术,已经在服务器端应用程序得到了广泛应用,比如Web服务器、数据库服务器、代理服务器和网络游戏服务器等。尤其面对海量并发连接时,线程池对能显著提高系统整体性能。该文提出了一种轻量级的、高效的线程池设计方案,在实际的网络服务器应用应用取得了非常好的效果。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号