共查询到17条相似文献,搜索用时 137 毫秒
1.
分布式Web服务器系统(DWSS)是提高网络服务性能的一种前沿方式.文中分析了当前几种在客户端、主机端提升Web服务性能的方法,提出了一种基于内容的缓存代理协作分布式Web服务器系统模型.该模型可以有效减少用户获得响应的时间,避免了服务器之间不必要的通信,提高了Web服务器整体的性能. 相似文献
2.
3.
分布式Web服务器系统(DWSS)是提高网络服务性能的一种前沿方式。文中分析了当前几种在客户端、主机端提升Web服务性能的方法,提出了一种基于内容的缓存代理协作分布式Web服务器系统模型。该模型可以有效减少用户获得响应的时间,避免了服务器之间不必要的通信,提高了Web服务器整体的性能。 相似文献
4.
集中管理式Web缓存系统及性能分析 总被引:5,自引:0,他引:5
共享缓存文件是减少网络通信量和服务器负载的重要方法,本文在介绍Web Caching技术及流行的Web缓存通信协议ICP的基础上,提出了一种集中管理式Web缓存系统,该系统通过将用户的HTTP请求,按照一定的算法分发到系统中某一合适的缓存服务器上,从而消除了缓存系统内部服务器之间庞大的通信开销及缓存处理负担,减少了缓存内容的冗余度.通过分析,证明了集中管理式Web缓存系统比基于ICP的简单缓存系统具有缓存效率高、处理开销低、延迟小等优点,并且该系统具有良好的可扩展性. 相似文献
5.
6.
7.
8.
一种基于分散协作的Web缓存集群体系结构 总被引:1,自引:0,他引:1
Web对象缓存技术是一种减少Web访问通信量和访问延迟的重要手段,该文通过分析现有的各种Web缓存系统,提出了一种基于分散协作的Web缓存集群体系结构。该体系结构克服了集中式系统需要额外配备一台管理服务器的缺陷,消除了管理服务器瓶颈失效造成系统瘫痪的危险,减少由于管理服务器带来的延迟;同时消除了分散系统的缓存不命中情况下的多级转发的延迟和缓存内容重叠,提高了资源利用率和系统效率,具有良好的可扩展性和健壮性。 相似文献
9.
10.
在对服务器集群Web QoS控制基础上,综合考虑请求内容和各服务器性能以及当前整个集群负载平衡状况,设计了一种基于L4/L7双层分配的混合负载平衡调度策略,算法引入了一个反馈环节动态地改变Web服务器的权值,通过负载平衡程度的阈值进行判断,选择不同的调度策略,从而提高了Web集群系统的性能。 相似文献
11.
12.
一种Web集群系统下的QoS控制策略 总被引:2,自引:0,他引:2
本文提出了一种在Web集群环境下的QoS控制策略。集群前端分配器以会话单位来分配访问请求,保证后端服务器间负载的合理均衡分配,同时后端服务器采用基于会话的QoS控制策略,实现对集群系统的服务质量控制。 相似文献
13.
基于内容分发的Cache集群系统 总被引:1,自引:0,他引:1
文章设计和实现了大规模基于内容请求分发的系统。一个VS/DR或VS/TUN调度器作为系统的单一入口点,将请求分发给一组KTCPVS分发器,KTCPVS分发器通过使用中心的内容位置服务来确定处理该请求的后端服务器,然后,再将该请求分发到后端服务器,并提出基于局部性的最小负载调度算法。最后将该体系结构应用到大规模Cache集群系统中加以验证。 相似文献
14.
动态Web Cache系统研究与探讨 总被引:1,自引:0,他引:1
在国内外Web Cache技术研究的基础上,针对日趋严峻的Web信息系统的性能问题,提出构建动态Web Cache系统提高系统性能,并介绍了动态Web Cache的优点和缺点、实现动态Web Cache系统的关键技术,对于有关核心问题将作进一步的探讨。最后对这一技术的应用前景进行了探讨。 相似文献
15.
《Computer Networks》2002,38(1):75-97
We describe the design, implementation and performance of a high-performance Web server accelerator which runs on an embedded operating system and improves Web server performance by caching data. It can serve Web data at rates an order of magnitude higher than that which would be achieved by a high-performance Web server running on similar hardware under a conventional operating system such as Unix or NT. The superior performance of our system results in part from its highly optimized communications stack. In order to maximize hit rates and maintain updated caches, our accelerator provides an API which allows application programs to explicitly add, delete, and update cached data. The API allows our accelerator to cache dynamic as well as static data. We describe how our accelerator can be scaled to multiple processors to increase performance and availability. The basic design alternatives include a content router or a TCP router (without content routing) in front of a set of Web cache accelerator nodes, with the cache memory distributed across the accelerator nodes. Content-based routing reduces cache node CPU cycles but can make the front-end router a bottleneck. With the TCP router, a request for a cached object may initially be sent to the wrong cache node; this results in larger cache node CPU cycles, but can provide a higher aggregate throughput, because the TCP router becomes a bottleneck at a higher throughput than the content router. We quantify the throughput ranges in which different designs are preferable. We also examine a combination of content-based and TCP routing techniques. In addition, we present statistics from critical deployments of our accelerator for improving performance at highly accessed Sporting and Event Web sites hosted by IBM. 相似文献
16.
Performance evaluation of Web proxy cache replacement policies 总被引:10,自引:0,他引:10
The continued growth of the World-Wide Web and the emergence of new end-user technologies such as cable modems necessitate the use of proxy caches to reduce latency, network traffic and Web server loads. In this paper we analyze the importance of different Web proxy workload characteristics in making good cache replacement decisions. We evaluate workload characteristics such as object size, recency of reference, frequency of reference, and turnover in the active set of objects. Trace-driven simulation is used to evaluate the effectiveness of various replacement policies for Web proxy caches. The extended duration of the trace (117 million requests collected over 5 months) allows long term side effects of replacement policies to be identified and quantified.
Our results indicate that higher cache hit rates are achieved using size-based replacement policies. These policies store a large number of small objects in the cache, thus increasing the probability of an object being in the cache when requested. To achieve higher byte hit rates a few larger files must be retained in the cache. We found frequency-based policies to work best for this metric, as they keep the most popular files, regardless of size, in the cache. With either approach it is important that inactive objects be removed from the cache to prevent performance degradation due to pollution. 相似文献
17.
Web caches are traditionally organised in a simple tree like hierarchy. In this paper, a new architecture is proposed, where federations of caches are distributed globally, caching data partially. The advantages of the proposed system are that contention on global caches is reduced, while at the same time improving the scalability of the system since extra cache resources can be added on the fly. Among other topics discussed in this papers, is the scalability of the proposed system, the algorithms used to control the federation of Web caches and the approach used to identify the potential Web cache partners. In order to obtain a successful collaborative Web caching system, the formation of federations must be controlled by an algorithm that takes the dynamics of the Internet traffic into consideration. We use the history of Web cache access in order to determine how federations should be formed. Initial performance results of a simulation of a number of nodes are promising. 相似文献