首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
查询结果缓存可以对查询结果的文档标识符集合或者实际的返回页面进行缓存,以提高用户查询的响应速度,相应的缓存形式可以分别称之为标识符缓存或页面缓存。对于固定大小的内存,标识符缓存可以获得更高的命中率,而页面缓存可以达到更高的响应速度。该文根据用户查询访问的时间局部性和空间局部性,提出了一种新颖的基于时空局部性的层次化结果缓存机制。首先,该机制将固定大小的结果缓存划分为两层:页面缓存和标识符缓存。对于用户提交的查询,该机制会首先使用第一层的页面缓存进行应答,如果未能命中,则继续尝试使用第二层的标识符缓存。实验显示这种层次化的缓存机制较传统的仅依赖于单一缓存形式的机制,在平均查询响应时间上,取得了可观的性能提升:例如,相对单纯的页面缓存,平均达到9%,最好情况下达到11%。其次,该机制在标识符缓存的基础上,设计了一种启发式的预取策略,对用户查询检索的空间局部性进行挖掘。实验显示,这种预取策略的融合,能进一步促进检索系统性能的有效提升,从而最终建立起一套时空完备的、有效的结果缓存机制。  相似文献   

2.
Web对象访问特征模拟器的设计与实现   总被引:2,自引:0,他引:2  
石磊  陶永才 《计算机仿真》2006,23(1):133-136
Web缓存是一个提高Web性能非常有效的方法,它可以位于网络的不同位置:客户端,代理服务器端,服务器端。研究表明Web缓存命中率可以达到30%-50%。Web缓存在应用中最大的问题就是Web缓存管理,研究Web访问特征是有效进行Web缓存管理的基础。Web日志生成模拟器对于研究Web缓存系统有很大地帮助,目前有两种方法模拟生成Web访问日志:日志驱动方法,数学模拟方法。日志驱动方法利用对历史日志进行变换来模拟生成新的日志,数学模拟方法在充分研究Ⅵ协对象访问特征的基础上,通过建立数学模型来模拟生成Web日志。该文通过分析Web对象访问特征,采用数学模拟方法分别模拟了Web对象高频区及低频区流行度特征,Web对象大小重尾分布特征,Web访问的时间局部性特征;设计并实现了一个Web日志模拟生成器WEBSIM。该模拟器不仅可以模拟生成Web对象访问日志,而且具有较大的灵活性,为进一步研究Web缓存技术和预取技术提供依据。  相似文献   

3.
Distributed high-speed caching uses replication strategies to improve the parallel service efficiency in data-intensive services in cloud-based environments. This paper proposes a replication strategy based on the access pattern of tile in order to optimize load balancing for large-scale user access in cloud-based WebGISs. First, this strategy considers the access bias and repeatability involved in tile access, to cache tiles with higher popularities, and obtains a higher cache hit rate under limited distributed caches. Second, on the basis of the uneven distribution and temporal local changes of tile access, it generates hot tile replicas with a seat allocation scheme to efficiently allocate the distributed cache. Finally, it balances the load for tile access requests based on the spatial correlation and spatial locality in tile access patterns to achieve fast data extraction. Experimental results indicate that the proposed strategy can be adopted to handle numerous WebGISs tile requests in a cloud-based environment.  相似文献   

4.
A new cache architecture based on temporal and spatial locality   总被引:5,自引:0,他引:5  
A data cache system is designed as low power/high performance cache structure for embedded processors. Direct-mapped cache is a favorite choice for short cycle time, but suffers from high miss rate. Hence the proposed dual data cache is an approach to improve the miss ratio of direct-mapped cache without affecting this access time. The proposed cache system can exploit temporal and spatial locality effectively by maximizing the effective cache memory space for any given cache size. The proposed cache system consists of two caches, i.e., a direct-mapped cache with small block size and a fully associative spatial buffer with large block size. Temporal locality is utilized by caching candidate small blocks selectively into the direct-mapped cache. Also spatial locality can be utilized aggressively by fetching multiple neighboring small blocks whenever a cache miss occurs. According to the results of comparison and analysis, similar performance can be achieved by using four times smaller cache size comparing with the conventional direct-mapped cache.And it is shown that power consumption of the proposed cache can be reduced by around 4% comparing with the victim cache configuration.  相似文献   

5.
语义缓存的最小权值项LWI替换策略   总被引:5,自引:0,他引:5  
在客户-服务器数据库系统中,语义缓存是基于客户查询语义相关建立的一种客户缓存.语义缓存的内容由以往查询的结果以及相应的描述构成.针对语义缓存的特征,提出语义缓存下最小权值项LWI(1east weight itern)替换策略.该策略由缓存项投影属性的访问频率和缓存项与查询的条件匹配情况,结合数据访问的时间局部性考虑决定缓存项的权值,替换最小权值项.通过性能分析实验,在语义缓存中,基于LWI替换策略的系统性能要优于基于传统LRU和LFU替换策略的系统性能.  相似文献   

6.
Proxy caching of large multimedia objects on the edge of the Internet has become increasingly important for reducing network latency. For a large media object, such as a two-hour video, treating the whole media as a single object for caching is not appropriate. In this paper, we study three media segmentation approaches to proxy caching: fixed, pyramid, and skyscraper. Blocks of a media stream are grouped into various segments for cache management. The cache admission and replacement policies attach different caching priorities to individual segments, taking into account the access frequency of the media object and the segment distance from the start of the media. These caching policies give preferential treatment to the beginning segments. As such, most user requests can be quickly played back from the proxy servers without delay. Event-driven simulations are conducted to evaluate the segmentation approaches and compare them with whole media caching. The results show that: 1) compared with whole media caching, segmentation-based caching is more effective not only in increased byte-hit ratio but also in lowered fraction of requests that requires delayed start; 2) pyramid segmentation, where segment size increases exponentially, is the best segmentation approach; and 3) segmentation-based caching is especially advantageous when the cache size is limited, when the set of hot media objects changes over time, when the media file size is large, and when there are a large number of distinct media objects.  相似文献   

7.
The Internet now offers more than just simple information to the users. Decision makers can now issue analytical, as opposed to transactional, queries that involve massive data (such as, aggregations of millions of rows in a relational database) in order to identify useful trends and patterns. Such queries are often referred to as On-Line-Analytical Processing (OLAP). Typically, pages carrying query results do not exhibit temporal locality and, therefore, are not considered for caching at Internet proxies. In OLAP processing, this is a major problem as the cost of these queries is significantly larger than that of the transactional queries. This paper proposes a technique to reduce the response time for OLAP queries originating from geographically distributed private LANs and issued through the Web toward a central data warehouse (DW) of an enterprise. An active caching scheme is introduced that enables the LAN proxies to cache some parts of the data, together with the semantics of the DW, in order to process queries and construct the resulting pages. OLAP queries arriving at the proxy are either satisfied locally or from the DW, depending on the relative access costs. We formulate a cost model for characterizing the respective latencies, taking into consideration the combined effects of both common Web access and query processing. We propose a cache admittance and replacement algorithm that operates on a hybrid Web-OLAP input, outperforming both pure-Web and pure-OLAP caching schemes.  相似文献   

8.
In information-centric networking, in-network caching has the potential to improve network efficiency and content distribution performance by satisfying user requests with cached content rather than downloading the requested content from remote sources. In this respect, users who request, download, and keep the content may be able to contribute to in-network caching by sharing their downloaded content with other users in the same network domain (i.e., user-assisted in-network caching). In this paper, we examine various aspects of user-assisted in-network caching in the hopes of efficiently utilizing user resources to achieve in-network caching. Through simulations, we first show that user-assisted in-network caching has attractive features, such as self-scalable caching, a near-optimal cache hit ratio (that can be achieved when the content is fully cached by the in-network caching) based on stable caching, and performance improvements over in-network caching. We then examine the caching strategy of user-assisted in-network caching. We examine three caching strategies based on a centralized server that maintains all content availability information and informs each user of what to cache. We also examine three caching strategies based on each user’s content availability information. We first show that the caching strategy affects the distribution of upload overhead across users and the number of cache hits in each segment. One interesting observation is that, even with a small storage space (i.e., 0.1% of the content size per user), the centralized and distributed approaches improve the cache hit ratio by 50% and 45%, respectively. With an overall view of caching information, the centralized approach can achieve a higher cache hit ratio than the distributed approach. Based on this observation, we discuss a distributed approach with a larger view of caching information than the distributed approach and, through simulations, confirm that a larger view leads to a higher cache hit ratio. Another interesting observation is that the random distributed strategy yields comparable performance to more complex strategies.  相似文献   

9.
Web预取技术综述   总被引:11,自引:0,他引:11  
Web预取是减少用户访问延时、提高网络服务质量的关键技术之一,近年来已成为国内外的研究热点.通过利用WWW访问的空间局部性,Web预取使缓存机制从时间局部性向空间局部性扩展.归纳了Web预取技术的分类,概括和比较了不同类别的优势和局限性,给出了预取模型的基本框架及每部分的主要功能,并对各种评价标准进行了详细介绍.同时,深入分析和探讨了现有的几种典型预取算法,系统地比较了这些算法的优缺点.最后从在线性、协作预取、动态流行度、划分用户会话和基于语义与基于路径相结合等方面指出了Web预取技术的研究方向.  相似文献   

10.
In a mobile computing environment, database servers disseminate information to multiple mobile clients via wireless channels. Due to the low bandwidth and low reliability of wireless channels, it is important for a mobile client to cache its frequently accessed database items into its local storage. This improves performance of database queries and improves availability of database items for query processing during disconnection. In this paper, we investigate issues on caching granularity, coherence strategy, and replacement policy of caching mechanisms for a mobile environment utilizing point-to-point communication paradigm.We first illustrate that page-based caching is not suitable in the mobile context due to the lack of locality among database items. We propose three different levels of caching granularity: attribute caching, object caching, and hybrid caching, a hybrid approach of attribute and object caching. Next, we show that existing coherence strategies are inappropriate due to frequent disconnection in a mobile environment, and propose a cache coherence strategy, based on the update patterns of database items. Via a detail simulation model, we examine the performance of various levels of caching granularity with our cache coherence strategy. We observe, in general, that hybrid caching could achieve a better performance. Finally, we propose several cache replacement policies that can adapt to the access patterns of database items. For each given caching granularity, we discover that our replacement policies outperform conventional ones in most situations.  相似文献   

11.
移动查询缓存处理的研究   总被引:5,自引:0,他引:5  
客户缓存为提高客户/服务器数据库系统整体性能以及客户方数据可用性提供了有效途径。移动环境下网络资源的贫乏使客户缓存的作用更为重要,语义缓存是基于客户查询语义相关建立的一类缓存,提出一个基于语义缓存的客户缓存机制,给出缓存的内容组织,提出缓存项合并策略;然后讨论了基于语义缓存的查询处理策略;最后,模拟结果表明该客户缓存机制能够提高分布式、特别是移动环境下客户服务器数据库系统的性能。  相似文献   

12.
针对NDN(命名数据网络)中确定性缓存和概率性缓存各自特点,提出一种确定性缓存和概率性缓存相结合的混合式NDN缓存策略(HDP)。基于区域划分的思想,在网络边缘采用基于热度的确定性缓存策略,在网络核心采用基于缓存收益和内容热度的概率性缓存策略,从而将两种缓存策略的优势相结合,进一步提高NDN缓存性能。实验表明该策略与现有NDN缓存方法相比,能有效提高缓存服务率和命中率,并有助于降低内容访问延迟,改善用户体验。  相似文献   

13.
Fragmental Proxy Caching for Streaming Multimedia Objects   总被引:1,自引:0,他引:1  
In this paper, a fragmental proxy-caching scheme that efficiently manages the streaming multimedia data in proxy cache is proposed to improve the quality of streaming multimedia services. The novel data-fragmentation method in this scheme not only provides finer granularity caching units to allow more effective cache replacement, but also offers a unique and natural way of handling the interactive VCR functions in the proxy-caching environment. Furthermore, a cache-replacement scheme, based on user request arrival rates for different multimedia objects and the playback rates of these objects, is proposed to address the drawbacks in existing cache-replacement schemes, most of which consider only the user access frequencies in their cache-replacement decisions. In this cache-replacement scheme, a sliding history window is employed to monitor the dynamic user request arrivals, and a tunable-victimization procedure is used to provide an excellent method of managing the cached multimedia data in accordance with different quality-of-service requirements of the streaming multimedia applications. Performance studies demonstrate that the fragmental proxy-caching scheme significantly outperforms other caching schemes, in terms of byte-hit ratio and the number of delayed starts and can be tuned to either maximize the byte-hit ratio or minimize the number of delayed starts  相似文献   

14.
现有以信息为中心的网络(Information-Centric Network,ICN)缓存的内容的时空分布不够合理,存在无效缓存及同质化缓存等问题。对此,文中提出一种基于内容流行度和社团重要度的ICN缓存与替换策略,其结合内容流行度与节点的社团重要度来选择缓存节点,把具有不同流行度的内容分散缓存在社团重要度不同的节点上,使缓存内容的空间分布趋于合理,从而增加了缓存内容的多样性。同时,基于社团局部流行度来替换缓存内容的策略更有利于优化缓存内容的时间分布,实现缓存内容空间分布的动态调整。实验结果表明,所提策略能有效地减少用户请求的平均响应时间,提高缓存对象的命中率和全网的缓存差异度。  相似文献   

15.
Web caching is a widely deployed technique to reduce the load to web servers and to reduce the latency for web browsers. Peer-to-Peer (P2P) web caching has been a hot research topic in recent years as it can create scalable and robust designs for decentralized internet-scale applications. However, many P2P web caching systems suffer expensive overheads such as lookup and publish messages, and lack locality awareness. In this paper, we present the development of a locality aware cache diffusion system that makes use of routing table locality, aggregation, and soft state to overcome these limitations. The analysis and experiments show that our cache diffusion system reduces the amount of information processed by nodes, reduces the number of index messages sent by nodes, and improves the locality of cache pointers.  相似文献   

16.
代理Web Cache性能分析   总被引:3,自引:0,他引:3  
采用WebCache技术提高当前Internet性能已成为一个主流的研究领域,其功能原理就象处理器和文件系统中的多级高速缓存一样。大规模Web高速缓存系统已成为许多国家Internet基础设施的重要组成部分。该文从三个不同访问规模的代理WebCache的跟踪日志出发,分析了WebCache的用户访问模式、Cache命中率、Cache服务器处理延迟等统计特征,提出基于分布式共享RAM和外存储结合的两级协同WebCache集群技术,可以提供可扩展的高性能并行Web高速缓存服务。  相似文献   

17.
网络化缓存是命名数据网络实现对信息的高效获取,有效降低互联网骨干网络流量的关键技术.网络化缓存将缓存作为普适的功能添加到每个网络节点.用户需要获取信息时,缓存有该内容的任意网络节点(例如路由器)接收到用户请求后都可直接向用户返回相应内容,提升用户请求响应效率.然而,命名数据网络采用泛在缓存使得内容发布者到用户的传输路径上的各节点对内容进行重复并无差别缓存,造成数据冗余、内容缓存无差别对待问题.为此,提出一种基于内容类型的隔跳概率缓存机制.首先根据业务特征(例如时延要求、带宽占用)将内容划分为4种类型:动态类、实时类、大数据类、以及小数据类;其次构造隔跳待定缓存策略,将数据存储在非连续的传输节点上,从空间上减少冗余缓存;最后针对不同内容提供差异化缓存服务:无缓存、网络边缘概率缓存、网络次边缘概率缓存、以及网络核心概率缓存策略,从而进一步降低冗余数据,同时提高用户获取内容的效率.实验结果表明,该机制能够减少冗余缓存,降低用户请求内容时延.  相似文献   

18.
Design of servers to meet the quality of service (QoS) requirements of interactive video-on-demand (VOD) systems is challenging. Recognizing the increasing use of these systems in a wide range of applications, as well as the stringent service demands expected from them, several design alternatives have been proposed to improve server throughput. A buffer management technique, called interval caching, is one such solution which exploits the temporal locality of requests to the same movie and tries to serve requests from the cache, thereby enhancing system throughput.In this paper, we present a comprehensive mathematical model for analyzing the performance of interactive video servers that use interval caching. The model takes into account the representative workload parameters of interactive servers employing interval caching and calculates the expected number of cached streams as an indication of the improvement in server capacity due to interval caching. Especially, user interactions, which sensitively affect the performance of interval caching, are realistically reflected in our model for an accurate analysis. A statistical admission control technique has also been developed based on this model. Using this model as a design tool, we apply the model to measure the impact of different VCR operations on client requests and rejection probability, as well as the effect of cache size.  相似文献   

19.
针对空间信息网络(space information network,SIN)中卫星节点缓存容量有限,且卫星高速移动使得星间链路时变,导致地面用户内容访问延迟增大的问题,提出一种基于人工蜂群算法的空间信息网络缓存决策策略(satellite improved artificial bee colony,SIABC)。首先根据低轨卫星节点链路切换的周期性和可预知性,建立网络分区模型,对空间信息网络的卫星节点进行分区,在此基础上,建立区域节点协作缓存模型,使得整个网络区域内的卫星节点有选择性地缓存不同流行度的内容,同时兼顾区域内节点的协作缓存,从而流行度高的内容缓存在网络边缘。仿真结果表明,与现有缓存机制相比,该缓存策略能够明显提高内容的平均缓存命中率,并显著降低用户的内容访问时延。  相似文献   

20.
缓存技术能有效的节省网络带宽,减少用户的访问延迟.在分布式缓存系统中,一个值得研究的问题是如何根据用户的请求动态的进行缓存部署,使得系统的收益最大.描述了缓存部署问题并建立了优化模型,在此基础上提出一种新的协作缓存部署算法,该算法利用对象的热度、网络距离,以及系统中各节点接收的请求和系统缓存分布信息,依次对请求路径上的节点进行缓存部署决策,同时该算法将计算分布到请求路径的各个节点上进行.仿真结果表明,该算法具有比LRU和Graph算法更高的缓存命中率和更低的访问延迟.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号