首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
移动数据库的缓存技术使客户机通过缓存少量的服务器数据项,便能拥有较高的处理速度,而且,它也是解决移动数据库频繁断接性问题的关键技术。文章研究了缓存管理策略在移动数据库中的应用,重点研究了缓存管理策略三大关键问题:缓存粒度、缓存一致性以及缓存替换,提出了一种基于统计与优先级原则解决缓存一致性问题的新思路。  相似文献   

2.
缓存技术在移动数据库中的应用研究   总被引:6,自引:1,他引:6  
缓存技术是提高教据访问性能的经典技术,已经在计算技术的很多方面得到了成功的应用.在移动计算环境中,由于设备和网络资源的限制,缓存将发挥更加重要的作用.研究了移动数据库中的客户端缓存管理技术,重点讨论了缓存管理的缓存粒度、缓存一致性策略和缓存替换策略3个关键问题.  相似文献   

3.
The mobile computing environment is receiving increasing attention recently. We consider a mobile environment in which a collection of mobile clients accesses a stationary database server via a wireless channel. Due to the limited bandwidth of a wireless channel and the instability of the wireless network, caching of frequently accessed data items in a client's local storage becomes especially important for improving the performance and data availability of data access queries. In this paper, we discuss the limitations of existing caching mechanisms in a mobile environment and investigate issues that need to be addressed. We propose an adaptive caching model that could cope with the nature of a mobile environment and the low-bandwidth wireless media, supporting fast data access. We describe the adaptive cache replacement and refresh mechanisms; explain the implementation in the context of object-oriented databases; and illustrate the results of some exploratory experiments to demonstrate the feasibility of the mechanisms.  相似文献   

4.
在移动计算环境中,移动性和断接性是它的最主要的两个特点.为了支持断接操作,数据时常需要先存储到移动客户机中.缓存技术成为移动计算环境中的重要技术.本文研究了基于广播技术中客户机端缓存的管理技术,讨论了缓存管理的粒度问题、缓存一致性策略和缓存替换策略三个关键问题.  相似文献   

5.
The diversity of services delivered over wireless channels has increased people's desire in ubiquitously accessing these services from their mobile devices. However, a ubiquitous mobile computing environment faces several challenges such as scarce bandwidth, limited energy resources, and frequent disconnection of the server and mobile devices. Caching frequently accessed data is an effective technique to improve the network performance because it reduces the network congestion, the query delay, and the power consumption. When caching is used, maintaining cache consistency becomes a major challenge since data items that are updated on the server should be also updated in the cache of the mobile devices. In this paper we propose a new cache invalidation scheme called Selective Adaptive Sorted (SAS) cache invalidation strategy that overcomes the false invalidation problem that exists in most of the invalidation strategies found in the literature. The performance of the proposed strategy is evaluated and compared with the selective cache invalidation strategy and the updated invalidation report startegy found in the literature. Results showed that a significant cost reduction can be obtained with the proposed strategy when measuring performance metrics such as delay, bandwidth, and energy.  相似文献   

6.
Data caching at mobile clients is an important technique for improving the performance of wireless data dissemination systems. However, variable data sizes, data updates, limited client resources, and frequent client disconnections make cache management a challenge. We propose a gain-based cache replacement policy, Min-SAUD, for wireless data dissemination when cache consistency must be enforced before a cached item is used. Min-SAUD considers several factors that affect cache performance, namely, access probability, update frequency, data size, retrieval delay, and cache validation cost. The paper employs stretch as the major performance metric since it accounts for the data service time and, thus, is fair when items have different sizes. We prove that Min-SAUD achieves optimal stretch under some standard assumptions. Moreover, a series of simulation experiments have been conducted to thoroughly evaluate the performance of Min-SAUD under various system configurations. The simulation results show that, in most cases, the Min-SAUD replacement policy substantially outperforms two existing policies, namely, LRU and SAIU.  相似文献   

7.
An inherent limitation in mobile data access is due to the unreliable and low bandwidth wireless communication channel. Caching of useful database items from database server in local storage of mobile clients is effective in reducing data access latency and wireless bandwidth consumption. In the event of disconnection, cached data can also serve the purpose of partial query processing. In this paper, we present the implementation and evaluate a new caching mechanism for object-oriented database systems in a mobile environment called MODEC. MODEC possesses the capabilities of performing caching at multiple granularities and adapting to changes in data access pattern, providing improved performance through tolerating limited inconsistency to read-only transactions. This caching capabilities is supported via standard ODMG modeling constructs. The prototype of MODEC is implemented using ODE database. Empirical system performance results are obtained from experiments on the prototype with data from a real-life database. The results are validated against results obtained via detailed simulation studies on MODEC. Both sets of results are found to be consistent and are in favor of our MODEC mechanism in providing a feasible solution to the mobile data access problem under the constraints in a mobile environment.  相似文献   

8.
Location Dependent Information Services (LDISs), through which mobile clients can access location sensitive data such as weather information, traffic reports, and local news, are gaining increasing popularity in recent years. Due to limited client power and intermittent connectivity, caching is an important approach to improve the performance of LDISs. In this paper, we propose a cache replacement policy called Location Dependent Cooperative Caching (LDCC). Unlike existing location dependent cache replacement policies, the LDCC strategy applies a prediction model to approximate client movement behavior and a probabilistic transition model to analyze the communication cost. These models are used in the design of a cache replacement policy to improve system overall performance. Simulation results demonstrate that the proposed strategy significantly outperforms existing caching policies in providing LDIS in mobile ad hoc networks.  相似文献   

9.
In the mobile wireless computing environment of the future, a large number of users, equipped with low-powered palmtop machines, will query databases over wireless communication channels. Palmtop-based units will often be disconnected for prolonged periods of time, due to battery power saving measures; palmtops also will frequently relocate between different cells, and will connect to different data servers at different times. Caching of frequently accessed data items will be an important technique that will reduce contention on the narrow-bandwidth, wireless channel. However, cache individualization strategies will be severely affected by the disconnection and mobility of the clients. The server may no longer know which clients are currently residing under its cell, and which of them are currently on. We propose a taxonomy of different cache invalidation strategies, and study the impact of clients' disconnection times on their performance. We study ways to improve further the efficiency of the invalidation techniques described. We also describe how our techniques can be implemented over different network environments.  相似文献   

10.
This paper describes a mobility-aware dynamic database caching scheme for wireless mobile computing and communications. A mobile-floating agent scheme is proposed, in which caching techniques are cognizant of the mobile nature of mobile users and the location-sensitive nature of mobile systems. The mobile-floating agent maintains a second class cache in the fixed network and employs Barbara's invalidation reports broadcasting cache consistency strategies to maintain a dynamic cache consistent with the first class cache in the mobile client. The invalidation reports broadcasting scheme is combined with knowledge of the mobility behavior of each individual mobile user and broadcasts of invalidation reports only occur within the user's mobility area. The evaluation results show that, for a large system (200 cells), this scheme can reduce the system cost by more than 87%, for even highly mobile users, compared with a fully replicated database system.Recommended by: Daniel Barbara, Ravi Jain and Narayanan Krishnakumar  相似文献   

11.
Data caching on mobile clients is widely seen as an effective solution to improve system performance. In particular, cooperative caching, based on the idea of sharing and coordination of cache data among multiple users, can be particularly effective for information access in mobile ad hoc networks where mobile clients are moving frequently and network topology is changing dynamically. Most existing cache strategies perform replacement independently, and they seldom consider coordinated replacement and energy saving issues in the context of a mobile ad hoc network. In this paper, we analyse the impact of energy on designing a cache replacement policy and formulate the Energy-efficient COordinated cache Replacement Problem (ECORP) as a 0-1 knapsack problem. A dynamic programming algorithm called ECORP-DP and a heuristic algorithm called ECORP-Greedy are presented to solve the problem. Simulations, using both synthetic workload traces and real workload traces in our experiments, show that the proposed policies can significantly reduce energy consumption and access latency when compared to other replacement policies.  相似文献   

12.
移动环境下缓存弱一致性的研究   总被引:5,自引:0,他引:5  
在移动环境下,客户缓存为提高客户一服务器数据库系统的整体性能提供了有效途径。缓存与服务方数据的同步策略是缓存研究的重要内容。移动环境下,考虑到网络的带宽、开销和可靠性等因素,客户有可能允许缓存维护弱一致性,即允许缓存与服务方数据存在偏差。本文针对基于语义的缓存,给出客户限定偏差范围的方法,并且提出基于有效期的缓存同步算法。  相似文献   

13.
Web caching proxy servers are essential for improving web performance and scalability, and recent research has focused on making proxy caching work for database-backed web sites. In this paper, we explore a new proxy caching framework that exploits the query semantics of HTML forms. We identify two common classes of form-based queries from real-world database-backed web sites, namely, keyword-based queries and function-embedded queries. Using typical examples of these queries, we study two representative caching schemes within our framework: (i) traditional passive query caching, and (ii) active query caching, in which the proxy cache can service a request by evaluating a query over the contents of the cache. Results from our experimental implementation show that our form-based proxy is a general and flexible approach that efficiently enables active caching schemes for database-backed web sites. Furthermore, handling query containment at the proxy yields significant performance advantages over passive query caching, but extending the power of the active cache to do full semantic caching appears to be less generally effective.  相似文献   

14.
断接下查询的缓存处理   总被引:5,自引:0,他引:5  
吴婷婷  章文嵩  周兴铭 《计算机学报》2003,26(10):1393-1399
移动环境下,由于无线网络可靠性低、费用高,移动主机本身受电源、资源等方面的限制,移动主机经常会主动或被动地处于断接,即没有网络连接的状态.为了提高断接时移动客户对数据的访问能力,有效利用移动缓存,该文提出断接下基于语义缓存的查询处理QPID算法.该算法的主要思路是先找出缓存中与当前查询相关的缓存项,再通过对相关项数据的进一步处理获得缓存中满足查询的结果.试验表明,基于QPID算法的查询处理可以更好地满足断接下客户的查询请求.  相似文献   

15.
《Parallel Computing》2007,33(7-8):497-520
In this paper, we present a multi-query optimization framework based on the concept of active semantic caching. The framework permits the identification and transparent reuse of data and computation in the presence of multiple queries (or query batches) that specify user-defined operators and aggregations originating from scientific data-analysis applications. We show how query scheduling techniques, coupled with intelligent cache replacement policies, can further improve the performance of query processing by leveraging the active semantic caching operators. We also propose a methodology for functionally decomposing complex queries in terms of primitives so that multiple reuse sites are exposed to the query optimizer, to increase the amount of reuse. The optimization framework and the database system implemented with it are designed to be efficient irrespective of the underlying parallel and/or distributed machine configuration. We present experimental results highlighting the performance improvements obtained by our methods using real scientific data-analysis applications on multiple parallel and distributed processing configurations (e.g., single symmetric multiprocessor (SMP) machine, cluster of SMP nodes, and a Grid computing configuration).  相似文献   

16.
Exploiting Regularities in Web Traffic Patterns for Cache Replacement   总被引:2,自引:0,他引:2  
Cohen  Kaplan 《Algorithmica》2002,33(3):300-334
Abstract. Caching web pages at proxies and in web servers' memories can greatly enhance performance. Proxy caching is known to reduce network load and both proxy and server caching can significantly decrease latency. Web caching problems have different properties than traditional operating systems caching, and cache replacement can benefit by recognizing and exploiting these differences. We address two aspects of the predictability of traffic patterns: the overall load experienced by large proxy and web servers, and the distinct access patterns of individual pages. We formalize the notion of ``cache load' under various replacement policies, including LRU and LFU, and demonstrate that the trace of a large proxy server exhibits regular load. Predictable load allows for improved design, analysis, and experimental evaluation of replacement policies. We provide a simple and (near) optimal replacement policy when each page request has an associated distribution function on the next request time of the page. Without the predictable load assumption, no such online policy is possible and it is known that even obtaining an offline optimum is hard. For experiments, predictable load enables comparing and evaluating cache replacement policies using partial traces , containing requests made to only a subset of the pages. Our results are based on considering a simpler caching model which we call the interval caching model . We relate traditional and interval caching policies under predictable load, and derive (near)-optimal replacement policies from their optimal interval caching counterparts.  相似文献   

17.
Finding replacement candidates for accommodating a new object is an important research issue in web caching. Due to the new emerging factors in the transcoding proxy and the aggregate effect of caching multiple versions of the same multimedia object, this problem becomes more important and complex as audio and video applications have proliferated over the Internet, especially in the environment of mobile computing systems. This paper addresses coordinated cache replacement in transcoding proxies. First, we propose an original model which determines cache replacement candidates on all candidate nodes in a coordinated fashion with the objective of minimizing the total cost loss for linear topology. We formulate this problem as an optimization problem and present a low-cost optimal solution for deciding cache replacement candidates. Second, we extend this problem to solve the same problem for tree networks. Finally, we conduct extensive simulations to evaluate the performance of our solutions by comparing with existing models.  相似文献   

18.
Many network applications requires access to most up-to-date information. An update event makes the corresponding cached data item obsolete, and cache hits due to obsolete data items become simply useless to those applications. Frequently accessed but infrequently updated data items should get higher preference while caching, and infrequently accessed but frequently updated items should have lower preference. Such items may not be cached at all or should be evicted from the cache to accommodate items with higher preference. In wireless networks, remote data access is typically more expensive than in wired networks. Hence, an efficient caching scheme considers both data access and update patterns can better reduce data transmissions in wireless networks. In this paper, we propose a step-wise optimal update-based replacement policy, called the Update-based Step-wise Optimal (USO) policy, for wireless data networks to optimize transmission cost by increasing effective hit ratio. Our cache replacement policy is based on the idea of giving preference to frequently accessed but infrequently updated data, and is supported by an analytical model with quantitative analysis. We also present results from our extensive simulations. We demonstrate that (1) the analytical model is validated by the simulation results and (2) the proposed scheme outperforms the Least Frequently Used (LFU) scheme in terms of effective hit ratio and communication cost.  相似文献   

19.
现有大多数内容缓存算法需要对内容流行度的准确估计,这在动态移动网络环境中是较难实现的。提出考虑内容异构5G无线网络云对边混合缓存策略,设计优化了内容缓存位置,其可以是原始内容服务器、云单元(CUs)和基站(BSs)。采用Lyapunov优化方法解决了NP-hard缓存控制问题与CU缓存和BS缓存控制决策之间的紧密耦合问题,有助于改善和识别网络体系结构的层次性和Cus缓存与BSs缓存之间的隶属关系,同时新的分层网络架构能够通过机会性地开发以云为中心和以边缘为中心的缓存来提高内容缓存性能,支持高平均请求的内容数据速率。采用李雅普诺夫优化技术,可实现恒定分数的容量区域的所有到达率的有限服务延迟,进而实现缓存数据的快速读取。仿真结果显示,所提缓存策略在平均端到端服务延迟和负载降低率方面具有较为显著的优势。  相似文献   

20.
代理缓存一致性策略和替换策略的研究   总被引:5,自引:1,他引:5  
针对代理缓存的一致性策略和替换策略还没有很好地结合起来,从而影响了代理缓存系统的整体性能的现状,分别探讨了基于Internet的代理缓存一致性策略和替换策略的处理流程、性能评价指标和研究现状,进而给出将这两种策略结合起来的一致性一替换算法的处理流程,并提出陈旧命中比是其主要性能评价指标,能很好地衡量代理缓存的各种算法的优劣和代理缓存系统的整体性能。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号