共查询到20条相似文献,搜索用时 15 毫秒
1.
MobiStore is a P2P data store for decentralized mobile computing, designed to achieve high availability and load balance. As P2P platforms, mobile devices connected to the Internet through WiFi or cellular networks are different from wired devices in two main aspects: (1) higher churn due to mobility, weak wireless signals, or battery constraints, and (2) significant variability in bandwidth and latency based on the point of attachment. These problems affect the stored content availability and skew the content serving load over the peers. MobiStore structures the mobile P2P network into clusters of redundant peers. The topology uses both algorithmically-defined and random edges among the peers of different clusters. The routing information is updated using a gossip-based protocol. Thus, MobiStore achieves, with high probability, O(1) lookup operations despite high churn and link variability. Inside the clusters, all peers replicate the content, which improves the content availability. Furthermore, based on the current load, MobiStore dynamically changes the number of peers inside the clusters and routes content request to randomly selected peers. These two dynamic techniques along with using consistent hashing to map content to peers balance the load over the peers. While some of these techniques are well known, the main contribution is on the novel ways of applying them to design and implement an efficient mobile P2P data store. Simulation results show MobiStore achieves an availability, i.e., lookup success rate, between 12–48 % higher than two baseline systems built over the MR-Chord and Chord P2P protocols; and it reduces the latency up to 9 times compared to these protocols. Finally, the results show MobiStore adapts to churn and workload to evenly distribute the requests across clusters and peers better than both baseline solutions. 相似文献
3.
Most peer-to-peer lookup schemes keep a small amount of routing state per node, typically logarithmic in the number of overlay nodes. This design assumes that routing information at each member node must be kept small so that the bookkeeping required to respond to system membership changes is also small, given that aggressive membership dynamics are expected. As a consequence, lookups have high latency as each lookup requires contacting several nodes in sequence. In this paper, we question these assumptions by presenting a peer-to-peer routing algorithm with small lookup paths. Our algorithm, called “OneHop,” maintains full information about the system membership at each node, routing in a single hop whenever that information is up to date and in a small number of hops otherwise. We show how to disseminate information about membership changes quickly enough so that nodes maintain accurate complete membership information. We also present analytic bandwidth requirements for our scheme that demonstrate that it could be deployed in systems with hundreds of thousands of nodes and high churn. We validate our analytic model using a simulated environment and a real implementation. Our results confirm that OneHop is able to achieve high efficiency, usually reaching the correct node directly 99 percent of the time. 相似文献
4.
Gossip-based protocols provide a simple, scalable, and robust way to disseminate messages in large-scale systems. In such
protocols, messages are spread in an epidemic manner. Gossiping may take place between nodes using push, pull, or a combination.
Push-based systems achieve reasonable latency and high resilience to failures but may impose an unnecessarily large redundancy
and overhead on the system. At the other extreme, pull-based protocols impose a lower overhead on the network at the price
of increased latencies. A few hybrid approaches have been proposed—typically pushing control messages and pulling data—to
avoid the redundancy of high-volume content and single-source streams. Yet, to the best of our knowledge, no other system
intermingles push and pull in a multiple-senders scenario, in such a way that data messages of one help in carrying control
messages of the other and in adaptively adjusting its rate of operation, further reducing overall cost and improving both
on delays and robustness. In this paper, we propose an efficient generic push-pull dissemination protocol, P ulp, which combines the best of both worlds. P ulp exploits the efficiency of push approaches, while limiting redundant messages and therefore imposing a low overhead, as pull
protocols do. P ulp leverages the dissemination of multiple messages from diverse sources: by exploiting the push phase of messages to transmit
information about other disseminations, P ulp enables an efficient pulling of other messages, which themselves help in turn with the dissemination of pending messages.
We deployed P ulp on a cluster and on PlanetLab. Our results demonstrate that P ulp achieves an appealing trade-off between coverage, message redundancy, and propagation delay. 相似文献
5.
The phenomenon of system churn degrades the lookup performance of distributed hash table (DHT) systems greatly. To handle the churn, a number of approaches have been proposed to date. However, there is a lack of theoretical analysis to direct how to make design choices under different churn rates and how to configure their parameters optimally. In this paper, we analytically study three important aspects on optimizing DHT lookup performance under churn, i.e. lookup strategy, lookup parallelism and lookup key replication. Our objective is to build a theoretical basis for designers to make better design choices in the future. We first compare the performance of two representative lookup strategies—recursive routing and iterative routing—and explore the existence of better alternatives. Then we study the effectiveness of lookup parallelism in systems with different churn rates and show how to select the optimal degree of parallelism. Owing to the importance of key replication on lookup performance, we also analyze the reliability of the replicated key under two different replication policies, and show how to perform proper configuration. Besides the analytical study, our results are also validated by simulation, and Kad is taken as a case to show the meaningfulness of our analysis. Copyright © 2007 John Wiley & Sons, Ltd. 相似文献
6.
Since its invention, the Web has evolved into the largest multimedia repository that has ever existed. This evolution is a direct result of the explosion of user-generated content, explained by the wide adoption of social network platforms. The vast amount of multimedia content requires effective management and retrieval techniques. Nevertheless, Web multimedia retrieval is a complex task because users commonly express their information needs in semantic terms, but expect multimedia content in return. This dissociation between semantics and content of multimedia is known as the semantic gap. To solve this, researchers are looking beyond content-based or text-based approaches, integrating novel data sources. New data sources can consist of any type of data extracted from the context of multimedia documents, defined as the data that is not part of the raw content of a multimedia file. The Web is an extraordinary source of context data, which can be found in explicit or implicit relation to multimedia objects, such as surrounding text, tags, hyperlinks, and even in relevance-feedback. Recent advances in Web multimedia retrieval have shown that context data has great potential to bridge the semantic gap. In this article, we present the first comprehensive survey of context-based approaches for multimedia information retrieval on the Web. We introduce a data-driven taxonomy, which we then use in our literature review of the most emblematic and important approaches that use context-based data. In addition, we identify important challenges and opportunities, which had not been previously addressed in this area. 相似文献
7.
This paper presents a new computational model for real-time systems, called the clocked transition system ( CTS) model. The CTS model is a development of our previous timed transition model, where some of the changes are inspired by the model of timed automata. The new model leads to a simpler style of temporal specification and verification, requiring no extension of the temporal
language. We present verification rules for proving safety a nd liveness properties of clocked transition systems. All rules
are associated with verification diagrams. The verification of response properties requires adjustments of the proof rules developed for untimed systems, reflecting the fact that progress in the
real time systems is ensured by the progress of time and not by fairness. The style of the verification rules is very close
to the verification style of untimed systems which allows the (re)use of verification methods and tools, developed for u ntimed
reactive systems, for proving all interesting properties of real-time systems.
We conclude with the presentation of a branching-time based approach for verifying that an arbitrary given CTS is non-zeno.
Finally, we present an extension of the model and the invariance proof rule for hybrid systems.
Received: 23 September 1998 / 7 June 1999 相似文献
8.
Randomized DHT-based Peer-to-Peer (P2P) systems grant nodes certain flexibility in selecting their overlay neighbors, leading to irregular overlay structures but to better overall performance in terms of path latency, static resilience and local convergence. However, routing in the presence of overlay irregularity is challenging. In this paper, we propose a novel routing protocol, RASTER, that approximates shortest overlay routes between nodes in randomized DHTs. Unlike previously proposed routing protocols, RASTER encodes and aggregates routing information. Its simple bitmap-encoding scheme together with the proposed RASTER routing algorithm enable a performance edge over current overlay routing protocols. RASTER provides a forwarding overhead of merely a small constant number of bitwise operations, a routing performance close to optimal, and a better resilience to churn. RASTER also provides nodes with the flexibility to adjust the size of the maintained routing information based on their storage/processing capabilities. The cost of storing and exchanging encoded routing information is manageable and grows logarithmically with the number of nodes in the system. 相似文献
9.
Geometric model fitting is a typical chicken-&-egg problem: data points should be clustered based on geometric proximity to
models whose unknown parameters must be estimated at the same time. Most existing methods, including generalizations of RANSAC, greedily search for models with most inliers (within a threshold) ignoring overall classification of points. We formulate
geometric multi-model fitting as an optimal labeling problem with a global energy function balancing geometric errors and
regularity of inlier clusters. Regularization based on spatial coherence (on some near-neighbor graph) and/or label costs is NP hard.
Standard combinatorial algorithms with guaranteed approximation bounds (e.g. α-expansion) can minimize such regularization energies over a finite set of labels, but they are not directly applicable to
a continuum of labels, e.g. R2{\mathcal{R}}^{2} in line fitting. Our proposed approach ( PEaRL) combines model sampling from data points as in RANSAC with iterative re-estimation of inliers and models’ parameters based on a global regularization functional. This technique
efficiently explores the continuum of labels in the context of energy minimization. In practice, PEaRL converges to a good quality local minimum of the energy automatically selecting a small number of models that best explain
the whole data set. Our tests demonstrate that our energy-based approach significantly improves the current state of the art
in geometric model fitting currently dominated by various greedy generalizations of RANSAC. 相似文献
10.
Ontologies represent domain concepts and relations in a form of semantic network. Many research works use ontologies in the
information matchmaking and retrieval. This trend is further accelerated by the convergence of various information sources
supported by ontologies. In this paper, we propose a novel multi-modality ontology model that integrates both the low-level
image features and the high-level text information to represent image contents for image retrieval. By embedding this ontology
into an image retrieval system, we are able to realize intelligent image retrieval with high precision. Moreover, benefiting
from the soft-coded ontology model, this system has good flexibility and can be easily extended to the larger domains. Currently,
our experiment is conducted on the animal domain canine. An ontology has been built based on the low-level features and the domain knowledge of canine. A prototype retrieval system is set up to assess the performance. We compare our experiment results with traditional text-based
image search engine and prove the advantages of our approach. 相似文献
11.
In this paper, we present a query-driven indexing/retrieval strategy for efficient full text retrieval from large document collections distributed within a structured P2P network. Our indexing strategy is based on two important properties: (1) the generated distributed index stores posting lists for carefully chosen indexing term combinations that are frequently present in user queries, and (2) the posting lists containing too many document references are truncated to a bounded number of their top-ranked elements. These two properties guarantee acceptable latency and bandwidth requirements, essentially because the number of indexing term combinations remains scalable and the posting lists transmitted during retrieval never exceed a constant size. A novel index update mechanism efficiently handles adding of new documents to the document collection. Thus, the generated distributed index corresponds to a constantly evolving query-driven indexing structure that efficiently follows current information needs of the users and changes in the document collection.We show that the size of the index and the generated indexing/retrieval traffic remains manageable even for Web-size document collections at the price of a marginal loss in precision for rare queries. Our theoretical analysis and experimental results provide convincing evidence about the feasibility of the query-driven indexing strategy for large scale P2P text retrieval. 相似文献
12.
Alternating-time temporal logic ( atl) is a logic for reasoning about open computational systems and multi-agent systems. It is well known that atl model checking is linear in the size of the model. We point out, however, that the size of an atl model is usually exponential in the number of agents. When the size of models is defined in terms of states and agents rather than transitions, it turns out that the problem is (1) Δ
3
P
-complete for concurrent game structures, and (2) Δ
2
P
-complete for alternating transition systems. Moreover, for “Positive atl” that allows for negation only on the level of propositions, model checking is (1) Σ
2
P
-complete for concurrent game structures, and (2) NP-complete for alternating transition systems. We show a nondeterministic polynomial reduction from checking arbitrary alternating
transition systems to checking turn-based transition systems, We also discuss the determinism assumption in alternating transition
systems, and show that it can be easily removed.
In the second part of the paper, we study the model checking complexity for formulae of atl
with imperfect information ( atl
ir
). We show that the problem is Δ
2
P
-complete in the number of transitions and the length of the formula (thereby closing a gap in previous work of Schobbens
in Electron. Notes Theor. Comput. Sci. 85(2), 2004). Then, we take a closer look and use the same fine structure complexity measure as we did for atl with perfect information. We get the surprising result that checking formulae of atl
ir
is also Δ
3
P
-complete in the general case, and Σ
2
P
-complete for “Positive atl
ir
”. Thus, model checking agents’ abilities for both perfect and imperfect information systems belongs to the same complexity
class when a finer-grained analysis is used. 相似文献
13.
We present and analyze a simple and general scheme to build a churn (fault)-tolerant structured Peer-to-Peer (P2P) network. Our scheme shows how to “convert” a static network into a dynamic distributed hash table(DHT)-based P2P network such that all the good properties of the static network are guaranteed with high probability (w.h.p). Applying our scheme to a cube-connected cycles network, for example, yields a O(log N) degree connected network, in which every search succeeds in O(log N) hops w.h.p., using O(log N) messages, where N is the expected stable network size. Our scheme has an constant storage overhead (the number of nodes responsible for servicing a data item) and an O(log N) overhead (messages and time) per insertion and essentially no overhead for deletions. All these bounds are essentially optimal. While DHT schemes with similar guarantees are already known in the literature, this work is new in the following aspects: (1) It presents a rigorous mathematical analysis of the scheme under a general stochastic model of churn and shows the above guarantees; (2) The theoretical analysis is complemented by a simulation-based analysis that validates the asymptotic bounds even in moderately sized networks and also studies performance under changing stable network size; (3) The presented scheme seems especially suitable for maintaining dynamic structures under churn efficiently. In particular, we show that a spanning tree of low diameter can be efficiently maintained in constant time and logarithmic number of messages per insertion or deletion w.h.p. 相似文献
14.
In this paper, we present clear and formal definitions of ranking factors that should be concerned in opinion retrieval and
propose a new opinion retrieval model which simultaneously combines the factors from the generative modeling perspective.
The proposed model formally unifies relevance-based ranking with subjectivity detection at the document level by taking multiple
ranking factors into consideration: topical relevance, subjectivity strength, and opinion-topic relatedness. The topical relevance measures how strongly a document relates to a given topic, and the subjectivity strength indicates
the likelihood that the document contains subjective information. The opinion-topic relatedness reflects whether the subjective
information is expressed with respect to the topic of interest. We also present the universality of our model by introducing
the model’s derivations that represent other existing opinion retrieval approaches. Experimental results on a large-scale
blog retrieval test collection demonstrate that not only are the individual ranking factors necessary in opinion retrieval
but they cooperate advantageously to produce a better document ranking when used together. The retrieval performance of the
proposed model is comparable to that of previous systems in the literature. 相似文献
15.
We describe a system which supports dynamic user interaction with multimedia information using content-based hypermedia navigation
techniques, specialising in a technique for navigation of musical content. The model combines the principles of open hypermedia, whereby hypermedia link information is maintained by a link service, with content-based retrieval techniques in which a database is queried based on a feature of the multimedia content; our approach could be described as
‘content-based retrieval of hypermedia links’. The experimental system focuses on temporal media and consists of a set of
component-based navigational hypermedia tools. We propose the use of melodic pitch contours in this context and we present
techniques for storing and querying contours, together with experimental results. Techniques for integrating the contour database
with open hypermedia systems are also discussed. 相似文献
16.
Machine readable dictionaries ( Mrds) contain knowledge about language and the world essential for tasks in natural language processing ( Nlp). However, this knowledge, collected and recorded by lexicographers for human readers, is not presented in a manner for Mrds to be used directly for Nlp tasks. What is badly needed are machine tractable dictionaries ( Mtds): Mrds transformed into a format usable for Nlp. This paper discusses three different but related large-scale computational methods to transform Mrds into Mtds. The Mrd used is The Longman Dictionary of Contemporary English ( Ldoce). The three methods differ in the amount of knowledge they start with and the kinds of knowledge they provide. All require some handcoding of initial information but are largely automatic. Method I, a statistical approach, uses the least handcoding. It generates relatedness networks for words in Ldoce and presents a method for doing partial word sense disambiguation. Method II employs the most handcoding because it develops and builds lexical entries for a very carefully controlled defining vocabulary of 2,000 word senses (1,000 words). The payoff is that the method will provide an Mtd containing highly structured semantic information. Method III requires the handcoding of a grammar and the semantic patterns used by its parser, but not the handcoding of any lexical material. This is because the method builds up lexical material from sources wholly within Ldoce. The information extracted is a set of sources of information, individually weak, but which can be combined to give a strong and determinate linguistic data base. 相似文献
17.
In this study, a semi-empirical modified vegetation backscattering model was developed to retrieve leaf area index (LAI) based on multi-temporal Radarsat-2 data and ground observations collected in China. This model combined the contribution of the vegetation and bare soil at the pixel level by adding vegetation coverage and the influence of bare soil on the total backscatter coefficients. Then, a lookup table algorithm was applied to calculate the value of vegetation water content and retrieve the LAI based on the linear relationship between the vegetation water content and LAI. The results indicated that the modified model was effective in evaluating and reproducing the total backscatter coefficients. Meanwhile, the LAI retrieval was well conducted with coefficient of determination ( R2) and root mean square error (RMSE) of 89% and 0.19 m 2 m ?2, respectively. Additionally, this method offers insight into the required application accuracy of LAI retrieval in the agricultural regions. 相似文献
18.
We use the U ppaal model checker for timed automata to verify the Timing-Sync time-synchronization protocol for sensor networks (TPSN), the
clock-synchronization algorithm of Lenzen, Locher and Wattenhofer (LLW) for general distributed systems, and the clock-thread
technique of the software monitoring with controllable overhead algorithm (SMCO). Clock-synchronization algorithms such as
TPSN, LLW, and SMCO must be able to perform arithmetic on clock values to calculate clock drift and network propagation delays.
They must also be able to read the value of a local clock and assign it to another local clock. Such operations are not directly
supported by the theory of timed automata. To overcome this formal-modeling obstacle, we augment the U ppaal specification language with the integer clock-derived type. Integer clocks, which are essentially integer variables that are periodically incremented by a global pulse
generator, greatly facilitate the encoding of the operations required to synchronize clocks as in the TPSN, LLW, and SMCO
protocols. With these integer-clock-based models in hand, we use U ppaal to verify a number of key correctness properties, including network-wide time synchronization, bounded clock skew, bounded
overhead skew, and absence of deadlock. We also use the U ppaal Tracer tool to illustrate how integer clocks can be used to capture clock drift and resynchronization during protocol execution. 相似文献
19.
In formal verification, we verify that a system is correct with respect to a specification. Cases like antecedent failure
can make a successful pass of the verification procedure meaningless. Vacuity detection can signal such “meaningless” passes
of the specification, and indeed vacuity checks are now a standard component in many commercial model checkers.
We address two dimensions of vacuity: the computational effort and the information that is given to the user. As for the first
dimension, we present several preliminary vacuity checks that can be done without the design itself, which implies that some
information can be found with a significantly smaller effort. As for the second dimension, we present algorithms for deriving
two types of information that are not provided by standard vacuity checks, assuming
for a model M and formula φ: (a) behaviors that are possibly missing from M (or wrongly restricted by the environment) (b) the largest subset of occurrences of literals in φ that can be replaced with false simultaneously without falsifying φ in M. The complexity of each of these problems is proven. Overall this extra information can lead to tighter specifications and
more guidance for finding errors.
A preliminary version of this article was published in the proceedings of MEMOCODE 2007 [ 18]. 相似文献
20.
In this paper we present a coherent framework for symbolic model checking of linear-time temporal logic ( ltl) properties over finite state reactive systems, taking full fairness constraints into consideration. We use the computational
model of a fair discrete system ( fds) which takes into account both justice (weak fairness) and compassion (strong fairness). The approach presented here reduces the model-checking problem into the question of whether a given fds is feasible (i.e. has at least one computation).
The contribution of the paper is twofold: On the methodological level, it presents a direct self-contained exposition of full
ltl symbolic model checking without resorting to reductions to either μ-calculus or ctl. On the technical level, it extends previous methods by dealing with compassion at the algorithmic level instead of either
adding it to the specification, or transforming compassion to justice.
Finally, we extend ctl∗ with past operators, and show that the basic symbolic feasibility algorithm presented here, can be used to model check an
arbitrary ctl∗ formula over an fds with full fairness constraints.
This research was supported in part by an infra-structure grant from the Israeli Ministry of Science and Art, a grant from
the U.S.-Israel Binational Science Foundation, and a gift from Intel. 相似文献
|