共查询到20条相似文献,搜索用时 15 毫秒
1.
Qijun Zhao 《Pattern recognition》2007,40(4):1334-1341
The tamper-proof of web pages is of great importance. Some watermarking schemes have been reported to solve this problem. However, both these watermarking schemes and the traditional hash methods have a problem of increasing file size. In this paper, we propose a novel watermarking scheme for the tamper-proof of web pages, which is free of this embarrassment. For a web page, the proposed scheme generates watermarks based on the principal component analysis (PCA) technique. The watermarks are then embedded into the web page through the upper and lower cases of letters in HTML tags. When a watermarked web page is tampered, the extracted watermarks can detect the modifications to the web page, thus we can keep the tampered one from being published. Extensive experiments are performed on the proposed scheme and the results show that the proposed scheme can be a feasible and efficient tool for the tamper-proof of web pages. 相似文献
2.
Joonho KwonAuthor Vitae 《Data & Knowledge Engineering》2012,71(1):69-91
Recently, there has been growing interest in developing web services composition search systems. Current solutions have the drawback of including redundant web services in the results. In this paper, we proposed a non-redundant web services composition search system called NRC, which is based on a two-phase algorithm. In the NRC system, the Link Index is built over web services according to their connectivity. In the forward phase, the candidate compositions are efficiently found by searching the Link Index. In the backward phase, the candidate compositions decomposed into several non-redundant web services compositions by using the concept of tokens. Results of experiments involving data sets with different characteristics show the performance benefits of the NRC techniques in comparison to state-of-the-art composition approaches. 相似文献
3.
本文以德州电业局创一流管理信息系统为例,详细介绍了企业Web站点资料上传,查询的设计与实现方法。 相似文献
4.
This paper provides a transparent and speculative algorithm for content based web page prefetching. The algorithm relies on a profile based on the Internet browsing habits of the user. It aims at reducing the perceived latency when the user requests a document by clicking on a hyperlink. The proposed user profile relies on the frequency of occurrence for selected elements forming the web pages visited by the user. These frequencies are employed in a mechanism for the prediction of the user’s future actions. For the anticipation of an adjacent action, the anchored text around each of the outbound links is used and weights are assigned to these links. Some of the linked documents are then prefetched and stored in a local cache according to the assigned weights. The proposed algorithm was tested against three different prefetching algorithms and yield improved cache–hit rates given a moderate bandwidth overhead. Furthermore, the precision of accurately inferring the user’s preference is evaluated through the recall–precision curves. Statistical evaluation testifies that the achieved recall–precision performance improvement is significant. 相似文献
5.
对HITS(Hyperlink-Induced Topic Search)算法进行研究,克服其主体漂移和扩大化等缺点,改进并提出建立一个新的基于链接分析的Web检索结果的相关度排序方法,并把它应用到一个专题搜索引擎中. 相似文献
6.
Josiane Xavier Parreira Carlos Castillo Debora Donato Sebastian Michel Gerhard Weikum 《The VLDB Journal The International Journal on Very Large Data Bases》2008,17(2):291-313
We present Juxtaposed approximate PageRank (JXP), a distributed algorithm for computing PageRank-style authority scores of
Web pages on a peer-to-peer (P2P) network. Unlike previous algorithms, JXP allows peers to have overlapping content and requires
no a priori knowledge of other peers’ content. Our algorithm combines locally computed authority scores with information obtained
from other peers by means of random meetings among the peers in the network. This computation is based on a Markov-chain state-lumping
technique, and iteratively approximates global authority scores. The algorithm scales with the number of peers in the network
and we show that the JXP scores converge to the true PageRank scores that one would obtain with a centralized algorithm. Finally,
we show how to deal with misbehaving peers by extending JXP with a reputation model.
Partially supported by the EU within the 6th Framework Programme under contract 001907 “Dynamically Evolving, Large Scale
Information Systems” (DELIS). 相似文献
7.
This paper illustrates a new approach to automatic re-authoring of web pages for rendering on small-screen devices. The approach
is based on automatic detection of the device type and screen size from the HTTP request header to render a desktop web page
or a transformed one for display on small screen devices, for example, PDAs. Known algorithms (transforms) are employed to
reduce the size of page elements, to hide parts of the text, and to transform tables into text while preserving the structural
format of the web page. The system comprises a preprocessor that works offline and a just-in-time handler that responds to
HTTP requests. The preprocessor employs Cascading Style Sheets (CSS) to set default attributes for the page and prepares it
for the handler. The latter is responsible for downsizing graphical elements in the page, converting tables to text, and inserting
visibility attributes and JavaScript code to allow the user of the client device to interact with the page and cause parts
of the text to disappear or reappear. A system was developed that implements the approach and was used it to collect performance
results and conduct usability testing. The importance of the approach lies in its ability to display hidden parts of the web
page without having to revisit the server, thus reducing user wait times considerably, saving battery power, and cutting down
on wireless network traffic. 相似文献
8.
9.
This study reported an investigation of eighth graders’ (14-year-olds) web searching strategies and outcomes, and then analyzed their correlations with students’ web experiences, epistemological beliefs, and the nature of searching tasks. Eighty-seven eighth graders were asked to fill out a questionnaire for probing epistemological beliefs (from positivist to constructivist-oriented views) and finished three different types of searching tasks. Their searching process was recorded by screen capture software and answers were reviewed by two expert teachers based on their accuracy, richness and soundness. Five quantitative indicators were used to assess students’ searching strategies: number of keywords, visited pages, maximum depth of exploration, refinement of keyword, and number of words used in the first keyword. The main findings derived from this study suggested that, students with richer web experiences could find more correct answers in “close-ended” search tasks. In addition, students with better metacognitive skills such as keyword refinement tended to achieve more successful searching outcomes in such tasks. However, in “open-ended” tasks, where questions were less certain and answers were more elaborated, students who had more advanced epistemological beliefs, concurring with a constructivist view, had better searching outcomes in terms of their soundness and richness. This study has concluded that epistemological beliefs play an influential role in open-ended Internet learning environments. 相似文献
10.
The limited display size of current small Internet devices is becoming a serious obstacle to information access. In this paper,
we introduce a Document REpresentation for Scalable Structures (DRESS) to help information providers make composite documents,
typically web pages, scalable in both logic and layout structure to support effective information acquisition in heterogeneous
environments. Through this novel document representation structure based on binary slicing trees, the document can dynamically
adapt its presentation according to display sizes by maximizing the information throughput to users. We discuss the details
of this structure with its key attributes. An automatic approach for generating this structure for existing web pages is also
presented. A branch-and-bound algorithm and a capacity ratio-based slicing method are proposed to select proper content representation
and aesthetic document layouts respectively. A set of user study experiments have been carried out and the results show that
compared with the thumbnail-based approach, the DRESS-based interface can reduce browsing time by 23.5%.
This work was performed when the second and the third authors were visiting students at Microsoft Research Asia. 相似文献
11.
Mona TaghaviAuthor Vitae Nikita SchmidtAuthor Vitae 《Computer Standards & Interfaces》2012,34(1):162-170
This study presents an analysis of users' queries directed at different search engines to investigate trends and suggest better search engine capabilities. The query distribution among search engines that includes spawning of queries, number of terms per query and query lengths is discussed to highlight the principal factors affecting a user's choice of search engines and evaluate the reasons of varying the length of queries. The results could be used to develop long to short term business plans for search engine service providers to determine whether or not to opt for more focused topic specific search offerings to gain better market share. 相似文献
12.
13.
Search engines are increasingly efficient at identifying the best sources for any given keyword query, and are often able to identify the answer within the sources. Unfortunately, many web sources are not trustworthy, because of erroneous, misleading, biased, or outdated information. In many cases, users are not satisfied with the results from any single source. In this paper, we propose a framework to aggregate query results from different sources in order to save users the hassle of individually checking query-related web sites to corroborate answers. To return the best answers to the users, we assign a score to each individual answer by taking into account the number, relevance and originality of the sources reporting the answer, as well as the prominence of the answer within the sources, and aggregate the scores of similar answers. We conducted extensive qualitative and quantitative experiments of our corroboration techniques on queries extracted from the TREC Question Answering track and from a log of real web search engine queries. Our results show that taking into account the quality of web pages and answers extracted from the pages in a corroborative way results in the identification of a correct answer for a majority of queries. 相似文献
14.
Predicting the goals of internet users can be extremely useful in e-commerce, online entertainment, and many other internet-based applications. One of the crucial steps to achieve this is to classify internet queries based on available features, such as contextual information, keywords and their semantic relationships. Beyond these methods, in this paper we propose to mine user interaction activities to predict the intent of the user during a navigation session. However, since in practice it is necessary to use a suitable mix of all such methods, it is important to exploit all the mentioned features in order to properly classify users based on their common intents. To this end, we have performed several experiments aiming to empirically derive a suitable classifier based on the mentioned features. 相似文献
15.
R. M. Alguliev R. M. Alyguliev F. F. Yusifov 《Automatic Control and Computer Sciences》2007,41(6):320-331
In the present article an approach to automatic determination of a user’s sphere of interests is proposed. The approach is based on a method involving clustering of documents which the user is interested in. The process of clustering of documents is reduced to a problem of discrete optimization for which quadratic-and linear-type models are proposed. Identification of interests makes it possible to determine the context of a request without any effort on the user’s part. Different methods are proposed for determining the context of a request. An ant algorithm for solving a quadratic-type discrete optimization problem is also proposed in the present study. 相似文献
16.
Isabelle Etcheverry Thierry Baccino Patrice Terrier Jean-Claude Marquié Mustapha Mojahid 《Computers in human behavior》2012
Recent research suggests that older Internet users seem to find it more difficult to locate navigation links than to find information content in web pages. One possibility is that older Internet users’ visual exploration of web pages is more linear in nature, even when this type of processing is not appropriate for the task. In the current study, the eye movements of young and older Internet users were recorded using an ecological version of the web pages or a discursive version designed to induce a linear exploration. The older adults found more targets when performing content-oriented compared to navigation-oriented searches, thus replicating previous results. Moreover, they performed less well than young people only when required to locate navigation links and tended to outperform the younger participants in content-oriented searches. Although the type of search task and type of web page resulted in different visual strategies, little or no support was found for the hypothesis that older participants explore web pages in a more linear way in cases where this strategy was not appropriate. The main conclusion is that differences in visual exploration do not seem to mediate the specific difficulty older adults experience in navigation-oriented searches in web pages. 相似文献
17.
Ole-Christoffer Granmo B. John Oommen 《Engineering Applications of Artificial Intelligence》2011,24(7):1238-1251
We consider the problem of polling web pages as a strategy for monitoring the world wide web. The problem consists of repeatedly polling a selection of web pages so that changes that occur over time are detected. In particular, we consider the case where we are constrained to poll a maximum number of web pages per unit of time, and this constraint is typically dictated by the governing communication bandwidth, and by the speed limitations associated with the processing. Since only a fraction of the web pages can be polled within a given unit of time, the issue at stake is one of determining which web pages are to be polled, and we attempt to do it in a manner that maximizes the number of changes detected. We solve the problem by first modelling it as a stochastic nonlinear fractional knapsack problem. We then present an online learning automata (LA) system, namely, the hierarchy of twofold resource allocation automata (H-TRAA), whose primitive component is a twofold resource allocation automaton (TRAA). Both the TRAA and the H-TRAA have been proven to be asymptotically optimal. Finally, we demonstrate empirically that the H-TRAA provides orders of magnitude faster convergence compared to the learning automata knapsack game (LAKG) which represents the state-of-the-art for this problem. Further, in contrast to the LAKG, the H-TRAA scales sub-linearly. Based on these results, we believe that the H-TRAA has also tremendous potential to handle demanding real-world applications, particularly those which deal with the world wide web. 相似文献
18.
19.
张春明 《网络安全技术与应用》2014,(9):211-212
企业网站若想取得好的营销效果,就要提高自己在搜索引擎上的排名.而排名的好坏需要站长的主动出击,以及营销团队与网站策划开发与维护团队的紧密配合.方法主要有主动登陆搜索引擎、被动等待搜索引擎的收集以及积极的利用搜索引擎的推广功能.多管齐下才能取得更好的效果. 相似文献
20.
Although many web development methods exist, they are rarely used by practitioners. The work reported here, seeks to explain
why this might be so, and suggests that, for many, the perceived benefits may be outweighed by the difficulty or effort required
to learn the method. In attempting to gauge the utility of methods the authors undertook a 2-year study of 23 small web development
projects, attempting to use a range of published (academic) methods. Of the 23 projects we found only one case where the developer
continued to use an academic web development method throughout the lifecycle. The ability to understand a method and/or its
techniques was repeatedly cited as the reason for its abandonment. Our findings also indicate a number of key areas, relating
to terminology, completeness, and guidance, where existing methods may be failing their intended users. In attempting to further
our understanding of web development methods we completed a comprehensive survey of web development methods; covering 52 web
development methods, encompassing a range of different research communities, and drawing upon 63 different sources. Our findings
here shed some light upon the confusion of methods for the would-be user. In summary, the findings are that, although there
is much of value in a variety of methods, method choice is somewhat bewildering for the newcomer to the field, and many methods
are incomplete in some dimension. By providing this work we hope to go some way towards supporting the software engineering
community, in both academia and industry, in their understanding of the quality issues that exist with the take up and use
of web development methods.
相似文献
Jonathan VincentEmail: |