共查询到20条相似文献,搜索用时 140 毫秒
1.
垂直搜索是针对某一个行业的专业搜索引擎,是搜索引擎的细分和延伸,是对网页库中的某类专门的信息进行一次整合.定向分字段抽取出需要的数据进行处理后再以某种形式返回给用户。相对通用搜索引擎的信息量大、查询不准确、深度不够等提出来的新的搜索引擎服务模式。通过针对某一特定领域、某一特定人群或某一特定需求提供的有一定价值的信息和相关服务。 相似文献
2.
3.
4.
针对搜索引擎存在的问题和刑罚信息检索的特定需要,提出了一个元搜索引擎模型,通过元搜索引擎为用户提供统一的访问服务,运用聚类方法对搜索引擎的返回结果进行快速分类,并通过用户的行为反应探视用户的兴趣,动态调整聚类结果,帮助用户快速检索到相关信息.提高了检索的有效性、查准率和查全率,实现了刑罚信息的个性化检索. 相似文献
5.
在巨大的Internet/Web信息中很难积极地搜索到准确的信息,搜索引擎技术解决了用户检索Web信息困难的问题,而现有的搜索引擎返回的信息却并不总令用户满意。文中在对MAS理论调研的基础上,提出一个基于MAS的搜索引擎的模型,并将其与著名的Google搜索引擎作比较和分析。 相似文献
6.
7.
如果你需要在WEB上查找某一 类的消息,就一定会用到搜索引擎。但有时返回的搜索结果多达数十以至上百。想全面浏览搜索的结果或者保存搜索结果,就不得不反复点击Next(下一页)以及Save(保存)。烦琐的重复劳动,加上为等待网页打开而耗费的时间,一定使你烦恼不堪。现在我为你推荐一个好工具──网络雪貂(WebFerret),它会为你免去这些烦恼。 一、简介 WebFerrt是一个WEB搜索软件,通过内置的数个著名搜索引擎的搜索代码,可在同一时间向数个著名的搜索引擎如 Yahoo、 AltaVista、 In… 相似文献
8.
9.
一个基于MAS的搜索引擎模型 总被引:1,自引:0,他引:1
在巨大的Internet/Web信息中很难积极地搜索到准确的信息,搜索引擎技术解决了用户检索Web信息困难的问题,而现有的搜索引擎返回的信息却并不总令用户满意。文中在对MAS理论调研的基础上,提出一个基于MAS的搜索引擎的模型。并将其与著名的Google搜索引擎作比较和分析。 相似文献
10.
基于用户兴趣的搜索引擎 总被引:1,自引:0,他引:1
随着WWW的出现和发展,Internet上出现的信息迅速增长.如何从大量的信息中获取有用的信息,正成为信息领域的关键技术.传统的搜索引擎没有考虑不同用户的兴趣,因此搜索出来的结果往往无法满足不同用户的特定需求.提出一种用户兴趣模型,能够有效表示用户兴趣,并对传统搜索引擎的搜索结果进行匹配度计算,从而将符合用户兴趣的结果返回给用户.基于这种模型开发了一个基于用户兴趣的法律领域的搜索引擎MyLaw. 相似文献
11.
集成搜索引擎的文本数据库选择 总被引:8,自引:0,他引:8
用户需要检索的信息往往分散存储在多个搜索多个搜索引擎各自的数据库里,对普通用户而言,访问多个搜索引擎并从返回的结果中分辨出确实有网页是一件费时费力的工作,集成搜索引擎则可以提供给用户一个同时记问多个搜索引擎人集成环境,集成搜索引擎能将其接收到的用户查询提交给底层的多个搜索引擎进行搜索,作为一种搜索工具,集成搜索引擎具有如WEB查询覆盖面比传统引擎更大,引警有更好的可扩展性等优点,讨论了解决集成搜索引擎的数据库选择问题的多种技术,针对用户提交的查询要求,通过数据库选择可以选定最有可能返回有用信息的底层搜索引擎。 相似文献
12.
The Internet is one of the most important sources of knowledge in the present time. It offers a huge volume of information which grows dramatically every day. Web search engines (e.g. Google, Yahoo…) are widely used to find specific data among that information. However, these useful tools also represent a privacy threat for the users: the web search engines profile them by storing and analyzing all the searches that they have previously submitted. To address this privacy threat, current solutions propose new mechanisms that introduce a high cost in terms of computation and communication. In this paper, we propose a new scheme designed to protect the privacy of the users from a web search engine that tries to profile them. Our system uses social networks to provide a distorted user profile to the web search engine. The proposed protocol submits standard queries to the web search engine; thus it does not require any change in the server side. In addition to that, this scheme does not require the server to collaborate with the users. Our protocol improves the existing solutions in terms of query delay. Besides, the distorted profiles still allow the users to get a proper service from the web search engines. 相似文献
13.
因特网的迅速发展对传统的爬行器和搜索引擎提出了巨大的挑战。各种针对特定领域、特定人群的搜索引擎应运而生。Web主题信息搜索系统(网络蜘蛛)是主题搜索引擎的最主要的部分,它的任务是将搜集到的符合要求的Web页面返回给用户或保存在索引库中。Web 上的信息资源如此广泛,如何全面而高效地搜集到感兴趣的内容是网络蜘蛛的研究重点。提出了基于网页分块技术的主题爬行,实验结果表明,相对于其它的爬行算法,提出的算法具有较高的效率、爬准率、爬全率及穿越隧道的能力。 相似文献
14.
15.
随着互联网的快速发展,网络信息呈现海量和多元化的趋势。如何为互联网用户快速、准确地提取其所需信息,已成为搜索引擎面临的首要问题。传统的通用搜索引擎虽然能够在较大的信息范围内获取目标,但在某些特定领域无法给用户提供专业而深入的信息。提出基于SVM分类的主题爬虫技术,其将基于文字内容和部分链接信息的主题相关度预测算法、SVM分类算法和HITS算法相结合,解决了特定信息检索的难题。实验结果表明,使用基于SVM分类算法的爬取策略,能够较好地区分主题相关网页和不相关网页,提高了主题相关网页的收获率和召回率,进而提高了搜索引擎的检索效率。 相似文献
16.
Search engines retrieve and rank Web pages which are not only relevant to a query but also important or popular for the users.
This popularity has been studied by analysis of the links between Web resources. Link-based page ranking models such as PageRank
and HITS assign a global weight to each page regardless of its location. This popularity measurement has shown successful
on general search engines. However unlike general search engines, location-based search engines should retrieve and rank higher
the pages which are more popular locally. The best results for a location-based query are those which are not only relevant
to the topic but also popular with or cited by local users. Current ranking models are often less effective for these queries
since they are unable to estimate the local popularity. We offer a model for calculating the local popularity of Web resources
using back link locations. Our model automatically assigns correct locations to the links and content and uses them to calculate
new geo-rank scores for each page. The experiments show more accurate geo-ranking of search engine results when this model
is used for processing location-based queries. 相似文献
17.
18.
Chakrabarti S. Dom B.E. Kumar S.R. Raghavan P. Rajagopalan S. Tomkins A. Gibson D. Kleinberg J. 《Computer》1999,32(8):60-67
The Web is a hypertext body of approximately 300 million pages that continues to grow at roughly a million pages per day. Page variation is more prodigious than the data's raw scale: taken as a whole, the set of Web pages lacks a unifying structure and shows far more authoring style and content variation than that seen in traditional text document collections. This level of complexity makes an “off-the-shelf” database management and information retrieval solution impossible. To date, index based search engines for the Web have been the primary tool by which users search for information. Such engines can build giant indices that let you quickly retrieve the set of all Web pages containing a given word or string. Experienced users can make effective use of such engines for tasks that can be solved by searching for tightly constrained key words and phrases. These search engines are, however, unsuited for a wide range of equally important tasks. In particular, a topic of any breadth will typically contain several thousand or million relevant Web pages. How then, from this sea of pages, should a search engine select the correct ones-those of most value to the user? Clever is a search engine that analyzes hyperlinks to uncover two types of pages: authorities, which provide the best source of information on a given topic; and hubs, which provide collections of links to authorities. We outline the thinking that went into Clever's design, report briefly on a study that compared Clever's performance to that of Yahoo and AltaVista, and examine how our system is being extended and updated 相似文献
19.
Eleni Kosta Christos Kalloniatis Lilian Mitrou Evangelia Kavakli 《Requirements Engineering》2011,16(1):47-54
Nowadays, Internet users are depending on various search engines in order to be able to find requested information on the
Web. Although most users feel that they are and remain anonymous when they place their search queries, reality proves otherwise.
The increasing importance of search engines for the location of the desired information on the Internet usually leads to considerable
inroads into the privacy of users. A heated debate is currently ongoing at European level regarding the question if search
engine providers that are established outside the European Union are covered by the European data protection framework and
the obligations it imposes on entities that process personal data. The scope of this paper is to examine the applicability
of the European data protection legislation to non-EU-based search engine providers and to study the main privacy issues with
regard to search engines, such as the character of search logs, their anonymisation and their retention period. Ixquick, a
privacy-friendly meta-search engine, will be presented as an alternative to privacy intrusive existing practices of search
engines. 相似文献