Dissertations / Theses on the topic 'Web Search engines'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Web Search engines.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Blaauw, Pieter. "Search engine poisoning and its prevalence in modern search engines." Thesis, Rhodes University, 2013. http://hdl.handle.net/10962/d1002037.
Full textShen, Yipeng. "Meta-search and distributed search systems /." View Abstract or Full-Text, 2002. http://library.ust.hk/cgi/db/thesis.pl?COMP%202002%20SHEN.
Full textIncludes bibliographical references (leaves 138-144). Also available in electronic version. Access restricted to campus users.
Kalinov, Pavel. "Intelligent Web Exploration." Thesis, Griffith University, 2012. http://hdl.handle.net/10072/365635.
Full textThesis (PhD Doctorate)
Doctor of Philosophy (PhD)
School of Information and Communication Technology
Science, Environment, Engineering and Technology
Full Text
Zhang, Lu Jansen Bernard J. "A branding model for web search engines." [University Park, Pa.] : Pennsylvania State University, 2009. http://etda.libraries.psu.edu/theses/approved/WorldWideIndex/ETD-3996/index.html.
Full textErola, Cañellas Arnau. "Contributions to privacy in web search engines." Doctoral thesis, Universitat Rovira i Virgili, 2013. http://hdl.handle.net/10803/130934.
Full textWeb Search Engines collects and stores information about their users in order to tailor their services better to their users' needs. Nevertheless, while receiving a personalized attention, the users lose the control over their own data. Search logs can disclose sensitive information and the identities of the users, creating risks of privacy breaches. In this thesis we discuss the problem of limiting the disclosure risks while minimizing the information loss. The first part of this thesis focuses on the methods to prevent the gathering of information by WSEs. Since search logs are needed in order to receive an accurate service, the aim is to provide logs that are still suitable to provide personalization. We propose a protocol which uses a social network to obfuscate users' profiles. The second part deals with the dissemination of search logs. We propose microaggregation techniques which allow the publication of search logs, providing $k$-anonymity while minimizing the information loss.
Lewandowski, Dirk. "Web Searching, Search Engines and Information Retrieval." ISO Press, 2005. http://hdl.handle.net/10150/106395.
Full textPetit, Albin. "Introducing privacy in current web search engines." Thesis, Lyon, 2017. http://www.theses.fr/2017LYSEI016/document.
Full textDuring the last few years, the technological progress in collecting, storing and processing a large quantity of data for a reasonable cost has raised serious privacy issues. Privacy concerns many areas, but is especially important in frequently used services like search engines (e.g., Google, Bing, Yahoo!). These services allow users to retrieve relevant content on the Internet by exploiting their personal data. In this context, developing solutions to enable users to use these services in a privacy-preserving way is becoming increasingly important. In this thesis, we introduce SimAttack an attack against existing protection mechanism to query search engines in a privacy-preserving way. This attack aims at retrieving the original user query. We show with this attack that three representative state-of-the-art solutions do not protect the user privacy in a satisfactory manner. We therefore develop PEAS a new protection mechanism that better protects the user privacy. This solution leverages two types of protection: hiding the user identity (with a succession of two nodes) and masking users' queries (by combining them with several fake queries). To generate realistic fake queries, PEAS exploits previous queries sent by the users in the system. Finally, we present mechanisms to identify sensitive queries. Our goal is to adapt existing protection mechanisms to protect sensitive queries only, and thus save user resources (e.g., CPU, RAM). We design two modules to identify sensitive queries. By deploying these modules on real protection mechanisms, we establish empirically that they dramatically improve the performance of the protection mechanisms
Selberg, Erik Warren. "Towards comprehensive Web search /." Thesis, Connect to this title online; UW restricted, 1999. http://hdl.handle.net/1773/6873.
Full textMtshontshi, Lindiwe. "Evaluation and comparison of search engines." Thesis, Stellenbosch : Stellenbosch University, 2004. http://hdl.handle.net/10019.1/49955.
Full textENGLISH ABSTRACT: A growing body of studies is developing approaches to evaluate human interaction with Web search engines. Measuring the information retrieval effectiveness of World Wide Web search engines is costly because of the human relevance judgements involved. However, both for business enterprises and people it is important to know the most effective Web search engine, since such search engines help their users find a higher number of relevant Web pages with less effort. Furthermore, this information can be used for several practical purposes. This study does not attempt to describe all the currently available search engines, but provides a comparison of some, which are deemed to be among the most useful. It concentrates on search engines and their characteristics only. The goal is to help a new user get the most useful "hits" when using the various tools.
AFRIKAANSE OPSOMMING: Al hoe meer studies word gedoen om benaderings te ontwikkel vir die evaluasie van menslike interaksie met Web-soekenjins. Om te meet hoe effektief 'n soekenjin inligting op die Wêreldwye Web kan opspoor, is duur vanweë die mens se relevansiebeoordeling wat daarby betrokke is. Dit is egter belangrik dat die bestuurders van sake-ondememings en ander mense sal weet watter die mees doeltreffende soekenjins is, aangesien sulke soekenjins hulle gebruikers help om 'n hoër aantal relevante Webblaaie met minder inspanning te vind. Hierdie inligting kan ook gebruik word om 'n paar praktiese doelwitte te verwesenlik. Daar word nie gepoog om al die soekenjins wat tans beskikbaar is, te beskryf nie, maar sommige van die soekenjins wat as die nuttigste beskou word, word vergelyk. Daar word alleenlik op soekenjins en hulle kenmerke gekonsentreer. Die doel is om die nuwe gebruiker te help om die nuttigste inligting te verkry deur gebruik te maak van verskeie hulpmiddels.
Chakravarty, Rupak, and Sukhwinder Randhawa. "ACADEMIC SEARCH ENGINES: LIBRARIAN'S FRIEND,RESEARCHER'S DELIGHT." INFLIBNET Centre, 2006. http://hdl.handle.net/10150/105232.
Full textZhao, Hongkun. "Automatic wrapper generation for the extraction of search result records from search engines." Diss., Online access via UMI:, 2007.
Find full textCosta, Miguel. "SIDRA: a Flexible Web Search System." Master's thesis, Department of Informatics, University of Lisbon, 2004. http://hdl.handle.net/10451/13914.
Full textBian, Jiang. "Contextualized web search: query-dependent ranking and social media search." Diss., Georgia Institute of Technology, 2010. http://hdl.handle.net/1853/37246.
Full textPinkerton, Brian. "WebCrawler : finding what people want /." Thesis, Connect to this title online; UW restricted, 2000. http://hdl.handle.net/1773/6890.
Full textHicks, Janette M. "Search algorithms for discovery of Web services." Diss., Online access via UMI:, 2005. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&res_dat=xri:pqdiss&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&rft_dat=xri:pqdiss:1425747.
Full textChen, Yan. "Enhanced Web Search Engines with Query-Concept Bipartite Graphs." Digital Archive @ GSU, 2010. http://digitalarchive.gsu.edu/cs_diss/54.
Full textMartins, Flávio Nuno Fernandes. "Improving search engines with open Web-based SKOS vocabularies." Master's thesis, Faculdade de Ciências e Tecnologia, 2012. http://hdl.handle.net/10362/8745.
Full textThe volume of digital information is increasingly larger and even though organiza-tions are making more of this information available, without the proper tools users have great difficulties in retrieving documents about subjects of interest. Good infor-mation retrieval mechanisms are crucial for answering user information needs. Nowadays, search engines are unavoidable - they are an essential feature in docu-ment management systems. However, achieving good relevancy is a difficult problem particularly when dealing with specific technical domains where vocabulary mismatch problems can be prejudicial. Numerous research works found that exploiting the lexi-cal or semantic relations of terms in a collection attenuates this problem. In this dissertation, we aim to improve search results and user experience by inves-tigating the use of potentially connected Web vocabularies in information retrieval en-gines. In the context of open Web-based SKOS vocabularies we propose a query expan-sion framework implemented in a widely used IR system (Lucene/Solr), and evaluated using standard IR evaluation datasets. The components described in this thesis were applied in the development of a new search system that was integrated with a rapid applications development tool in the context of an internship at Quidgest S.A.
Fundação para a Ciência e Tecnologia - ImTV research project, in the context of the UTAustin-Portugal collaboration (UTA-Est/MAI/0010/2009); QSearch project (FCT/Quidgest)
Mendoza, Rocha Marcelo Gabriel. "Query log mining in search engines." Tesis, Universidad de Chile, 2007. http://www.repositorio.uchile.cl/handle/2250/102877.
Full textLa Web es un gran espacio de información donde muchos recursos como documentos, imágenes u otros contenidos multimediales pueden ser accesados. En este contexto, varias tecnologías de la información han sido desarrolladas para ayudar a los usuarios a satisfacer sus necesidades de búsqueda en la Web, y las más usadas de estas son los motores de búsqueda. Los motores de búsqueda permiten a los usuarios encontrar recursos formulando consultas y revisando una lista de respuestas. Uno de los principales desafíos para la comunidad de la Web es diseñar motores de búsqueda que permitan a los usuarios encontrar recursos semánticamente conectados con sus consultas. El gran tamaño de la Web y la vaguedad de los términos más comúnmente usados en la formulación de consultas es un gran obstáculo para lograr este objetivo. En esta tesis proponemos explorar las selecciones de los usuarios registradas en los logs de los motores de búsqueda para aprender cómo los usuarios buscan y también para diseñar algoritmos que permitan mejorar la precisión de las respuestas recomendadas a los usuarios. Comenzaremos explorando las propiedades de estos datos. Esta exploración nos permitirá determinar la naturaleza dispersa de estos datos. Además presentaremos modelos que nos ayudarán a entender cómo los usuarios buscan en los motores de búsqueda. Luego, exploraremos las selecciones de los usuarios para encontrar asociaciones útiles entre consultas registradas en los logs. Concentraremos los esfuerzos en el diseño de técnicas que permitirán a los usuarios encontrar mejores consultas que la consulta original. Como una aplicación, diseñaremos métodos de reformulación de consultas que ayudarán a los usuarios a encontrar términos más útiles mejorando la representación de sus necesidades. Usando términos de documentos construiremos representaciones vectoriales para consultas. Aplicando técnicas de clustering podremos determinar grupos de consultas similares. Usando estos grupos de consultas, introduciremos métodos para recomendación de consultas y documentos que nos permitirán mejorar la precisión de las recomendaciones. Finalmente, diseñaremos técnicas de clasificación de consultas que nos permitirán encontrar conceptos semánticamente relacionados con la consulta original. Para lograr esto, clasificaremos las consultas de los usuarios en directorios Web. Como una aplicación, introduciremos métodos para la manutención automática de los directorios.
Ali, Halil, and hali@cs rmit edu au. "Effective web crawlers." RMIT University. CS&IT, 2008. http://adt.lib.rmit.edu.au/adt/public/adt-VIT20081127.164414.
Full textHenriksson, Adam. "Alternative Search : From efficiency to experience." Thesis, Umeå universitet, Institutionen Designhögskolan, 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-97836.
Full textSearch Engines, Interaction Design
Young, Joel D. "Exploiting the structure of the web for spidering /." View online version; access limited to Brown University users, 2005. http://wwwlib.umi.com/dissertations/fullcit/3174706.
Full textAdya, Kaushik. "An implicit-feedback based ranking methodology for Web search engines /." Available to subscribers only, 2005. http://proquest.umi.com/pqdweb?did=1079672381&sid=1&Fmt=2&clientId=1509&RQT=309&VName=PQD.
Full textAsghariaghamashhadi, Mohammad Bagher <1988>. "Using of Trademarks in Keyword Advertising in Web Search Engines." Doctoral thesis, Alma Mater Studiorum - Università di Bologna, 2018. http://amsdottorato.unibo.it/8298/1/FINAL%20VERSION%20OF%20THESIS.pdf.
Full textNeethling, Riaan. "Search engine optimisation or paid placement systems-user preference /." Thesis, [S.l. : s.n.], 2007. http://dk.cput.ac.za/cgi/viewcontent.cgi?article=1076&context=td_cput.
Full text李莉華 and Lei-wah Lee. "On improving the relevancy ranking algorithm in web search engine." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2000. http://hub.hku.hk/bib/B31222973.
Full textLee, Lei-wah. "On improving the relevancy ranking algorithm in web search engine /." Hong Kong : University of Hong Kong, 2000. http://sunzi.lib.hku.hk/hkuto/record.jsp?B21607448.
Full textLakshmi, Shriram. "Web-based search engine for Radiology Teaching File." [Gainesville, Fla.] : University of Florida, 2002. http://purl.fcla.edu/fcla/etd/UFE0000559.
Full textDeng, Lin. "Mining user preference using SPY voting for search engine personalization /." View abstract or full-text, 2006. http://library.ust.hk/cgi/db/thesis.pl?COMP%202006%20DENG.
Full textLin, Jason. "WebSearch: A configurable parallel multi-search web browser." CSUSB ScholarWorks, 1999. https://scholarworks.lib.csusb.edu/etd-project/1948.
Full textÖfverman, Jakob. "Information Presentation in Search Engines on Mobile Devices." Thesis, Linnaeus University, School of Computer Science, Physics and Mathematics, 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-7945.
Full textThis thesis discusses the possibilities to visualise the presentation layer of a search engine on a mobile device in an alternative way. Previous work in the area has shown that the use of text-based-lists can be problematic when accessed on a device with a limited display. In the scope of the thesis and in order to tackle the current problems when displaying the results a literature review was carried out. The findings of the review formed the basis for a requirement definition on which a mock-up was developed. The mock-up was then evaluated and tested during a usability test where a number of users got to experience the alternative presentation layer that uses a visualisation technique called tree- map. The results from the test show that the mock-up could be seen as a alternative to the current presentation of results. The mock-up also shows that a future implementation could also include the use of categories and sorting of information in order to provide content with a meaning.
Gopinathan-Leela, Ligon, and n/a. "Personalisation of web information search: an agent based approach." University of Canberra. Information Sciences & Engineering, 2005. http://erl.canberra.edu.au./public/adt-AUC20060728.120849.
Full textZuze, Herbert. "The crossover point between keyword rich website text and spamdexing." Thesis, Cape Peninsula University of Technology, 2011. http://hdl.handle.net/20.500.11838/1767.
Full textWith over a billion Internet users surfing the Web daily in search of information, buying, selling and accessing social networks, marketers focus intensively on developing websites that are appealing to both the searchers and the search engines. Millions of webpages are submitted each day for indexing to search engines. The success of a search engine lies in its ability to provide accurate search results. Search engines’ algorithms constantly evaluate websites and webpages that could violate their respective policies. For this reason some websites and webpages are subsequently blacklisted from their index. Websites are increasingly being utilised as marketing tools, which result in major competition amongst websites. Website developers strive to develop websites of high quality, which are unique and content rich as this will assist them in obtaining a high ranking from search engines. By focusing on websites of a high standard, website developers utilise search engine optimisation (SEO) strategies to earn a high search engine ranking. From time to time SEO practitioners abuse SEO techniques in order to trick the search engine algorithms, but the algorithms are programmed to identify and flag these techniques as spamdexing. Search engines do not clearly explain how they interpret keyword stuffing (one form of spamdexing) in a webpage. However, they regard spamdexing in many different ways and do not provide enough detail to clarify what crawlers take into consideration when interpreting the spamdexing status of a website. Furthermore, search engines differ in the way that they interpret spamdexing, but offer no clear quantitative evidence for the crossover point of keyword dense website text to spamdexing. Scholars have indicated different views in respect of spamdexing, characterised by different keyword density measurements in the body text of a webpage. This raised several fundamental questions that form the basis of this research. This research was carried out using triangulation in order to determine how the scholars, search engines and SEO practitioners interpret spamdexing. Five websites with varying keyword densities were designed and submitted to Google, Yahoo! and Bing. Two phases of the experiment were done and the results were recorded. During both phases almost all of the webpages, including the one with a 97.3% keyword density, were indexed. The aforementioned enabled this research to conclusively disregard the keyword stuffing issue, blacklisting and any form of penalisation. Designers are urged to rather concentrate on usability and good values behind building a website. The research explored the fundamental contribution of keywords to webpage indexing and visibility. Keywords used with or without an optimum level of measurement of richness and poorness result in website ranking and indexing. However, the focus should be on the way in which the end user would interpret the content displayed, rather than how the search engine would react towards the content. Furthermore, spamdexing is likely to scare away potential clients and end users instead of embracing them, which is why the time spent on spamdexing should rather be used to produce quality content.
Visser, Eugene Bourbon. "Fusing website usability variables and on-page search engine optimisation elements." Thesis, Cape Peninsula University of Technology, 2011. http://hdl.handle.net/20.500.11838/1407.
Full textIt was concluded in the literature review that small- to medium-sized enterprises (SMME) should prioritise utilising the websites on the Internet, as it provides a low cost infrastructure, unlocking opportunities and allowing small- to medium-sized enterprises to market to the international customer, promoting business activities in a low-risk environment. However, visitors do not know that they do not know, meaning a need for facilitation exists between the Internet user in terms of the information required and the information available on the Internet. Search engines (governed by their organic ranking algorithms) were created for this very purpose, to facilitate users in finding relevant information on the Internet in the shortest time possible. Search engines interpret and evaluate any given indexed web page from a targeted keywords perspective, indicating that web pages must be optimised from a search engine perspective. However, the elements search engines perceive to be important may not always be aligned with what website visitors perceive to be important. Anything on the web page that may remotely impede the visitors’ experience could be detrimental as alternative website options are but a click away. An example would be the excessive use of content on a given web page. The search engine may find the excessive content useful as it may provide contextual interpretation of the web page. However, the excessive content may impede a visitor’s website interaction as it is estimated that the average visitors will often view a web page for 45-60 seconds and read a maximum of 200 words only. During the process of identifying the contradictory search engine optimisation (SEO) elements and website usability (WU) attributes, three journal articles were written, with two journal articles following their own research methodologies and the third journal article utilising all the research results in order to create the fused SEO and WU model. Journal Article 1: Two websites were used as part of the experiment: • Control Website (CW): http://www.copywriters.co.za • Experimental Website (EW): http://www.copywriters.co.za/ppc/. The CW is an existing website with no special emphasis applied to SEO and/or WU. The EW was developed by implementing the WU attributes and ignoring all contradictory SEO elements. In order to ensure integrity of the experiment, search engines were denied access to the EW. The traffic sources for the CW were search engines (organic) traffic, as well as direct and referrer traffic.
King, John D. "Search engine content analysis." Queensland University of Technology, 2008. http://eprints.qut.edu.au/26241/.
Full textKing, John Douglas. "Search engine content analysis." Thesis, Queensland University of Technology, 2008. https://eprints.qut.edu.au/26241/1/John_King_Thesis.pdf.
Full textWeldeghebriel, Zemichael Fesahatsion. "Evaluating and comparing search engines in retrieving text information from the web." Thesis, Stellenbosch : Stellenbosch University, 2004. http://hdl.handle.net/10019.1/53740.
Full textENGLISH ABSTRACT: With the introduction of the Internet and the World Wide Web (www), information can be easily accessed and retrieved from the web using information retrieval systems such as web search engines or simply search engines. There are a number of search engines that have been developed to provide access to the resources available on the web and to help users in retrieving relevant information from the web. In particular, they are essential for finding text information on the web for academic purposes. But, how effective and efficient are those search engines in retrieving the most relevant text information from the web? Which of the search engines are more effective and efficient? So, this study was conducted to see how effective and efficient search engines are and to see which search engines are most effective and efficient in retrieving the required text information from the web. It is very important to know the most effective and efficient search engines because such search engines can be used to retrieve a higher number of the most relevant text web pages with minimum time and effort. The study was based on nine major search engines, four search queries and relevancy judgments as relevant/partly-relevanUnon-relevant. Precision and recall were calculated based on the experimental or test results and these were used as basis for the statistical evaluation and comparisons of the retrieval effectiveness of the nine search engines. Duplicated items and broken links were also recorded and examined separately and were used as an additional measure of search engine effectiveness. A response time was also recorded and used as a base for the statistical evaluation and comparisons of the retrieval efficiency of the nine search engines. Additionally, since search engines involve indexing and searching in the information retrieval processes from the web, this study first discusses, from the theoretical point of view, how the indexing and searching processes are performed in an information retrieval environment. It also discusses the influences of indexing and searching processes on the effectiveness and efficiency of information retrieval systems in general and search engines in particular in retrieving the most relevant text information from the web.
AFRIKAANSE OPSOMMING: Met die koms van die Internet en die Wêreldwye Web (www) is inligting maklik bekombaar. Dit kan herwin word deur gebruik te maak van inligtingherwinningsisteme soos soekenjins. Daar is 'n hele aantal sulke soekenjins wat ontwikkel is om toegang te verleen tot die hulpbronne beskikbaar op die web en om gebruikers te help om relevante inligting vanaf die web in te win. Dit is veral noodsaaklik vir die verkryging van teksinligting vir akademiese doeleindes. Maar hoe effektief en doelmatig is die soekenjins in die herwinning van die mees relevante teksinligting vanaf die web? Watter van die soekenjins is die effektiefste? Hierdie studie is onderneem om te kyk watter soekenjins die effektiefste en doelmatigste is in die herwinning van die nodige teksinligting. Dit is belangrik om te weet watter soekenjin die effektiefste is want so 'n enjin kan gebruik word om 'n hoër getal van die mees relevante tekswebblaaie met die minimum van tyd en moeite te herwin. Heirdie studie is baseer op die sewe hoofsoekenjins, vier soektogte, en toepasliksheidsoordele soos relevant /gedeeltelik relevant/ en nie- relevant. Presiesheid en herwinningsvermoë is bereken baseer op die eksperimente en toetsresultate en dit is gebruik as basis vir statistiese evaluasie en vergelyking van die herwinningseffektiwiteit van die nege soekenjins. Gedupliseerde items en gebreekte skakels is ook aangeteken en apart ondersoek en is gebruik as bykomende maatstaf van effektiwiteit. Die reaksietyd is ook aangeteken en is gebruik as basis vir statistiese evaluasie en die vergelyking van die herwinningseffektiwiteit van die nege soekenjins. Aangesien soekenjins betrokke is by indeksering en soekprosesse, bespreek hierdie studie eers uit 'n teoretiese oogpunt, hoe indeksering en soekprosesse uitgevoer word in 'n inligtingherwinningsomgewing. Die invloed van indeksering en soekprosesse op die doeltreffendheid van herwinningsisteme in die algemeen en veral van soekenjins in die herwinning van die mees relevante teksinligting vanaf die web, word ook bespreek.
Kritzinger, Wouter Thomas. "Development of a search engine marketing model using the application of a dual strategy." Thesis, Cape Peninsula University of Technology, 2017. http://hdl.handle.net/20.500.11838/2603.
Full textAny e-commerce venture using a website as main shop-front should invest in marketing their website. Previous empirical evidence shows that most Search Engine Marketing (SEM) spending (approximately 82%) is allocated to Pay Per Click (PPC) campaigns while only 12% was spent on Search Engine Optimisation (SEO). The remaining 6% of the total spending was allocated to other SEM strategies. No empirical work was found on how marketing expenses compare when used solely for either the one or the other of the two main types of SEM. In this study, a model will be designed to guide the development of a dual SEM strategy.
Mbikiwa, Fernie Neo. "Search engine exclusion policies: implications on indexing e-commerce websites." Thesis, [S.l. : s.n.], 2005. http://dk.cput.ac.za/cgi/viewcontent.cgi?article=1080&context=td_cput.
Full textKritzinger, Wouter Thomas. "The effect webpage body keywords location has on ranking in search engines results : an empirical study /." Thesis, Click here for online access, 2005. http://dk.cput.ac.za/cgi/viewcontent.cgi?article=1077&context=td_cput.
Full textSaunders, Tana. "Evaluation of Internet search tools instrument design." Thesis, Stellenbosch : Stellenbosch University, 2004. http://hdl.handle.net/10019.1/49957.
Full textENGLISH ABSTRACT: This study investigated Internet search tools / engines to identify desirable features that can be used as a benchmark or standard to evaluate web search engines. In the past, the Internet was thought of as a big spider's web, ultimately connecting all the bits of information. It has now become clear that this is not the case, and that the bow tie analogy is more accurate. This analogy suggests that there is a central core of well-connected pages, with links IN and OUT to other pages, tendrils and orphan pages. This emphasizes the importance of selecting a search tool that is well connected and linked to the central core. Searchers must take into account that not all search tools search the Invisible Web and this will reflect on the search tool selected. Not all information found on the Web and Internet is reliable, current and accurate, and Web information must be evaluated in terms of authority, currency, bias, purpose of the Web site, etc. Different kinds of search tools are available on the Internet, such as search engines, directories, library gateways, portals, intelligent agents, etc. These search tools were studied and explored. A new categorization for online search tools consisting of Intelligent Agents, Search Engines, Directories and Portals / Hubs is suggested. This categorization distinguishes the major differences between the 21 kinds of search tools studied. Search tools / engines consist of spiders, crawlers, robots, indexes and search tool software. These search tools can be further distinguished by their scope, internal or external searches and whether they search Web pages or Web sites. Most search tools operate within a relationship with other search tools, and they often share results, spiders and databases. This relationship is very dynamic. The major international search engines have identifiable search features. The features of Google, Yahoo, Lycos and Excite were studied in detail. Search engines search for information in different ways, and present their results differently. These characteristics are critical to the Recall/Precision ratio. A well-planned search strategy will improve the Precision/Recall ratio and consider the web-user capabilities and needs. Internet search tools/engines is not a panacea for all information needs, and have pros and cons. The Internet search tool evaluation instrument was developed based on desirable features of the major search tools, and is considered a benchmark or standard for Internet search tools. This instrument, applied to three South African search tools, provided insight into the capabilities of the local search tools compared to the benchmark suggested in this study. The study concludes that the local search engines compare favorably with the major ones, but not enough so to use them exclusively. Further research into this aspect is needed. Intelligent agents are likely to become more popular, but the only certainty in the future of Internet search tools is change, change, and change.
AFRIKAANSE OPSOMMING: Hierdie studie het Internetsoekinstrumente/-enjins ondersoek met die doel om gewenste eienskappe te identifiseer wat as 'n standaard kan dien om soekenjins te evalueer. In die verlede is die Internet gesien as 'n groot spinnerak, wat uiteindelik al die inligtingsdeeltjies verbind. Dit het egter nou duidelik geword dat dit glad nie die geval is nie, en dat die strikdas analogie meer akkuraat is. Hierdie analogie stel voor dat daar 'n sentrale kern van goed gekonnekteerde bladsye is, met skakels IN en UIT na ander bladsye, tentakels en weesbladsye. Dit beklemtoon die belangrikheid om die regte soekinstrument te kies, naamlik een wat goed gekonnekteer is, en geskakel is met die sentrale kern van dokumente. Soekers moet in gedagte hou dat nie alle soekenjins in die Onsigbare Web soek nie, en dit behoort weerspieël te word in die keuse van die soekinstrument. Nie alle inligting wat op die Web en Internet gevind word is betroubaar, op datum en akkuraat nie, en Web-inligting moet geëvalueer word in terme van outoriteit, tydigheid, vooroordeel, doel van die Webruimte, ens. Verskillende soorte soekinstrumente is op die Internet beskikbaar, soos soekenjins, gidse, biblioteekpoorte, portale, intelligente agente, ens. Hierdie soekinstrumente is bestudeer en verken. 'n Nuwe kategorisering vir aanlyn soekinstrumente bestaande uit Intelligente Agente, Soekinstrumente, Gidse en Portale/Middelpunte word voorgestel. Hierdie kategorisering onderskei die hoofverskille tussen die 21 soorte soekinstrumente wat bestudeer is. Soekinstrumente/-enjins bestaan uit spinnekoppe, kruipers, robotte, indekse en soekinstrument sagteware. Hierdie soekinstrumente kan verder onderskei word deur hulle omvang, interne of eksterne soektogte en of hulle op Webbladsye of Webruimtes soek. Die meeste soekinstrumente werk in verhouding met ander soekinstrumente, en hulle deel dikwels resultate, spinnekoppe en databasisse. Hierdie verhouding is baie dinamies. Die hoof internasionale soekenjins het soekeienskappe wat identifiseerbaar is. Die eienskappe van Google, Yahoo en Excite is in besonderhede bestudeer. Soekenjins soek op verskillende maniere na inligting, en lê hulle resultate verskillend voor. Hierdie karaktereienskappe is krities vir die Herwinning/Presisie verhouding. 'n Goedbeplande soekstrategie sal die Herwinning/Presisie verhouding verbeter. Internet soekinstrumente/-enjins is nie die wondermiddel vir alle inligtingsbehoeftes nie, en het voor- en nadele. Die Internet soekinstrument evalueringsmeganisme se ontwikkeling is gebaseer op gewenste eienskappe van die hoof soekinstrumente, en word beskou as 'n standaard vir Internet soekinstrumente. Hierdie instrument, toegepas op drie Suid-Afrikaanse soekenjins, het insae verskaf in die doeltreffendheid van die plaaslike soekinstrumente soos vergelyk met die standaard wat in hierdie studie voorgestel word. In die studie word tot die slotsom gekom dat die plaaslike soekenjins gunstig vergelyk met die hoof soekenjins, maar nie genoegsaam sodat hulle eksklusief gebruik kan word nie. Verdere navorsing oor hierdie aspek is nodig. Intelligente Agente sal waarskynlik meer gewild word, maar die enigste sekerheid vir die toekoms van Internet soekinstrumente is verandering, verandering en nogmaals verandering.
Homoceanu, Silviu [Verfasser], and Wolf-Tilo [Akademischer Betreuer] Balke. "What Search Engines Can’t Do. Holistic Entity Search on Web Data / Silviu Homoceanu ; Betreuer: Wolf-Tilo Balke." Braunschweig : Technische Universität Braunschweig, 2015. http://d-nb.info/1175819921/34.
Full textPetit, Albin [Verfasser], Lionel [Akademischer Betreuer] Brunie, and Michael [Akademischer Betreuer] Granitzer. "Introducing Privacy in Current Web Search Engines / Albin Petit ; Lionel Brunie, Michael Granitzer." Passau : Universität Passau, 2017. http://d-nb.info/1129450872/34.
Full textZamir, Oren Eli. "Clustering web documents : a phrase-based method for grouping search engine results /." Thesis, Connect to this title online; UW restricted, 1999. http://hdl.handle.net/1773/6884.
Full textNgindana, Mongezi. "Visibility of e-commerce websites to search engines : a comparison between text-based and graphic-based hyperlinks /." Thesis, Click here for online access, 2006. http://dk.cput.ac.za/cgi/viewcontent.cgi?article=1081&context=td_cput.
Full textChiou, Bo-Yun. "Google takes on China a cross-cultural analysis of internet service design /." Muncie, Ind. : Ball State University, 2009. http://cardinalscholar.bsu.edu/643.
Full textLi, Ping 1965. "Doctoral students’ mental models of a web search engine : an exploratory study." Thesis, McGill University, 2007. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=94181.
Full textCette recherche préliminaire examine les facteurs qui peuvent influencer les modèles mentaux d’un groupe spécifique d’utilisateurs d’un moteur de recherche sur le Web: Google, mesurés selon l’étendue de leur réussite.Une échelle de cette réussite en suivant un modèle mental a été constituée en adaptant les modèles présentés par Borgman, Dimitroff et Saxon, incluant la perception (1) de la nature du moteur de recherche sur le Web, (2) des caractéristiques de la recherche propres à ce moteur, (3) de l’interaction entre le chercheur et le moteur de recherche. A l’aide de cette échelle, le niveau de réussite par un sujet donné utilisant un modèle mental a été déterminé en fonction du nombre de composantes des deux premières parties de l’échelle décrites et du niveau d’interaction entre le sujet et le moteur Google, tel que révélé par ses recherches. Le choix des facteurs a été fondé sur des études précédentes portant sur les différences individuelles entre les chercheurs d’information, comprenant le degré d’expérience d’une telle recherche par l’utilisateur, son style cognitif, son style d’apprentissage, ses aptitudes techniques, la formation reçue, la discipline et le sexe. Seize étudiants en doctorat ayant l’anglais comme première langue ont participé à cette étude. Des entretiens individuels semi-dirigés ont permis de déterminer le niveau de réussite des étudiants suivant leur modèle mental, ainsi que leur expérience de la recherche, la formation reçue, la discipline et le sexe. Une observation technique directe a été utilisée pour observer l’interaction réelle des étudiants avec Google. Des tests standardisés ont été administrés pour déterminer le style cognitif des étudiants, leur style d’apprentissage et leurs aptitudes techniques. fr
Romero, Tris Cristina. "Client-side privacy-enhancing technologies in web search." Doctoral thesis, Universitat Rovira i Virgili, 2014. http://hdl.handle.net/10803/284036.
Full textLos motores de búsqueda (en inglés, Web Search Engines -WSEs-) son herramientas que permiten a los usuarios localizar información específica en Internet. Uno de los objetivos de los WSEs es devolver los resultados que mejor coinciden con los intereses de cada usuario. Para ello, los WSEs recogen y analizan el historial de búsqueda de los usuarios para construir perfiles. Como resultado, un usuario que envía una cierta consulta recibirá los resultados más interesantes en las primeras posiciones. Aunque ofrecen un servicio muy útil, también representan una amenaza para la privacidad de sus usuarios. Los perfiles se construyen a partir del historial de consultas y otros datos relacionados que pueden contener información privada y personal. Para evitar esta amenaza de privacidad, es necesario establecer mecanismos de protección de privacidad de motores de búsqueda. En la actualidad, existen varias soluciones en la literatura para proporcionar privacidad a estos usuarios. Uno de los objetivos de este trabajo es examinar las soluciones existentes, analizando sus diferencias y las ventajas y desventajas de cada propuesta. Después, basándonos en el estado del arte actual, presentamos nuevas propuestas que protegen la privacidad de los usuarios. Más concretamente, esta tesis doctoral propone tres protocolos que preservan la privacidad de los usuarios en las búsquedas web. La idea general es distribuir a los usuarios en grupos donde intercambian sus consultas, como método de ofuscación para ocultar las consultas reales de cada usuario. El primer protocolo distribuido que proponemos se centra en reducir el tiempo de espera de la consulta, es decir, el tiempo que cada miembro del grupo tiene que esperar para recibir los resultados de la consulta. El segundo protocolo propuesto mejora anteriores propuestas porque resiste ataques internos, mejorando propuestas similares en términos de cómputo y comunicación. La tercera propuesta es un protocolo P2P, donde los usuarios se agrupan según sus preferencias. Esto permite ofuscar los perfiles de los usuarios pero conservando a sus intereses generales. En consecuencia, el WSE es capaz de clasificar mejor los resultados de sus consultas.
Web search engines (WSEs) are tools that allow users to locate specific information on the Internet. One of the objectives of WSEs is to return the results that best match the interests of each user. For this purpose, WSEs collect and analyze users’ search history in order to build profiles. Consequently, a profiled user who submits a certain query will receive the results which are more interesting for her in the first positions. Although they offer a very useful service, they also represent a threat for their users’ privacy. Profiles are built from past queries and other related data that may contain private and personal information. In order to avoid this privacy threat, it is necessary to provide privacy-preserving mechanisms that protect users. Nowadays, there exist several solutions that intend to provide privacy in this field. One of the goals of this work is to survey the current solutions, analyzing their differences and remarking the advantages and disadvantages of each approach. Then, based on the current state of the art, we present new proposals that protect users’ privacy. More specifically, this dissertation proposes three different privacy-preserving multi-party protocols for web search. A multi-party protocol for web search arranges users into groups where they exchange their queries. This serves as an obfuscation method to hide the real queries of each user. The first multi-party protocol that we propose focuses on reducing the query delay. This is the time that every group member has to wait in order to receive the query results. The second proposed multi-party protocol improves current literature because it is resilient against internal attacks, outperforming similar proposals in terms of computation and communication. The third proposal is a P2P protocol, where users are grouped according to their preferences. This allows to obfuscate users’ profiles but conserving their general interests. Consequently, the WSE is able to better rank the results of their queries.
González-Caro, Cristina. "Supervised Identification of the User Intent of Web Search Queries." Doctoral thesis, Universitat Pompeu Fabra, 2011. http://hdl.handle.net/10803/51300.
Full textA medida que la Web sigue creciendo, tanto en tamaño como en complejidad, la búsqueda Web llega a ser un servicio ubicuo que permite a los usuarios encontrar todo tipo de información, recursos y actividades. Sin embargo, así como la Web evoluciona también lo hacen las necesidades de los usuarios. Hoy en día, los usuarios tienen intereses más complejos que van más allá de las tradicionales consultas informacionales. Por lo tanto, es importante para los motores de búsqueda Web, no solo continuar respondiendo efectivamente las consultas informacionales y navegacionales, sino también identificar y proveer resultados precisos para los nuevos tipos de consultas. El objetivo de esta tesis es analizar el impacto de la intención de la consulta en el comportamiento de búsqueda de los usuarios. Para lograr esto, primero estudiamos el comportamiento de usuarios con diferentes intenciones en las páginas de resultados de motores de búsqueda (SERP). Nuestro estudio muestra que la intención de la consulta afecta todo el proceso de decisión en la SERP. Los usuarios con diferentes intenciones prefieren resultados de búsqueda diferentes (orgánicos, patrocinados), miran diferentes áreas de interés (título, snippet, URL, imagen) y se concentran en resultados con diferente posición en el ranking. Identificar automáticamente la intención de la consulta aportaría elementos valiosos que permitirán a los sistemas de búsqueda adaptar sus resultados a los comportamientos cambiantes del usuario. Por esto, esta tesis propone un método para identificar automáticamente la intención detrás de la consulta. Nuestra hipótesis es que el rendimiento de la clasificación de consultas basada en facetas simples puede ser mejorado con la introducción de ejemplos multi-faceta en el proceso de aprendizaje. Por lo tanto, estudiamos un grupo amplio de facetas e investigamos si la combinación de facetas puede mejorar su predictibilidad. Nuestros resultados muestran que esta idea puede mejorar significativamente la calidad de la clasificación. Dado que la mayoría de trabajos previos están orientados al estudio de facetas individuales, estos resultados son un primer paso hacia un modelo integrado de clasificación de la intención de la consulta.
Morrison, Patrick Jason. "Tagging and Searching: Search Retrieval Effectiveness of Folksonomies on the Web." [Kent, Ohio] : Kent State University, 2007. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=kent1177305096.
Full textTitle from PDF t.p. (viewed July 2, 2007). Advisor: David B. Robins. Keywords: information retrieval, search engine, social bookmarking, tagging, folksonomy, Internet, World Wide Web. Includes survey instrument. Includes bibliographical references (p. 137-141).
Chambers, Rickard. "Search engine strategies : a model to improve website visibility for SMME website /." Thesis, Click here for online access, 2005. http://dk.cput.ac.za/cgi/viewcontent.cgi?article=1079&context=td_cput.
Full text