To see the other types of publications on this topic, follow the link: Pages.

Dissertations / Theses on the topic 'Pages'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Pages.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Sanoja, Vargas Andrés. "Segmentation de pages web, évaluation et applications." Thesis, Paris 6, 2015. http://www.theses.fr/2015PA066004/document.

Full text
Abstract:
Les pages web sont devenues plus complexes que jamais, principalement parce qu'elles sont générées par des systèmes de gestion de contenu (CMS). Il est donc difficile de les analyser, c'est-à-dire d'identifier et classifier automatiquement les différents éléments qui les composent. La segmentation de pages web est une des solutions à ce problème. Elle consiste à décomposer une page web en segments, visuellement et sémantiquement cohérents, appelés blocs. La qualité d'une segmentation est mesurée par sa correction et sa généricité, c'est-à-dire sa capacité à traiter des pages web de différents types. Notre recherche se concentre sur l'amélioration de la segmentation et sur une mesure fiable et équitable de la qualité des segmenteurs. Nous proposons un modèle pour la segmentation ainsi que notre segmenteur Block-o-Matic (BoM). Nous définissons un modèle d'évaluation qui prend en compte le contenu ainsi que la géométrie des blocs pour mesurer la correction d'un segmenteur par rapport à une vérité de terrain. Ce modèle est générique, il permet de tester tout algorithme de segmentation et observer ses performances sur différents types de page. Nous l'avons testé sur quatre segmenteurs et quatre types de pages. Les résultats montrent que BOM surpasse ses concurrents en général et que la performance relative d'un segmenteur dépend du type de page. Enfin, nous présentons deux applications développées au dessus de BOM. Pagelyzer compare deux versions de pages web et décide si elles sont similaires ou pas. C'est la principale contribution de notre équipe au projet européen Scape (FP7-IP). Nous avons aussi développé un outil de migration de pages HTML4 vers le nouveau format HTML5<br>Web pages are becoming more complex than ever, as they are generated by Content Management Systems (CMS). Thus, analyzing them, i.e. automatically identifying and classifying different elements from Web pages, such as main content, menus, among others, becomes difficult. A solution to this issue is provided by Web page segmentation which refers to the process of dividing a Web page into visually and semantically coherent segments called blocks.The quality of a Web page segmenter is measured by its correctness and its genericity, i.e. the variety of Web page types it is able to segment. Our research focuses on enhancing this quality and measuring it in a fair and accurate way. We first propose a conceptual model for segmentation, as well as Block-o-Matic (BoM), our Web page segmenter. We propose an evaluation model that takes the content as well as the geometry of blocks into account in order to measure the correctness of a segmentation algorithm according to a predefined ground truth. The quality of four state of the art algorithms is experimentally tested on four types of pages. Our evaluation framework allows testing any segmenter, i.e. measuring their quality. The results show that BoM presents the best performance among the four segmentation algorithms tested, and also that the performance of segmenters depends on the type of page to segment.We present two applications of BoM. Pagelyzer uses BoM for comparing two Web pages versions and decides if they are similar or not. It is the main contribution of our team to the European project Scape (FP7-IP). We also developed a migration tool of Web pages from HTML4 format to HTML5 format in the context of Web archives
APA, Harvard, Vancouver, ISO, and other styles
2

Sanoja, Vargas Andrés. "Segmentation de pages web, évaluation et applications." Electronic Thesis or Diss., Paris 6, 2015. http://www.theses.fr/2015PA066004.

Full text
Abstract:
Les pages web sont devenues plus complexes que jamais, principalement parce qu'elles sont générées par des systèmes de gestion de contenu (CMS). Il est donc difficile de les analyser, c'est-à-dire d'identifier et classifier automatiquement les différents éléments qui les composent. La segmentation de pages web est une des solutions à ce problème. Elle consiste à décomposer une page web en segments, visuellement et sémantiquement cohérents, appelés blocs. La qualité d'une segmentation est mesurée par sa correction et sa généricité, c'est-à-dire sa capacité à traiter des pages web de différents types. Notre recherche se concentre sur l'amélioration de la segmentation et sur une mesure fiable et équitable de la qualité des segmenteurs. Nous proposons un modèle pour la segmentation ainsi que notre segmenteur Block-o-Matic (BoM). Nous définissons un modèle d'évaluation qui prend en compte le contenu ainsi que la géométrie des blocs pour mesurer la correction d'un segmenteur par rapport à une vérité de terrain. Ce modèle est générique, il permet de tester tout algorithme de segmentation et observer ses performances sur différents types de page. Nous l'avons testé sur quatre segmenteurs et quatre types de pages. Les résultats montrent que BOM surpasse ses concurrents en général et que la performance relative d'un segmenteur dépend du type de page. Enfin, nous présentons deux applications développées au dessus de BOM. Pagelyzer compare deux versions de pages web et décide si elles sont similaires ou pas. C'est la principale contribution de notre équipe au projet européen Scape (FP7-IP). Nous avons aussi développé un outil de migration de pages HTML4 vers le nouveau format HTML5<br>Web pages are becoming more complex than ever, as they are generated by Content Management Systems (CMS). Thus, analyzing them, i.e. automatically identifying and classifying different elements from Web pages, such as main content, menus, among others, becomes difficult. A solution to this issue is provided by Web page segmentation which refers to the process of dividing a Web page into visually and semantically coherent segments called blocks.The quality of a Web page segmenter is measured by its correctness and its genericity, i.e. the variety of Web page types it is able to segment. Our research focuses on enhancing this quality and measuring it in a fair and accurate way. We first propose a conceptual model for segmentation, as well as Block-o-Matic (BoM), our Web page segmenter. We propose an evaluation model that takes the content as well as the geometry of blocks into account in order to measure the correctness of a segmentation algorithm according to a predefined ground truth. The quality of four state of the art algorithms is experimentally tested on four types of pages. Our evaluation framework allows testing any segmenter, i.e. measuring their quality. The results show that BoM presents the best performance among the four segmentation algorithms tested, and also that the performance of segmenters depends on the type of page to segment.We present two applications of BoM. Pagelyzer uses BoM for comparing two Web pages versions and decides if they are similar or not. It is the main contribution of our team to the European project Scape (FP7-IP). We also developed a migration tool of Web pages from HTML4 format to HTML5 format in the context of Web archives
APA, Harvard, Vancouver, ISO, and other styles
3

Draheim, Dirk. "Typed server pages." [S.l. : s.n.], 2002. http://www.diss.fu-berlin.de/2003/57/index.html.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Törnell, Mattias. "Richard Pages Sound." Thesis, Linnéuniversitetet, Institutionen för musik och bild (MB), 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-27178.

Full text
Abstract:
Fem låtar från Richard Pages skiva Peculiar Life har analyserats med syfte att ta reda på vad det är som ger dem dess sound. Analyserna har gjorts genom läsning av bookleten till skivan, informationssökning på internet, genom att titta på videodagbok från inspelning av skivan och genom att lyssna på låtarna. Analyserna har delats upp i kategorierna skriven information, video, utrustning och låtanalyser. Analyserna har visat att alla led i produktionen har stor betydelse för slutresultatet. Informationen som samlats in har använts till att återskapa soundet genom att komponera, arrangera, producera, spela in och mixa fem egna låtar baserat på analyserna. En undersökning gjordes med en grupp bestående av 66 gymnasieelever på estetiskt program med syfte att ta reda på om de egenkomponerade låtarna har samma sound som referensmaterialet. Ljudklipp från referenslåtarna och de egenkomponerade låtarna spelades upp tätt efter varandra varpå eleverna fick besvara en lyssnarenkät som tagits fram. De kategorier som betygsattes i enkäten var harmonier, arrangering och produktion. På enkäten fanns en förklaring av kategorierna. Resultatet från enkäten har sammanställts och analyserats och visar på ett delvis lyckat genomförande inom samtliga bedömningskategorier. Det första låtparet fick högt betyg, det andra låtparet fick medelbetyg och det tredje låtparet högt betyg.
APA, Harvard, Vancouver, ISO, and other styles
5

Ahlsén, Emely. "Open spaces - blank pages." Thesis, Konstfack, Ädellab/Metallformgivning, 2008. http://urn.kb.se/resolve?urn=urn:nbn:se:konstfack:diva-3227.

Full text
Abstract:
I stand in front of the remnants of a building. Braces and beams create a silhouette against the sky. The walls and roof are missing, but the windows remain, and the sheet metal doors. The structure frames the emptiness. This former building became a metaphor for man. Vulnerable in the face of a sudden turn of events, we lack protection against the world around us. Only our skeleton, muscles and naked skin hold us together. And just as eyes are said to reflect the soul, the leftover windows provide a hint of the former building’s life. The windows became the key to this building’s history. We are dependent on our buildings, just as the buildings are dependent upon us. In times of change, our social structures support us. Our life cycles are the course of events that allow us to move on, change, and be unique. Marks are left with the passage of time. Facades can be open, opened, forced, closed, covered, empty, or ready to be filled. Facades can be the time that passes. We are facades in transition, but stable in our foundations.Nothing is really alive until there is something missing, and the opportunity to place ourselves into a context arises. And in that aspect, the emptiness following the fall of the Twin Towers is therefore more telling than the buildings themselves.Everything we see in a building tells a story. In this essay, the story of Berlin’s House of Parliament is told. I reflect on how man mirrors his present in a building that is in a process of transition and how we are forced to confront ourselves with its history. I consider how a new dimension is created when something is opened, handing over the inside to reveal its past. How time can alter a symbol’s meaning. Today the dome on the House of Parliament represents democracy, as opposed to one hundred years ago, when it stood for power.Facades are in constant change.Does a facade need to be affected by transformation, deconstruction and destruction so that we can understand its whole, or are there other ways to penetrate its surface? Is the structure’s ability to evolve andadapt that which fascinates us and allows us to see its possibilities?Both man and building need “change” in order to illustrate their past and interpret their present. This is most apparent in the moment that the "facade cracks". In the vacuum that is created, it becomes obvious that the present is not forever. To exist in the present, is to be in a state of change.
APA, Harvard, Vancouver, ISO, and other styles
6

Motlow, Elizabeth Handley. "Pages from my library /." Online version of thesis, 1985. http://hdl.handle.net/1850/8754.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Self, Lance. "APPLYING INTERACTIVE WEB PAGES." International Foundation for Telemetering, 2001. http://hdl.handle.net/10150/607689.

Full text
Abstract:
International Telemetering Conference Proceedings / October 22-25, 2001 / Riviera Hotel and Convention Center, Las Vegas, Nevada<br>Visitors to web pages are, in most cases, restricted to viewing information the page designer has anticipated they will be interested in viewing. Many times this is adequate, but there are instances where the visitor wants the information they view to be based on selections they choose. The Air Force Research Laboratory (AFRL) Space Vehicles Directorate anticipates selected customers will have a need to view very large data sets that vary from the satellite payload to the satellite state of health1, and will require controlling what they view in an “ad hoc” manner. In response, AFRL is using Java Server Pages developed within the data center to bring interactive and dynamic web page content to these customers.
APA, Harvard, Vancouver, ISO, and other styles
8

Wang, Yi. "Hierarchhical classification of web pages." Thesis, University of Macau, 2008. http://umaclib3.umac.mo/record=b1943013.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Bastard, Irène. "De proches en pages, de pages en proches : Exploration et réception des informations en ligne." Thesis, Paris, ENST, 2015. http://www.theses.fr/2015ENST0064/document.

Full text
Abstract:
Pourquoi publier un article d'actualité sur son mur Facebook ? Que signifie ce geste par rapport à d'autres formes d'interaction? Et qu'est-ce qu'apporte cette pratique aux multiples prises de la réception des médias? Cette recherche s'intéresse à l'activité de partage d'information en ligne, c'est-à-dire aux interactions numériques qui recourent à une information. Si les artefacts du web semblent généraliser le partage comme expression, force est de constater que les médias ne peuvent observer qu'une faible activité autour de leurs contenus. Le partage d'information n'est que la partie émergée de l'iceberg des sociabilités et des pratiques médiatiques. Un focus est réalisé sur les pratiques juvéniles. Les adolescents rappellent la prédominance du cadre interactionnel sur Facebook et expliquent que le dispositif permet de nombreuses stratégies. Dans ce contexte, les actualités ne trouvent une place qu'à condition d'avoir un sens social. Partager une information revient à dire sa réception du contenu, c'est du moins comme cela que le geste est lu. Enfin, cette recherche utilise les traces Facebook d'un large échantillon d'enquêtés grâce à l'application Algopol. Les médias forment une référence limitée mais stable parmi les liens publiés par les internautes. Les individus partagent de préférence certaines sources de médias, reproduisant en ligne des références a priori acquises hors ligne. Les enquêtés qui citent le même type de contenus ont de plus des similitudes sociodémographiques. Le partage d'information en ligne se développe à partir des pratiques médiatiques dans l'interstice des relations, entre exploration personnelle et réception des informations<br>Why do people post news on Facebook? What does a post mean compared to other forms ofexpression? And how does this activity contribute to media appropriation? This research focuses on sharing news, meaning an online interaction which uses information. This activity seems to be the social solution to supporting a person's online exploration. “Who shares what with whom?” is the question asked in this research through multiple approaches, both qualitative and quantitative. The first part is an exploration of online artefacts, counts of what media was shared online, and a few interviews. These diverse points of view show that sharing news enhances public expression of information, but only for some users. Sharing news is a limited part of online media consumption and online interaction. The second part of the research looks in detail at teenagers’ practices on Facebook. For young people on Facebook, all expression has to be social and online artefacts are a way to tell somebody one's position on an issue or on a relationship. News may be there to open interactions. The last part of the research deals with a large dataset of individual Facebook data, available thanks to Algopol Application. Analysis of the diversity of publications shows that the news people share online takes into account their offline media preferences and social activities. Sharing news is an online way of expression based on informations’ reception
APA, Harvard, Vancouver, ISO, and other styles
10

Bastard, Irène. "De proches en pages, de pages en proches : Exploration et réception des informations en ligne." Electronic Thesis or Diss., Paris, ENST, 2015. http://www.theses.fr/2015ENST0064.

Full text
Abstract:
Pourquoi publier un article d'actualité sur son mur Facebook ? Que signifie ce geste par rapport à d'autres formes d'interaction? Et qu'est-ce qu'apporte cette pratique aux multiples prises de la réception des médias? Cette recherche s'intéresse à l'activité de partage d'information en ligne, c'est-à-dire aux interactions numériques qui recourent à une information. Si les artefacts du web semblent généraliser le partage comme expression, force est de constater que les médias ne peuvent observer qu'une faible activité autour de leurs contenus. Le partage d'information n'est que la partie émergée de l'iceberg des sociabilités et des pratiques médiatiques. Un focus est réalisé sur les pratiques juvéniles. Les adolescents rappellent la prédominance du cadre interactionnel sur Facebook et expliquent que le dispositif permet de nombreuses stratégies. Dans ce contexte, les actualités ne trouvent une place qu'à condition d'avoir un sens social. Partager une information revient à dire sa réception du contenu, c'est du moins comme cela que le geste est lu. Enfin, cette recherche utilise les traces Facebook d'un large échantillon d'enquêtés grâce à l'application Algopol. Les médias forment une référence limitée mais stable parmi les liens publiés par les internautes. Les individus partagent de préférence certaines sources de médias, reproduisant en ligne des références a priori acquises hors ligne. Les enquêtés qui citent le même type de contenus ont de plus des similitudes sociodémographiques. Le partage d'information en ligne se développe à partir des pratiques médiatiques dans l'interstice des relations, entre exploration personnelle et réception des informations<br>Why do people post news on Facebook? What does a post mean compared to other forms ofexpression? And how does this activity contribute to media appropriation? This research focuses on sharing news, meaning an online interaction which uses information. This activity seems to be the social solution to supporting a person's online exploration. “Who shares what with whom?” is the question asked in this research through multiple approaches, both qualitative and quantitative. The first part is an exploration of online artefacts, counts of what media was shared online, and a few interviews. These diverse points of view show that sharing news enhances public expression of information, but only for some users. Sharing news is a limited part of online media consumption and online interaction. The second part of the research looks in detail at teenagers’ practices on Facebook. For young people on Facebook, all expression has to be social and online artefacts are a way to tell somebody one's position on an issue or on a relationship. News may be there to open interactions. The last part of the research deals with a large dataset of individual Facebook data, available thanks to Algopol Application. Analysis of the diversity of publications shows that the news people share online takes into account their offline media preferences and social activities. Sharing news is an online way of expression based on informations’ reception
APA, Harvard, Vancouver, ISO, and other styles
11

Li, Yong. "Topic-based segmentation of web pages." Thesis, University of Macau, 2005. http://umaclib3.umac.mo/record=b1445895.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Chambers, Cynthia R. "Turning Pages Together: Inclusive Book Clubs." Digital Commons @ East Tennessee State University, 2013. https://dc.etsu.edu/etsu-works/3867.

Full text
APA, Harvard, Vancouver, ISO, and other styles
13

Yang, Xiaoyu. "Querying Web pages with database query languages." Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1999. http://www.collectionscanada.ca/obj/s4/f2/dsk2/ftp01/MQ39899.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
14

MacGregor, Fiona M. "Agents of change : women creating web pages /." Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1999. http://www.collectionscanada.ca/obj/s4/f2/dsk1/tape10/PQDD_0019/MQ54934.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
15

Janevski, Angel. "UniversityIE: Information Extraction From University Web Pages." UKnowledge, 2000. http://uknowledge.uky.edu/gradschool_theses/217.

Full text
Abstract:
The amount of information available on the web is growing constantly. As a result, theproblem of retrieving any desired information is getting more difficult by the day. Toalleviate this problem, several techniques are currently being used, both for locatingpages of interest and for extracting meaningful information from the retrieved pages.Information extraction (IE) is one such technology that is used for summarizingunrestricted natural language text into a structured set of facts. IE is already being appliedwithin several domains such as news transcripts, insurance information, and weatherreports. Various approaches to IE have been taken and a number of significant resultshave been reported.In this thesis, we describe the application of IE techniques to the domain of universityweb pages. This domain is broader than previously evaluated domains and has a varietyof idiosyncratic problems to address. We present an analysis of the domain of universityweb pages and the consequences of having them input to IE systems. We then presentUniversityIE, a system that can search a web site, extract relevant pages, and processthem for information such as admission requirements or general information. TheUniversityIE system, developed as part of this research, contributes three IE methods anda web-crawling heuristic that worked relatively well and predictably over a test set ofuniversity web sites.We designed UniversityIE as a generic framework for plugging in and executing IEmethods over pages acquired from the web. We also integrated in the system a genericweb crawler (built at the University of Kentucky) and ported to Java and integrated anexternal word lexicon (WordNet) and a syntax parser (Link Grammar Parser).
APA, Harvard, Vancouver, ISO, and other styles
16

Santini, Marina. "Automatic identification of genre in Web pages." Thesis, University of Brighton, 2007. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.439208.

Full text
APA, Harvard, Vancouver, ISO, and other styles
17

Self, Lance. "SERVING INTERACTIVE WEB PAGES TO TechSat21 CUSTOMERS." International Foundation for Telemetering, 2001. http://hdl.handle.net/10150/607668.

Full text
Abstract:
International Telemetering Conference Proceedings / October 22-25, 2001 / Riviera Hotel and Convention Center, Las Vegas, Nevada<br>TechSat21 is an innovative satellite program sponsored by the Air Force Research Laboratory Space Vehicles Directorate and the Air Force Office of Scientific Research. Its mission is to control a cluster of satellites that, when combined, create a “virtual satellite” with which to conduct various experiments in sparse aperture sensing and formation flying. Because TechSat21 customers have a need to view very large data sets that vary from the payload to the satellite state of health1 a modern viewing method using Java Server Pages and Active Server Pages is being developed to meet these interactive dynamic demands.
APA, Harvard, Vancouver, ISO, and other styles
18

Siva, Sahithi Pokala. "Design and delivery : functional colour web pages." Thesis, University of Liverpool, 2001. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.343620.

Full text
APA, Harvard, Vancouver, ISO, and other styles
19

Papacharissi, Zozo-Angeliki. "The personal utility of individual home pages /." Digital version accessible at:, 2000. http://wwwlib.umi.com/cr/utexas/main.

Full text
APA, Harvard, Vancouver, ISO, and other styles
20

Frese, Anette. "Barocke Titelgraphik am Beispiel der Verlagsstadt Köln, 1570-1700 : Funktion, Sujet, Typologie /." Köln : Böhlau Verl, 1989. http://catalogue.bnf.fr/ark:/12148/cb35526429m.

Full text
APA, Harvard, Vancouver, ISO, and other styles
21

Wei, Chenjie. "Using Automated Extraction of the Page Component Hierarchy to Customize and Adapt Web Pages to Mobile Devices." The Ohio State University, 2012. http://rave.ohiolink.edu/etdc/view?acc_num=osu1338348757.

Full text
APA, Harvard, Vancouver, ISO, and other styles
22

Simioni, Angela <1995&gt. "Testing clarity and performance of two landing pages: the impact of people's emotions and perceptions on conversions and pages' performance." Master's Degree Thesis, Università Ca' Foscari Venezia, 2019. http://hdl.handle.net/10579/15764.

Full text
Abstract:
CRO, Conversion Rate Optimization, is that field of study which concerns about the ability to truly understand customers and their needs, translating the knowledge acquired about it into possible and increasing revenues. Landing pages enter in the world of CRO and usability, since are used as a place where to land potential customers; due to their importance, marketers must follow some best practice which can increase the probability a visitor converts. Anyway, not only visible and conscious elements matter, but also subconscious and emotional triggers have a great influence on the conversion rate. The thesis aims to study, through a research over two versions of a landing pages, how elements composing a landing page can deviate or contribute to the overall clarity and perception of the offer proposed; by using the five second test technique and an emotional matrix to detach respondents emotions, useful patterns emerge, also quantitatively proving the meaningful change from the first to the second version.
APA, Harvard, Vancouver, ISO, and other styles
23

Wagner, Jodi. "Development of dynamically-generated pages on a website." [Denver, Colo.] : Regis University, 2006. http://165.236.235.140/lib/JWagner2006.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
24

JUNIOR, EUGENIO PACELLI FERREIRA DIAS. "USING REINFORCEMENT LEARNING ON WEB PAGES REVISITING PROBLEM." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2012. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=19637@1.

Full text
Abstract:
PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO<br>No ambiente da Internet, as informações que desejamos frequentemente encontram-se em diferentes localidades. Algumas aplicações, para funcionarem corretamente, precisam manter cópias locais de parte dessas informações. Manter a consistência e a atualidade de uma base de dados, mais especificamente um conjunto de cópias de páginas web, é uma tarefa que vem sendo sistematicamente estudada. Uma abordagem possível a esse problema é a aplicação de técnicas de aprendizado por reforço, que utiliza técnicas de programação dinâmica e análise estocástica para obter uma boa política de agendamento de atualizações das cópias de páginas web. O presente trabalho tem por finalidade validar o uso de técnicas de aprendizado por reforço no problema em questão, assim como encontrar aspectos do problema que possam ser úteis na modelagem da solução empregada.<br>In the Internet, the information we desire is usually spread over different locations. For some applications, it is necessary to maintain local copies of this information. Keeping consistency as well as freshness of a data base, or more specifically a set of internet web pages, is a task systematically studied. An approach to this problem is the use of reinforcement learning techniques, using dynamic programming and stochastic analysis to obtain a good rescheduling policy for the web pages copies. This work is proposed to validate the use of reinforcement learning techniques over this problem, as well as finding features of the problem useful to model the developed solution.
APA, Harvard, Vancouver, ISO, and other styles
25

Eraslan, Sukru. "Eye tracking scanpath trend analysis on Web pages." Thesis, University of Manchester, 2016. https://www.research.manchester.ac.uk/portal/en/theses/eye-tracking-scanpath-trend-analysis-on-web-pages(ebe4f6de-308a-4c5a-885a-2ba7af04e4b7).html.

Full text
Abstract:
Web pages are typically comprised of different kinds of visual elements such as menus, headers and footers. To improve user experience, eye tracking has been widely used to investigate how users interact with such elements. In particular, eye movement sequences, called scanpaths, have been analysed to understand the path that people follow in terms of these elements. However, individual scanpaths are typically complicated and they are related to specific users, and therefore any processing done with those scanpaths will be specific to individuals and will not be representative of multiple users. Therefore, those scanpaths should be clustered to provide a general direction followed by users. This direction will allow researchers to better understand user interactions with web pages, and then improve the design of the pages accordingly. Existing research tends to provide a very short scanpath which is not representative for understanding user behaviours. This thesis introduces a new algorithm for clustering scanpaths, called Scanpath Trend Analysis (STA). In contrast to existing research, in STA, if a particular element is not shared by all users but it gets at least the same attention as the fully shared elements, it is included in the resulting scanpath. Thus, this algorithm provides a richer understanding of how users interact with web pages. The STA algorithm was evaluated with a series of eye tracking studies where the web pages used were automatically segmented into their visual elements by using different approaches. The results show that the outputs of the STA algorithm are significantly more similar to the inputted scanpaths in comparison with the outputs of other existing work, and this is not limited to a particular segmentation approach. The effects of the number of users were also investigated on the STA algorithm as the number of users required for scanpath analysis has not been studied in depth in the literature. The results show the possibility to reach the same results with a smaller group of users. The research presented in this thesis should be of value to eye tracking researchers, to whom the STA algorithm has been made available to analyse scanpaths, and to behaviour analysis researchers, who can use the algorithm to understand user behaviours on web pages, and then design, develop and present the pages accordingly.
APA, Harvard, Vancouver, ISO, and other styles
26

Nathan, Vikram. "Measuring time to interactivity for modern Web pages." Thesis, Massachusetts Institute of Technology, 2018. http://hdl.handle.net/1721.1/115749.

Full text
Abstract:
Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2018.<br>Cataloged from PDF version of thesis.<br>Includes bibliographical references (pages 53-56).<br>Web pages continually strive for faster loading times to improve user experience. However, a good metric for "page load time" is elusive. In particular, we contend that modern web pages should be evaluated with respect to interactivity: a page should be considered loaded when the user can fully interact with all visible content. However, existing metrics fail to accurately measure interactivity. On one hand, "page load time", the most widely used metric, overestimates the time to full interactivity by requiring that all content on a page has been both fetched and evaluated, including content below-the-fold that is not immediately visible to the user. Newer metrics like Above-the-Fold Time and Speed Index solve this problem by focusing primarily on above-the-fold content; however, these metrics only evaluate the time at which a page is fully visible to the user, disregarding page functionality, and thus interactivity. In this thesis, we define a new metric called Ready Index, which explicitly captures interactivity. Defining the metric is straightforward, but measuring it is not, since web developers do not explicitly annotate the parts of a page that support user interaction. To solve this problem, we introduce Vesper, a tool which rewrites a page's source code to automatically discover the page's interactive state. Armed with Vesper, we compare Ready Index to prior load time metrics like Speed Index. We find that, across a variety of network conditions, prior metrics underestimate or overestimate the true load time for a page by between 24% and 64%. Additionally, we introduce a tool that optimizes a page for Ready Index and is able to decrease the median time to page interactivity by between 29% and 32%.<br>by Vikram Nathan.<br>S.M.
APA, Harvard, Vancouver, ISO, and other styles
27

Caraponale, Vanessa Ferrer. "Traduções de Pages d\'Atelier de Francis Ponge." Universidade de São Paulo, 2017. http://www.teses.usp.br/teses/disponiveis/8/8146/tde-03052018-094557/.

Full text
Abstract:
Esta dissertação tem por objetivo propor o acesso a uma parte ainda pouco conhecida da obra de Francis Ponge por meio da tradução de quinze textos selecionados, extraídos da antologia de textos inéditos e manuscritos intitulada Pages dAtelier 1917 1982, em diferentes formatos e escritos em diferentes períodos, aqui organizados em torno de quatro temas centrais: Reflexões sobre a literatura, Autorreflexões poéticas, Objetos textuais e Em torno do surrealismo. Considerando a abordagem pretendida, buscamos apresentar as principais características da obra do escritor francês, a saber, as questões da linguagem, a materialidade das palavras e a ausência de formas, além de contextualizar sua obra, discutindo suas influências e trazendo alguns dos diálogos estabelecidos em seus textos com escritores e filósofos franceses. Além disso, o trabalho traz a discussão da tradução e da antologia como espaços de reescrita, importante para se pensar a recepção de Ponge no Brasil. Por se tratar de uma tradução comentada, trazemos ainda uma discussão sobre os tipos de notas possíveis nessa prática.<br>This dissertation aims at allowing the access to a little known portion of Francis Ponges works by means of the translation of fifteen selected texts, extracted from his anthology of unedited works and manuscripts titled Pages dAtelier 1917-1982, written in different formats and periods, that are here organized around four central themes: Reflections on literature, Poetic self-reflections, Textual objects and On Surrealism. Considering the intended approach, we seek to present the main characteristics of the french writers work, namely, language issues, the materiality of the words and the absence of form, as well as contextualizing his work, discussing its influences and bringing some of the dialogues established in his texts with french writers and philosophers. In addition also, this work brings a debate about the translation and the anthology as places of rewriting, something that is important when we think about Ponges reception in Brazil. Because it is a commented translation, we bring as well a discussion on the possible types of notes in this kind of practice.
APA, Harvard, Vancouver, ISO, and other styles
28

Alves, Ricardo João de Freitas. "Declarative approach to data extraction of web pages." Master's thesis, Faculdade de Ciências e Tecnologia, 2009. http://hdl.handle.net/10362/5822.

Full text
Abstract:
Thesis submitted to Faculdade de Ciências e Tecnologia of the Universidade Nova de Lisboa, in partial fulfilment of the requirements for the degree of Master in Computer Science<br>In the last few years, we have been witnessing a noticeable WEB evolution with the introduction of significant improvements at technological level, such as the emergence of XHTML, CSS,Javascript, and Web2.0, just to name ones. This, combined with other factors such as physical expansion of the Web, as well as its low cost, have been the great motivator for the organizations and the general public to join, with a consequent growth in the number of users and thus influencing the volume of the largest global data repository. In consequence, there was an increasing need for regular data acquisition from the WEB, and because of its frequency, length or complexity, it would only be viable to obtain through automatic extractors. However, two main difficulties are inherent to automatic extractors. First, much of the Web's information is presented in visual formats mainly directed for human reading. Secondly, the introduction of dynamic webpages, which are brought together in local memory from different sources, causing some pages not to have a source file. Therefore, this thesis proposes a new and more modern extractor, capable of supporting the Web evolution, as well as being generic, so as to be able to be used in any situation, and capable of being extended and easily adaptable to a more particular use. This project is an extension of an earlier one which had the capability of extractions on semi-structured text files. However it evolved to a modular extraction system capable of extracting data from webpages, semi-structured text files and be expanded to support other data source types. It also contains a more complete and generic validation system and a new data delivery system capable of performing the earlier deliveries as well as new generic ones. A graphical editor was also developed to support the extraction system features and to allow a domain expert without computer knowledge to create extractions with only a few simple and intuitive interactions on the rendered webpage.
APA, Harvard, Vancouver, ISO, and other styles
29

LUQUE, PAZ FRANCISCO. "Grammaires et grammaticalite. (pages critiques de syntaxe espagnole)." Paris 3, 1988. http://www.theses.fr/1989PA030085.

Full text
Abstract:
Puisque le critere de grammaticalite semble universellement aller de pair avec la norme grammaticale qui pretend l'instituer, norme a laquelle la langue paraitrait devoir etre infeodee dans toute production de discours, nous voulons montrer, en analysant certains aspects de la syntaxe espagnole dans les manuels les plus representatifs de la production grammaticale depuis un demi-siecle, particulierement en france, mais aussi en espagne, comment beaucoup de regles sur lesquelles repose la norme du "bon usage" que chaque grammairien etablit, se revelent non seulement contradictoires entre elles, selon le manuel etudie, mais aussi et surtout comment la grammaire de la langue et les grammaires normatives ne coincident pas dans l'acceptabilite de nombreux enonces, au point que, sans tenir compte du fait majeur que la langue se joue tres souvent des ornieres que l'on voudrait lui imposer -puisque a l'evidence la langue et la norme ne suivent pas necessairement le meme destin-, il convient de se demander si erasme n'avait pas raison, qui disait: "il y a autant de grammaires que de grammairiens"<br>As the criterion for grammatical acceptability always seems to go along with the grammatical norm from which it proceeds (and to which any speech-utterance is obviously supposed to be subordinated), we inted to show, through a critical analysis of some aspects of spanish syntax as it has been introduced and studied in the best-known and most often used grammar-books both in france and spain for half-a-century, not only how many grammar-rules, which have been put forward by their respective authors on the basis of their own language-acceptabilities, are self-contradictory (within each separate book) but how all these normative grammars themselves differ from the actual language-system at work in many accepted speech-occurences. In other words and despite the well-established fact that any language tends to elude the limits that one should like to impose upon it, the blatant diversity of points of view among grammarians makes one wonder if erasmus was not right when he mentioned that "there are as many grammars as grammarians"
APA, Harvard, Vancouver, ISO, and other styles
30

Weng, Daiyue. "Extracting structured data from Web query result pages." Thesis, Queen's University Belfast, 2016. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.709858.

Full text
Abstract:
A rapidly increasing number of Web databases are now become accessible via their HTML form- based query interfaces only. Comparing various services or products from a number of web sites in a specific domain is time-consuming and tedious. There is a demand for value-added Web applications that integrate data from multiple sources. To facilitate the development of such applications, we need to develop techniques for automating the process of providing integrated access to a multitude of database-driven Web sites, and integrating data from their underlying databases. This presents three challenges, namely query form extraction, query form matching and translation, and Web query result extraction. In this thesis, 1 focus on Web query result extraction, which aims to extract structured data encoded in semi-structured HTML pages, and return extracted data in relational tables. 1 begin by reviewing the existing approaches for Web query result extraction. 1 categorize them based on their degree of automation, i.e. manual, semi-automatic and fully automatic approaches. For each category, every approach will be described in terms of its technical features, followed by an analysis listing the advantages and limitations of the approach. The literature review leads to my proposed approaches, which resolve the Web data extraction problem, i.e. Web data record extraction, Web data alignment and Web data annotation. Each approach is presented in a chapter which includes the methodology, experiment and related work. The last chapter concludes the thesis.
APA, Harvard, Vancouver, ISO, and other styles
31

Chambers, Cynthia R. "Turning Pages Together: An Inclusive Book Club Program." Digital Commons @ East Tennessee State University, 2012. https://dc.etsu.edu/etsu-works/3870.

Full text
APA, Harvard, Vancouver, ISO, and other styles
32

Chambers, Cynthia R., and M. Sheffield. "Turning Pages Together: An Inclusive Book Club Project." Digital Commons @ East Tennessee State University, 2012. https://dc.etsu.edu/etsu-works/3872.

Full text
APA, Harvard, Vancouver, ISO, and other styles
33

Chambers, Cynthia R. "Turning Pages Together: Supporting Literacy and Social Engagement." Digital Commons @ East Tennessee State University, 2011. https://dc.etsu.edu/etsu-works/3848.

Full text
APA, Harvard, Vancouver, ISO, and other styles
34

Gkoutos, Georgios Vasileios. "Internet-based resource discovery in chemo-bioinformatics." Thesis, Imperial College London, 2002. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.268851.

Full text
APA, Harvard, Vancouver, ISO, and other styles
35

Jenkins, Charlotte. "Automatic classification and metadata generation for world-wide web resources." Thesis, University of Wolverhampton, 2002. http://hdl.handle.net/2436/89094.

Full text
Abstract:
The aims of this project are to investigate the possibility and potential of automatically classifying Web documents according to a traditional library classification scheme and to investigate the extent to which automatic classification can be used in automatic metadata generation on the web. The Wolverhampton Web Library (WWLib) is a search engine that classifies UK Web pages according to Dewey Decimal Classification (DDC). This search engine is introduced as an example application that would benefit from an automatic classification component such as that described in the thesis. Different approaches to information resource discovery and resource description on the Web are reviewed, as are traditional Information Retrieval (IR) techniques relevant to resource discovery on the Web. The design, implementation and evaluation of an automatic classifier, that classifies Web pages according to DDC, is documented. The evaluation shows that automatic classification is possible and could be used to improve the performance of a search engine. This classifier is then extended to perform automatic metadata generation using the Resource Description Framework (RDF) and Dublin Core. A proposed RDF data model, schema and automatically generated RDF syntax are documented. Automatically generated RDF metadata describing a range of automatically classified documents is shown. The research shows that automatic classification is possible and could potentially be used to enable context sensitive browsing in automated web search engines. The classifications could also be used in generating context sensitive metadata tailored specifically for the search engine domain.
APA, Harvard, Vancouver, ISO, and other styles
36

Karimian, Sichani Mandana. "Entwicklung und Implementierung einer statistischen Auswertung für die Dekanatsdatenbank der Fakultät Elektrotechnik und Informationstechnik." [S.l. : s.n.], 2000. http://www.bsz-bw.de/cgi-bin/xvms.cgi?SWB8862164.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Bou, Issa Youssef. "Accessibilité aux informations visuelles des pages web pour les non-voyants." Toulouse 3, 2010. http://thesesups.ups-tlse.fr/1274/.

Full text
Abstract:
Le principal objectif de cette thèse est d'améliorer l'accessibilité à l'information visuelle des pages web pour les non voyants. En se basant d'une part sur le concept des Images de Pages (IdP) et sur le Modèle d'Architecture Textuelle (MAT) qui permettent de représenter la dimension visuelle d'un objet textuel, et d'autre part, sur le modèle RDF (Resource Description Framework) qui permet de décrire les pages web, nous avons conçu un nouveau modèle appelé MAP-RDF (Modèle d’Architecture des Pages web). Ce modèle permet de représenter les informations qui restent « cachées » aux non-voyants tels que la structure de la page, les groupements visuels et leurs propriétés ainsi que les relations associant ces groupements. Toutes ces informations sont traduites par la suite dans les termes d'un langage tactile que nous avons élaboré. Nous avons également réalisé une évaluation des Images de Pages Web tactiles auprès de sujets non-voyants. Les résultats étant très prometteurs, nous avons pu développer un prototype qui permet d'une part, d'annoter une page web suivant notre modèle, et d'autre part de générer automatiquement l'affichage en tactile de la page web répartie selon trois niveaux de granularité d'Images de Pages que nous avons identifiés expérimentalement<br>The main objective of my thesis is to enhance the accessibility to the visual information in web pages for the blind users. Based on the concepts of IdP and MAT that offer a representation of the visual aspect of a textual object and on the RDF model that facilitates the description of the web pages, we conceived a new model that we called MAP-RDF. This model offers a representation of the visual information that remains hidden to the blind, like the visual structure of the page, the visual groupings, their properties and the relations that associate these groupings. The totality of this information is translated into the terms of a tactile language that we developed. We also fulfilled an evaluation of the tactile Images of Pages (IdP) with blind persons. The results being very promising, we could develop a prototype that permits the annotation of web pages according to our model. It can also generate automatically a tactile view of the web page in which the visual information is distributed into three levels of granularity that we defined experimentally
APA, Harvard, Vancouver, ISO, and other styles
38

Kennedy, Alistair. "Automatic genre classification of home pages on the web /." Halifax, N.S. : Dalhousie University, 2004. http://torch.cs.dal.ca/%7Ekennedy/genre%5Fthesis.pdf.

Full text
Abstract:
Thesis (B.C.S.)--Dalhousie University, Halifax.<br>"Submitted in partial fulfillment of the requirements for the degree of bachelor of computer science with honours at Dalhousie University, Halifax, Nova Scotia, April 2004." Includes bibliographical references (p. 33-35). Also available in PDF via the World Wide Web.
APA, Harvard, Vancouver, ISO, and other styles
39

Sonon, Marcia S. "IT-21 compliant controlled access to Internet web pages." Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 1998. http://handle.dtic.mil/100.2/ADA355950.

Full text
Abstract:
Thesis (M.S. in Systems Engineering) Naval Postgraduate School, September 1998.<br>Thesis advisor(s): Gus K. Lott. "September 1998." Includes bibliographical references (p. 103-104). Also available online.
APA, Harvard, Vancouver, ISO, and other styles
40

Wildgruber, Robert. "Proteomanalyse von Saccharomyces cerevisiae zur Initialisierung der PAGES-Datenbank." [S.l. : s.n.], 2002. http://deposit.ddb.de/cgi-bin/dokserv?idn=965678806.

Full text
APA, Harvard, Vancouver, ISO, and other styles
41

Dash, Suvendu Kumar. "Context-based metrics for evaluating changes to web pages." Thesis, Texas A&M University, 2003. http://hdl.handle.net/1969.1/79.

Full text
Abstract:
The web provides a lot of fluid information but this information changes, moves, and even disappears over time. Bookmark lists, portals, and paths are collections where the building blocks are web pages, which are susceptible to these changes. A lot of research, both in industry and in academia, focuses on organizing this vast amount of data. In this thesis, I present context-based algorithms for measuring changes to a document. The methods proposed use other documents in a collection as the context for evaluating changes in the web pages. These metrics will be used in maintaining paths as the individual pages in paths change. This approach will enhance the evaluations of change made by the currently existing Path Manager, in the Walden's Paths project that is being developed in the Center for the Study of Digital Libraries at Texas A&M University.
APA, Harvard, Vancouver, ISO, and other styles
42

Williams, Nicola Kim. "Objective predictors of subjective aesthetic ratings of web pages." Thesis, Keele University, 2015. http://eprints.keele.ac.uk/2324/.

Full text
Abstract:
This research is concerned with the effect of visual stimulus on decision-­‐making and opinions, what visual aspects of a page affect very early impressions of web sites, and how this relates to computational methods of prediction and evaluation of web pages. The aim of this study was to discover whether there are identifiable visual attributes of web pages that can be used to predict subjective opinions. This was explored through three separate studies. These consisted of two correlational studies and a categorisation task. Participants were gained through convenience and snowball sampling, and the materials reviewed were two distinct sets of web pages. Cards sorts, laddering and an online data collection tool were used to gather the information. Both qualitative and quantitative analysis was used to explore the information. The visual attributes found to correlate with subjective opinions were inconsistent across the two correlational studies. Study One had a number of limitations that may have contributed to this inconsistency. Concrete findings were that levels of encouragement and discouragement influenced by web pages are on two distinct scales, as, although there is a negative correlation between them, a large number of pages were rated poorly on both scales. The similarity between the card sort and questionnaire results had consistent findings for predictors of low-­‐rated web pages. The findings from the cards sorts also show that users are able to make preference judgements of web pages without being able to understand the content. An application of the findings regarding prediction of low-­‐rated pages would be to create web design optimisation system, enabling web pages to be reviewed computationally. Although this should never replace user testing, it may provide an economical alternative during the early stages of design.
APA, Harvard, Vancouver, ISO, and other styles
43

Bloomfield, Mikahila T. "Impulse cues on the Facebook pages of apparel retailers." Thesis, University of Delaware, 2015. http://pqdtopen.proquest.com/#viewpdf?dispub=1585141.

Full text
Abstract:
<p> This study examined impulse cues on the Facebook pages of apparel retailers by extending previous research by Dawson &amp; Kim (2010), which examined impulse cues on the websites of top apparel retailers. The purpose of this study was to determine whether a relationship exists between impulse cues (Facebook content that entices consumers to buy) and web sales. The percentage of Facebook traffic website referrals for apparel retailers was also examined. The potentially influential role of impulse cues on web sales is highlighted in this study with distinct marketing implications for apparel retailers. When retailers use impulse cues on Facebook as part of their marketing strategy, they increase the potential for user engagement. The purpose of increasing user engagement is to increase conversion to web sales. The Facebook posts of the top apparel retailers were examined to identify the number of times impulse cues were used over a 30-day period in 2012. The top apparel retailers were extracted from InternetRetailer.com's 2012 Top 500 list. The Top 500 list ranks Internet retailers of all categories (electronics, books, apparel, beauty, etc.) using 2011 web sales data. Additional impulse cue categories were added to Dawson &amp; Kim's original impulse cue categories. The relationship between Facebook likes, comments, and shares and web sales of the retailers' was also analyzed.</p>
APA, Harvard, Vancouver, ISO, and other styles
44

FILHO, FRANCISCO BENJAMIM. "RANKING OF WEB PAGES BY LEARNING MULTIPLE LATENT CATEGORIES." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2012. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=19540@1.

Full text
Abstract:
PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO<br>O crescimento explosivo e a acessibilidade generalizada da World Wide Web (WWW) levaram ao aumento da atividade de pesquisa na área da recuperação de informação para páginas Web. A WWW é um rico e imenso ambiente em que as páginas se assemelham a uma comunidade grande de elementos conectada através de hiperlinks em razão da semelhança entre o conteúdo das páginas, a popularidade da página, a autoridade sobre o assunto e assim por diante, sabendo-se que, em verdade, quando um autor de uma página a vincula à outra, está concebendo-a como importante para si. Por isso, a estrutura de hiperlink da WWW é conhecida por melhorar significativamente o desempenho das pesquisas para além do uso de estatísticas de distribuição simples de texto. Nesse sentido, a abordagem Hyperlink Induced Topic Search (HITS) introduz duas categorias básicas de páginas Web, hubs e autoridades, que revelam algumas informações semânticas ocultas a partir da estrutura de hiperlink. Em 2005, fizemos uma primeira extensão do HITS, denominada de Extended Hyperlink Induced Topic Search (XHITS), que inseriu duas novas categorias de páginas Web, quais sejam, novidades e portais. Na presente tese, revisamos o XHITS, transformando-o em uma generalização do HITS, ampliando o modelo de duas categorias para várias e apresentando um algoritmo eficiente de aprendizagem de máquina para calibrar o modelo proposto valendo-se de múltiplas categorias latentes. As descobertas aqui expostas indicam que a nova abordagem de aprendizagem fornece um modelo XHITS mais preciso. É importante registrar, por fim, que os experimentos realizados com a coleção ClueWeb09 25TB de páginas da WWW, baixadas em 2009, mostram que o XHITS pode melhorar significativamente a eficácia da pesquisa Web e produzir resultados comparáveis aos do TREC 2009/2010 Web Track, colocando-o na sexta posição, conforme os resultados publicados.<br>The rapid growth and generalized accessibility of the World Wide Web (WWW) have led to an increase in research in the field of the information retrieval for Web pages. The WWW is an immense and prodigious environment in which Web pages resemble a huge community of elements. These elements are connected via hyperlinks on the basis of similarity between the content of the pages, the popularity of a given page, the extent to which the information provided is authoritative in relation to a given field etc. In fact, when the author of a Web page links it to another, s/he is acknowledging the importance of the linked page to his/her information. As such the hyperlink structure of the WWW significantly improves research performance beyond the use of simple text distribution statistics. To this effect, the HITS approach introduces two basic categories of Web pages, hubs and authorities which uncover certain hidden semantic information using the hyperlink structure. In 2005, we made a first extension of HITS, called Extended Hyperlink Induced Topic Search (XHITS), which inserted two new categories of Web pages, which are novelties and portals. In this thesis, we revised the XHITS, transforming it into a generalization of HITS, broadening the model from two categories to various and presenting an efficient machine learning algorithm to calibrate the proposed model using multiple latent categories. The findings we set out here indicate that the new learning approach provides a more precise XHITS model. It is important to note, in closing, that experiments with the ClueWeb09 25TB collection of Web pages, downloaded in 2009, demonstrated that the XHITS is capable of significantly improving Web research efficiency and producing results comparable to those of the TREC 2009/2010 Web Track.
APA, Harvard, Vancouver, ISO, and other styles
45

AMORIM, EVELIN CARVALHO FREIRE DE. "NCE: AN ALGORITHM FOR CONTENT EXTRACTION IN NEWS PAGES." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2009. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=31449@1.

Full text
Abstract:
PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO<br>COORDENAÇÃO DE APERFEIÇOAMENTO DO PESSOAL DE ENSINO SUPERIOR<br>FUNDAÇÃO DE APOIO À PESQUISA DO ESTADO DO RIO DE JANEIRO<br>PROGRAMA DE EXCELENCIA ACADEMICA<br>BOLSA NOTA 10<br>A extração de entidades de páginas web é comumente utilizada para melhorar a qualidade de muitas tarefas realizadas por máquinas de busca como detecção de páginas duplicadas e ranking. Essa tarefa se torna ainda mais relevante devido ao crescente volume de informação da internet com as quais as máquinas de busca precisam lidar. Existem diversos algoritmos para detecção de conteúdo na literatura, alguns orientados a sites e outros que utilizam uma abordagem mais local e são chamados de algoritmos orientados a páginas. Os algoritmos orientados a sites utilizam várias páginas de um mesmo site para criar um modelo que detecta o conteúdo relevante da página. Os algoritmos orientados a páginas detectam conteúdo avaliando as características de cada página, sem comparar com outras páginas. Neste trabalho apresentamos um algoritmo, chamado NCE ( News Content Extractor), orientado a página e que se propõe a realizar extração de entidades em páginas de notícias. Ele utiliza atributos de uma árvore DOM para localizar determinadas entidades de uma página de notícia, mais especificamente, o título e o corpo da notícia. Algumas métricas são apresentadas e utilizadas para aferir a qualidade do NCE. Quando comparado com outro método baseado em página e que utiliza atributos visuais, o NCE se mostrou superior tanto em relação à qualidade de extração quanto no que diz respeito ao tempo de execução.<br>The entity extraction of web pages is commonly used to enhance the quality of tasks performed by search engines, like duplicate pages and ranking. The relevance of entity extraction is crucial due to the fact that search engines have to deal with fast growning volume of information on the web. There are many algorithms that detect entities in the literature, some using site level strategy and others using page level strategy. The site level strategy uses many pages from the same site to create a model that extracts templates. The page level strategy creates a model to extract templates according to features of the page. Here we present an algorithm, called NCE (News Content Extractor), that uses a page level strategy and its objective is to perform entity extraction on news pages. It uses features from a DOM tree to search for certain entities, namely, the news title and news body. Some measures are presented and used to evaluate how good NCE is. When we compare NCE to a page level algorithm that uses visual features, NCE shows better execution time and extraction quality.
APA, Harvard, Vancouver, ISO, and other styles
46

Walsh, Thomas. "Automatic identification of presentation failures in responsive web pages." Thesis, University of Sheffield, 2018. http://etheses.whiterose.ac.uk/22104/.

Full text
Abstract:
With the increasing number and variety of devices being used to access the World Wide Web, providing a good browsing experience to all users, regardless of device, is a critical task. To do this, many web developers now use responsive web design (RWD) to build web pages that provide a bespoke layout tailored to the specific characteristics of the device in use, normally the viewport width. However, implementing responsive web pages is an error-prone task, as web page elements can behave in unpredictable ways as the viewport expands and contracts. This leads to presentation failures — errors in the visual appearance of the web page. As well-designed responsive web pages can have an array of benefits, identifying presentation failures quickly and accurately is an important task. Unfortunately, current approaches to detecting presentation failures in web pages are insufficient. The huge number of different viewport widths that require support makes thorough checking of the layout on all devices infeasible. Furthermore, the current range of developer tools only provide limited support for testing responsive web pages. This thesis tackles these problems by making the following contributions. First, it proposes the responsive layout graph (RLG), a model of the dynamic layout of modern responsive web pages. Then, it explores how the RLG can be used to automatically detect potentially unseen side-effects of small changes to the source code of a web page. Next, it investigates the detection of several common types of layout failures, leveraging implicit oracle information in place of an explicit oracle. Experiments showed both the approach for detecting potentially unseen side-effects and the approach for identifying common types of layout failure to be highly effective. The manual effort required by the user is further reduced by an approach that automatically grouped related failures together. Finally, a case study of 33 real-world responsive layout failures investigates how difficult such failures are to fix. These approaches have all been implemented into a software tool, ReDeCheck, which helps web developers create better responsive web pages.
APA, Harvard, Vancouver, ISO, and other styles
47

Abreu, Luís Pedro Borges. "Morphing Web Pages to Preclude Web Page Tampering Threats." Master's thesis, 2016. https://repositorio-aberto.up.pt/handle/10216/90184.

Full text
Abstract:
O número de utilizadores da Internet continua a aumentar todos os anos e a Internet é cada vez mais uma ferramenta diária na vida de cada indivíduo, utilizada como instrumento de trabalho ou de entretenimento. Contudo, ao navegar na Internet, os utilizadores tornam-se possíveis alvos de ataques informáticos um vez que efetuam transações de dados, muitas vezes privados e sensíveis, com servidores remotos.Entre os diferentes ataques informáticos existentes, destaca-se o ataque MitB que deu origem ao tema desta dissertação. Os ataques MitB são realizados com recurso a Malware instalado e em execução nos computadores dos utilizadores, que tem acesso às informações das janelas dos navegadores de Internet - por exemplo através de bibliotecas de funções do sistema operativo ou até recorrendo a extensões dos navegadores de Internet. Estes ataques utilizam âncoras do DOM para identificar as secções de uma página web onde pretendem atacar - recolhendo dados ou modificando a própria página. O resultado do ataque será diretamente influenciado pela capacidade do Malware em identificar os pontos de ataque numa determinada página web.O Polimorfismo é um conceito geral que pode ser aplicado a páginas web como uma ferramenta para neutralizar e derrotar este tipo de ataques informáticos, tal como foi documentado pela empresa Shape Security, Inc. em 2014. Aplicando técnicas de polimorfismo a páginas web, as respostas de um servidor serão textualmente diferentes entre si, mas o resultado visual apresentado ao utilizador será sempre o mesmo. Concretamente, os valores dos atributos estáticos e a estrutura dos documentos HTML poderão ser modificados no servidor, criando assim versões polimorfas de uma página web. Estas transformações podem ser realizadas em tempo real no servidor ou pré-calculadas. Desta forma, nunca dois documentos HTML serão textualmente iguais, tornando as páginas em alvos em movimento, dificultando os ataques MitB. Este nível de proteção é necessário uma vez que todas as alterações da página realizadas pelo atacante são locais e portanto difíceis de detectar pelas estruturas de segurança e controlo implementadas nos servidores dos fornecedores dos serviços.Neste dissertação, pretende-se desenvolver uma ferramenta baseada em polimorfismo para proteger páginas web e os seus utilizadores dos ataques MitB que recorrem a âncoras do DOM. Esta ferramenta será avaliada em precisão e eficiência. A precisão será avaliada comparando a lista de erros e de avisos gerada pela navegador de Internet para a página original e as várias versões polimorfas produzidas, enquanto que a eficiência será calculada recorrendo a tentativas automáticas de alterações indevidas das páginas protegidas com a ferramenta desenvolvida.<br>The number of Internet users keeps growing every year. Moreover, the Internet is becoming a daily tool, which impacts the individual's lives used either as a work tool or for entertainment purposes. However, by using it, people become possible targets for cyber attacks as they keep exchanging data, sometimes sensitive and private data, with remote servers.Among all the different attacks types, MitB is the reason behind the genesis of this thesis subject. MitB attacks are performed by a computer program running on user's computer that is commonly known as Malware, which has access to what happens inside a browser window. It can be a system library or even a browser extension programmed to, automatically, misrepresent the source code of the client-side server response, and other information stored in user's browsers. They rely on markup and DOM anchors to identify sections of a web page to attack. The end result of an attack will be dictated by the malware's ability to successfully identify the right location on the web page to perform the attack.Polymorphism is a broad concept that can be applied to web pages as a tool to both neutralize and defeat such kind of attacks, as documented by Shape Security, Inc. in 2014. Applying polymorphic techniques to web pages, the server response will be textually different between requests, but the visual display to the user will always be the same. That is, the values of static attributes and the structure of HTML documents may be modified on the server immediately before responses are sent off, creating a polymorphic version of the web page, or by pre-building this new versions on the server to decrease the real time computational costs. Therefore, no two HTML documents will be textually the same, turning web pages in somehow a moving target against MitB attacks. This level of protection is necessary since all changes are made locally, client side, making their detection difficult by control and security structures implemented on the service provider's servers.In this thesis, we aim to develop a tool based on polymorphism to protect web pages and users from MitB attacks based on markup and DOM anchors. This tool will be evaluated by accuracy and efficiency. The first metric will be evaluated by recording and comparing the list of errors and warnings generated by original web pages and by their polymorphic versions created with our tool. The efficiency will be evaluated by running automated attempts for tampering web pages protected by our tool.
APA, Harvard, Vancouver, ISO, and other styles
48

Abreu, Luís Pedro Borges. "Morphing Web Pages to Preclude Web Page Tampering Threats." Dissertação, 2016. https://repositorio-aberto.up.pt/handle/10216/90184.

Full text
Abstract:
O número de utilizadores da Internet continua a aumentar todos os anos e a Internet é cada vez mais uma ferramenta diária na vida de cada indivíduo, utilizada como instrumento de trabalho ou de entretenimento. Contudo, ao navegar na Internet, os utilizadores tornam-se possíveis alvos de ataques informáticos um vez que efetuam transações de dados, muitas vezes privados e sensíveis, com servidores remotos.Entre os diferentes ataques informáticos existentes, destaca-se o ataque MitB que deu origem ao tema desta dissertação. Os ataques MitB são realizados com recurso a Malware instalado e em execução nos computadores dos utilizadores, que tem acesso às informações das janelas dos navegadores de Internet - por exemplo através de bibliotecas de funções do sistema operativo ou até recorrendo a extensões dos navegadores de Internet. Estes ataques utilizam âncoras do DOM para identificar as secções de uma página web onde pretendem atacar - recolhendo dados ou modificando a própria página. O resultado do ataque será diretamente influenciado pela capacidade do Malware em identificar os pontos de ataque numa determinada página web.O Polimorfismo é um conceito geral que pode ser aplicado a páginas web como uma ferramenta para neutralizar e derrotar este tipo de ataques informáticos, tal como foi documentado pela empresa Shape Security, Inc. em 2014. Aplicando técnicas de polimorfismo a páginas web, as respostas de um servidor serão textualmente diferentes entre si, mas o resultado visual apresentado ao utilizador será sempre o mesmo. Concretamente, os valores dos atributos estáticos e a estrutura dos documentos HTML poderão ser modificados no servidor, criando assim versões polimorfas de uma página web. Estas transformações podem ser realizadas em tempo real no servidor ou pré-calculadas. Desta forma, nunca dois documentos HTML serão textualmente iguais, tornando as páginas em alvos em movimento, dificultando os ataques MitB. Este nível de proteção é necessário uma vez que todas as alterações da página realizadas pelo atacante são locais e portanto difíceis de detectar pelas estruturas de segurança e controlo implementadas nos servidores dos fornecedores dos serviços.Neste dissertação, pretende-se desenvolver uma ferramenta baseada em polimorfismo para proteger páginas web e os seus utilizadores dos ataques MitB que recorrem a âncoras do DOM. Esta ferramenta será avaliada em precisão e eficiência. A precisão será avaliada comparando a lista de erros e de avisos gerada pela navegador de Internet para a página original e as várias versões polimorfas produzidas, enquanto que a eficiência será calculada recorrendo a tentativas automáticas de alterações indevidas das páginas protegidas com a ferramenta desenvolvida.<br>The number of Internet users keeps growing every year. Moreover, the Internet is becoming a daily tool, which impacts the individual's lives used either as a work tool or for entertainment purposes. However, by using it, people become possible targets for cyber attacks as they keep exchanging data, sometimes sensitive and private data, with remote servers.Among all the different attacks types, MitB is the reason behind the genesis of this thesis subject. MitB attacks are performed by a computer program running on user's computer that is commonly known as Malware, which has access to what happens inside a browser window. It can be a system library or even a browser extension programmed to, automatically, misrepresent the source code of the client-side server response, and other information stored in user's browsers. They rely on markup and DOM anchors to identify sections of a web page to attack. The end result of an attack will be dictated by the malware's ability to successfully identify the right location on the web page to perform the attack.Polymorphism is a broad concept that can be applied to web pages as a tool to both neutralize and defeat such kind of attacks, as documented by Shape Security, Inc. in 2014. Applying polymorphic techniques to web pages, the server response will be textually different between requests, but the visual display to the user will always be the same. That is, the values of static attributes and the structure of HTML documents may be modified on the server immediately before responses are sent off, creating a polymorphic version of the web page, or by pre-building this new versions on the server to decrease the real time computational costs. Therefore, no two HTML documents will be textually the same, turning web pages in somehow a moving target against MitB attacks. This level of protection is necessary since all changes are made locally, client side, making their detection difficult by control and security structures implemented on the service provider's servers.In this thesis, we aim to develop a tool based on polymorphism to protect web pages and users from MitB attacks based on markup and DOM anchors. This tool will be evaluated by accuracy and efficiency. The first metric will be evaluated by recording and comparing the list of errors and warnings generated by original web pages and by their polymorphic versions created with our tool. The efficiency will be evaluated by running automated attempts for tampering web pages protected by our tool.
APA, Harvard, Vancouver, ISO, and other styles
49

"FLETT preliminary pages." Thesis, 2013. http://hdl.handle.net/10388/ETD-2013-08-1209.

Full text
Abstract:
This creative thesis examines and explores the point of first-contact on Turtle Island: the Eastern Coast of what is now known as Canada, now named Newfoundland and Labrador, through a braided poetry/prose narrative of three different cultures and times. The Vikings landed near L’Anse-Aux-Meadows in the early 1000‘s, the Beothuks were last seen in the 18th century, and today’s cod fishery has been on a moratorium since the late 1980‘s/early 1990‘s. An assortment of the few recorded Beothuk words which have survived are also incorporated into the manuscript. ***Because this is a creative work, only the preliminary pages are available on the internet. The manuscript itself, and all publication rights, remains in the possession of the author.***
APA, Harvard, Vancouver, ISO, and other styles
50

Chean, Chao-Nan, and 陳昭男. "Detection of Page Type, Time, and Key Terms of Web Pages." Thesis, 2003. http://ndltd.ncl.edu.tw/handle/49481238888740466198.

Full text
Abstract:
碩士<br>國立中正大學<br>資訊工程研究所<br>91<br>With the rapid growth of WWW, the amount of online resources is getting richer. Modern search engines not only provide general search service for web pages, but domain-specified or type-specified search service to meet users'' need. To be able to provide type-specified search service, one needs to build up an automatic mechanism for type detection. By statistical analysis of the web pages, we find out some features which are appropriate for type detection. We also propose a scoring method to evaluate which type the web page belongs. Sometimes, the time information described in the content of the web page may be different from the last modified time of the web page. We define some rules to detect the time information from the web page. When extracting key terms, three features are calculated for each term in the web page. They are: location, which is the term''s first appearance; emphatic tag, whether the term is emphasized by some kinds of HTML tag or not; and TFIDF, a generality measure of a term''s frequency in a web page.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography