To see the other types of publications on this topic, follow the link: Web-logs.

Dissertations / Theses on the topic 'Web-logs'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 33 dissertations / theses for your research on the topic 'Web-logs.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Rao, Rashmi Jayathirtha. "Modeling learning behaviour and cognitive bias from web logs." The Ohio State University, 2017. http://rave.ohiolink.edu/etdc/view?acc_num=osu1492560600002105.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Lam, Yin-wan, and 林燕雲. "Senior secondary students use of web-logs in writing Chinese." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2006. http://hub.hku.hk/bib/B37198361.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Chiara, Ramon. ""Aplicação de técnicas de data mining em logs de servidores web"." Universidade de São Paulo, 2003. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-19012004-093205/.

Full text
Abstract:
Com o advento da Internet, as empresas puderam mostrar-se para o mundo. A possibilidade de colocar um negócio na World Wide Web (WWW) criou um novo tipo de dado que as empresas podem utilizar para melhorar ainda mais seu conhecimento sobre o mercado: a sequência de cliques que um usuário efetua em um site. Esse dado pode ser armazenado em uma espécie de Data Warehouse para ser analisado com técnicas de descoberta de conhecimento em bases de dados. Assim, há a necessidade de se realizar pesquisas para mostrar como retirar conhecimento a partir dessas sequências de cliques. Neste trabalho são discutidas e analisadas algumas das técnicas utilizadas para atingir esse objetivo. é proposta uma ferramenta onde os dados dessas sequências de cliques são mapeadas para o formato atributo-valor utilizado pelo Sistema Discover, um sistema sendo desenvolvindo em nosso Laboratório para o planejamento e execução de experimentos relacionados aos algoritmos de aprendizado utilizados durante a fase de Mineração de Dados do processo de descoberta de conhecimento em bases de dados. Ainda, é proposta a utilização do sistema de Programação Lógica Indutiva chamado Progol para extrair conhecimento relacional das sessões de sequências de cliques que caracterizam a interação de usuários com as páginas visitadas no site. Experimentos iniciais com a utilização de uma sequência de cliques real foram realizados usando Progol e algumas das facilidades já implementadas pelo Sistema Discover.
APA, Harvard, Vancouver, ISO, and other styles
4

Holmes, Ashley Joyce. "Web logs in the Post-Secondary Writing Classroom: A Study of Purposes." NCSU, 2005. http://www.lib.ncsu.edu/theses/available/etd-03222005-205901/.

Full text
Abstract:
In the past few decades, education research has been thriving in the areas of computers and new technologies. Often, teachers turn to what is popular in the technological world for new ideas to use in their classrooms. One such technology that has become extremely popular in Web culture is Web logs, now most often referred to as ?weblogs,? or simply ?blogs.? The present work seeks to further research on weblogs in education by identifying the various ways in which current post-secondary writing course teachers are using them in their courses. This definitional study attempts to answer the question: for what educational, or non-educational, purposes are weblogs in post-secondary writing courses being used? The study looks at the way educators claim to be using weblogs in their courses based on how they explain their blog assignments to students (either on a course syllabus or course blog posting). Adding depth to the analysis, the study also explores survey responses from thirty-two college writing teachers across the country. The eleven main uses for weblogs in writing courses that this study identifies are as follows: 1) as a public space with a broad audience, 2) to post student work, 3) as a journal, 4) to reflect on course-related assignments, 5) for student discussion and interaction, 6) to explore and share ideas, as well as brainstorm, 7) to engage with and respond to assigned readings, 8) for collaborative projects, 9) to link to Web materials, 10) to ask and answer questions related to the course, and 11) to discuss topics not necessarily related to the course. After compiling data as to these current uses of weblogs in college writing courses, this researcher explores the implications of these uses, offering suggestions and drawing conclusions as to how the new technology of weblogs has impacted and will impact college level writing courses.
APA, Harvard, Vancouver, ISO, and other styles
5

Villalobos, Luengo César Alexis. "Análisis de archivos Logs semi-estructurados de ambientes Web usando tecnologías Big-Data." Tesis, Universidad de Chile, 2016. http://repositorio.uchile.cl/handle/2250/140417.

Full text
Abstract:
Magíster en Tecnologías de la Información
Actualmente el volumen de datos que las empresas generan es mucho más grande del que realmente pueden procesar, por ende existe un gran universo de información que se pierde implícito en estos datos. Este proyecto de tesis logró implementar tecnologías Big Data capaces de extraer información de estos grandes volúmenes de datos existentes en la organización y que no eran utilizados, de tal forma de transformarlos en valor para el negocio. La empresa elegida para este proyecto se dedicada al pago de cotizaciones previsionales de forma electrónica por internet. Su función es ser el medio por el cual se recaudan las cotizaciones de los trabajadores del país. Cada una de estas cotizaciones es informada, rendida y publicada a las instituciones previsionales correspondientes (Mutuales, Cajas de Compensación, AFPs, etc.). Para realizar su función, la organización ha implementado a lo largo de sus 15 años una gran infraestructura de alto rendimiento orientada a servicios web. Actualmente esta arquitectura de servicios genera una gran cantidad de archivos logs que registran los sucesos de las distintas aplicaciones y portales web. Los archivos logs tienen la característica de poseer un gran tamaño y a la vez no tener una estructura rigurosamente definida. Esto ha causado que la organización no realice un eficiente procesamiento de estos datos, ya que las actuales tecnologías de bases de datos relaciones que posee no lo permiten. Por consiguiente, en este proyecto de tesis se buscó diseñar, desarrollar, implementar y validar métodos que sean capaces de procesar eficientemente estos archivos de logs con el objetivo de responder preguntas de negocio que entreguen valor a la compañía. La tecnología Big Data utilizada fue Cloudera, la que se encuentra en el marco que la organización exige, como por ejemplo: Que tenga soporte en el país, que esté dentro de presupuesto del año, etc. De igual forma, Cloudera es líder en el mercado de soluciones Big Data de código abierto, lo cual entrega seguridad y confianza de estar trabajando sobre una herramienta de calidad. Los métodos desarrollados dentro de esta tecnología se basan en el framework de procesamiento MapReduce sobre un sistema de archivos distribuido HDFS. Este proyecto de tesis probó que los métodos implementados tienen la capacidad de escalar horizontalmente a medida que se le agregan nodos de procesamiento a la arquitectura, de forma que la organización tenga la seguridad que en el futuro, cuando los archivos de logs tengan un mayor volumen o una mayor velocidad de generación, la arquitectura seguirá entregando el mismo o mejor rendimiento de procesamiento, todo dependerá del número de nodos que se decidan incorporar.
APA, Harvard, Vancouver, ISO, and other styles
6

Vasconcelos, Leandro Guarino de. "Uma abordagem para mineração de logs para apoiar a construção de aplicações web adaptativas." Instituto Nacional de Pesquisas Espaciais (INPE), 2017. http://urlib.net/sid.inpe.br/mtc-m21b/2017/07.24.15.06.

Full text
Abstract:
Atualmente, há mais de 1 bilhão de web sites disponíveis. Neste enorme hiperespaço, há muitos web sites que fornecem o mesmo conteúdo ou serviço. Portanto, quando um usuário não encontra o que está procurando ou enfrenta dificuldades na interação, ele tende a procurar em outro web site. Para suprir as necessidades dos usuários atuais da Web, web sites adaptativos têm sido propostos. As abordagens de adaptação existentes geralmente adaptam o conteúdo das páginas de acordo com o interesse do usuário. Entretanto, a adaptação da estrutura da interface para atender às necessidades do usuário ainda necessita ser explorada. Nesta tese, uma abordagem é proposta para analisar o comportamento do usuário de aplicações Web durante a navegação, explorando a mineração de logs de cliente, chamada RUM (em inglês, Real-time Usage Mining). Nesta abordagem, as ações do usuário são coletadas na interface da aplicação e processadas de forma síncrona. Assim, a RUM é capaz de detectar problemas de usabilidade e padrões de comportamento para o usuário ativo, enquanto ele navega na aplicação. A fim de facilitar a implantação, a RUM fornece um toolkit que permite à aplicação consumir informações sobre o comportamento do usuário. Usando o toolkit, os desenvolvedores podem codificar adaptações que são automaticamente disparadas em resposta aos dados fornecidos pelo toolkit. Experimentos foram realizados em diferentes web sites para demonstrar a eficiência da abordagem em apoiar adaptações na interface que aprimoram a experiência do usuário.
Currently, there are more than 1 billion websites available. In this huge hyperspace, there are many websites that provide exactly the same content or service. Therefore, when the user does not find what she is looking for easily or she faces difficulties during the interaction, she tends to search for another website. In order to fullfil the needs and preferences of todays web users, adaptive websites have been proposed. Existing adaptation approaches usually adapt the content of pages according to the user interest. However, the adaptation of the interface structure in order to meet user needs and preferences is still incipient. In this thesis, an approach is proposed to analyze the user behavior of Web applications during navigation, exploring the mining of client logs, called RUM (Real-time Usage Mining). In this approach, user actions are collected in the applications interface and processed synchronously. Thus, RUM is able to detect usability problems and behavioral patterns for the current application user, while she is browsing the application. In order to facilitate its deployment, RUM provides a toolkit which allows the application to consume information about the user behavior. By using this toolkit, developers are able to code adaptations that are automatically triggered in response to the data provided by the toolkit. Experiments were conducted on different websites to demonstrate the efficiency of the approach in order to support interface adaptations that improve the user experience.
APA, Harvard, Vancouver, ISO, and other styles
7

Tanasa, Doru. "Web usage mining : contributions to intersites logs preprocessing and sequential pattern extraction with low support." Nice, 2005. http://www.theses.fr/2005NICE4019.

Full text
Abstract:
Le Web Usage Mining (WUM), domaine de recherche assez récent, correspond au processus d’extraction des connaissances à partir des données (ECD) appliquées aux données d’usage sur le Web. Il comporte trois étapes principales : le prétraitement des données, la découverte des schémas et l’analyse des résultats. La quantité des données d’usage à analyser ainsi que leur faible qualité (en particulier l’absence de structuration) sont les principaux problèmes en WUM. Les algorithmes classiques de fouille de données appliquées sur ces données donnent généralement des résultats décevants en termes de pratiques des internautes. Dans cette thèse, nous apportons deux contributions importantes pour un processus WUM, implémentées dans notre boîte à outils Axislogminer. D’abord, nous proposons une méthodologie générale de prétraitement des logs Web dont l’originalité consiste dans le fait qu’elle prend en compte l’aspect multi-sites du WUM. Nous proposons dans notre méthodologie quatre étapes distinctes : la fusion des fichiers logs, le nettoyage, la structuration et l’agrégation des données. Notre deuxième contribution vise à la découverte à partir d’un fichier log prétraité de grande taille, des comportements minoritaires correspondant à des motifs séquentiels de très faible support. Pour cela, nous proposons une méthodologie générale visant à diviser le fichier log prétraité en sous-logs, se déclinant selon trois approches d’extraction de motifs séquentiels au support faible (séquentielle, itérative et hiérarchique). Celles-ci ont été implémentées dans des méthodes concrètes hybrides mettant en jeu des algorithmes de classification et d’extraction de motifs séquentiels
The Web use mining (WUM) is a rather research field and it corresponds to the process of knowledge discovery from databases (KDD) applied to the Web usage data. It comprises three main stages : the pre-processing of raw data, the discovery of schemas and the analysis (or interpretation) of results. The quantity of the web usage data to be analysed and its low quality (in particular the absence of structure) are the principal problems in WUM. When applied to these data, the classic algorithms of data mining, generally, give disappointing results in terms of behaviours of the Web sites users (E. G. Obvious sequential patterns, stripped of interest). In this thesis, we bring two significant contributions for a WUM process, both implemented in our toolbox, the Axislogminer. First, we propose a complete methodology for pre-processing the Web logs whose originality consists in its intersites aspect. We propose in our methodology four distinct steps : the data fusion, data cleaning, data structuration and data summarization. Our second contribution aims at discovering from a large pre-processed log file the minority behaviours corresponding to the sequential patterns with low support. For that, we propose a general methodology aiming at dividing the pre-processed log file into a series of sub-logs. Based on this methodology, we designed three approaches for extracting sequential patterns with low support (the sequential, iterative and hierarchical approaches). These approaches we implemented in hybrid concrete methods using algorithms of clustering and sequential pattern mining
APA, Harvard, Vancouver, ISO, and other styles
8

Allam, Amir Ali. "Measuring the use of online corporate annual reports through the analysis of web server logs." Thesis, University of Birmingham, 2005. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.633067.

Full text
Abstract:
The current study investigates a novel area of accounting research; the use of online annual reports by corporate Websites' visitors. This study is of a cross-disciplinary nature as it involves knowledge from different fields including accounting, mass communication and computer science. It is argued that the examination of communication theory may provide insights into how best to enhance the value relevance of accounting information. The Internet, as a new means of communication, has many possible effects on the way accounting information is disseminated and the way its usage can be investigated. Traditional methods, including questionnaires and interviews, have been used to study the usage of annual reports. This study explores and applies a novel method, Web Server Log Analysis, to study how online annual reports are accessed by users. The study investigated the Web Server Log Files of six companies and found that the most accessed sections of online annual reports are the Notes to the Financial Statements, the Chairman Statement, and the Profit and Loss Account. Narrative sections were accessed more frequently compared to the financial sections of the annual report. In addition, the non-statutory sections were found to be more accessed by users compared to the statutory ones.
APA, Harvard, Vancouver, ISO, and other styles
9

Tanasa, Doru. "Fouille de données d'usage du Web : Contributions au prétraitement de logs Web Intersites et à l'extraction des motifs séquentiels avec un faible support." Phd thesis, Université de Nice Sophia-Antipolis, 2005. http://tel.archives-ouvertes.fr/tel-00178870.

Full text
Abstract:
Les quinze dernières années ont été marquées par une croissance exponentielle du domaine du Web tant dans le nombre de sites Web disponibles que dans le nombre d'utilisateurs de ces sites. Cette croissance a généré de très grandes masses de données relatives aux traces d'usage duWeb par les internautes, celles-ci enregistrées dans des fichiers logs Web. De plus, les propriétaires de ces sites ont exprimé le besoin de mieux comprendre leurs visiteurs afin de mieux répondre à leurs attentes. Le Web Usage Mining (WUM), domaine de recherche assez récent, correspond justement au processus d'extraction des connaissances à partir des données (ECD) appliqué aux données d'usage sur le Web. Il comporte trois étapes principales : le prétraitement des données, la découverte des schémas et l'analyse (ou l'interprétation) des résultats. Un processus WUM extrait des patrons de comportement à partir des données d'usage et, éventuellement, à partir d'informations sur le site (structure et contenu) et sur les utilisateurs du site (profils). La quantité des données d'usage à analyser ainsi que leur faible qualité (en particulier l'absence de structuration) sont les principaux problèmes en WUM. Les algorithmes classiques de fouille de données appliqués sur ces données donnent généralement des résultats décevants en termes de pratiques des internautes (par exemple des patrons séquentiels évidents, dénués d'intérêt). Dans cette thèse, nous apportons deux contributions importantes pour un processus WUM, implémentées dans notre bo^³te à outils AxisLogMiner. Nous proposons une méthodologie générale de prétraitement des logs Web et une méthodologie générale divisive avec trois approches (ainsi que des méthodes concrètes associées) pour la découverte des motifs séquentiels ayant un faible support. Notre première contribution concerne le prétraitement des données d'usage Web, domaine encore très peu abordé dans la littérature. L'originalité de la méthodologie de prétraitement proposée consiste dans le fait qu'elle prend en compte l'aspect multi-sites du WUM, indispensable pour appréhender les pratiques des internautes qui naviguent de fa»con transparente, par exemple, sur plusieurs sites Web d'une même organisation. Outre l'intégration des principaux travaux existants sur ce thème, nous proposons dans notre méthodologie quatre étapes distinctes : la fusion des fichiers logs, le nettoyage, la structuration et l'agrégation des données. En particulier, nous proposons plusieurs heuristiques pour le nettoyage des robots Web, des variables agrégées décrivant les sessions et les visites, ainsi que l'enregistrement de ces données dans un modèle relationnel. Plusieurs expérimentations ont été réalisées, montrant que notre méthodologie permet une forte réduction (jusqu'à 10 fois) du nombre des requêtes initiales et offre des logs structurés plus riches pour l'étape suivante de fouille de données. Notre deuxième contribution vise la découverte à partir d'un fichier log prétraité de grande taille, des comportements minoritaires correspondant à des motifs séquentiels de très faible support. Pour cela, nous proposons une méthodologie générale visant à diviser le fichier log prétraité en sous-logs, se déclinant selon trois approches d'extraction de motifs séquentiels au support faible (Séquentielle, Itérative et Hiérarchique). Celles-ci ont été implémentées dans des méthodes concrètes hybrides mettant en jeu des algorithmes de classification et d'extraction de motifs séquentiels. Plusieurs expérimentations, réalisées sur des logs issus de sites académiques, nous ont permis de découvrir des motifs séquentiels intéressants ayant un support très faible, dont la découverte par un algorithme classique de type Apriori était impossible. Enfin, nous proposons une boite à outils appelée AxisLogMiner, qui supporte notre méthodologie de prétraitement et, actuellement, deux méthodes concrètes hybrides pour la découverte des motifs séquentiels en WUM. Cette boite à outils a donné lieu à de nombreux prétraitements de fichiers logs et aussi à des expérimentations avec nos méthodes implémentées.
APA, Harvard, Vancouver, ISO, and other styles
10

Mantella, Dana G. ""Pro-ana" Web-log uses and gratifications towards understanding the pro-anorexia paradox." unrestricted, 2007. http://etd.gsu.edu/theses/available/etd-04182007-194043/.

Full text
Abstract:
Thesis (M.A.)--Georgia State University, 2007.
Cynthia Hoffner, committee chair; Jaye Atkinson, Mary Ann Romski, committee members. Electronic text (90 p.) : digital, PDF file. Title from file title page. Description based on contents viewed Dec. 14, 2007. Includes bibliographical references (p. 67-74).
APA, Harvard, Vancouver, ISO, and other styles
11

Stomeo, Carlo. "Applying Machine Learning to Cyber Security." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2018. http://amslaurea.unibo.it/17303/.

Full text
Abstract:
Intrusion Detection Systems (IDS) nowadays are a very important part of a system. In the last years many methods have been proposed to implement this kind of security measure against cyber attacks, including Machine Learning and Data Mining based. In this work we discuss in details the family of anomaly based IDSs, which are able to detect never seen attacks, paying particular attention to adherence to the FAIR principles. This principles include the Accessibility and the Reusability of software. Moreover, as the purpose of this work is the assessment of what is going on in the state of the art we have selected three approaches, according to their reproducibility and we have compared their performances with a common experimental setting. Lastly real world use case has been analyzed, resulting in the proposal of an usupervised ML model for pre-processing and analyzing web server logs. The proposed solution uses clustering and outlier detection techniques to detect attacks in an unsupervised way.
APA, Harvard, Vancouver, ISO, and other styles
12

Vàllez, Letrado Mari. "Exploración de procedimientos semiautomáticos para el proceso de indexación en el entorno web." Doctoral thesis, Universitat Pompeu Fabra, 2015. http://hdl.handle.net/10803/359393.

Full text
Abstract:
La ingente cantidad de información que existe actualmente hace necesario el desarrollo de herramientas, métodos y procesos que faciliten el acceso a la misma. Especialmente, se requieren sistemas de información que sean eficientes y precisos. Las técnicas de indexación cuentan con una larga tradición en este ámbito. Sin embargo, su aplicación a gran escala y en el contexto de la Web no siempre es viable por la magnitud y la heterogeneidad de la información presente en ella. En esta tesis se presentan dos propuestas para facilitar el proceso de indexación de documentos en Internet. La primera se caracteriza por el uso de técnicas de indexación semiautomáticas basadas en aspectos de posicionamiento web, que se aplican a través de una herramienta propia denominada DigiDoc MetaEdit. La segunda propone un modelo para la actualización de vocabularios controlados a partir del procesamiento de los logs de las búsquedas formuladas por los usuarios en los buscadores.
La ingent quantitat d'informació que hi ha actualment fa necessari el desenvolupament d'eines, mètodes i processos que facilitin l'accés a la mateixa. Especialment, es requereixen sistemes d'informació que siguin eficients i precisos. Les tècniques d'indexació compten amb una llarga tradició en aquest àmbit. No obstant això, la seva aplicació a gran escala i en el context de la web no sempre és viable per la magnitud i heterogeneïtat de la informació present en ella. En aquesta tesi es presenten dues propostes per a facilitar el procés d'indexació de documents a Internet. La primera es caracteritza per l'ús de tècniques d'indexació semiautomàtiques basades en aspectes de posicionament web, i que s'apliquen a través d'una eina pròpia anomenada DigiDoc MetaEdit. La segona proposa un model per a l'actualització de vocabularis controlats a partir del processament dels logs de les cerques formulades pels usuaris als cercadors.
The vast amount of information that currently exists necessitates the development of tools, methods and processes that facilitate access to it. In particular, information systems that are efficient and accurate are required. Indexing techniques have a long tradition of promoting the improvement of these systems. However, its application on a large scale and in the context of the Web is not always feasible because of the magnitude and diversity of the information in it. This thesis presents two proposals to facilitate the process of indexing documents on the Internet. The first is characterized by the use of semi-automatic indexing techniques based on aspects of SEO, and applied through a proprietary tool called DigiDoc MetaEdit. The second proposes a model for updating controlled vocabularies from the processing of logs of searches made by users on search engines.
APA, Harvard, Vancouver, ISO, and other styles
13

Belaud, Lydie. "Une approche ergonomique des sites marchands sur internet : de la perception au comportement des consommateurs." Phd thesis, Université de Bourgogne, 2011. http://tel.archives-ouvertes.fr/tel-00681182.

Full text
Abstract:
L'environnement très concurrentiel de la toile détermine le contexte de la recherche. Attirer un internaute n'est pas simple et de surcroît, cela ne suffit pas puisqu'il faut ensuite qu'il devienne un consommateur fidèle qui recommandera le site marchand. La littérature a identifié certaines variables ergonomiques isolées ou semi-groupées qui auraient une influence sur l'intention comportementale, en termes d'intention d'achat, de retour, ou en termes d'intention de recommander le site, mais aucune approche satisfaisante du construit ergonomie, considéré dans sa globalité, n'est recensée dans la littérature marketing. Ce fut l'un des enjeux de cette recherche doctorale puisque nous nous sommes demandé quelle pouvait être l'influence de la perception de l'ergonomie, comprise de façon holistique, sur l'intention comportementale de l'internaute. Des allers-retours entre la littérature et les études exploratoires nous ont conduits vers des questions de recherche complémentaires. En effet, évaluer l'influence de la perception de l'ergonomie requiert de prendre en considération l'orientation motivationnelle de l'internaute, les efforts sous-jacents au traitement d'un stimulus visuel étant en partie liés à la motivation d'un individu. Par ailleurs, nous avons voulu dépasser la seule perspective cognitive pour considérer une vision plus expérientielle de la consommation. Enfin, nous avons étudié la relation qu'il pouvait y avoir entre le comportement sur le site traduit par les données logs et l'intention comportementale. Ainsi, le paradigme Personne x Objet x Situation nous a paru le mieux adapté à cette recherche doctorale en ce qu'il stipule que pour déterminer les sources ou les causes d'un phénomène il faut à la fois prendre en compte les caractéristiques de l'objet étudié, les caractéristiques de l'individu ainsi que les caractéristiques situationnelles. La circonscription et la mesure de la perception de l'ergonomie, un construit jusque-là non clairement délimité en comportement du consommateur, est l'un des apports théoriques majeurs de cette recherche. Nous avons créé un instrument de mesure court, aux propriétés psychométriques très satisfaisantes et qui est valable pour tout type de site. Le modèle structurel a montré que la perception de l'ergonomie d'un site n'aurait pas d'influence sur l'intention comportementale en ce sens où la relation serait totalement médiatisée par la dimension hédonico sensorielle. D'un point de vue managérial, cette recherche doctorale insiste sur une révision des standards ergonomiques tels qu'ils sont définis actuellement, conjointement à une meilleure prise en compte de la dimension hédonico sensorielle. Le gestionnaire de l'entreprise virtuelle s'interrogera sur un ré-enchantement du site marchand, ainsi que sur le degré de réenchantement de ce site, ce qui passera par une véritable réflexion stratégique quant au positionnement de l'enseigne et un juste équilibre entre standards ergonomiques et variables expérientielles. D'un point de vue méthodologique, nous avons souligné la complémentarité expert/utilisateur et toute la richesse d'une multiangulation des sources de données. Nous nous sommes appuyés sur des méthodologies issues d'autres disciplines comme l'oculométrie traditionnellement utilisée en IHM et qui reste peu répandue en comportement du consommateur. Nous avons montré pourquoi l'utilisation des données eye tracking pouvait se révéler particulièrement intéressante pour le chercheur, la manipulation de variables liées au design étant alors évaluée par une réponse quasi physiologique comme le temps passé sur une zone. Enfin, nous avons proposé une nouvelle méthodologie de collecte des logs avec l'utilisation du logiciel Netobserve, un outil particulièrement intéressant pour des sites non administrés par le chercheur et qui rend une relative autonomie dans le choix du terrain de recherche.
APA, Harvard, Vancouver, ISO, and other styles
14

Lam, Yin-wan. "Senior secondary students use of web-logs in writing Chinese a case study = Xianggang gao zhong xue sheng zhong wen wang shang ri zhi xie zuo ge an yan jiu /." Click to view the E-thesis via HKUTO, 2006. http://sunzi.lib.hku.hk/hkuto/record/B37198361.

Full text
APA, Harvard, Vancouver, ISO, and other styles
15

Pettersson, Albin, and Robin Rogne. "Webbplats för översikt av loggar." Thesis, Karlstads universitet, Fakulteten för hälsa, natur- och teknikvetenskap (from 2013), 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:kau:diva-42974.

Full text
Abstract:
Detta är en avhandling om hur vi utfört ett projekt som ger vår kund Ninetech en enkel översikt av loggar. Letande av specifik information såsom loggar i stora system kan vara tidskrävande. Så var fallet för företaget Ninetech. Företagets vilja var därav att ha en webbplats för att presentera loggar i ett översiktligt format. Webbplatsens syfte var att underlätta och minska tiden vid letande efter loggar. Att förkorta tiden är viktigt eftersom loggarna innehåller värdefull information som företaget använder för att lösa supportärenden. Webbplatsen som skapats är en fullständig produkt som levererats till företaget och körs i produktion. I projektet har vi haft en nära kontakt med kund, därav har en agil metod tillämpats. Som resultat har webbplatsen tillfört behovet att finna loggar snabbt och enkelt. Webbplatsen har blivit en del av Ninetechs vardagliga arbete. Störst fokus i avhandlingen ägnas åt vyn som presenterar loggar.
This is a dissertation on how we performed a project that gives our customer Ninetech a simple overview of logs. Searching of specific information such as logs in large systems can be time consuming. This was the case for the company Ninetech. Therefore, the company’s will was to have a website to present the logs in a simple format. The website’s purpose was to facilitate and reduce the time when searching for logs. It is important to shorten the time because the logs contain valuable information that the company uses to resolve support issues. The website created is a complete product delivered to the company and is currently running in production. In the project, we have had a close contact with the customer, hence has an agile method applied. As a result, the website has brought the need to find logs quickly and easily. The website has become a part of Ninetech's everyday work. The main focus of the dissertation is devoted to the view that presents logs.
APA, Harvard, Vancouver, ISO, and other styles
16

Kilic, Sefa. "Clustering Frequent Navigation Patterns From Website Logs Using Ontology And Temporal Information." Master's thesis, METU, 2012. http://etd.lib.metu.edu.tr/upload/12613979/index.pdf.

Full text
Abstract:
Given set of web pages labeled with ontological items, the level of similarity between two web pages is measured using the level of similarity between ontological items of pages labeled with. Using similarity measure between two pages, degree of similarity between two sequences of web page visits can be calculated as well. Using clustering algorithms, similar frequent sequences are grouped and representative sequences are selected from these groups. A new sequence is compared with all clusters and it is assigned to most similar one. Representatives of the most similar cluster can be used in several real world cases. They can be used for predicting and prefetching the next page user will visit or for helping the navigation of user in the website. They can also be used to improve the structure of website for easier navigation. In this study the effect of time spent on each web page during the session is analyzed.
APA, Harvard, Vancouver, ISO, and other styles
17

Nunes, José Manuel Rodrigues. "Visualização de interação em cenários de comunicação humano-computador." Doctoral thesis, Universidade de Aveiro, 2017. http://hdl.handle.net/10773/23153.

Full text
Abstract:
Doutoramento em Informação e Comunicação em Plataformas Digitais
Os contextos infocomunicacionais suportados em mediação tecnológica (no contexto geral da comunicação mediada por computador) estão a tornar-se cada vez mais presentes nas atividades do dia a dia de um número crescente de indivíduos e instituições. Especificamente, as tecnologias e serviços da internet/web têm uma presença marcante nas instituições, um pouco por todo lado. Os sites internos das instituições (vulgo intranets) são desenvolvidos de acordo com as estratégias de comunicação internas, refletindo os fluxos internos de informação e respetivos serviços de comunicação que lhe estão associados. Um problema emergente tem a ver com a gestão destas plataformas infocomunicacionais internas (intranets) e da relação com os seus interlocutores externos (extranets), ambas em crescimento constante. Os especialistas em comunicação organizacional sentem a falta de ferramentas que lhes permita analisar (padrões de atividade e comportamento) e perceber o que realmente se está a passar dentro da instituição. De facto, estes instrumentos tendem a basear-se em métricas de cariz técnico clássico, em muitos casos, para afinações técnicas e não para gerir a comunicação organizacional ou para a análise e gestão da informação. Esta tese centra-se na conceção e avaliação destas ferramentas de análise e diagnóstico para que possam contribuir para um desenvolvimento destas infraestruturas sofisticadas e, consequentemente, melhorar a eficiência dos processos infocomunicacionais que lhes são intrínsecos. Um dos problemas está na identificação dos desajustes utilizador-sistema ao nível da interação humano-computador, que têm que ser completamente identificados, e os problemas prontamente apresentados à equipa que procede ao desenho e desenvolvimento das plataformas infocomunicacionais. O sistema tem que servir a organização e manter com eficácia os seus padrões de fluxo de informação e respetivas tarefas. O conceito sistémico de feedback apresenta-se aqui como fundamental e necessariamente eficiente para a rigorosa identificação de problemas na plataforma infocomunicacional de uma determinada instituição. As propostas apresentadas demonstram capacidade de diagnosticar problemas estruturais e de conteúdo a dois níveis: ao nível da própria interface dos serviços infocomunicacionais e ao nível da estrutura interna, ou de organização relacional de informação. Os serviços de diagnóstico apresentados baseiam-se em pressupostos de análise contextual fortemente suportados em técnicas de análise visual e revelam, através de algumas experiências de cariz empírico, conseguir dar resposta ao desafio lançado por esta tese.
Technologically mediated info-communicational scenarios are becoming more and more pervasive in the day-to-day activity of a growing number of individuals and institutions. Specifically, internet/web technologies and services have a strong presence in institutions worldwide. Internal web sites (also known as intranets) are developed in compliance with internal communication strategies, reflecting internal information, workflow and related communication services. An emerging problem concerns the management of these constantly growing internal info-communicational platforms (intranets) and its external counterparts (extranets). Organizational communication specialists lack efficient tools to analyze (activity and behavioral patterns) and understand what is really going on inside the institutions. In fact, these instruments tend to be based on classical technical metrics, in most situations, for technical tuning and not for organizational communication and information analysis. This thesis is focused on the conception and evaluation of these diagnostic tools in order to contribute to the development of these sophisticated infrastructures and, consequently, improve the efficiency of their internal info-communicational processes. One of the issues lies in identifying user-system mismatch at the human-computer interaction level, which must be thoroughly identified, and the problems pinpointed to the design team. The system must serve the organization and adapt perfectly to its internal communication strategies, sustaining efficiently its information and workflow patterns. Efficient feedback instruments are fundamental to identify info-communicational platform problems inside an institution. The offered proposals demonstrate the ability to diagnose structural and content issues at two levels: at the level of its own info-communication services interface, and at the level of the internal structure or relational layout of information. The presented diagnostic services are based upon assumed contextual analysis, strongly supported in visual assessment methods, and manage to provide a response to the challenge issued by this thesis, through some empirical experiments.
APA, Harvard, Vancouver, ISO, and other styles
18

Nassopoulos, Georges. "Deducing Basic Graph Patterns from Logs of Linked Data Providers." Thesis, Nantes, 2017. http://www.theses.fr/2017NANT4110/document.

Full text
Abstract:
Conformément aux principes de Linked Data, les fournisseurs de données ont publié des milliards de faits en tant que données RDF. Exécuter les requêtes SPARQL sur les endpoints SPARQL ou les serveurs Triple Pattern Fragments (TPF) permet de consommer facilement des données du Linked Data. Cependant, le traitement des requêtes SPARQL fédérées, tout comme le traitement des requêtes TPF, décompose la requête initiale en de nombreuses sous-requêtes. Les fournisseurs de données ne voient alors que les sous-requêtes et la requête initiale n’est connue que des utilisateurs finaux. La connaissance des requêtes exécutées est fondamentale pour les fournisseurs, afin d’assurer un contrôle de l’utilisation des données, d’optimiser le coût des réponses aux requêtes, de justifier un retour sur investissements, d’améliorer l’expérience utilisateur ou de créer des modèles commerciaux à partir de tendances d’utilisation. Dans cette thèse, nous nous concentrons sur l’analyse des logs d’exécution des serveurs TPF et des endpoints SPARQL pour extraire les Basic Graph Patterns (BGP) des requêtes SPARQL exécutées. Le principal défi pour l’extraction des BGPs est l’exécution simultanée des requêtes SPARQL. Nous proposons deux algorithmes : LIFT et FETA. Sous certaines conditions, nous constatons que LIFT et FETA sont capables d’extraire des BGPs avec une bonne précision et un bon rappel
Following the principles of Linked Data, data providers published billions of facts as RDF data. Executing SPARQL queries over SPARQL endpoints or Triple Pattern Fragments (TPF) servers allow to easily consume Linked Data. However, federated SPARQL query processing and TPF query processing decompose the initial query into subqueries. Consequently, the data providers only see subqueries and the initial query is only known by end users. Knowing executed SPARQL queries is fundamental for data providers, to ensure usage control, to optimize costs of query answering, to justify return of investment, to improve the user experience or to create business models of usage trends. In this thesis, we focus on analyzing execution logs of TPF servers and SPARQL endpoints to extract Basic Graph Patterns (BGP) of executed SPARQL queries. The main challenge to extract BGPs is the concurrent execution of SPARQL queries. We propose two algorithms: LIFT and FETA. LIFT extracts BGPs of executed queries from a single TPF server log. FETA extracts BGPs of federated queries from a log of a set of SPARQL endpoints. For experiments, we run LIFT and FETA on synthetic logs and real logs. LIFT and FETA are able to extract BGPs with good precision and recall under certain conditions
APA, Harvard, Vancouver, ISO, and other styles
19

Soto, Mu?oz Leonardo Humberto. "Desarrollo de modelo de negocio para un gestor de logs para aplicaciones desarrolladas en la nube (cloud)." Tesis, Universidad de Chile, 2014. http://repositorio.uchile.cl/handle/2250/131997.

Full text
Abstract:
Tesis para optar al grado de Mag?ster en Gesti?n para la Globalizaci?n
Este trabajo presenta el desarrollo de un modelo de negocios para BeautifulLogs , un gestor de logs de aplicaciones (m?viles o web) desarrolladas para una infraestructura conocida como la nube (cloud computing). El desarrollo del modelo de negocio sigue una metodolog?a basada en los m?todos Lean Startup y Customer Development, enfocada en validar con clientes reales los supuestos de un modelo de negocios y reflejando los avances y el estado de dicho modelo en un Canvas de Modelo de Negocios. Cada validaci?n se conforma de un experimento con un resultado esperado en el caso en que el supuesto sea correcto. Durante el desarrollo de este proceso para el modelo de negocios de BeautifulLogs se encontr? un espacio en el mercado (ya existente) de este tipo de herramientas, nicho basado en necesidades avanzadas que no est? siendo cubierto por las soluciones actuales. Entre estas necesidades se cuenta la necesidad de almacenar la informaci?n por m?s tiempo que las soluciones actuales (que promedian dos semanas), incrementar la potencia de las funciones de b?squeda y agregar elementos de m?tricas y anal?ticas inferidas a partir de los mismos logs. En base a estas necesidades se establecieron propuestas de valor empaquetadas en planes con un precio validado por los propios usuarios encuestados. El modelo de negocio sigue el esquema SaaS: Software como servicio por el cual el cliente paga una mensualidad (o anualidad) a cambio de su uso. Si bien los costos variables para poder ofrecer las soluciones ofrecidas en los planes resultaron ser relativamente elevados (para el mercado SaaS), existe un saludable margen entre el ingreso variable promedio por usuario (USD 143,33) y el costo variable por usuario proyectado (USD 68,13). Por otra parte, los costos de adquisici?n, la conversi?n de usuarios y el crecimiento neto de usuarios se presentan como las variables claves que determinar?n la rentabilidad del negocio, que en un escenario moderado proyecta ventas por casi un mill?n de d?lares en el octavo trimestre de vida del negocio. En su forma actual, el modelo requiere una inversi?n inicial de USD 350.000 a ocupar principalmente en desarrollo de la plataforma, as? como tambi?n en capital de trabajo. La recomendaci?n a seguir es levantar una fracci?n de esa inversi?n para efectuar una validaci?n de clientes mediante un producto m?nimo viable, que permita bajar el riesgo presentado por la variaci?n de las variables claves mencionadas anteriormente.
APA, Harvard, Vancouver, ISO, and other styles
20

Bednář, Martin. "Automatické testování projektu JavaScript Restrictor." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2020. http://www.nusl.cz/ntk/nusl-432879.

Full text
Abstract:
The aim of the thesis was to design, implement and evaluate the results of automatic tests for the JavaScript Restrictor project, which is being developed as a web browser extension. The tests are divided into three levels - unit, integration, and system. The Unit Tests verify the behavior of individual features, the Integration Tests verify the correct wrapping of browser API endpoints, and the System Tests check that the extension does not suppress the desired functionality of web pages. The System Tests are implemented for parallel execution in a distributed environment which has succeeded in achieving an almost directly proportional reduction in time with respect to the number of the tested nodes. The benefit of this work is detection of previously unknown errors in the JavaScript Restrictor extension and provision of the necessary information that allowed to fix some of the detected bugs.
APA, Harvard, Vancouver, ISO, and other styles
21

Hsiao, Kuang-Yu, and 蕭廣佑. "Fuzzy Data Mining on Web Logs." Thesis, 2004. http://ndltd.ncl.edu.tw/handle/74688406448388466191.

Full text
Abstract:
碩士
南台科技大學
資訊管理系
92
With the improvement of technology, the Internet has been becoming an important part of everyday life. Governmental institutions and enterprises propose to advertise and marketing through webs. With the traveling records of browsers, one can analyze the preference of browsers, further understand the demands of consumers, and promote the advertising and marketing. In this study, we utilize Maximum Forward Reference algorithm to find the travel pattern of browsers from web logs. Simultaneously, experts are asked to evaluate the fuzzy importance weightings for different webs. At last, we employ fuzzy data mining technique that combines Apriori algorithm with fuzzy weights to determine the associate rules. From the yielded association rules, one can be accurately aware the information consumers need and which webs they prefer. This is important to governmental institutions and enterprises. Enterprises can find the commercial opportunities and improve the design of webs by means of this study. Governmental institutions can realize the needs of people from the obtained association rules, make the promotion of policy more efficiently, and provide better service quality.
APA, Harvard, Vancouver, ISO, and other styles
22

Weng, Hong-yang, and 翁弘彥. "Exploring Web Logs on Internet Communications." Thesis, 2013. http://ndltd.ncl.edu.tw/handle/68202977758871794235.

Full text
Abstract:
碩士
世新大學
資訊管理學研究所(含碩專班)
101
In recent years, the globe use Internet the population and penetration rate continuously improve with the rapid development of the Internet. Not only the population is growth who is using Internet, but also people use Internet habits of proportion in their daily lives are increasing. With the development of the Internet, Web-based business market is growing. The companies also have to invest in the network market, the rise of e-commerce, whether it is online store, online advertising and so continue to increase. Since the rise of social networking sites to share will be faster of everything. So many people want to use of the convenience of this share distribution, valuable information or products quickly spread to consumers. Therefore, these research uses the sequential pattern mining by data mining for users to browse the website analyze the data. After the first site browsing will next visit the site to find more people visit the site of the reference node with a wider spread site of opinion leaders. And to find the relevance between the websites use the people matrix. Therefore, this paper will use these methods using social network analysis. To find new channels of advertising placement that through the results of this analysis enables companies.
APA, Harvard, Vancouver, ISO, and other styles
23

Lin, Ching-Nan, and 林慶南. "Enhancement of Web Sites Security Utilizing Web Logs Mining." Thesis, 2002. http://ndltd.ncl.edu.tw/handle/34333759103494653266.

Full text
Abstract:
碩士
中原大學
電子工程研究所
90
Abstract The problem of information security on the Web has become an important research issue recently. Because the Backdoors or information leak of scripts in Common Getaway Interface(CGI)is hidden inadvertently or premeditated by programmers, these problems cause enterprise’s information to be gotten illegally, and can’t be detected by security tools easily. Besides, Internet grows fast to encourage the important research of Web mining. Therefore, in order to detect Backdoor or information leak of CGI scripts that the some security tools can’t detect and to avoid damage of enterprises, we propose a log data mining to enhance the security of Web servers. First, we combine Web application log data with Web log data to solve the problems in Web log. Then, our method uses the density-based clustering algorithm to mine some abnormal Web log and Web application log data. The obtained information can help system administrator detecting the Backdoor or information leakages in programs more easily. Moreover, the mined information can help system administrator detecting the problem of CGI scripts from on-line Web site log data.
APA, Harvard, Vancouver, ISO, and other styles
24

Wang, Tseng-Pu, and 王曾甫. "Clustering Customers Based on Web-Browsing Logs." Thesis, 2012. http://ndltd.ncl.edu.tw/handle/76727886739977180393.

Full text
Abstract:
碩士
世新大學
資訊管理學研究所(含碩專班)
100
Therefore the income composition of enterprises at current internet industry, advertisement takes up to most of portion out of it. In tradition, promotional activities usually use basic reading/hearing users information (head count Mathematics) to classify target users group. However simply use head count Mathematics to classify target users group does not consider the differences of hobbies and interests of peoples which might cause the dynamic changes of people’s behavior. Therefore considering user’s behavior characteristic to proceed uer’s classification, which can make enterprise to understand internet user’s tendency as well as raise it’s overall marketing coverage rate. This research uses the database of website browsing behaviors recorded by Genesis Market Research Consulting Corporation to proceed research, also to use the seven kinds of consumer’s behavior researched and analyzed by Yahoo U.S.A., and eventually to establish the groups models of internet users in accordance with foregoing research.
APA, Harvard, Vancouver, ISO, and other styles
25

Caldera, Amithalal, University of Western Sydney, of Science Technology and Environment College, and School of Computing and Information Technology. "Effectively capturing user sessions on the Web using Web server logs." 2005. http://handle.uws.edu.au:8081/1959.7/11206.

Full text
Abstract:
The usage of Web sites has been of interest to Web administrators and researchers ever since the Web started. Analysis of Web site usage data helps to understand the behaviour of its users, which is very important, as many important decisions can be made based on it. The user behaviour may be deduced by knowing all the activities each user does from the time s/he starts a session on the Web site until s/he leaves it, which is collectively called a user session. As Web server logs explicitly record the browsing behaviour of site users and are readily and economically available, this thesis explores the use of Web server logs in capturing user sessions on Web. In order to protect users’ privacy, the standard Web server logs in general do not record the user identities or similar measures to uniquely identify the users. This thesis concentrates on heuristic strategies to infer user sessions. The heuristics exploit the background knowledge of user navigational behaviour recorded in the standard Web server logs without requiring additional information through cookies, logins and session ids. They identify relationships that may exist among the log data and make use of them to assess whether requests registered by the Web server can belong to the same individual and whether these requests were performed during the same visit. Researchers have proposed several heuristics, which were adversely affected by proxy servers, caching and undefined referrers. The thesis proposes new heuristics, which effectively address all the limitations, thus extending the work in this field. It also introduces a set of measures to quantify the performance of the heuristics and uses them to investigate their efficiency based on logs from three Web sites and makes recommendations for the Web sites to devise their own heuristics. The investigation has shown satisfactory results and the new heuristics are applicable to wider range of Web sites.
Doctor of Philosophy (PhD)
APA, Harvard, Vancouver, ISO, and other styles
26

林逸塵. "Collection and retrieval of suicide information in web logs." Thesis, 2006. http://ndltd.ncl.edu.tw/handle/6wm3pm.

Full text
APA, Harvard, Vancouver, ISO, and other styles
27

Tsai, Tsung-chou, and 蔡聰洲. "Integrating Data Warehousing and Data Mining for Web Logs Analysis." Thesis, 2001. http://ndltd.ncl.edu.tw/handle/27623832729147888571.

Full text
Abstract:
碩士
國立交通大學
資訊管理所
89
Because of the highly mature network technologies and the explosive growth of Internet, the numbers of Internet users increase substantially. In order to keep good relationships with customers and raise customers’ satisfaction, more and more companies start to provide their customers easy and fast retrieval of business information and services through Internet, such as WWW. Consequently, Web logs on the Web servers keep growing as time passing by. The Web logs are not useless and not just wasting memory space. Instead, they are important sources from which we can retrieve useful information. Effective analysis of Web logs is helpful to promote enterprise competence. In this thesis, integrated system architecture for analyzing Web logs is proposed. It integrates Data mining technology to provide recommendations for Web users and Data Warehousing technology to provide decision information for managers. A test Web site simulating a company selling computer peripherals has been set up to verify the proposed architecture. The results are analyzed and presented in the thesis. Finally, some comparisons and discussions with other related papers are presented.
APA, Harvard, Vancouver, ISO, and other styles
28

""Aplicação de técnicas de data mining em logs de servidores web"." Tese, Biblioteca Digital de Teses e Dissertações da USP, 2003. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-19012004-093205/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
29

Hong, Rong Zong, and 洪榮宗. "A Study on Security Enhancement of Web Sites Utilizing Logs Analysis." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/00011799081038754784.

Full text
Abstract:
碩士
長庚大學
資訊管理學研究所
97
The application of Internet becomes popular. The services provided by the website include those related to foods, clothing, living, transportation, education and entertainment. However, we usually hear that some websites have been hacked or attacked. The events led to information security incidents. The research issue of this thesis is : Can we use the server log to enhance the security in the website ? This paper presents a method that uses server logs as the source for analysis to enhance the website security. The server logs analyzed in the thesis include Fedora Code Operation System log, Apache server log, Snort log, IPTABLES log. As a result, we can prevent some potential known attacks. For those unknown attacks, we can identify potential threats and thus protect the security of the web server.
APA, Harvard, Vancouver, ISO, and other styles
30

Tang, Ran. "AN APPROACH FOR IDENTIFYING SERVICE COMPOSITION PATTERNS FROM EXECUTION LOGS." Thesis, 2010. http://hdl.handle.net/1974/6114.

Full text
Abstract:
Different types of web resources, such as Web Services, HTTP-based APIs and websites, can be located on the web to provide various services, such as information access and online banking. Such services are the basic building blocks to compose more complex functionality that cannot be achieved by a single service. Many service-oriented applications can be composed to fulfill similar functional requirements. Among various applications, a set of services can be frequently used together to deliver a unique functionality. Such set of services are called a service composition pattern. More specifically, a service composition pattern consists of a set of services and the control flow among the services. The reuse of the service composition patterns can facilitate the composition of new applications, improve existing applications and optimize maintenance process of services. To facilitate the identification and reuse of service composition patterns, we propose an approach that mines the service composition patterns from execution logs produced by service-oriented applications during runtime. Since the execution logs can be produced by heterogeneous web resources, we propose a unified description schema to describe various web resources in order to identify functionally similar services of different types. This helps reveal complete service composition patterns. Then we identify frequently associated services using Apriori algorithm and heuristics. Finally, we recover the control flow among the services using the event graph and process mining techniques. The effectiveness of the approach is evaluated through two case studies. The result shows that the unified description schema facilitates the identification of similar services of different types and our approach can effectively identify service composition patterns.
Thesis (Master, Electrical & Computer Engineering) -- Queen's University, 2010-09-29 18:08:07.55
APA, Harvard, Vancouver, ISO, and other styles
31

Cordes, Christopher Sean. "Blogging the future: Theory and use of web logs to enhance library information services." 2004. http://hdl.handle.net/10150/105509.

Full text
Abstract:
Digital resources are becoming a common medium to address patron and library staff needs. There are a number of means and applications used to satisfy information demands. These include digital reference materials, and the reference persons that provide access to these objects, information and instruction web sites and, multi-media applications and displays. These applications provide a fairly thorough means for meeting the information demands of faculty, staff, and patrons. But there are some information needs relating to complex, implicit, or specialized knowledge that arenâ t readily addressed end-to-end by typical information publishing and knowledge management and instructional methods. This paper uses information theory principles to provide a framework for identifying some of the limitations of current information delivery methods in terms of their relevance to modern library reference and information services. In addition there is an explanation of blog technology, with suggestions for enhancing library information services. Last, some cautions for using blogs is discussed.
APA, Harvard, Vancouver, ISO, and other styles
32

Jiang, Jyun-Yu, and 姜俊宇. "Improving Ranking Consistency for Web Search by Leveraging a Knowledge Base and Search Logs." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/67605146002402152486.

Full text
Abstract:
碩士
國立臺灣大學
資訊工程學研究所
103
In this paper, we propose a new idea called ranking consistency in web search. Relevance ranking is one of the biggest problems in creating an effective web search system. Given some queries with similar search intents, conventional approaches typically only optimize ranking models by each query separately. Hence, there are inconsistent rankings in modern search engines. It is expected that the search results of different queries with similar search intents should preserve ranking consistency. The aim of this paper is to learn consistent rankings in search results for improving the relevance ranking in web search. We then propose a re-ranking model aiming to simultaneously improve relevance ranking and ranking consistency by leveraging knowledge bases and search logs. To the best of our knowledge, our work offers the first solution to improving relevance rankings with ranking consistency. Extensive experiments have been conducted using the Freebase knowledge base and the large-scale query-log of a commercial search engine. The experimental results show that our approach significantly improves relevance ranking and ranking consistency. Two user surveys on Amazon Mechanical Turk also show that users are sensitive and prefer the consistent ranking results generated by our model.
APA, Harvard, Vancouver, ISO, and other styles
33

Borges, Eurico Alexandre Teixeira. "Sistemas de Data Webhousing : análise, desenho, implementação e exploração de sistemas reais." Master's thesis, 2004. http://hdl.handle.net/1822/2787.

Full text
Abstract:
Dissertação de mestrado em Informática, especialidade de Sistemas Distribuídos, Comunicações por Computador e Arquitectura de Computadores
A Web tem-se tornado um dos espaços mais apelativos para as organizações como forma de divulgação das suas actividades, promoção dos seus produtos e serviços e desenvolvimento de actividades comerciais. Todavia, os visitantes de um sítio Web podem facilmente saltar para um sítio da concorrência caso não encontrem rapidamente aquilo que procuram, ou se tiverem qualquer outro motivo que não seja do seu agrado. Conhecer os visitantes e garantir que os produtos, serviços ou informação são aqueles que eles procuram é imperativo. É por isso que as organizações têm tentado analisar vários tipos de questões relacionadas, por exemplo, com a forma como os clientes procuram os produtos, onde abandonam o sítio e porquê, qual a frequência de visitas dos seus clientes, quais os produtos ou serviços que mais interesse despertaram nos visitantes, enfim tudo o que possa contribuir para a melhoria do sítio e para manter ou atrair novos clientes. Todos os movimentos e selecções dos utilizadores de um sítio Web podem ser acompanhados através dos “cliques“ que vão fazendo ao longo do seu processo de interacção com as diversas páginas Web. A esta sequência de “diques” dá-se o nome de clickstream. Será a partir dos dados registados pelo servidor Web sobre as selecções do utilizador que se poderá iniciar o estudo das suas iterações e comportamento. Contudo, o registo mantido pelos servidores Web forma apenas um esqueleto que terá de ser enriquecido com os registos dos vários componentes e sistemas que suportam o seu funcionamento. Este tipo de integração e conciliação de dados num único repositório é, tradicionalmente, feito no seio de um Data Warehouse que, pelo acréscimo dos dados de dlickstream, se torna num Data Webhouse. Todo o processo de extracção, transformação e integração no Data Webhouse é, no entanto, dificultado pelo volume, incomplitude e heterogeneidade dos dados e pela própria tecnologia utilizada no ambiente Web. Nesta dissertação, é apresentado e descrito um modelo dimensional para um Data Webhouse para análise de um sítio Web comercial. São estudadas e apresentadas algumas das suas fontes de dados bem como técnicas que podem ser utilizadas para eliminar ou reduzir os problemas existentes nos dados de clickstream. É descrito todo o desenvolvimento e implementação do processo de extracção, limpeza, transformação e integração de dados no Data Webhouse com especial relevo para as tarefas de clickstream - a identificação de utilizadores e agentes automáticos e a reconstrução de sessões. É apresentado o Webuts — Web Usage Tracking Statistics, um protótipo de um sistema de apoio à decisão para acompanhamento e análise estatística das actividades dos utilizadores de um sítio Web e onde se incorporam alguns dos elementos, técnicas, princípios e práticas descritas.
The Web is becoming one of the most appeallng environments for the many organisations as a means of promoting its businesses and activities as well as a commercialisation channel. However, a Web user can easily leave one organisation’s Web site for its competitors if he doesn’t find what he is looking for or if he finds something unpleasant on one organisation’s site. To know the site’s users and making sure that the products, services or information the site is providing is what the users want is nowadays a must. That is why many organisations have started to study how their web site users browse the site, where are they leaving the site and why, how frequently do their users return, what products and services are most appealing and, in general terms, everything that may be used to improve the Web site and attract new users. Every user moves may be tracked by retaining the clicks selections they do on the different Web pages during their visit. This flow of clicks is now called clickstream. It is the data logged by the Web server on the user’s selections that will enable the organisation to study their moves and behaviour. However, the Web server log only keeps the bare bones of the user’s activity. This data will have to be enriched with data collected by other systems designed to provide the Web site with contents or additional functionalities. Traditionally, the gathering and integration of data from heterogeneous data sources is done inside a Data Warehouse. By adding clickstream data to it we are creating a Data Webhouse. However, Web technology, the data volume, its heterogeneity and incompleteness will create difficulties in the process of extracting, transforming and loading data into the Data Webhouse. In this document we present a dimensional model for a Data Webhouse whose purpose is to analyse a commercial Web site. Several data sources are presented and analised in detail. Some of the techniques used to eliminate or reduce clickstream data problems are also described. The Data Webhouse extraction, cleaning, transformation and loading process is described and special attention is paid to clickstream processing tasks such as user and robot identification and user session reconstruction. A new decision support system prototype, named Webuts - Web Usage Tracking Statistics, is presented. This system’s purpose is to track and analyse a Web site users’ moves and actitivities as well as generate some statistical data on the Web site operation. Its operation is based on a Data Webhouse and its development incorporated some of the elements, techniques and best practices studied and described.
Sonae, Indústria Consultoria e Gestão - Departamento de Sistemas de Informação
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography