Tesis sobre el tema "Documental database model"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte los 28 mejores tesis para su investigación sobre el tema "Documental database model".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Explore tesis sobre una amplia variedad de disciplinas y organice su bibliografía correctamente.
Jansson, Jens, Alexandar Vukosavljevic y Ismet Catovic. "Performance comparison between multi-model, key-value and documental NoSQL database management systems". Thesis, Högskolan i Skövde, Institutionen för informationsteknologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-19857.
Texto completoGavryliuk, Olga. "Nástroj pro správu dokumentů v managementu projektů". Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2019. http://www.nusl.cz/ntk/nusl-403822.
Texto completoGifreu, Castells Arnau. "El Documental interactiu com a nou gènere audiovisual : estudi de l'aparició del nou gènere, aproximació a la seva definició i proposta de taxonomia i d'un model d'anàlisi a efectes d'avaluació, disseny i producció". Doctoral thesis, Universitat Pompeu Fabra, 2013. http://hdl.handle.net/10803/119606.
Texto completoThe objective of this doctoral thesis is to show and demonstrate what an interactive documentary consists of. In recent years this interactive medium has been proposed to be a new audiovisual genre with its own specific characteristics. This research discusses how interactive digital media have affected the logics of audiovisual production, exhibition and reception of the traditional documentary genre. In recent years, these new dynamics have transformed the communication processes and the nature of the different actors involved, giving rise to a new genre called “interactive documentary”. From a sample of significant cases we have developed an original conceptual framework with the aim of characterizing the interactive documentary as a specific audiovisual genre, while differentiating it from the conventional forms of documentary and interactive non-fiction. The first part of the study establishes the precedents and the context in which the interactive audiovisual environment is developed. Based on a general overview of the documentary genre and interactive digital media, we describe the main context in which this format is created, the Web, as well as the types of products that it includes, interactive applications. In this section we also introduce the main interactive non-fiction genres, characterizing them and differentiating them among each other and among audiovisual non-fiction genres. The second part discusses and justifies the reasons why we can consider the interactive documentary as a new audiovisual genre with its own specific characteristics. To place it in an appropriate context, we situate and describe the emergence, consolidation and diversification of this new genre, based on which we formulate an original definition, carry out an extensive characterization and propose a taxonomy. We also focus on the transformation of the production and dissemination processes according to the new logics and current dynamics. In order to establish and define the conceptual framework specifically and precisely we have developed a categorization model for analyzing the new genre, which is also useful for creating and producing this type of documentary. In addition, we evaluate the state-of-the-art of the interactive documentary in relation to the topics discussed, the access platform and the user's experience. The conclusions chapter summarizes the main original contributions of the research, affirming the relevance of identifying the interactive documentary as a new and promising audiovisual genre.
Lau, Ho Lam. "The development of the nested relational sequence model to support XML databases /". View Abstract or Full-Text, 2002. http://library.ust.hk/cgi/db/thesis.pl?COMP%202002%20LAU.
Texto completoIncludes bibliographical references (leaves 87-96). Also available in electronic version. Access restricted to campus users.
Hirzalla, Naél B. "Media processing and retrieval model for multimedia documents databases". Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1998. http://www.collectionscanada.ca/obj/s4/f2/dsk2/tape17/PQDD_0013/NQ28346.pdf.
Texto completoHirzalla, Nael B. "Media processing and retrieval model for multimedia documents databases". Thesis, University of Ottawa (Canada), 1998. http://hdl.handle.net/10393/4541.
Texto completoEl, Abri Marwa. "Probabilistic relational models learning from graph databases". Thesis, Nantes, 2018. http://www.theses.fr/2018NANT4019/document.
Texto completoHistorically, Probabilistic Graphical Models (PGMs) are a solution for learning from uncertain and flat data, also called propositional data or attributevalue representations. In the early 2000s, great interest was addressed to the processing of relational data which includes a large number of objects participating in different relations. Probabilistic Relational Models (PRMs) present an extension of PGMs to the relational context. With the rise of the internet, numerous technological innovations and web applications are driving the dramatic increase of various and complex data. Consequently, Big Data has emerged. Several types of data stores have been created to manage this new data, including the graph databases. Recently there has been an increasing interest in graph databases to model objects and interactions. However, all PRMs structure learning use wellstructured data that are stored in relational databases. Graph databases are unstructured and schema-free data stores. Edges between nodes can have various signatures. Since, relationships that do not correspond to an ER model could be depicted in the database instance. These relationships are considered as exceptions. In this thesis, we are interested by this type of data stores. Also, we study two kinds of PRMs namely, Direct Acyclic Probabilistic Entity Relationship (DAPER) and Markov Logic Networks (MLNs). We propose two significant contributions. First, an approach to learn DAPERs from partially structured graph databases. A second approach consists to benefit from first-order logic to learn DAPERs using MLN framework to take into account the exceptions that are dropped during DAPER learning. We are conducting experimental studies to compare our proposed methods with existing approaches
McElroy, Jonathan David. "Automatic Document Classification in Small Environments". DigitalCommons@CalPoly, 2012. https://digitalcommons.calpoly.edu/theses/682.
Texto completoMurphy, Brian R. "Order-sensitive XML query processing over relational sources". Link to electronic thesis, 2003. http://www.wpi.edu/Pubs/ETD/Available/etd-0505103-123753.
Texto completoKeywords: computation pushdown; XML; order-based Xquery processing; relational database; ordered SQL queries; data model mapping; XQuery; XML data mapping; SQL; XML algebra rewrite rules; XML document order. Includes bibliographical references (p. 64-67).
Hahn, Henrik. "Expressive sampling synthesis. Learning extended source-filter models from instrument sound databases for expressive sample manipulations". Thesis, Paris 6, 2015. http://www.theses.fr/2015PA066564/document.
Texto completoWithin this thesis an imitative sound synthesis system will be introduced that is applicable to most quasi-harmonic instruments. The system bases upon single-note recordings that represent a quantized version of an instrument's possible timbre space with respect to its pitch and intensity dimension. A transformation method then allows to render sound signals with continuous values of the expressive control parameters which are perceptually coherent with its acoustic equivalents. A parametric instrument model is therefore presented based on an extended source-filter model with separate manipulations of a signal’s harmonic and residual components. A subjective evaluation procedure will be shown to assess a variety of transformation results by a direct comparison with unmodified recordings to determine how perceptually close the synthesis results are regarding their respective acoustic correlates
Broadbent, Robert Emer. "A Functional Framework for Content Management". BYU ScholarsArchive, 2009. https://scholarsarchive.byu.edu/etd/1737.
Texto completoLi, Haizhou. "Modeling and verification of probabilistic data-aware business processes". Thesis, Clermont-Ferrand 2, 2015. http://www.theses.fr/2015CLF22563/document.
Texto completoThere is a wide range of new applications that stress the need for business process models that are able to handle imprecise data. This thesis studies the underlying modelling and analysis issues. It uses as formal model to describe process behaviours a labelled transitions system in which transitions are guarded by conditions defined over a probabilistic database. To tackle verification problems, we decompose this model to a set of traditional automata associated with probabilities named as world-partition automata. Next, this thesis presents an approach for testing probabilistic simulation preorder in this context. A complexity analysis reveals that the problem is in 2-exptime, and is exptime-hard, w.r.t. expression complexity while it matches probabilistic query evaluation w.r.t. data-complexity. Then P-LTL and P-CTL model checking methods are studied to verify this model. In this context, the complexity of P-LTL and P-CTL model checking is in exptime. Finally a prototype called ”PRODUS” which is a modeling and verification tool is introduced and we model a realistic scenario in the domain of GIS (graphical information system) by using our approach
Suchý, Petr. "Realizace webového modulárního systému". Master's thesis, Vysoké učení technické v Brně. Fakulta elektrotechniky a komunikačních technologií, 2008. http://www.nusl.cz/ntk/nusl-217328.
Texto completoMoumen, Chiraz. "Une méthode d'optimisation hybride pour une évaluation robuste de requêtes". Thesis, Toulouse 3, 2017. http://www.theses.fr/2017TOU30070/document.
Texto completoThe quality of an execution plan generated by a query optimizer is highly dependent on the quality of the estimates produced by the cost model. Unfortunately, these estimates are often imprecise. A body of work has been done to improve estimate accuracy. However, obtaining accurate estimates remains very challenging since it requires a prior and detailed knowledge of the data properties and run-time characteristics. Motivated by this issue, two main optimization approaches have been proposed. A first approach relies on single-point estimates to choose an optimal execution plan. At run-time, statistics are collected and compared with estimates. If an estimation error is detected, a re-optimization is triggered for the rest of the plan. At each invocation, the optimizer uses specific values for parameters required for cost calculations. Thus, this approach can induce several plan re-optimizations, resulting in poor performance. In order to avoid this, a second approach considers the possibility of estimation errors at the optimization time. This is modelled by the use of multi-point estimates for each error-prone parameter. The aim is to anticipate the reaction to a possible plan sub-optimality. Methods in this approach seek to generate robust plans, which are able to provide good performance for several run-time conditions. These methods often assume that it is possible to find a robust plan for all expected run-time conditions. This assumption remains unjustified. Moreover, the majority of these methods maintain without modifications an execution plan until the termination. This can lead to poor performance in case of robustness violation at run-time. Based on these findings, we propose in this thesis a hybrid optimization method that aims at two objectives : the production of robust execution plans, particularly when the uncertainty in the used estimates is high, and the correction of a robustness violation during execution. This method makes use of intervals of estimates around error-prone parameters. It produces execution plans that are likely to perform reasonably well over different run-time conditions, so called robust plans. Robust plans are then augmented with what we call check-decide operators. These operators collect statistics at run-time and check the robustness of the current plan. If the robustness is violated, check-decide operators are able to make decisions for plan modifications to correct the robustness violation without a need to recall the optimizer. The results of performance studies of our method indicate that it provides significant improvements in the robustness of query processing
Ozturk, Aybuke. "Design, Implementation and Analysis of a Description Model for Complex Archaeological Objects". Thesis, Lyon, 2018. http://www.theses.fr/2018LYSE2048/document.
Texto completoCeramics are one of the most important archaeological materials to help in the reconstruction of past civilizations. Information about complex ceramic objects is composed of textual, numerical and multimedia data, which induce several research challenges addressed in this thesis. From a technical perspective, ceramic databases have different file formats, access protocols and query languages. From a data perspective, ceramic data are heterogeneous and experts have differentways of representing and storing data. There is no standardized content and terminology, especially in terms of description of ceramics. Moreover, data navigation and observation are difficult. Data integration is also difficult due to the presence of various dimensions from distant databases, which describe the same categories of objects in different ways.Therefore, the research project presented in this thesis aims to provide archaeologists and archaeological scientists with tools for enriching their knowledge by combining different information on ceramics. We divide our work into two complementary parts: (1) Modeling of Complex Archaeological Data and (2) Clustering Analysis of Complex Archaeological Data. The first part of this thesis is dedicated to the design of a complex archaeological database model for the storage of ceramic data. This database is also used to source a data warehouse for doing online analytical processing (OLAP). The second part of the thesis is dedicated to an in-depth clustering (categorization) analysis of ceramic objects. To do this, we propose a fuzzy approach, where ceramic objects may belong to more than one cluster (category). Such a fuzzy approach is well suited for collaborating with experts, by opening new discussions based on clustering results.We contribute to fuzzy clustering in three sub-tasks: (i) a novel fuzzy clustering initialization method that keeps the fuzzy approach linear; (ii) an innovative quality index that allows finding the optimal number of clusters; and (iii) the Multiple Clustering Analysis approach that builds smart links between visual, textual and numerical data, which assists in combining all types ofceramic information. Moreover, the methods we propose could also be adapted to other application domains such as economy or medicine
Harward, Gregory Brent. "Suitability of the NIST Shop Data Model as a Neutral File Format for Simulation". Diss., CLICK HERE for online access, 2005. http://contentdm.lib.byu.edu/ETD/image/etd899.pdf.
Texto completoGiffard-Roisin, Sophie. "Personnalisation non-invasive de modèles électrophysiologiques cardiaques à partir d'électrogrammes surfaciques". Thesis, Université Côte d'Azur (ComUE), 2017. http://www.theses.fr/2017AZUR4092/document.
Texto completoThe objective of this thesis is to use non-invasive data (body surface potential mapping, BSPM) to personalise the main parameters of a cardiac electrophysiological (EP) model for predicting the response to cardiac resynchronization therapy (CRT). CRT is a clinically proven treatment option for some heart failures. However, these therapies are ineffective in 30% of the treated patients and involve significant morbidity and substantial cost. The precise understanding of the patient-specific cardiac function can help to predict the response to therapy. Until now, such methods required to measure intra-cardiac electrical potentials through an invasive endovascular procedure which can be at risk for the patient. We developed a non-invasive EP model personalisation based on a patient-specific simulated database and machine learning regressions. First, we estimated the onset activation location and a global conduction parameter. We extended this approach to multiple onsets and to ischemic patients by means of a sparse Bayesian regression. Moreover, we developed a reference ventricle-torso anatomy in order to perform an common offline regression and we predicted the response to different pacing conditions from the personalised model. In a second part, we studied the adaptation of the proposed method to the input of 12-lead electrocardiograms (ECG) and the integration in an electro-mechanical model for a clinical use. The evaluation of our work was performed on an important dataset (more than 25 patients and 150 cardiac cycles). Besides having comparable results with state-of-the-art ECG imaging methods, the predicted BSPMs show good correlation coefficients with the real BSPMs
Milea, Dominique. "Usage et mésusage dans la prescription des antidépresseurs : l’apport des bases de données". Thesis, Lyon 1, 2010. http://www.theses.fr/2010LYO10322/document.
Texto completoConsidering the large increase in antidepressant use observed in the last decades, the question of good or poor usage of antidepressants is of importance. The three studies presented in this report consisted into a descriptive analysis of the increase in the use of antidepressants in different countries, and thereafter analyse the underlying mechanisms. Our analysis of sales database indicates that the increase in antidepressants volume differ from country to country, but follows a classical scheme with rapid increase until stabilisation to a certain threshold and that national policies to ensure expenditure limitations has a different impact on the volume depending on the maturity of the market. The analyses performed afterwards on US claims databases, and show that if improper use of antidepressants is a reality, it does not seem to correspond to a massive off-label use but rather to an incorrect use of antidepressants: treatment without precise diagnosis, isolated prescription which suggest undue treatment, treatment for too short periods of time with regards to international guidelines which recommend at least 6 month to avoid relapse or recurrence of the disease. The analyses we have conducted on databases can serve as a basis for the development of a reliable tool to evaluate drug consumption (sales databases) or drug use (claims databases) : simple and rapid use of sales databases to get a mapping of consumption in their environment, and more complex and time consuming use of claims database to gain an understanding of practice. Used in routine, the sales monitoring tools could play a role to alert public health services on abnormal evolution of drug consumption and allow the initiation of more refined analyses either on large claims databases or through prospective cohort of patients in order to characterise the practice and gain an understanding on the drivers of drug use
Bertelsmeier, Cleo. "Biologie des invasions de fourmis dans un contexte de changement climatique". Thesis, Paris 11, 2013. http://www.theses.fr/2013PA112358/document.
Texto completoClimate change and biological invasions are both among the greatest threats to biodiversity and their impacts might increase by the end of the century. Among invasive species, ants are a prominent group due to their negative impacts on native species, ecosystem processes, human and animal health, agro-ecosystems and the economy. The objective of this thesis was to forecast future ant invasions – especially in the light of on-going climate change, which is generally thought to favour invasive species by removing thermal barriers. I used three complementary approaches to study the potential of different ant species to invade in the future: species distribution modelling, behavioural experiments and the analysis of a database of ecological traits. I modelled suitable area for 15 of the worst invasive ant species, both currently and with predicted climate change, globally, regionally and within the world’s 34 biodiversity hotspots. Surprisingly, the potential distribution of only five species was predicted to increase (up to 35.8%) with climate change, with most declining by up to 63.3%. The ant invasion hotspots are predominantly in tropical and subtropical regions of South America, Africa, Asia and Oceanic islands, and particularly correspond with biodiversity hotspots. Contrary to general expectations, climate change and biological invasions will not systematically act synergistically for ants. In addition, I found that the impacts of climate change can change over time and even reverse the trend of the impact (i.e., an increase instead of a decrease or vice versa). However, ant invasions will likely remain as a major global problem, especially where invasion hotspots coincide with biodiversity hotspots. The species distribution models have identified large potentially overlapping distributions of several invasive ants. In the future, these species may arrive simultaneously in the same regions and compete with each other. In a series of experiments, I tested behavioural differences among 7 highly invasive ant species (Anoplolepis gracilipes, Paratrechina longicornis, Myrmica rubra, Linepithema humile, Lasius neglectus, Wasmannia auropunctata and Pheidole megacephala). I discovered two different behavioural strategies among invasive ants. Interactions at the colony level, exhibited more complex demographic processes and more variability. Further, I investigated resource competition and differences in resource exploitation. I found significant differences among species, with competitive abilities that were negatively correlated with behavioural dominance. This series of experiments suggests that the ‘mechanisms’ of invasiveness are more complex than previously thought and that different invasive ant species may use different behavioural strategies. Since there are more than 250 exotic species of ants, it would be interesting to identify potential future invaders. In order to identify traits associated with invasiveness in ants, I set up a database with 2193 ant species and 24 ecological characteristics. I performed a preliminary analysis of trait differences between native and invasive ants that shows clearly different clusters of invasive and native species, with exotic species in between. These results could be used as a basis to construct a predictive model of future ant invasions. The different methods used (models, experiments, database) are complementary in that they explore different aspects of the future ant invasions which are likely to be influenced by on-going climate change. The worst invaders of tomorrow may not be the same as today and similarly, areas most at risk are likely to change
Zhang, Zebin. "Intégration des méthodes de sensibilité d'ordre élevé dans un processus de conception optimale des turbomachines : développement de méta-modèles". Thesis, Ecully, Ecole centrale de Lyon, 2014. http://www.theses.fr/2014ECDL0047/document.
Texto completoThe turbomachinery optimal design usually relies on some iterative methods with either experimental or numerical evaluations that can lead to high cost due to numerous manipulations and intensive usage of CPU. In order to limit the cost and shorten the development time, the present thesis work proposes to integrate a parameterization method and the meta-modelization method in an optimal design cycle of an axial low speed turbomachine. The parameterization, realized by the high order sensitivity study of Navier-Stokes equations, allows to construct a parameterized database that contains not only the evaluations results, but also the simple and cross derivatives of objectives as a function of parameters. Enriched information brought by the derivatives are utilized during the meta-model construction, particularly by the Co-Kriging method employed to couple several databases. Compared to classical methods that are without derivatives, the economic benefit of the proposed method lies in the use of less reference points. Provided the number of reference points is small, chances are a unique point presenting at one or several dimensions, which requires a hypothesis on the error distribution. For those dimensions, the Co-Kriging works like a Taylor extrapolation from the reference point making the most of its derivatives. This approach has been experimented on the construction of a meta-model for a conic hub fan. The methodology recalls the coupling of databases based on two fan geometries and two operating points. The precision of the meta-model allows to perform an optimization with help of NSGA-2, one of the optima selected reaches the maximum efficiency, and another covers a large operating range. The optimization results are eventually validated by further numerical simulations
Mater, Gautier. "Caractérisation des mesures d’exposition à des produits chimiques dans les bases de données françaises COLCHIC et SCOLA pour la prévention des maladies professionnelles". Thesis, Université de Lorraine, 2016. http://www.theses.fr/2016LORR0210/document.
Texto completoTwo occupational exposure databases of occupational exposures to chemicals, COLCHIC and SCOLA, coexist in France with different objectives (prevention and compliance). Little is known about their representativeness of exposures in the general population. We explored to what extent COLCHIC and SCOLA adequately reflect occupational exposures in France. After performing a descriptive and comparative analysis, associations between exposure levels and ancillary information were explored for each agent, separately for COLCHIC and SCOLA and in a common dataset, using statistical modelling. Modelling results were synthesized across agents using Meta analysis. COLCHIC and SCOLA contain, respectively, 929 700 (670 chemicals) and 429 104 records (105). Three predictors "Sample Time", "Personal protective equipment" and "Year" are strongly associated with exposure levels across a large majority of chemicals in both databases, and 3 others are specific to each one. Exposure levels are in average twice higher in COLCHIC compared to SCOLA in 2007, but become comparable from 2012-2015. COLCHIC and SCOLA are an important source of information. Inclusion of descriptors associated with exposure levels in our study and the use of predictive methods should help to improve their interpretation
Réau, Manon. "Importance des données inactives dans les modèles : application aux méthodes de criblage virtuel en santé humaine et environnementale". Thesis, Paris, CNAM, 2019. http://www.theses.fr/2019CNAM1251/document.
Texto completoVirtual screening is widely used in early stages of drug discovery and to build toxicity prediction models. Commonly used protocols include an evaluation of the performances of different tools on benchmarking databases before applying them for prospective studies. The content of benchmarking tools is a critical point; most benchmarking databases oppose active data to putative inactive due to the scarcity of published inactive data in the literature. Nonetheless, experimentally validated inactive data also bring information. Therefore, we constructed the NR-DBIND, a database dedicated to nuclear receptors that contains solely experimentally validated active and inactive data. The importance of the integration of inactive data in docking and pharmacophore models construction was evaluated using the NR-DBIND data. Virtual screening protocols were used to resolve the potential binding mode of small molecules on FXR, NRP-1 et TNF⍺
von, Wenckstern Michael. "Web applications using the Google Web Toolkit". Master's thesis, Technische Universitaet Bergakademie Freiberg Universitaetsbibliothek "Georgius Agricola", 2013. http://nbn-resolving.de/urn:nbn:de:bsz:105-qucosa-115009.
Texto completoDiese Diplomarbeit beschreibt die Erzeugung desktopähnlicher Anwendungen mit dem Google Web Toolkit und die Umwandlung klassischer Java-Programme in diese. Das Google Web Toolkit ist eine Open-Source-Entwicklungsumgebung, die Java-Code in browserunabhängiges als auch in geräteübergreifendes HTML und JavaScript übersetzt. Vorgestellt wird der Großteil des GWT Frameworks inklusive des Java zu JavaScript-Compilers sowie wichtige Sicherheitsaspekte von Internetseiten. Um zu zeigen, dass auch komplizierte graphische Oberflächen mit dem Google Web Toolkit erzeugt werden können, wird das bekannte Brettspiel Agricola mittels Model-View-Presenter Designmuster implementiert. Zur Ermittlung der richtigen Technologie für das nächste Webprojekt findet ein Vergleich zwischen dem Google Web Toolkit und JavaServer Faces statt
Vavrek, Michal. "Evolution Management in NoSQL Document Databases". Master's thesis, 2018. http://www.nusl.cz/ntk/nusl-383112.
Texto completoWen, Lu Bo y 盧博文. "An Implementation of a Electronic Document Database Workflow Model". Thesis, 1996. http://ndltd.ncl.edu.tw/handle/51255992251956219790.
Texto completo國立臺灣科技大學
電子工程學系
84
The thesis proposed a three layers workflow model. The pro- posed model helps group members in various location to complete traditional workfolw process with simple and efficient way. The application layer is responsible for the user interface, including the presentation of form and document, the view of the workflow database, and the workflow design tools. The workflow layer is based on Client/Server architecture. The client application translates the workflow operations into various type of workflow message, and the server application interprets these workflow message and excutes actions ( such as regist, download, apply, route, approve, reject , and file). The network layer uses the MAPI to transfer message,and uses the ODBC to access the workflow database. These features allow the workflow to be in progress without time and location limi- tation. To solve the traditional workflow problem,the workflow model in chapter 3 is implemented with these feathers: single-form, multi-view and routing-slip with no conditional branch routing mechanism. And in chapter 4, we also proposed an advanced work- flow model which is based on internet with more powerful features included multi-form and multi-view. It uses routing- script with conditional branch routing mechanism to simplify the workflow design, and also uses integrated-document which transfers the data in workflow database to useful information to suport the decision making in enterprise.
Liu, Min-Jang y 劉民樟. "A Data Exchange Model between Hierarchical Databases and XML Documents". Thesis, 2004. http://ndltd.ncl.edu.tw/handle/44233746885124759243.
Texto completo朝陽科技大學
資訊管理系碩士班
92
In recent years, data exchange is popular among enterprises and organizations. The data may be structured or unstructured such as purchase orders, product catalogs, official documents and so on. XML, proposed by W3C, has been extensively applied to many areas, especially for database applications. Most researches focused on data exchange between relational databases and XML documents. However, the hierarchical database is still being used in some businesses. In this paper, we propose a model for enterprises or organizations to perform data exchange between hierarchical databases and XML documents. The XML Schema is used to describe the data type of exchanged data to ensure that the original data type would not lose during data conversion. In this model, four modules are used to translate the data in hierarchical databases into XML documents and vise versa. Experiments are performed to analyze the performance of the proposed model.
"Parameter free document stream classification". Thesis, 2006. http://library.cuhk.edu.hk/record=b6074286.
Texto completoFor the problem of bursty topics identification, PFreeBT adopts an approach, in which we term it as feature-pivot clustering approach. Given a document stream, PFreeBT first identifies a set of bursty features from there. The identification process is based on computing the probability distributions. According to the patterns of the bursty features and two newly defined concepts (equivalent and map-to), a set of bursty topics can be extracted.
For the problem of constructing a reliable classifier, we formulate it as a partially supervised classification problem. In this classification problem, only a few training examples are labeled as positive (P). All other training examples (U) are remained unlabeled. Here, U is mixed with the negative examples (N) and some other positive examples (P'). Existing techniques that tackle this problem all focus on finding N from U. None of them attempts to extract P' from U. In fact, it is difficult to succeed as the topics in U are diverse and the features in there are sparse. In this dissertation, PNLH is proposed for extracting a high quality of P' and N from U.
In this dissertation, two heuristics, PFreeBT and PNLH, are proposed to tackle the aforementioned problems. PFreeBT aims at identifying the bursty topics in a document stream, whereas PNLH aims at constructing a reliable classifier for a given bursty topic. It is worth noting that both heuristics are parameter free. Users do not need to provide any parameter explicitly. All of the required variables can be computed base on the given document stream automatically.
In this information overwhelming century, information becomes ever more pervasive. A new class of data-intensive application arises where data is modeled best as an open-ended stream. We call such kind of data as data stream. Document stream is a variation of data stream, which consists of a sequence of chronological ordered documents. A fundamental problem of mining document streams is to extract meaningful structure from there, so as to help us to organize the contents systematically. In this dissertation, we focus on such a problem. Specifically, this dissertation studies two problems: to identify the bursty topics in a document stream and to construct a classifiers for the bursty topics. A bursty topic is one of the topics resides in the document stream, such that a large number of documents would be related to it during a bounded time interval.
Fung Pui Cheong Gabriel.
"August 2006."
Adviser: Jeffrey Xu Yu.
Source: Dissertation Abstracts International, Volume: 68-03, Section: B, page: 1720.
Thesis (Ph.D.)--Chinese University of Hong Kong, 2006.
Includes bibliographical references (p. 122-130).
Electronic reproduction. Hong Kong : Chinese University of Hong Kong, [2012] System requirements: Adobe Acrobat Reader. Available via World Wide Web.
Electronic reproduction. [Ann Arbor, MI] : ProQuest Information and Learning, [200-] System requirements: Adobe Acrobat Reader. Available via World Wide Web.
Abstracts in English and Chinese.
School code: 1307.
Huang, Chih-Yen y 黃致彥. "A Operation Model of Secure XML-RDB API and Optimal Querying of Encrypted XML Documents Stored in Relational Database". Thesis, 2009. http://ndltd.ncl.edu.tw/handle/19901935255867816873.
Texto completo國立花蓮教育大學
學習科技研究所
97
With the prevalence of the Internet, XML(Extensible Markup Language) became the standard of data exchange on the Internet. Due to the usual use of XML documents for the data exchange or data transfer protocol between heterogeneous platforms, how to manage and query XML documents efficiently become the key-point when the storage of XML documents increases. At present, there are two typical ways supported by relational database to store XML documents: shred and native XML. Furthermore, the common use of XML to transmit information on the Internet raises the concern of the security of XML documents. So far, in order to improve the security of XML documents, W3C organisation also proposes the suggestions and formats of the encryption and signature for the secured XML documents. This dissertation is an attempt at designing a model of the transfer between XML and relational database. After completely storing the secured XML into the relational database, with the API providing the secured maintenance and operation, this model can provide the best efficiency of querying data.