Dissertations / Theses on the topic 'Data Mining Techniques'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Data Mining Techniques.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Tong, Suk-man Ivy. "Techniques in data stream mining." Click to view the E-thesis via HKUTO, 2005. http://sunzi.lib.hku.hk/hkuto/record/B34737376.
Full textTong, Suk-man Ivy, and 湯淑敏. "Techniques in data stream mining." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2005. http://hub.hku.hk/bib/B34737376.
Full textBurgess, Martin. "Transformation techniques in data mining." Thesis, University of East Anglia, 2004. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.410093.
Full textAl-Hashemi, Idrees Yousef. "Applying data mining techniques over big data." Thesis, Boston University, 2013. https://hdl.handle.net/2144/21119.
Full textThe rapid development of information technology in recent decades means that data appear in a wide variety of formats — sensor data, tweets, photographs, raw data, and unstructured data. Statistics show that there were 800,000 Petabytes stored in the world in 2000. Today’s internet has about 0.1 Zettabytes of data (ZB is about 1021 bytes), and this number will reach 35 ZB by 2020. With such an overwhelming flood of information, present data management systems are not able to scale to this huge amount of raw, unstructured data—in today’s parlance, Big Data. In the present study, we show the basic concepts and design of Big Data tools, algorithms, and techniques. We compare the classical data mining algorithms to the Big Data algorithms by using Hadoop/MapReduce as a core implementation of Big Data for scalable algorithms. We implemented the K-means algorithm and A-priori algorithm with Hadoop/MapReduce on a 5 nodes Hadoop cluster. We explore NoSQL databases for semi-structured, massively large-scaling of data by using MongoDB as an example. Finally, we show the performance between HDFS (Hadoop Distributed File System) and MongoDB data storage for these two algorithms.
Zhou, Wubai. "Data Mining Techniques to Understand Textual Data." FIU Digital Commons, 2017. https://digitalcommons.fiu.edu/etd/3493.
Full textXIAO, XIN. "Data Mining Techniques for Complex User-Generated Data." Doctoral thesis, Politecnico di Torino, 2016. http://hdl.handle.net/11583/2644046.
Full textYu, Congcong. "Parallelizing ensemble techniques for data mining." Thesis, National Library of Canada = Bibliothèque nationale du Canada, 2001. http://www.collectionscanada.ca/obj/s4/f2/dsk3/ftp04/MQ59416.pdf.
Full textCID, DANTE JOSE ALEXANDRE. "DATA MINING WITH ROUGH SETS TECHNIQUES." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2000. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=7244@1.
Full textEsta dissertação investiga a utilização de Rough Sets no processo de descoberta de conhecimento em Bancos de Dados (KDD - Knowledge Discovery in Databases). O objetivo do trabalho foi avaliar o desempenho da técnica de Rough Sets na tarefa de Classificação de Dados. A Classificação é a tarefa da fase de Mineração de Dados que consiste na descoberta de regras de decisão, ou regras de inferência, que melhor representem um grupo de registros do banco de dados. O trabalho consistiu de cinco etapas principais: estudo sobre o processo de KDD; estudo sobre as técnicas de Rough Sets aplicadas à mineração de dados; análise de ferramentas de mineração de dados do mercado; evolução do projeto Bramining; e a realização de alguns estudos de caso para avaliar o Bramining. O estudo sobre o caso KDD abrangeu todas as suas fases: transformação, limpeza, seleção, mineração de dados e pós-processamento. O resultado obtido serviu de base para o aprimoramento do projeto Bramining. O estudo sobre as técnicas de Rough Sets envolveu a pesquisa de seus conceitos e sua aplicabilidade no contexto de KDD. A teoria de Rough Sets foi apresentada por Zdzislaw Pawlak no início dos anos 80 como uma abordagem matemática para a análise de dados vagos e imprecisos. Este estudo permitiu sua aplicação na ferramenta de mineração de dados desenvolvida. A análise de ferramentas de mineração de dados do mercado abrangeu o estudo e testes de aplicativos baseados em diferentes técnicas, enriquecimento a base de comparação utilizada na avaliação da pesquisa. A evolução do projeto Bramining consistiu no aprimoramento do ambiente KDD desenvolvido em estudos anteriores, passando a incluir técnica de Rough Sets em seu escopo. Os estudos de caso foram conduzidos paralelamente com o uso de Bramining e de outras ferramentas existentes, para efeito de comparação. Os índices apresentados pelo Bramining nos estudos de caso foram considerados, de forma geral, equivalentes aos do software comercial, tendo ambos obtidos regras de boa qualidade na maioria dos casos. O Bramining, entretanto, mostrou-se mais completo para o processo de KDD, graças às diversas opções nele disponíveis para preparação dos dados antes da fase de mineração. Os resultados obtidos comprovaram, através da aplicação desenvolvida, a adequação dos conceitos de Rough Sets à tarefa de classificação de dados. Alguns pontos frágeis da técnica foram identificados, como a necessidade de um mecanismo de apoio para a redução de atributos e a dificuldade em trabalhar com atributos de domínio contínuo. Porém, ao se inserir a técnica em um ambiente mais completo de KDD, como o Bramining, estas deficiências foram sanadas. As opções de preparação da base que o Bramining disponibiliza ao usuário para executar, em particular, a redução e a codificação de atributos permitem deixar os dados em estado adequado à aplicação de Rough Sets. A mineração de dados é uma questão bastante relevante nos dias atuais, e muitos métodos têm sido propostos para as diversas tarefas que dizem respeito a esta questão. A teoria de Rough Sets não mostrou significativas vantagens ou desvantagens em relação a outras técnicas já consagradas, mas foi de grande valia comprovar que há caminhos alternativos para o processo de descoberta de conhecimento.
This dissertation investigates the application of Rough Sets to the process of KDD - Knowledge Discovery in Databases. The main goal of the work was to evaluate the performance of Rough Sets techniques in solving the classification problem. Classification is a task of the Data Mining step in KDD Process that performs the discovery of decision rules that best represent a group of registers in a database. The work had five major steps: study of the KDD process; study of Rough Sets techniques applied to data mining; evaluation of existing data mining tools; development of Bramining project; and execution of some case studies to evaluate Bramining. The study of KDD process included all its steps: transformation, cleaning, selection, data mining and post- processing. The results obtained served as a basis to the enhamcement of Bramining. The study of Rough Sets techniques included the research of theory´s concepts and its applicability at KDD context. The Rough Sets tehory has been introduced by Zdzislaw Pawlak in the early 80´s as a mathematical approach to the analysis of vague and uncertain data. This research made possible the implementation of the technique under the environment of the developed tool. The analysis of existing data mining tools included studying and testing of software based on different techniques, enriching the background used in the evaluation of the research. The evolution of Bramining Project consisted in the enhancement of the KDD environment developed in previous works, including the addition of Rough Sets techniques. The case studies were performed simultaneously with Bramining and a commercial minig tool, for comparison reasons. The quality of the knowledge generated by Bramining was considered equivalent to the results of commercial tool, both providing good decision rules for most of the cases. Nevertheless, Bramining proved to be more adapted to the complete KDD process, thanks to the many available features to prepare data to data mining step. The results achieved through the developed application proved the suitability of Rough Sets concepts to the data classification task. Some weaknesses of the technique were identified, like the need of a previous attribute reduction and the inability to deal with continuous domain data. But as the technique has been inserted in a more complete KDD environment like the Bramining Project, those weaknesses ceased to exist. The features of data preparation available in Bramining environment, particularly the reduction and attribute codification options, enable the user to have the database fairly adapted to the use of Rough Sets algorithms. Data mining is a very relevant issue in present days and many methods have been proposed to the different tasks involved in it. Compared to other techniques, Rough Sets Theory did not bring significant advantages or disadvantages to the process, but it has been of great value to show there are alternate ways to knowledge discovery.
Al-Bataineh, Hussien Suleiman. "Islanding Detection Using Data Mining Techniques." Thesis, North Dakota State University, 2015. https://hdl.handle.net/10365/27634.
Full textJANAKIRAMAN, KRISHNAMOORTHY. "ENTITY IDENTIFICATION USING DATA MINING TECHNIQUES." University of Cincinnati / OhioLINK, 2001. http://rave.ohiolink.edu/etdc/view?acc_num=ucin989852516.
Full textAliotta, Marco Antonio. "Data mining techniques on volcano monitoring." Doctoral thesis, Università di Catania, 2013. http://hdl.handle.net/10761/1364.
Full textDi, Silvestro Lorenzo Paolo. "Data Mining and Visual Analytics Techniques." Doctoral thesis, Università di Catania, 2014. http://hdl.handle.net/10761/1559.
Full textSchubert, Matthias. "Advanced Data Mining Techniques for Compound Objects." Diss., lmu, 2004. http://nbn-resolving.de/urn:nbn:de:bvb:19-27981.
Full textDutta, Ila. "Data Mining Techniques to Identify Financial Restatements." Thesis, Université d'Ottawa / University of Ottawa, 2018. http://hdl.handle.net/10393/37342.
Full textHamby, Stephen Edward. "Data mining techniques for protein sequence analysis." Thesis, University of Nottingham, 2010. http://eprints.nottingham.ac.uk/11498/.
Full textMAHOTO, NAEEM AHMED. "Data mining techniques for complex application domains." Doctoral thesis, Politecnico di Torino, 2013. http://hdl.handle.net/11583/2506368.
Full textZANONI, MARCO. "Data mining techniques for design pattern detection." Doctoral thesis, Università degli Studi di Milano-Bicocca, 2012. http://hdl.handle.net/10281/31515.
Full textOkafor, Anthony. "Entropy based techniques with applications in data mining." [Gainesville, Fla.] : University of Florida, 2005. http://purl.fcla.edu/fcla/etd/UFE0013113.
Full textTekieh, Mohammad Hossein. "Analysis of Healthcare Coverage Using Data Mining Techniques." Thèse, Université d'Ottawa / University of Ottawa, 2012. http://hdl.handle.net/10393/20547.
Full textAlmutairi, Abdulrazaq Z. "Improving intrusion detection systems using data mining techniques." Thesis, Loughborough University, 2016. https://dspace.lboro.ac.uk/2134/21313.
Full textSobolewska, Katarzyna-Ewa. "Web links utility assessment using data mining techniques." Thesis, Blekinge Tekniska Högskola, Avdelningen för programvarusystem, 2006. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-2936.
Full textakasha.kate@gmail.com
Kanellopoulos, Yiannis. "Supporting software systems maintenance using data mining techniques." Thesis, University of Manchester, 2008. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.496254.
Full textEspy, John. "Data mining techniques for constructing jury selection models." Thesis, California State University, Long Beach, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=1527548.
Full textJury selection can determine a case before it even begins. The goal is to predict whether a juror rules for the plaintiff or the defense in the medical malpractice trials that are conducted, and which variables are significant in predicting this. The data for the analysis were obtained from mock trials that simulated actual trials, with possible arguments from the defense and the plaintiff with ample discussion time. These mock trials were supplemented by surveys that attempted to capture the characteristics and attitudes of the mock juror and the case at hand. The data were modeled using the logistic regression as well as decision trees and neural networks techniques.
Siddiqui, Muazzam Ahmed. "HIGH PERFORMANCE DATA MINING TECHNIQUES FOR INTRUSION DETECTION." Master's thesis, University of Central Florida, 2004. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/4435.
Full textM.S.
School of Computer Science
Engineering and Computer Science
Computer Science
floyd, stuart. "Data Mining Techniques for Prognosis in Pancreatic Cancer." Digital WPI, 2007. https://digitalcommons.wpi.edu/etd-theses/671.
Full textKim, Hyunki. "Developing semantic digital libraries using data mining techniques." [Gainesville, Fla.] : University of Florida, 2005. http://purl.fcla.edu/fcla/etd/UFE0010105.
Full textWang, Qing. "Intelligent Data Mining Techniques for Automatic Service Management." FIU Digital Commons, 2018. https://digitalcommons.fiu.edu/etd/3883.
Full textJABEEN, SAIMA. "Document analysis by means of data mining techniques." Doctoral thesis, Politecnico di Torino, 2014. http://hdl.handle.net/11583/2537297.
Full textBetter, Marco L. "Data mining techniques for prediction and classification in discrete data applications." Connect to online resource, 2007. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:3273688.
Full textStorer, Jeremy J. "Computational Intelligence and Data Mining Techniques Using the Fire Data Set." Bowling Green State University / OhioLINK, 2016. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1460129796.
Full textTorre, Fabrizio. "3D data visualization techniques and applications for visual multidimensional data mining." Doctoral thesis, Universita degli studi di Salerno, 2014. http://hdl.handle.net/10556/1561.
Full textDespite modern technology provide new tools to measure the world around us, we are quickly generating massive amounts of high-dimensional, spatialtemporal data. In this work, I deal with two types of datasets: one in which the spatial characteristics are relatively dynamic and the data are sampled at different periods of time, and the other where many dimensions prevail, although the spatial characteristics are relatively static. The first dataset refers to a peculiar aspect of uncertainty arising from contractual relationships that regulate a project execution: the dispute management. In recent years there has been a growth in size and complexity of the projects managed by public or private organizations. This leads to increased probability of project failures, frequently due to the difficulty and the ability to achieve the objectives such as on-time delivery, cost containment, expected quality achievement. In particular, one of the most common causes of project failure is the very high degree of uncertainty that affects the expected performance of the project, especially when different stakeholders with divergent aims and goals are involved in the project...[edited by author]
XII n.s.
Li, Qiao. "Data mining and statistical techniques applied to genetic epidemiology." Thesis, University of East Anglia, 2010. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.533716.
Full textPalmer, Nathan Patrick. "Data mining techniques for large-scale gene expression analysis." Thesis, Massachusetts Institute of Technology, 2011. http://hdl.handle.net/1721.1/68493.
Full textCataloged from PDF version of thesis.
Includes bibliographical references (p. 238-256).
Modern computational biology is awash in large-scale data mining problems. Several high-throughput technologies have been developed that enable us, with relative ease and little expense, to evaluate the coordinated expression levels of tens of thousands of genes, evaluate hundreds of thousands of single-nucleotide polymorphisms, and sequence individual genomes. The data produced by these assays has provided the research and commercial communities with the opportunity to derive improved clinical prognostic indicators, as well as develop an understanding, at the molecular level, of the systemic underpinnings of a variety of diseases. Aside from the statistical methods used to evaluate these assays, another, more subtle challenge is emerging. Despite the explosive growth in the amount of data being generated and submitted to the various publicly available data repositories, very little attention has been paid to managing the phenotypic characterization of their samples (i.e., managing class labels in a controlled fashion). If sense is to be made of the underlying assay data, the samples' descriptive metadata must first be standardized in a machine-readable format. In this thesis, we explore these issues, specifically within the context of curating and analyzing a large DNA microarray database. We address three main challenges. First, we acquire a large subset of a publicly available microarray repository and develop a principled method for extracting phenotype information from freetext sample labels, then use that information to generate an index of the sample's medically-relevant annotation. The indexing method we develop, Concordia, incorporates pre-existing expert knowledge relating to the hierarchical relationships between medical terms, allowing queries of arbitrary specificity to be efficiently answered. Second, we describe a highly flexible approach to answering the question: "Given a previously unseen gene expression sample, how can we compute its similarity to all of the labeled samples in our database, and how can we utilize those similarity scores to predict the phenotype of the new sample?" Third, we describe a method for identifying phenotype-specific transcriptional profiles within the context of this database, and explore a method for measuring the relative strength of those signatures across the rest of the database, allowing us to identify molecular signatures that are shared across various tissues ad diseases. These shared fingerprints may form a quantitative basis for optimal therapy selection and drug repositioning for a variety of diseases.
by Nathan Patrick Palmer.
Ph.D.
Han, J., M. Kamber, and J. Pei. "Data mining: concepts and techniques." 2012. http://hdl.handle.net/10454/9053.
Full textVeltman, Lisa M. "Incident Data Analysis Using Data Mining Techniques." 2008. http://hdl.handle.net/1969.1/ETD-TAMU-2008-08-35.
Full textMantovani, Matteo. "Approximate Data Mining Techniques on Clinical Data." Doctoral thesis, 2020. http://hdl.handle.net/11562/1018039.
Full textHuang, Jen-Chieh, and 黃仁傑. "Using Data Mining Techniques to Support Data Retrieval." Thesis, 2001. http://ndltd.ncl.edu.tw/handle/41468540528588892817.
Full text中原大學
資訊管理研究所
89
The explosive growth of the Internet, and particular the World Wide Web, in recent years has put huge amounts of information at the disposal of anyone with access to the Internet. A problem facing information retrieval on the web is how to help user read easily. Clustering method is to group the data with similar features in clusters without needing predefined cluster labels. And Document Clustering or Document Classfy can help user read easily. In the past, clustering or classfy must extract keyword to descript this document. And keyword extract from document is a content-based clustering method. Data mining is a new technique that can discover something unknowledge from a mount of data. Recently some research is using data mining technique to find knowledge in the web named web mining. In our research, a new method for clustering Web Page is proposed. This method is using web mining technique to produce user-based clustering. A major advantage of this approach is that the relavency information is objectively reflected by the usage logs; frequent simultaneous visits to two unrelated documents should indicate that they are in fact closely related. By analysis sogi web’s log file, web page clustering is proposed, and the experiment result show that cluster having high percisioin.
Narvaez, Vilema Miryan Estela, Felice Crupi, and Fabrizio Angiulli. "Data mining techniques for large and complex data." Thesis, 2017. http://hdl.handle.net/10955/1875.
Full textDuring these three years of research I dedicated myself to the study and design of data mining techniques for large quantities of data. Particular attention was devoted to training set condensing techniques for the nearest-neighbor classification rule and to techniques for node anomaly detection in networks. The first part of this thesis was focused on the design of strategies to reduce the size of the subset extracted from condensing techniques and to their experimentation. The training set condensing techniques aim to determine a subset of the original training set having the property of allowing to correctly classify all the training set examples. The subset extracted from these techniques also known as consistent subset. The result of the research was the development of various strategies of subset selection, designed to determine during the training phase the most promising subset based on different methods of estimating test accuracy. Among them, the PACOPT strategy is based on Pessimistic Error Estimate (PEE) to estimate generalization as a trade-off between training set accuracy and model complexity. The experimental phase has had for reference the FCNN technique of condensation. Among the methods of condensation based on the nearest neighbor decision rule (NN rule), FCNN (for Fast Condensed NN) it is one of the most advantageous technique, particularly in terms of time performance. We showed that the designed selection strategies guarantee to preserve the accuracy of a consistent subset. We also demonstrated that the proposed selection strategies guarantee to significantly reduce the size of the model. Comparison with notable training-set reduction techniques for the NN rule witness for state-of-the-art performances of the here introduced strategies. The second part of the thesis is directed towards the design of analysis tools for network structured data. Anomaly detection is an area that has received much attention in recent years. It has a wide variety of applications, including fraud detection and network intrusion detection. The techniques focused on anomaly detection in static graphs assume that the networks do not change and are capable of representing only a single snapshot of data. As real-world networks are constantly changing, there has been a shift in focus to dynamic graphs, which evolve over time. We present a technique for node anomaly detection in networks where arcs are annotated with time of creation. The technique aims at singling out anomalies by taking simultaneously into account information concerning both the structure of the network and the order in which connections have been established. The latter information is obtained by timestamps associated with arcs. A set of temporal structures is induced by checking certain conditions on the order of arc appearance denoting different kinds of user behaviors. The distribution of these structures is computed for each node and used to detect anomalies. We point out that the approach here investigated is substantially different from techniques dealing with dynamic networks. Indeed, our aim is not to determine the points in time in which a certain portion of the networks (typically a community or a subgraph) exhibited a significant change, as usually done by dynamic-graph anomaly detection techniques. Rather, our primary aim is to analyze each single node by taking simultaneously into account its temporal footprint.
Università della Calabria
Guarascio, Massimo, Domenico Saccà, Giuseppe Manco, and Luigi Palopoli. "Data mining techniques for fraud detection." Thesis, 2014. http://hdl.handle.net/10955/419.
Full textMallick, Jnanaranjan, and Manmohan Tudu. "Study of various data mining techniques." Thesis, 2007. http://ethesis.nitrkl.ac.in/4254/1/%E2%80%9CStudy_of_various_data_mining_techniques.pdf.
Full textChuang, Tse-sheng, and 莊澤生. "Discovering Issue Networks Using Data Mining Techniques." Thesis, 2002. http://ndltd.ncl.edu.tw/handle/27215766204365017293.
Full text國立中山大學
資訊管理學系研究所
90
By means of data mining techniques development these days, the knowledge discovered by virtue of data mining has ranging from business application to fraud detection. However, too often, we see only the profit-making justification for investing in data mining while losing sight of the fact that they can help resolve issues of global or national importance. In this research, we propose the architecture for issue oriented information construction and knowledge discovery that related to political or public policy issues. In this architecture, we adopt issue networks as the description model and data mining as the core technique. This study is also performed and verified with prototype system constructing and case data analyzing. There are three main topics in our research. The issue networks information construction starts with text files information retrieving of specified issue from news reports. Keywords retrieved from news reports are converted into structuralized network nodes and presented in the form of issue networks. The second topic is the clustering of network actors. We adopt an issue-association clustering method to provide views of clustering of issue participators based on relations of issues. In third topic, we use specified link analysis method to compute the importance of actors and sub-issues. Our study concludes with performance evaluation via domain experts. We conduct recall, precision evaluation for first topic above and certainty, novelty, utility evaluation for others.
HUANG, JYUN-HAO, and 黃俊豪. "Design optimization process with data mining techniques." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/59498094123929998366.
Full text國立中興大學
機械工程學系所
97
Abstract This thesis incorporates data mining into optimization process. The useful information embedded in the data can be dug out to enhance the computational efficiency and produce better solutions. For unconstrained optimization problems, the data mining is used to find the design space that might contain the global solution. For constrained optimization problems, it is used to find the possible feasible regions. The Sequential quadratic programming (SQP) is then used to find the optimum solution in the identified areas to see whether the chance to find the global solution is increased. Based on test results, it shows that for high-noised multi-modal problems, data mining and SQP may not be able to find the global solution. Therefore, the evolutionary algorithm will be incorporated into the optimization process in the second part of this thesis. The evolutionary algorithm searches the design space using multiple points simultaneously. If the design space can be reduced, then the computational time spent will be reduced and the chance to find the global solution will be increased. In order to save the computational time for structural optimization problems, the artificial neural network is employed to get the approximate results of structural analyses. This thesis uses evolution strategy to search for the optimum solution in design space found by data mining. For structural optimization problems, neural networks are used to replace exact finite element analyses. The SQP is used in the last step to search for the exact optimum solution. Several test problems show that the proposed approach not only can find better solutions but also spends less computational time.
Shiau, Shu-Min, and 蕭書民. "Applying Data Mining Techniques to Products Test." Thesis, 2011. http://ndltd.ncl.edu.tw/handle/58276969863380870641.
Full text世新大學
資訊管理學研究所(含碩專班)
99
With the longer product development time, the cost usually will relative higher. And if the issues are found at the later of development process, then the cost will even higher just to pay for the correction.Therefore, product's functional testing in the development process is a necessary stage and through testing hope the problem can be detected at the earliest time.Because of effective testing, the issues can be found and correct right away before the product shipment.To let the customers have the highest quality and stability product is our goal.So, the product testing is extremely important task during the product manufacturing. Users require a long period of use and maintain stability for network products.In order to meet customers demand, the testing is very important during product development's period. This study is based on the network product testing data to do the survey and model building, and through the use to analysis its information association with rules and validation.The results show that the network product testing data can be used effectively on determine the analytes in the test plan.Information through different conditions and factors classification, you can see the relationship between product and test plan more clearly.If just focus on one data, then you will find that different type of products have its own relationship with the test plan.Network product testing can also be using network data survey techniques to obtain the accuracy and save more labor requirement.By using this way, the labor resource can be used more efficient and also let the data survey perform its best ability to create more profit to corporate, also by using tihs way to enhance the product satisfication on customers.
Chen, Po-Jung, and 陳柏融. "Selecting Test Items by Data Mining Techniques." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/31407843272135626158.
Full text世新大學
資訊管理學研究所(含碩專班)
97
One of the most important goals of tests designing is to pick items with most discrimination. In the past, most work assumed no dependent relations among test items so that the test papers are made by picking items with highest individual discriminations. But in reality, test items may relate to other items, the overall discrimination of a test paper can not be simply added-up. Hence, this study proposes a two-step method to design test papers by picking discriminative items combinations from the item bank. We first analyze the archival tests to discover substitute items as well as recognize discriminative test itemsets by using data mining technology. Then, the test items are recommended to complete the discriminative test paper. Finally, a real life case is used to testify the proposed method. These test data are provided by the Chinese Enterprise Planning Association (CERP) in Taiwan. The experimental results show the two-step method can complete the test design task efficiently. In addition, the newly composed test paper presents high discriminative since it is very close to the maximum discrimination under the assumption of items independence which are ideally generated by the Item Response Theory.
Cernaut, Oana-Maria. "Customer targeting models using data mining techniques." Master's thesis, 2019. http://hdl.handle.net/10773/30010.
Full textMestrado em Marketing
Lin, Yen-Tim, and 林彥廷. "Analysis of Spam Using Data Mining Techniques." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/02200975323671982267.
Full text華梵大學
資訊管理學系碩士班
96
Abstract For the development of novel technology, the application of the information technology becomes one of the main habits in daily life, for example, the use of e-mail. However, these followed by the intrusion of the spam mails which persecute the users and waste the network resources. So, the setting up of the e-mail filter system is an important topic of network security. The previous studies of the e-mail filter are mostly depended on the keywords or building up of black and white lists to identify the spam, but they are limited to the problem of wrongly classification the e-mails that may cause the missing of some important mails. This study is based on the characteristics of the spam. The spam mails are divided into 16 attributes with the technology of data mining, and use the voting method to classify the mails. The data are mainly collected from the mail box of the Huafan University. First, they are changed into the data format with 16 characteristics. Thereafter, the decision tree, back-propagation network (BPN) and support vector machine (SVM) are conducted to proceed the learning vote which can produce the required classified results. According to the simulation results, this application of the e-mail filter mechanism in this study can provide a satisfied outcome.
Chu, Yu-Hou, and 朱宇侯. "Using Data Mining Techniques for Vehicle Warranty." Thesis, 2005. http://ndltd.ncl.edu.tw/handle/64698339906754441668.
Full text中原大學
資訊管理研究所
93
The innovation of information technology has increased the development of information production and gathering. Lots of companies have setup information systems to collect the daily operation data and stored these data in the database. In order to enforce the competitive advantage for company. Therefore, how to smartly and automatically transfer data into useful information and knowledge becomes a very pioneering goal of data application. As a result, data mining gradually becomes important. The researches of data mining techniques have been developed very well in many fields. The application of data mining techniques in maintenance field being in paid much attention in these years. The project adapts Rough Set and Association Rules of data miming to develop a model based on RSFP ( Rough set and Frequent pattern list). The model sifts all attributes out to leave only t he most important ones. Then it comes to data miming which is used to get the last association rules. In the project, we take the example of reported data from the service station. We sift the minor attributes out by the theory of rough sets to gather the most important attributes. This project is an important theory to sift out attributes by means of rough sets. Furthermore, I use Frequent pattern list to mine association rules. We also apply the combination of both rough sets and Frequent pattern list to vehicle warranty fees and data. The result heightens the efficiency of sifting and miming of attributes and also helps the service station find out what rules they are interested in.
Kao, Ming-Jui, and 高明瑞. "Importers Value Discovery Using Data Mining Techniques." Thesis, 2010. http://ndltd.ncl.edu.tw/handle/85174480441730344807.
Full text元智大學
資訊管理學系
98
As the market of shipping by sea is flourishing, liner container carriers are raising their volume of freight and enhancing service quality in order to meet customers'' demands. However, the customers with higher contribution margin are very important to be acquired from a large number of customers in the global market. To determine those customers with higher contribution margin, this research combined RFM Model for measuring the value of customers and K-means algorithm for clustering analysis. Since Trans-Pacific trade of US import from the Far East is more indicative, we adopted US market data for this research. The results showed that the proposed clustering model has the essence of efficiency. The customers were divided into five categories, they are (A)Best , (B)Spender, (C)Frequent, (D)Uncertain and (E)Loss/negative. Among those five categories, the Best and Spender account for 75 % of overall Trans-Pacific trades. The corresponding marketing strategies for the above two groups are proposed in this research as well.
謝祖仁. "Using Data Mining Techniques in Patient Diagnosis." Thesis, 2012. http://ndltd.ncl.edu.tw/handle/93434524520345344958.
Full text國立彰化師範大學
資訊管理學系所
100
The medical industry is a service industry, except offering to the acute patient the service of seeking medical advice , The target of serving more is that often one fixes the chronic patient who goes to a doctor in the hospital, Age strata are mostly age patients on the middle and senior level, except that main disease is sought while seeking medical advice, will often be with other disease symptoms, such as the diabetes patient person often with high blood pressure, some disease symptoms of the eye, Must check to track too that treat to other medical training clinic while going to a doctor, if can offer a patient to seek medical advice medical care intact and convenient, to the patient seeking medical advice to the hospital, can save time, can offer a patient to combine as to hospital and intact medical treatment is looked after, further analyze the relevant information of disease that a patient seeks medical advice, the expectation can be found potential and having rule of relation nature, can even use to get the result of prevention from suffering from the diseases. Find result of study, at occupy all by certain rate no matter in number of times, women suffer from number and going to a doctor rate to suffer from the number of people and clinic and go to a doctoring etc. blood pressure disease, reveal that looks after the importance on at chronic disease. Women suffer from the number of people at multiple chronic diseases and is generally higher than men. Analyze, reveal with age level, have blood pressure these sector suffers from number to obviously increase in 40-64 year old. The Association rule analysis result of the chronic disease diagnosis: high blood pressure and Confidence that spends above and other 50% chronic disease such as vascular disease of the brain, diabetes, heart disease, other metabolism lack proper care and the intersection of immunity and unregulated illness and the intersection of gullet and the intersection of and stomach and disease of duodenum, etc. have high relation.
Renda, Alessandro. "Algorithms and techniques for data stream mining." Doctoral thesis, 2021. http://hdl.handle.net/2158/1235915.
Full text