Dissertations / Theses on the topic 'Neighbor selection'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 30 dissertations / theses for your research on the topic 'Neighbor selection.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Woerner, August Eric, and August Eric Woerner. "On the Neutralome of Great Apes and Nearest Neighbor Search in Metric Spaces." Diss., The University of Arizona, 2016. http://hdl.handle.net/10150/621578.
Full textBengtsson, Thomas. "Time series discrimination, signal comparison testing, and model selection in the state-space framework /." free to MU campus, to others for purchase, 2000. http://wwwlib.umi.com/cr/mo/fullcit?p9974611.
Full textKarginova, Nadezda. "Identification of Driving Styles in Buses." Thesis, Halmstad University, Intelligent systems (IS-lab), 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:hh:diva-4830.
Full textIt is important to detect faults in bus details at an early stage. Because the driving style affects the breakdown of different details in the bus, identification of the driving style is important to minimize the number of failures in buses.
The identification of the driving style of the driver was based on the input data which contained examples of the driving runs of each class. K-nearest neighbor and neural networks algorithms were used. Different models were tested.
It was shown that the results depend on the selected driving runs. A hypothesis was suggested that the examples from different driving runs have different parameters which affect the results of the classification.
The best results were achieved by using a subset of variables chosen with help of the forward feature selection procedure. The percent of correct classifications is about 89-90 % for the k-nearest neighbor algorithm and 88-93 % for the neural networks.
Feature selection allowed a significant improvement in the results of the k-nearest neighbor algorithm and in the results of the neural networks algorithm received for the case when the training and testing data sets were selected from the different driving runs. On the other hand, feature selection did not affect the results received with the neural networks for the case when the training and testing data sets were selected from the same driving runs.
Another way to improve the results is to use smoothing. Computing the average class among a number of consequent examples allowed achieving a decrease in the error.
FAIRBANKS, MICHAEL STEWART. "MINIMIZING CONGESTION IN PEER-TO-PEER NETWORKS UNDER THE PRESENCE OF GUARDED NODES." University of Cincinnati / OhioLINK, 2006. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1147362818.
Full textDong, Yingying. "Microeconometric Models with Endogeneity -- Theoretical and Empirical Studies." Thesis, Boston College, 2009. http://hdl.handle.net/2345/753.
Full textThis dissertation consists of three independent essays in applied microeconomics and econometrics. Essay 1 investigates the issue why individuals with health insurance use more health care. One obvious reason is that health care is cheaper for the insured. But additionally, having insurance can encourage unhealthy behavior via moral hazard. The effect of health insurance on medical utilization has been extensively studied; however, previous work has mostly ignored the effect of insurance on behavior and how that in turn affects medical utilization. This essay examines these distinct effects. The increased medical utilization due to reduced prices may help the insured maintain good health, while that due to increased unhealthy behavior does not, so distinguishing these two effects has important policy implications. A two-period dynamic forward-looking model is constructed to derive the structural causal relationships among the decision to buy insurance, health behaviors (drinking, smoking, and exercise), and medical utilization. The model shows how exogenous changes in insurance prices and past behaviors can identify the direct and indirect effects of insurance on medical utilization. An empirical analysis also distinguishes between intensive and extensive margins (e.g., changes in the number of drinkers vs. the amount of alcohol consumed) of the insurance effect, which turns out to be empirically important. Health insurance is found to encourage less healthy behavior, particularly heavy drinking, but this does not yield a short term perceptible increase in doctor or hospital visits. The effects of health insurance are primarily found at the intensive margin, e.g., health insurance may not cause a non-drinker to take up drinking, while it encourages a heavy drinker to drink even more. These results suggest that to counteract behavioral moral hazard, health insurance should be coupled with incentives that target individuals who currently engage in unhealthy behaviors, such as heavy drinkers. Essay 2 examines the effect of repeating kindergarten on the retained children's academic performance. Although most existing research concludes that grade retention generates no benefits for retainees' later academic performance, holding low achieving children back has been a popular practice for decades. Drawing on a recently collected nationally representative data set in the US, this paper estimates the causal effect of kindergarten retention on the retained children's later academic performance. Since children are observed being held back only when they enroll in schools that permit retention, this paper jointly models 1) the decision of entering a school allowing for kindergarten retention, 2) the decision of undergoing a retention treatment in kindergarten, and 3) children's academic performance in higher grades. The retention treatment is modeled as a binary choice with sample selection. The outcome equations are linear regressions including the kindergarten retention dummy as an endogenous regressor with a correlated random coefficient. A control function estimator is developed for estimating the resulting double-hurdle treatment model, which allows for unobserved heterogeneity in the retention effect. As a comparison, a nonparametric bias-corrected nearest neighbor matching estimator is also implemented. Holding children back in kindergarten is found to have positive but diminishing effects on their academic performance up to the third grade. Essay 3 proves the semiparametric identification of a binary choice model having an endogenous regressor without relying on outside instruments. A simple estimator and a test for endogeneity are provided based on this identification. These results are applied to analyze working age male's migration within the US, where labor income is potentially endogenous. Identification relies on the fact that the migration probability among workers is close to linear in age while labor income is nonlinear in age(when both are nonparametrically estimated). Using data from the PSID, this study finds that labor income is endogenous and that ignoring this endogeneity leads to downward bias in the estimated effect of labor income on the migration probability
Thesis (PhD) — Boston College, 2009
Submitted to: Boston College. Graduate School of Arts and Sciences
Discipline: Economics
Gopal, Kreshna. "Efficient case-based reasoning through feature weighting, and its application in protein crystallography." [College Station, Tex. : Texas A&M University, 2007. http://hdl.handle.net/1969.1/ETD-TAMU-1906.
Full textGashler, Michael S. "Advancing the Effectiveness of Non-Linear Dimensionality Reduction Techniques." BYU ScholarsArchive, 2012. https://scholarsarchive.byu.edu/etd/3216.
Full textHolsbach, Nicole. "Método de mineração de dados para diagnóstico de câncer de mama baseado na seleção de variáveis." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2012. http://hdl.handle.net/10183/76183.
Full textThis dissertation presents a data mining method for breast cancer (BC) diagnosis based on selected features. We first carried out a systematic literature review, and then suggested a method for feature selection and classification of observations, i.e., patients, into benign or malignant classes based on patients’ breast tissue measures. The proposed method relies on four operational steps: (i) split the original dataset into training and testing sets and apply PCA (Principal Component Analysis) on the training set; (ii) generate attribute importance indices based on PCA weights and percent of variance explained by the retained components; (iii) classify the training set using KNN (k-Nearest Neighbor) or DA (Discriminant Analysis) techniques, eliminate irrelevant features and compute the classification accuracy. Next, eliminate the feature with the lowest importance index, classify the dataset, and re-compute the accuracy. Continue such iterative process until one feature is left; and (iv) choose the subset of features yielding the maximum classification accuracy, and classify the testing set based on those features. When applied to the WBCD (Wisconsin Breast Cancer Database), the proposed method led to average 97.77% accurate classifications while retaining average 5.8 features. One variation of the proposed method is presented based on four different types of polynomial kernels aimed at remapping the original database; steps (i) to (iv) are then applied to such kernels. When applied to the WBCD, the proposed modification increased average accuracy to 98.09% while retaining average of 17.24 features from the 54 variables generated by the recommended kernel. The proposed method can assist the physician in making the diagnosis, selecting a smaller number of variables (involved in the decision-making) with greater accuracy, thereby obtaining the highest possible accuracy.
Ferrero, Carlos Andres. "Algoritmo kNN para previsão de dados temporais: funções de previsão e critérios de seleção de vizinhos próximos aplicados a variáveis ambientais em limnologia." Universidade de São Paulo, 2009. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-19052009-135128/.
Full textTreating data that contains sequential information is an important problem that arises during the data mining process. Time series constitute a popular class of sequential data, where records are indexed by time. The k-Nearest Neighbor - Time Series Prediction kNN-TSP method is an approximator for time series prediction problems. The main advantage of this approximator is its simplicity, and is often used in nonlinear time series analysis for prediction of seasonal time series. Although kNN-TSP often finds the best fit for nearly periodic time series forecasting, some problems related to how to determine its parameters still remain. In this work, we focus in two of these parameters: the determination of the nearest neighbours and the prediction function. To this end, we propose a simple approach to select the nearest neighbours, where time is indirectly taken into account by the similarity measure, and a prediction function which is not disturbed in the presence of patterns at different levels of the time series. Both parameters were empirically evaluated on several artificial time series, including chaotic time series, as well as on a real time series related to several environmental variables from the Itaipu reservoir, made available by Itaipu Binacional. Three of the most correlated limnological variables were considered in the experiments carried out on the real time series: water temperature, air temperature and dissolved oxygen. Analyses of correlation were also accomplished to verify if the predicted variables values maintain similar correlation as the original ones. Results show that both proposals, the one related to the determination of the nearest neighbours as well as the one related to the prediction function, are promising
Glawing, Henrik. "Measurement data selection and association in a collision mitigation system." Thesis, Linköping University, Department of Electrical Engineering, 2002. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-1233.
Full textToday many car manufactures are developing systems that help the driver to avoid collisions. Examples of this kind of systems are: adaptive cruise control, collision warning and collision mitigation / avoidance.
All these systems need to track and predict future positions of surrounding objects (vehicles ahead of the system host vehicle), to calculate the risk of a future collision. To validate that a prediction is correct the predictions must be correlated to observations. This is called the data association problem. If a prediction can be correlated to an observation, this observation is used for updating the tracking filter. This process maintains the low uncertainty level for the track.
From the work behind this thesis, it has been found that a sequential nearest- neighbour approach for the solution of the problem to correlate an observation to a prediction can be used to find the solution to the data association problem.
Since the computational power for the collision mitigation system is limited, only the most dangerous surrounding objects can be tracked and predicted. Therefore, an algorithm that classifies and selects the most critical measurements is developed. The classification into order of potential risk can be done using the measurements that come from an observed object.
Baggu, Gnanesh. "Efficient Approach for Order Selection of Projection-Based Model Order Reduction." Thesis, Université d'Ottawa / University of Ottawa, 2018. http://hdl.handle.net/10393/37967.
Full textLiang, Wen. "Integrated feature, neighbourhood, and model optimization for personalised modelling and knowledge discovery." Click here to access this resource online, 2009. http://hdl.handle.net/10292/749.
Full textSilva, Carlos Filipe Moreira e. "Contemporary electromagnetic spectrum reuse techniques: tv white spaces and D2D communications." reponame:Repositório Institucional da UFC, 2015. http://www.repositorio.ufc.br/handle/riufc/15899.
Full textSubmitted by Marlene Sousa (mmarlene@ufc.br) on 2016-03-29T12:13:57Z No. of bitstreams: 1 2015_tese_cfmsilva.pdf: 8667655 bytes, checksum: a7e392b18bd01de3be64a22b8ff9906b (MD5)
Approved for entry into archive by Marlene Sousa(mmarlene@ufc.br) on 2016-03-31T14:31:46Z (GMT) No. of bitstreams: 1 2015_tese_cfmsilva.pdf: 8667655 bytes, checksum: a7e392b18bd01de3be64a22b8ff9906b (MD5)
Made available in DSpace on 2016-03-31T14:31:46Z (GMT). No. of bitstreams: 1 2015_tese_cfmsilva.pdf: 8667655 bytes, checksum: a7e392b18bd01de3be64a22b8ff9906b (MD5) Previous issue date: 2015-12-12
Over the last years, the wireless broadband access has achieved a tremendous success. With that, the telecommunications industry has faced very important changes in terms of technology, heterogeneity, kind of applications, and massive usage (virtual data tsunami) derived from the introduction of smartphones and tablets; or even in terms of market structure and its main players/actors. Nonetheless, it is well-known that the electromagnetic spectrum is a scarce resource, being already fully occupied (or at least reserved for certain applications). Tra- ditional spectrum markets (where big monopolies dominate) and static spectrum management originated a paradoxal situation: the spectrum is occupied without actually being used! In one hand, with the global transition from analog to digital Television (TV), part of the spectrum previously licensed for TV is freed and geographically interleaved, originating the consequent Television White Spaces (TVWS); on the other hand, the direct communications between devices, commonly referred as Device-to-Device (D2D) communications, are attracting crescent attention by the scientific community and industry in order to overcome the scarcity problem and satisfy the increasing demand for extra capacity. As such, this thesis is divided in two main parts: (a) Spectrum market for TVWS: where a SWOT analysis for the use of TVWS is performed giving some highlights in the directions/actions that shall be followed so that its adoption becomes effective; and a tecno-economic evaluation study is done considering as a use-case a typical European city, showing the potential money savings that operators may reach if they adopt by the use of TVWS in a flexible market manner; (b) D2D communications: where a neighbor discovery technique for D2D communications is proposed in the single-cell scenario and further extended for the multi-cell case; and an interference mitigation algorithm based on the intelligent selection of Downlink (DL) or Uplink (UL) band for D2D communications underlaying cellular networks. A summary of the principal conclusions is as follows: (a) The TVWS defenders shall focus on the promotion of a real-time secondary spectrum market, where through the correct implementation of policies for protection ratios in the spectrum broker and geo-location database, incumbents are protected against interference; (b) It became evident that an operator would recover its investment around one year earlier if it chooses to deploy the network following a flexible spectrum market approach with an additional TVWS carrier, instead of the traditional market; (c) With the proposed neighbor discovery technique the time to detect all neighbors per Mobile Station (MS) is significantly reduced, letting more time for the actual data transmission; and the power of MS consumed during the discovery process is also reduced because the main processing is done at the Base Station (BS), while the MS needs to ensure that D2D communication is possible just before the session establishment; (d) Despite being a simple concept, band selection improves the gains of cellular communications and limits the gains of D2D communications, regardless the position within the cell where D2D communications happen, providing a trade-off between system performance and interference mitigation.
Nos últimos anos, o acesso de banda larga atingiu um grande sucesso. Com isso, a indústria das telecomunicações passou por importantes transformações em termos de tecnologia, heterogeneidade, tipo de aplicações e uso massivo (tsunami virtual de dados) em consequência da introdução dos smartphones e tablets; ou até mesmo na estrutura de mercado e os seus principais jogadores/atores. Porém, é sabido que o espectro electromagnético é um recurso limitado, estando já ocupado (ou pelo menos reservado para alguma aplicação). O mercado tradicional de espectro (onde os grandes monopólios dominam) e o seu gerenciamento estático contribuíram para essa situação paradoxal: o espectro está ocupado mas não está sendo usado! Por um lado, com a transição mundial da Televisão (TV) analógica para a digital, parte do espectro anteriormente licenciado para a TV é libertado e geograficamente multiplexado para evitar a interferência entre sinais de torres vizinhas, dando origem a «espaços em branco» na frequência da TV ou Television White Spaces (TVWS); por outro lado, as comunicações diretas entre usuários, designadas por comunicações diretas Dispositivo-a-Dispositivo (D2D), está gerando um crescente interesse da comunidade científica e indústria, com vista a ultrapassar o problema da escassez de espectro e satisfazer a crescente demanda por capacidade extra. Assim, a tese está dividida em duas partes principais: (a) Mercado de espectro eletromagnético para TVWS: onde é feita uma análise SWOT para o uso dos TVWS, dando direções/ações a serem seguidas para que o seu uso se torne efetivo; e um estudo tecno-econômico considerando como cenário uma típica cidade Europeia, onde se mostram as possíveis poupanças monetárias que os operadores conseguem obter ao optarem pelo uso dos TVWS num mercado flexível; (b) Comunicações D2D: onde uma técnica de descoberta de vizinhos para comunicações D2D é proposta, primeiro para uma única célula e mais tarde estendida para o cenário multi-celular; e um algoritmo de mitigação de interferência baseado na seleção inteligente da banda Ascendente (DL) ou Descendente (UL) a ser reusada pelas comunicações D2D que acontecem na rede celular. Um sumário das principais conclusões é o seguinte: (a) Os defensores dos TVWS devem-se focar na promoção do mercado secundário de espectro electromagnético, onde através da correta implementação de politicas de proteção contra a interferência no broker de espectro e na base de dados, os usuários primário são protegidos contra a interferência; (b) Um operador consegue recuperar o seu investimento aproximadamente um ano antes se ele optar pelo desenvolvimento da rede seguindo um mercado secundário de espectro com a banda adicional de TVWS, em vez do mercado tradicional; (c) Com a técnica proposta de descoberta de vizinhos, o tempo de descoberta por usuário é significativamente reduzido; e a potência consumida nesse processo é também ela reduzida porque o maior processamento é feito na Estação Rádio Base (BS), enquanto que o usuário só precisa de se certificar que a comunicação direta é possível; (d) A seleção de banda, embora seja um conceito simples, melhora os ganhos das comunicações celulares e limita os das comunicações D2D, providenciando um compromisso entre a performance do sistema e a mitigação de interferência.
Duan, Haoyang. "Applying Supervised Learning Algorithms and a New Feature Selection Method to Predict Coronary Artery Disease." Thèse, Université d'Ottawa / University of Ottawa, 2014. http://hdl.handle.net/10393/31113.
Full textHeidaripak, Samrend. "PREDICTION OF PUBLIC BUS TRANSPORTATION PLANNING BASED ON PASSENGER COUNT AND TRAFFIC CONDITIONS." Thesis, Mälardalens högskola, Akademin för innovation, design och teknik, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:mdh:diva-53408.
Full textCirincione, Antonio. "Algoritmi di Machine Learning per la Classificazione di Dati Inerziali." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2019.
Find full textWauthier, Kaitlyn E. ""Real? Hell, Yes, It's Real. It's Mexico": Promoting a US National Imaginary in the Works of William Spratling and Katherine Anne Porter." Bowling Green State University / OhioLINK, 2014. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1404248907.
Full textBílý, Ondřej. "Moderní řečové příznaky používané při diagnóze chorob." Master's thesis, Vysoké učení technické v Brně. Fakulta elektrotechniky a komunikačních technologií, 2011. http://www.nusl.cz/ntk/nusl-218971.
Full textHamad, Sofian. "Efficient route discovery for reactive routing." Thesis, Brunel University, 2013. http://bura.brunel.ac.uk/handle/2438/7634.
Full textDočekal, Martin. "Porovnání klasifikačních metod." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2019. http://www.nusl.cz/ntk/nusl-403211.
Full textSkalak, David Bingham. "Prototype selection for composite nearest neighbor classifiers." 1997. https://scholarworks.umass.edu/dissertations/AAI9737585.
Full textHsu, Shu-ming, and 許書銘. "A Reverse Nearest Neighbor Based Instance Selection Algorithm." Thesis, 2011. http://ndltd.ncl.edu.tw/handle/67336232094961102183.
Full text國立臺灣科技大學
資訊工程系
99
Data reduction is to extract a subset from a dataset. The advantage of data reduction is decreasing the requirement of storage. Using the subset as training data is possible to maintain classification accuracy; sometimes, it can be further improved because of eliminating noises. The key is how to choose representative samples while ignoring noises at the same time. Many instance selection algorithms are based on Nearest Neighbor decision rule (NN). Some of these algorithms select samples based on two strategies, incremental and decremental. The first type of algorithms selects some instances as samples and iteratively adds instances which do not have the same class label with their nearest sample to the sample set. The second type of algorithms gradually removes instances based on its own strategies. However, we propose an algorithm based on Reverse Nearest Neighbor (RNN), called Reverse Nearest Neighbor Reduction (RNNR). RNNR selects samples which can represent other instances in the same class. In addition, RNNR does not need to iteratively scan a dataset which takes much processing time. Experimental results show that RNNR generally achieves higher accuracy, selects fewer samples and takes less processing time than comparators.
Boyd, Bryan 1985. "Local Randomization in Neighbor Selection Improves PRM Roadmap Quality." Thesis, 2012. http://hdl.handle.net/1969.1/148341.
Full textYi-Cheng, Lin, and 林怡成. "A flexible training sample selection method for VQ trained nearest neighbor classifiers." Thesis, 1998. http://ndltd.ncl.edu.tw/handle/27783060845329646006.
Full textTsai, Yung-Hsun, and 蔡咏勳. "Variable Neighborhood Search and k-Nearest Neighbors Algorithmfor Feature Selection Problem." Thesis, 2018. http://ndltd.ncl.edu.tw/handle/js8nez.
Full text元智大學
工業工程與管理學系
107
In this study, we proposed Variable Neighborhood Search(VNS) algorithm to solve the problem of feature selection.By searching different solutions, we uses the k-Nearest Neighbors algorithm(kNN) to evaluate the classification performance of the solution.It is expected to pick out a subset of feature that can effectively classify the data with a small number of features. This study first conducts a series of parametric experiments on the nearest neighbor method (KNN) to find out the parameter set that can have good classification performance in most data sets. Then we tested the algorithm in data sets with different sizes. The experimental results compared with previous studies shows that the proposed method can achieve similar or even better classification performance in most data sets than in previous studies.
Aggarwal, Vinay Kumar [Verfasser]. "ISP-aided neighbour selection in peer-to-peer systems / vorgelegt von Vinay Kumar Aggarwal." 2008. http://d-nb.info/992512387/34.
Full textLee, Chien-Pang, and 李建邦. "The Study on Gene Selection and Sample Classification Based on Gene Expression Data Using Adaptive Genetic Algorithms / k-Nearest Neighbors Method." Thesis, 2006. http://ndltd.ncl.edu.tw/handle/01635740897987498234.
Full text國立中興大學
農藝學系所
94
Microarray technology has become a valuable tool for studying gene expression in recent years. The main difference between microarray and traditional methods is that microarray can measure thousands of genes at the same time. In the past, researchers always used parametric statistical methods to find the significant genes. However, microarray data often cannot obey some assumptions of parametric statistical methods, and type I error would be over expanded while each gene was tested for significance. Therefore, this research was expected to find a variable selection method without assumptions restriction to reduce the dimension of the data set. After using the proposed method, biologists can select the relevant genes according to the sub-gene set. In this study, adaptive genetic algorithms / k-nearest neighbors (AGA / KNN) was used to reduce the dimension of the data set, and it was based on genetic algorithms / k-nearest neighbors (GA / KNN) which was first described by Li et al.(2001a). Although AGA and KNN were well-developed, AGA / KNN was first used to analyze the microarray data. Since AGA was a machine learning tool and KNN was a nonparametric discrimination analysis, both of them could be used without assumptions restriction. There are three main differences between AGA/KNN and GA / KNN. Firstly, the encoding has become binary code, and each string included all genes. Secondly, the adaptive probabilities of crossover and mutation were added. Finally, the extinction and immigration strategy was added. Since GA can just find the near optimal solution, the best string of each run is often not the same. Here, AGA / KNN was repeated by many runs to solve that problem. Thus, lots of the best strings were saved. The frequency of gene was computed by those strings to reduce the dimension of the data set. In this study, an original colon data which is a high-density oligonucleotide chip (Alon et al., 1999) was analyzed. In addition, mice apo AI data which is a cDNA chip (Callow et al., 2000) was also used to compare the ability of gene selection of AGA / KNN and GA / KNN. Based on the results, it was found that AGA / KNN and GA / KNN could reduce the dimension of the data set and all samples could be classified correctly. But the accuracy of AGA / KNN was higher than that of GA / KNN, and it only took half CPU time of GA / KNN. Therefore, it was claimed that the performance of AGA / KNN should not be worse than that of GA / KNN. Finally, we suggested that when AGA / KNN was employed to analyze the microarray data, the top 50 and up to 100 most frequent genes were selected after AGA / KNN were repeated about 100 runs. Those selected genes should include relevant genes, and those selected genes could classify sample correctly.
Sousa, Diogo Macedo de. "Decision support service for Bewegen bike-sharing systems." Master's thesis, 2019. http://hdl.handle.net/10773/29670.
Full textOs sistemas de bike-sharing estão a tornar-se cada vez mais populares e a sua gestão mais complexa. O objetivo principal desta dissertação é o desenvolvimento de um serviço de suporte de decisão, baseado em métodos de aprendizagem automática, para os sistemas de bikesharing da empresa Bewegen. Um objetivo secundário é o desenvolvimento de um mecanismo de recolha sistemática de dados de utilização do sistema, necessários ao desenvolvimento e teste dos métodos de aprendizagem automática. O serviço de suporte de decisão tem dois objetivos. O primeiro objetivo é a previsão do número de bicicletas em cada estação com 30 minutos de antecedência, informação esta a disponilizar aos clientes do sistema de bike-sharing. O segundo objetivo é a previsão do número de bicicletas em cada estacão com 24 horas de antecedência, informação esta a disponilizar aos operadores do sistema no planeamento da distribuição das bicicletas pelas diferentes estacões. Para cumprir com estes objetivos foram implementados dois algoritmos de aprendizagem automática: uma rede neuronal e um algoritmo k-nearest neighbour. Os testes realizados mostram que os algoritmos baseados em redes neuronais obtém melhor desempenho nos dois objectivos. Os dados utilizados nos testes dos dois algoritmos são os dados históricos de um dos sistemas da Bewegen recolhidos desde 1 de janeiro de 2019 até 30 de abril de 2019.
Mestrado em Engenharia de Computadores e Telemática
Σαψάνης, Χρήστος. "Αναγνώριση βασικών κινήσεων του χεριού με χρήση ηλεκτρομυογραφήματος." Thesis, 2013. http://hdl.handle.net/10889/6420.
Full textThe aim of this work was to identify six basic movements of the hand using two systems. Being an interdisciplinary topic, there has been conducted studying in the anatomy of forearm muscles, biosignals, the method of electromyography (EMG) and methods of pattern recognition. Moreover, the signal contained enough noise and had to be analyzed, using EMD, to extract features and to reduce its dimensionality, using RELIEF and PCA, to improve the success rate of classification. The first part uses an EMG system of Delsys initially for an individual and then for six people with the average successful classification, for these six movements at rates of over 80%. The second part involves the construction of an autonomous system EMG using an Arduino microcontroller, EMG sensors and electrodes, which are arranged in an elastic glove. Classification results in this case reached 75% of success.
Molisse, Giulia. "Above ground biomass and carbon sequestration estimation -Implementation of a sentinel-2 based exploratory workflow." Master's thesis, 2021. http://hdl.handle.net/10362/113902.
Full textThis work presents a Sentinel-2 based exploratory work ow for the estimation of Above Ground Biomass (AGB) and Carbon Sequestration (CS) in a subtropical forest. In the last decades, remote sensing-based studies on AGB have been widely investigated alongside with a variety of sensors, features and Machine Learning (ML) algorithms. Up-to-date and reliable mapping of such measures have been increasingly required by international commitments under the climate convention as well as by sustainable forest management practices. The proposed approach consists of 5 major steps: 1) generation of several Vegetation Indices (VI), biophysical parameters and texture measures; 2) feature selection with Mean Decrease in Impurity (MDI), Mean Decrease in Accuracy (MDA), L1 Regularization (LASSO), and Principal Component Analysis (PCA); 3) feature selection testing with k-Nearest Neighbour (kNN), Random Forest (RF), Extreme Gradient Boosting (XGB), and Arti cial Neural Network (ANN); 4) hyper-parameters ne-tuning with Grid Search, Random Search and Bayesian Optimization; and nally, 5) model explanation with the SHapley Additive exPlanations (SHAP) package, which to this day has not been investigated in the context of AGB mapping. The following results were obtained: 1) MDI was chosen as the best performing feature selection method by the XGB and the Deep Neural Network (DNN), MDA was chosen by the RF and the kNN, while LASSO was chosen by the Shallow Neural Network (SNN) and the Linear Neural Network (LNN); 2) before hyper-parameters optimization, the Deep Neural Network (DNN) yielded the best performance with a Root Mean Squared Error (RMSE) of 42.30 t=ha; 3) after hyper-parameters ne-tuning with Bayesian Optimization, the XGB model yielded the best performance with a RMSE of 37.79 t=ha; 4) model explanation with SHAP allowed for a deeper understanding of the features impact on the model predictions. Finally, the predicted AGB throughout the study area showed an average value of 83 t=ha, ranging from 0 t=ha to 346.56 t=ha. The related CS was estimated by using a conversion factor of 0.47.