Academic literature on the topic 'Multi-window based ensemble learning'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Multi-window based ensemble learning.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Multi-window based ensemble learning"

1

Li, Hu, Ye Wang, Hua Wang, and Bin Zhou. "Multi-window based ensemble learning for classification of imbalanced streaming data." World Wide Web 20, no. 6 (March 8, 2017): 1507–25. http://dx.doi.org/10.1007/s11280-017-0449-x.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Abdillah, Abid Famasya, Cornelius Bagus Purnama Putra, Apriantoni Apriantoni, Safitri Juanita, and Diana Purwitasari. "Ensemble-based Methods for Multi-label Classification on Biomedical Question-Answer Data." Journal of Information Systems Engineering and Business Intelligence 8, no. 1 (April 26, 2022): 42–50. http://dx.doi.org/10.20473/jisebi.8.1.42-50.

Full text
Abstract:
Background: Question-answer (QA) is a popular method to seek health-related information and biomedical data. Such questions can refer to more than one medical entity (multi-label) so determining the correct tags is not easy. The question classification (QC) mechanism in a QA system can narrow down the answers we are seeking. Objective: This study develops a multi-label classification using the heterogeneous ensembles method to improve accuracy in biomedical data with long text dimensions. Methods: We used the ensemble method with heterogeneous deep learning and machine learning for multi-label extended text classification. There are 15 various single models consisting of three deep learning (CNN, LSTM, and BERT) and four machine learning algorithms (SVM, kNN, Decision Tree, and Naïve Bayes) with various text representations (TF-IDF, Word2Vec, and FastText). We used the bagging approach with a hard voting mechanism for the decision-making. Results: The result shows that deep learning is more powerful than machine learning as a single multi-label biomedical data classification method. Moreover, we found that top-three was the best number of base learners by combining the ensembles method. Heterogeneous-based ensembles with three learners resulted in an F1-score of 82.3%, which is better than the best single model by CNN with an F1-score of 80%. Conclusion: A multi-label classification of biomedical QA using ensemble models is better than single models. The result shows that heterogeneous ensembles are more potent than homogeneous ensembles on biomedical QA data with long text dimensions. Keywords: Biomedical Question Classification, Ensemble Method, Heterogeneous Ensembles, Multi-Label Classification, Question Answering
APA, Harvard, Vancouver, ISO, and other styles
3

Meng, Jinyu, Zengchuan Dong, Yiqing Shao, Shengnan Zhu, and Shujun Wu. "Monthly Runoff Forecasting Based on Interval Sliding Window and Ensemble Learning." Sustainability 15, no. 1 (December 21, 2022): 100. http://dx.doi.org/10.3390/su15010100.

Full text
Abstract:
In recent years, machine learning, a popular artificial intelligence technique, has been successfully applied to monthly runoff forecasting. Monthly runoff autoregressive forecasting using machine learning models generally uses a sliding window algorithm to construct the dataset, which requires the selection of the optimal time step to make the machine learning tool function as intended. Based on this, this study improved the sliding window algorithm and proposes an interval sliding window (ISW) algorithm based on correlation coefficients, while the least absolute shrinkage and selection operator (LASSO) method was used to combine three machine learning models, Random Forest (RF), LightGBM, and CatBoost, into an ensemble to overcome the preference problem of individual models. Example analyses were conducted using 46 years of monthly runoff data from Jiutiaoling and Zamusi stations in the Shiyang River Basin, China. The results show that the ISW algorithm can effectively handle monthly runoff data and that the ISW algorithm produced a better dataset than the sliding window algorithm in the machine learning models. The forecast performance of the ensemble model combined the advantages of the single models and achieved the best forecast accuracy.
APA, Harvard, Vancouver, ISO, and other styles
4

Shen, Zhiqiang, Zhankui He, and Xiangyang Xue. "MEAL: Multi-Model Ensemble via Adversarial Learning." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 4886–93. http://dx.doi.org/10.1609/aaai.v33i01.33014886.

Full text
Abstract:
Often the best performing deep neural models are ensembles of multiple base-level networks. Unfortunately, the space required to store these many networks, and the time required to execute them at test-time, prohibits their use in applications where test sets are large (e.g., ImageNet). In this paper, we present a method for compressing large, complex trained ensembles into a single network, where knowledge from a variety of trained deep neural networks (DNNs) is distilled and transferred to a single DNN. In order to distill diverse knowledge from different trained (teacher) models, we propose to use adversarial-based learning strategy where we define a block-wise training loss to guide and optimize the predefined student network to recover the knowledge in teacher models, and to promote the discriminator network to distinguish teacher vs. student features simultaneously. The proposed ensemble method (MEAL) of transferring distilled knowledge with adversarial learning exhibits three important advantages: (1) the student network that learns the distilled knowledge with discriminators is optimized better than the original model; (2) fast inference is realized by a single forward pass, while the performance is even better than traditional ensembles from multi-original models; (3) the student network can learn the distilled knowledge from a teacher model that has arbitrary structures. Extensive experiments on CIFAR-10/100, SVHN and ImageNet datasets demonstrate the effectiveness of our MEAL method. On ImageNet, our ResNet-50 based MEAL achieves top-1/5 21.79%/5.99% val error, which outperforms the original model by 2.06%/1.14%.
APA, Harvard, Vancouver, ISO, and other styles
5

Koohzadi, Maryam, Nasrollah Moghadam Charkari, and Foad Ghaderi. "Unsupervised representation learning based on the deep multi-view ensemble learning." Applied Intelligence 50, no. 2 (July 31, 2019): 562–81. http://dx.doi.org/10.1007/s10489-019-01526-0.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Shan, Shuo, Chenxi Li, Zhetong Ding, Yiye Wang, Kanjian Zhang, and Haikun Wei. "Ensemble learning based multi-modal intra-hour irradiance forecasting." Energy Conversion and Management 270 (October 2022): 116206. http://dx.doi.org/10.1016/j.enconman.2022.116206.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Aboneh, Tagel, Abebe Rorissa, and Ramasamy Srinivasagan. "Stacking-Based Ensemble Learning Method for Multi-Spectral Image Classification." Technologies 10, no. 1 (January 26, 2022): 17. http://dx.doi.org/10.3390/technologies10010017.

Full text
Abstract:
Higher dimensionality, Hughes phenomenon, spatial resolution of image data, and presence of mixed pixels are the main challenges in a multi-spectral image classification process. Most of the classical machine learning algorithms suffer from scoring optimal classification performance over multi-spectral image data. In this study, we propose stack-based ensemble-based learning approach to optimize image classification performance. In addition, we integrate the proposed ensemble learning with XGBoost method to further improve its classification accuracy. To conduct the experiment, the Landsat image data has been acquired from Bishoftu town located in the Oromia region of Ethiopia. The current study’s main objective was to assess the performance of land cover and land use analysis using multi-spectral image data. Results from our experiment indicate that, the proposed ensemble learning method outperforms any strong base classifiers with 99.96% classification performance accuracy.
APA, Harvard, Vancouver, ISO, and other styles
8

Kwon, Beom, and Sanghoon Lee. "Ensemble Learning for Skeleton-Based Body Mass Index Classification." Applied Sciences 10, no. 21 (November 4, 2020): 7812. http://dx.doi.org/10.3390/app10217812.

Full text
Abstract:
In this study, we performed skeleton-based body mass index (BMI) classification by developing a unique ensemble learning method for human healthcare. Traditionally, anthropometric features, including the average length of each body part and average height, have been utilized for this kind of classification. Average values are generally calculated for all frames because the length of body parts and the subject height vary over time, as a result of the inaccuracy in pose estimation. Thus, traditionally, anthropometric features are measured over a long period. In contrast, we controlled the window used to measure anthropometric features over short/mid/long-term periods. This approach enables our proposed ensemble model to obtain robust and accurate BMI classification results. To produce final results, the proposed ensemble model utilizes multiple k-nearest neighbor classifiers trained using anthropometric features measured over several different time periods. To verify the effectiveness of the proposed model, we evaluated it using a public dataset. The simulation results demonstrate that the proposed model achieves state-of-the-art performance when compared with benchmark methods.
APA, Harvard, Vancouver, ISO, and other styles
9

Krasnopolsky, Vladimir M., and Ying Lin. "A Neural Network Nonlinear Multimodel Ensemble to Improve Precipitation Forecasts over Continental US." Advances in Meteorology 2012 (2012): 1–11. http://dx.doi.org/10.1155/2012/649450.

Full text
Abstract:
A novel multimodel ensemble approach based on learning from data using the neural network (NN) technique is formulated and applied for improving 24-hour precipitation forecasts over the continental US. The developed nonlinear approach allowed us to account for nonlinear correlation between ensemble members and to produce “optimal” forecast represented by a nonlinear NN ensemble mean. The NN approach is compared with the conservative multi-model ensemble, with multiple linear regression ensemble approaches, and with results obtained by human forecasters. The NN multi-model ensemble improves upon conservative multi-model ensemble and multiple linear regression ensemble, it (1) significantly reduces high bias at low precipitation level, (2) significantly reduces low bias at high precipitation level, and (3) sharpens features making them closer to the observed ones. The NN multi-model ensemble performs at least as well as human forecasters supplied with the same information. The developed approach is a generic approach that can be applied to other multi-model ensemble fields as well as to single model ensembles.
APA, Harvard, Vancouver, ISO, and other styles
10

Kang, Xiangping, Deyu Li, and Suge Wang. "A multi-instance ensemble learning model based on concept lattice." Knowledge-Based Systems 24, no. 8 (December 2011): 1203–13. http://dx.doi.org/10.1016/j.knosys.2011.05.010.

Full text
APA, Harvard, Vancouver, ISO, and other styles
More sources

Dissertations / Theses on the topic "Multi-window based ensemble learning"

1

Ontañón, Villar Santi. "Ensemble Case Based Learning for Multi-Agent Systems." Doctoral thesis, Universitat Autònoma de Barcelona, 2005. http://hdl.handle.net/10803/3050.

Full text
Abstract:
Esta monografía presenta un marco de trabajo para el aprendizaje en un escenario de datos distribuidos y con control descentralizado. Hemos basado nuestro marco de trabajo en Sistemas Multi-Agente (MAS) para poder tener control descentralizado, y en Razonamiento Basado en Casos (CBR), dado que su naturaleza de aprendizaje perezoso lo hacen adecuado para sistemas multi-agentes dinámicos. Además, estamos interesados en agentes autónomos que funcionen como ensembles. Un ensemble de agentes soluciona problemas de la siguiente manera: cada agente individual soluciona el problema actual individualmente y hace su predicción, entonces todas esas predicciones se agregan para formar una predicción global. Así pues, en este trabajo estamos interesados en desarrollar estrategias de aprendizaje basadas en casos y en ensembles para sistemas multi-agente.
Concretamente, presentaremos un marco de trabajo llamado Razonamiento Basado en Casos Multi-Agente (MAC), una aproximación al CBR basada en agentes. Cada agente individual en un sistema MAC es capaz de aprender y solucionar problemas individualmente utilizando CBR con su base de casos individual. Además, cada base de
casos es propiedad de un agente individual, y cualquier información de dicha base de casos será revelada o compartida únicamente si el agente lo decide así. Por tanto, este marco de trabajo preserva la privacidad de los datos y la autonomía de los agentes para revelar información.
Ésta tesis se centra en desarrollar estrategias para que agentes individuales con capacidad de aprender puedan incrementar su rendimiento tanto cuando trabajan individualmente como cuando trabajan como un ensemble. Además, las decisiones en un sistema MAC se toman de manera descentralizada, dado que cada agente tiene autonomía de decisión. Por tanto, las técnicas desarrolladas en este marco de trabajo consiguen un incremento del rendimiento como resultado de decisiones individuales tomadas de manera descentralizada. Concretamente, presentaremos tres tipos de estrategias: estrategias para crear ensembles de agentes, estrategias para realizar retención de casos en sistemas multi-agente, y estrategias para realizar redistribución de casos.
This monograph presents a framework for learning in a distributed data scenario with decentralized decision making. We have based our framework in Multi-Agent Systems (MAS) in order to have decentralized decision making, and in Case-Based Reasoning (CBR), since the lazy learning nature of CBR is suitable for dynamic multi-agent systems. Moreover, we are interested in autonomous agents that collaboratively
work as ensembles. An ensemble of agents solves problems in the following way: each individual agent solves the problem at hand individually and makes its individual prediction, then all those predictions are aggregated to form a global prediction. Therefore, in this work we are interested in developing ensemble case based
learning strategies for multi-agent systems.
Specifically, we will present the Multi-Agent Case Based Reasoning (MAC) framework, a multi-agent approach to CBR. Each individual agent in a MAC system is capable of individually learn and solve problems using CBR with an individual case base. Moreover, each case base is owned and managed by an individual agent, and any information is disclosed or shared only if the agent decides so. Thus, this framework preserves the privacy of data, and the autonomy to disclose data.
The focus of this thesis is to develop strategies so that individual learning agents improve their performance both individually and as an ensemble. Moreover, decisions in the MAC framework are made in a decentralized way since each individual agent has decision autonomy. Therefore, techniques developed in this framework achieve an improvement of individual and ensemble performance as a result of individual decisions made in a decentralized way. Specifically, we will present three kind of strategies: strategies to form ensembles of agents, strategies to perform case retention in multi-agent systems, and strategies to perform case redistribution.
APA, Harvard, Vancouver, ISO, and other styles
2

Börthas, Lovisa, and Sjölander Jessica Krange. "Machine Learning Based Prediction and Classification for Uplift Modeling." Thesis, KTH, Matematisk statistik, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-266379.

Full text
Abstract:
The desire to model the true gain from targeting an individual in marketing purposes has lead to the common use of uplift modeling. Uplift modeling requires the existence of a treatment group as well as a control group and the objective hence becomes estimating the difference between the success probabilities in the two groups. Efficient methods for estimating the probabilities in uplift models are statistical machine learning methods. In this project the different uplift modeling approaches Subtraction of Two Models, Modeling Uplift Directly and the Class Variable Transformation are investigated. The statistical machine learning methods applied are Random Forests and Neural Networks along with the standard method Logistic Regression. The data is collected from a well established retail company and the purpose of the project is thus to investigate which uplift modeling approach and statistical machine learning method that yields in the best performance given the data used in this project. The variable selection step was shown to be a crucial component in the modeling processes as so was the amount of control data in each data set. For the uplift to be successful, the method of choice should be either the Modeling Uplift Directly using Random Forests, or the Class Variable Transformation using Logistic Regression. Neural network - based approaches are sensitive to uneven class distributions and is hence not able to obtain stable models given the data used in this project. Furthermore, the Subtraction of Two Models did not perform well due to the fact that each model tended to focus too much on modeling the class in both data sets separately instead of modeling the difference between the class probabilities. The conclusion is hence to use an approach that models the uplift directly, and also to use a great amount of control data in each data set.
Behovet av att kunna modellera den verkliga vinsten av riktad marknadsföring har lett till den idag vanligt förekommande metoden inkrementell responsanalys. För att kunna utföra denna typ av metod krävs förekomsten av en existerande testgrupp samt kontrollgrupp och målet är således att beräkna differensen mellan de positiva utfallen i de två grupperna. Sannolikheten för de positiva utfallen för de två grupperna kan effektivt estimeras med statistiska maskininlärningsmetoder. De inkrementella responsanalysmetoderna som undersöks i detta projekt är subtraktion av två modeller, att modellera den inkrementella responsen direkt samt en klassvariabeltransformation. De statistiska maskininlärningsmetoderna som tillämpas är random forests och neurala nätverk samt standardmetoden logistisk regression. Datan är samlad från ett väletablerat detaljhandelsföretag och målet är därmed att undersöka vilken inkrementell responsanalysmetod och maskininlärningsmetod som presterar bäst givet datan i detta projekt. De mest avgörande aspekterna för att få ett bra resultat visade sig vara variabelselektionen och mängden kontrolldata i varje dataset. För att få ett lyckat resultat bör valet av maskininlärningsmetod vara random forests vilken används för att modellera den inkrementella responsen direkt, eller logistisk regression tillsammans med en klassvariabeltransformation. Neurala nätverksmetoder är känsliga för ojämna klassfördelningar och klarar därmed inte av att erhålla stabila modeller med den givna datan. Vidare presterade subtraktion av två modeller dåligt på grund av att var modell tenderade att fokusera för mycket på att modellera klassen i båda dataseten separat, istället för att modellera differensen mellan dem. Slutsatsen är således att en metod som modellerar den inkrementella responsen direkt samt en relativt stor kontrollgrupp är att föredra för att få ett stabilt resultat.
APA, Harvard, Vancouver, ISO, and other styles
3

Xia, Junshi. "Multiple classifier systems for the classification of hyperspectral data." Thesis, Grenoble, 2014. http://www.theses.fr/2014GRENT047/document.

Full text
Abstract:
Dans cette thèse, nous proposons plusieurs nouvelles techniques pour la classification d'images hyperspectrales basées sur l'apprentissage d'ensemble. Le cadre proposé introduit des innovations importantes par rapport aux approches précédentes dans le même domaine, dont beaucoup sont basées principalement sur un algorithme individuel. Tout d'abord, nous proposons d'utiliser la Forêt de Rotation (Rotation Forest) avec différentes techiniques d'extraction de caractéristiques linéaire et nous comparons nos méthodes avec les approches d'ensemble traditionnelles, tels que Bagging, Boosting, Sous-espace Aléatoire et Forêts Aléatoires. Ensuite, l'intégration des machines à vecteurs de support (SVM) avec le cadre de sous-espace de rotation pour la classification de contexte est étudiée. SVM et sous-espace de rotation sont deux outils puissants pour la classification des données de grande dimension. C'est pourquoi, la combinaison de ces deux méthodes peut améliorer les performances de classification. Puis, nous étendons le travail de la Forêt de Rotation en intégrant la technique d'extraction de caractéristiques locales et l'information contextuelle spatiale avec un champ de Markov aléatoire (MRF) pour concevoir des méthodes spatio-spectrale robustes. Enfin, nous présentons un nouveau cadre général, ensemble de sous-espace aléatoire, pour former une série de classifieurs efficaces, y compris les arbres de décision et la machine d'apprentissage extrême (ELM), avec des profils multi-attributs étendus (EMaPS) pour la classification des données hyperspectrales. Six méthodes d'ensemble de sous-espace aléatoire, y compris les sous-espaces aléatoires avec les arbres de décision, Forêts Aléatoires (RF), la Forêt de Rotation (RoF), la Forêt de Rotation Aléatoires (Rorf), RS avec ELM (RSELM) et sous-espace de rotation avec ELM (RoELM), sont construits par multiples apprenants de base. L'efficacité des techniques proposées est illustrée par la comparaison avec des méthodes de l'état de l'art en utilisant des données hyperspectrales réelles dans de contextes différents
In this thesis, we propose several new techniques for the classification of hyperspectral remote sensing images based on multiple classifier system (MCS). Our proposed framework introduces significant innovations with regards to previous approaches in the same field, many of which are mainly based on an individual algorithm. First, we propose to use Rotation Forests with several linear feature extraction and compared them with the traditional ensemble approaches, such as Bagging, Boosting, Random subspace and Random Forest. Second, the integration of the support vector machines (SVM) with Rotation subspace framework for context classification is investigated. SVM and Rotation subspace are two powerful tools for high-dimensional data classification. Therefore, combining them can further improve the classification performance. Third, we extend the work of Rotation Forests by incorporating local feature extraction technique and spatial contextual information with Markov random Field (MRF) to design robust spatial-spectral methods. Finally, we presented a new general framework, Random subspace ensemble, to train series of effective classifiers, including decision trees and extreme learning machine (ELM), with extended multi-attribute profiles (EMAPs) for classifying hyperspectral data. Six RS ensemble methods, including Random subspace with DT (RSDT), Random Forest (RF), Rotation Forest (RoF), Rotation Random Forest (RoRF), RS with ELM (RSELM) and Rotation subspace with ELM (RoELM), are constructed by the multiple base learners. The effectiveness of the proposed techniques is illustrated by comparing with state-of-the-art methods by using real hyperspectral data sets with different contexts
APA, Harvard, Vancouver, ISO, and other styles
4

Pesaranghader, Ali. "A Reservoir of Adaptive Algorithms for Online Learning from Evolving Data Streams." Thesis, Université d'Ottawa / University of Ottawa, 2018. http://hdl.handle.net/10393/38190.

Full text
Abstract:
Continuous change and development are essential aspects of evolving environments and applications, including, but not limited to, smart cities, military, medicine, nuclear reactors, self-driving cars, aviation, and aerospace. That is, the fundamental characteristics of such environments may evolve, and so cause dangerous consequences, e.g., putting people lives at stake, if no reaction is adopted. Therefore, learning systems need to apply intelligent algorithms to monitor evolvement in their environments and update themselves effectively. Further, we may experience fluctuations regarding the performance of learning algorithms due to the nature of incoming data as it continuously evolves. That is, the current efficient learning approach may become deprecated after a change in data or environment. Hence, the question 'how to have an efficient learning algorithm over time against evolving data?' has to be addressed. In this thesis, we have made two contributions to settle the challenges described above. In the machine learning literature, the phenomenon of (distributional) change in data is known as concept drift. Concept drift may shift decision boundaries, and cause a decline in accuracy. Learning algorithms, indeed, have to detect concept drift in evolving data streams and replace their predictive models accordingly. To address this challenge, adaptive learners have been devised which may utilize drift detection methods to locate the drift points in dynamic and changing data streams. A drift detection method able to discover the drift points quickly, with the lowest false positive and false negative rates, is preferred. False positive refers to incorrectly alarming for concept drift, and false negative refers to not alarming for concept drift. In this thesis, we introduce three algorithms, called as the Fast Hoeffding Drift Detection Method (FHDDM), the Stacking Fast Hoeffding Drift Detection Method (FHDDMS), and the McDiarmid Drift Detection Methods (MDDMs), for detecting drift points with the minimum delay, false positive, and false negative rates. FHDDM is a sliding window-based algorithm and applies Hoeffding’s inequality (Hoeffding, 1963) to detect concept drift. FHDDM slides its window over the prediction results, which are either 1 (for a correct prediction) or 0 (for a wrong prediction). Meanwhile, it compares the mean of elements inside the window with the maximum mean observed so far; subsequently, a significant difference between the two means, upper-bounded by the Hoeffding inequality, indicates the occurrence of concept drift. The FHDDMS extends the FHDDM algorithm by sliding multiple windows over its entries for a better drift detection regarding the detection delay and false negative rate. In contrast to FHDDM/S, the MDDM variants assign weights to their entries, i.e., higher weights are associated with the most recent entries in the sliding window, for faster detection of concept drift. The rationale is that recent examples reflect the ongoing situation adequately. Then, by putting higher weights on the latest entries, we may detect concept drift quickly. An MDDM algorithm bounds the difference between the weighted mean of elements in the sliding window and the maximum weighted mean seen so far, using McDiarmid’s inequality (McDiarmid, 1989). Eventually, it alarms for concept drift once a significant difference is experienced. We experimentally show that FHDDM/S and MDDMs outperform the state-of-the-art by representing promising results in terms of the adaptation and classification measures. Due to the evolving nature of data streams, the performance of an adaptive learner, which is defined by the classification, adaptation, and resource consumption measures, may fluctuate over time. In fact, a learning algorithm, in the form of a (classifier, detector) pair, may present a significant performance before a concept drift point, but not after. We define this problem by the question 'how can we ensure that an efficient classifier-detector pair is present at any time in an evolving environment?' To answer this, we have developed the Tornado framework which runs various kinds of learning algorithms simultaneously against evolving data streams. Each algorithm incrementally and independently trains a predictive model and updates the statistics of its drift detector. Meanwhile, our framework monitors the (classifier, detector) pairs, and recommends the efficient one, concerning the classification, adaptation, and resource consumption performance, to the user. We further define the holistic CAR measure that integrates the classification, adaptation, and resource consumption measures for evaluating the performance of adaptive learning algorithms. Our experiments confirm that the most efficient algorithm may differ over time because of the developing and evolving nature of data streams.
APA, Harvard, Vancouver, ISO, and other styles
5

Engen, Vegard. "Machine learning for network based intrusion detection : an investigation into discrepancies in findings with the KDD cup '99 data set and multi-objective evolution of neural network classifier ensembles from imbalanced data." Thesis, Bournemouth University, 2010. http://eprints.bournemouth.ac.uk/15899/.

Full text
Abstract:
For the last decade it has become commonplace to evaluate machine learning techniques for network based intrusion detection on the KDD Cup '99 data set. This data set has served well to demonstrate that machine learning can be useful in intrusion detection. However, it has undergone some criticism in the literature, and it is out of date. Therefore, some researchers question the validity of the findings reported based on this data set. Furthermore, as identified in this thesis, there are also discrepancies in the findings reported in the literature. In some cases the results are contradictory. Consequently, it is difficult to analyse the current body of research to determine the value in the findings. This thesis reports on an empirical investigation to determine the underlying causes of the discrepancies. Several methodological factors, such as choice of data subset, validation method and data preprocessing, are identified and are found to affect the results significantly. These findings have also enabled a better interpretation of the current body of research. Furthermore, the criticisms in the literature are addressed and future use of the data set is discussed, which is important since researchers continue to use it due to a lack of better publicly available alternatives. Due to the nature of the intrusion detection domain, there is an extreme imbalance among the classes in the KDD Cup '99 data set, which poses a significant challenge to machine learning. In other domains, researchers have demonstrated that well known techniques such as Artificial Neural Networks (ANNs) and Decision Trees (DTs) often fail to learn the minor class(es) due to class imbalance. However, this has not been recognized as an issue in intrusion detection previously. This thesis reports on an empirical investigation that demonstrates that it is the class imbalance that causes the poor detection of some classes of intrusion reported in the literature. An alternative approach to training ANNs is proposed in this thesis, using Genetic Algorithms (GAs) to evolve the weights of the ANNs, referred to as an Evolutionary Neural Network (ENN). When employing evaluation functions that calculate the fitness proportionally to the instances of each class, thereby avoiding a bias towards the major class(es) in the data set, significantly improved true positive rates are obtained whilst maintaining a low false positive rate. These findings demonstrate that the issues of learning from imbalanced data are not due to limitations of the ANNs; rather the training algorithm. Moreover, the ENN is capable of detecting a class of intrusion that has been reported in the literature to be undetectable by ANNs. One limitation of the ENN is a lack of control of the classification trade-off the ANNs obtain. This is identified as a general issue with current approaches to creating classifiers. Striving to create a single best classifier that obtains the highest accuracy may give an unfruitful classification trade-off, which is demonstrated clearly in this thesis. Therefore, an extension of the ENN is proposed, using a Multi-Objective GA (MOGA), which treats the classification rate on each class as a separate objective. This approach produces a Pareto front of non-dominated solutions that exhibit different classification trade-offs, from which the user can select one with the desired properties. The multi-objective approach is also utilised to evolve classifier ensembles, which yields an improved Pareto front of solutions. Furthermore, the selection of classifier members for the ensembles is investigated, demonstrating how this affects the performance of the resultant ensembles. This is a key to explaining why some classifier combinations fail to give fruitful solutions.
APA, Harvard, Vancouver, ISO, and other styles
6

Yang, Ming-Han, and 楊明翰. "Improved Neural Network Based Acoustic Modeling Leveraging Multi-task Learning and Ensemble Learning for Meeting Speech Recognition." Thesis, 2016. http://ndltd.ncl.edu.tw/handle/99106158778715235461.

Full text
Abstract:
碩士
國立臺灣師範大學
資訊工程學系
104
This thesis sets out to explore the use of multi-task learning (MTL) and ensemble learning techniques for more accurate estimation of the parameters involved in neural network based acoustic models, so as to improve the accuracy of meeting speech recognition. Our main contributions are three-fold. First, we conduct an empirical study to leverage various auxiliary tasks to enhance the performance of multi-task learning on meeting speech recognition. Furthermore, we also study the synergy effect of combing multi-task learning with disparate acoustic models, such as deep neural network (DNN) and convolutional neural network (CNN) based acoustic models, with the expectation to increase the generalization ability of acoustic modeling. Second, since the way to modulate the contribution (weights) of different auxiliary tasks during acoustic model training is far from optimal and actually a matter of heuristic judgment, we thus propose a simple model adaptation method to alleviate such a problem. Third, an ensemble learning method is investigated to systematically integrate the various acoustic models (weak learners) trained with multi-task learning. A series of experiments have been carried out on the augmented multi-party interaction (AMI) and Mandarin meeting recording (MMRC) corpora, which seem to reveal the effectiveness of our proposed methods in relation to several existing baselines.
APA, Harvard, Vancouver, ISO, and other styles
7

(6790182), Francisco D. Vaca. "An Ensemble Learning Based Multi-level Network Intrusion Detection System for Wi-Fi Dominant Networks." Thesis, 2019.

Find full text
Abstract:
Today, networks contribute signicantly to everyone's life. The enormous usefulness of networks for various services and data storage motivates adversaries to launch attacks on them. Network Intrusion Detection Systems (NIDSs) are used as security measure inside the organizational networks to identify any intrusions and generate alerts for them. The idea of deploying an NIDS is quite known and has been studied and adopted in both academia and industry. However, most of the NIDS literature have emphasized to detect the attacks that originate externally in a wired network infrastructure. In addition, Wi-Fi and wired networks are treated the same for the NIDSs. The open infrastructure in Wi-Fi network makes it different from the wired network. Several internal attacks that could happen in a Wi-Fi network are not pos-
sible in a wired network. The NIDSs developed using traditional approaches may fail to identify these internal attacks.

The thesis work attempts to develop a Multi-Level Network Intrusion Detection System (ML-NIDS) for Wi-Fi dominant networks that can detect internal attacks specic to Wi-Fi networks as well as the generic network attacks that are independent of network infrastructure. In Wi-Fi dominant networks, Wi-Fi devices (stations) are prevalent at the edge of campus and enterprise networks and integrated with the fixed wired infrastructure at the access. The implementation is proposed for Wi-Fi dominant networks; nevertheless, it aims to work for the wired network as well. We develop the ML-NIDS using an ensemble learning method that combines several weak
learners to create a strong learner.

APA, Harvard, Vancouver, ISO, and other styles
8

Wang, Ye. "Robust Text Mining in Online Social Network Context." Thesis, 2018. https://vuir.vu.edu.au/38645/.

Full text
Abstract:
Text mining is involved in a broad scope of applications in diverse domains that mainly, but not exclusively, serve political, commercial, medical and academic needs. Along with the rapid development of the Internet technology in recent thirty years and the advent of online social media and network in a decade, text data is obliged to entail features of online social data streams, for example, the explosive growth, the constantly changing content and the huge volume. As a result, text mining is no longer merely oriented to textual content itself, but requires consideration of surroundings and combining theories and techniques of stream processing and social network analysis, which give birth to a wide range of applications used for understanding thoughts spread over the world , such as sentiment analysis, mass surveillance and market prediction. Automatically discovering sequences of words that represent appropriate themes in a collection of documents, topic detection closely associated with document clustering and classification. These two tasks play integral roles in revealing deep insight into the text content in the whole text mining framework. However, most existing detection techniques cannot adapt to the dynamic social context. This shows bottlenecks of detecting performance and deficiencies of topic models. In this thesis, we take aim at text data stream, investigating novel techniques and solutions for robust text mining to tackle arising challenges associated with the online social context by incorporating methodologies of stream processing, topic detection and document clustering and classification. In particular, we have advanced the state-of-theart by making the following contributions: 1. A Multi-Window based Ensemble Learning (MWEL) framework is proposed for imbalanced streaming data that comprehensively improves the classification performance. MWEL ensures that the ensemble classifier is maintained up to date and adaptive to the evolving data distribution by applying a multi-window monitoring mechanism and efficient updating strategy. 2. A semi-supervised learning method is proposed to detect latent topics from news streams and the corresponding social context with a constraint propagation scheme to adequately exploit the hidden geometrical structure as supervised information in given data space. A collective learning algorithm is proposed to integrate the textual content into the social context. A locally weighted scheme is afterwards proposed to seek an improvement of the algorithm stability. 3. A Robust Hierarchical Ensemble (RHE) framework is introduced to enhance the robustness of the topic model. It, on the one hand, reduces repercussions caused by outliers and noises, and on the other overcomes inherent defects of text data. RHE adapts to the changing distribution of text stream by constructing a flexible document hierarchy which can be dynamically adjusted. A discussion of how to extract the most valuable social context is conducted with experiments for the purpose of removing some noises from the surroundings and efficiency of the proposed.
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "Multi-window based ensemble learning"

1

Wang, Ye, Hu Li, Hua Wang, Bin Zhou, and Yanchun Zhang. "Multi-Window Based Ensemble Learning for Classification of Imbalanced Streaming Data." In Lecture Notes in Computer Science, 78–92. Cham: Springer International Publishing, 2015. http://dx.doi.org/10.1007/978-3-319-26187-4_6.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Körner, Christine, and Stefan Wrobel. "Multi-class Ensemble-Based Active Learning." In Lecture Notes in Computer Science, 687–94. Berlin, Heidelberg: Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/11871842_68.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Wang, Qing, and Liang Zhang. "Ensemble Learning Based on Multi-Task Class Labels." In Advances in Knowledge Discovery and Data Mining, 464–75. Berlin, Heidelberg: Springer Berlin Heidelberg, 2010. http://dx.doi.org/10.1007/978-3-642-13672-6_44.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Cervantes, Louie, Jung-sik Lee, and Jaewan Lee. "Agent-Based Approach to Distributed Ensemble Learning of Fuzzy ARTMAP Classifiers." In Agent and Multi-Agent Systems: Technologies and Applications, 805–14. Berlin, Heidelberg: Springer Berlin Heidelberg, 2007. http://dx.doi.org/10.1007/978-3-540-72830-6_84.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Minn, Sein, Michel C. Desmarais, and ShunKai Fu. "Refinement of a Q-matrix with an Ensemble Technique Based on Multi-label Classification Algorithms." In Adaptive and Adaptable Learning, 165–78. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-45153-4_13.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Wang, Tianhao, Tianrang Weng, Jiacheng Ji, Mingjun Zhong, and Baili Zhang. "A Text Multi-label Classification Scheme Based on Resampling and Ensemble Learning." In Advances in Artificial Intelligence and Security, 67–80. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-031-06761-7_6.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Yi, Minhan, Dandan Zhao, Chenglin Liao, and Hongpeng Yin. "MK-SCE:A Novel Multi-Kernel Based Self-adapt Concept Drift Ensemble Learning." In Lecture Notes in Electrical Engineering, 492–97. Singapore: Springer Singapore, 2021. http://dx.doi.org/10.1007/978-981-16-6324-6_50.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Wu, Qinggang, and Zhongchi Liu. "Constrained Energy Minimization for Hyperspectral Multi-target Detection Based on Ensemble Learning." In Advanced Data Mining and Applications, 406–16. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-030-95408-6_31.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Kang, Yuxin, Hansheng Li, Xin Han, Boju Pan, Yuan Li, Yan Jin, Qirong Bu, Lei Cui, Jun Feng, and Lin Yang. "Automated Tumor Proportion Scoring for Assessment of PD-L1 Expression Based on Multi-Stage Ensemble Strategy." In Machine Learning in Medical Imaging, 70–79. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-59861-7_8.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Saikia, Aditya, Anil Hazarika, Bikram Patir, and Amarprit Singh. "A Supervised Ensemble Subspace Learning Model Based on Multi-view Feature Fusion Employing Multi-template EMG Signals." In Studies in Big Data, 269–88. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-030-95239-6_10.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Multi-window based ensemble learning"

1

Xu, Junyi, Le Li, and Ming Ji. "Ensemble learning based multi-source information fusion." In The Second International Conference on Image, Video Processing and Artificial Intelligence, edited by Ruidan Su. SPIE, 2019. http://dx.doi.org/10.1117/12.2542941.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Li, Hu, Peng Zou, Weihong Han, Rongze Xia, and Fei Liu. "Ensemble multi-label learning based on neural network." In the Fifth International Conference. New York, New York, USA: ACM Press, 2013. http://dx.doi.org/10.1145/2499788.2499808.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Wu, Yonglin, Jun Zhang, Yue Wu, Gengxin Ning, and Cui Yang. "Speech Enhancement Based on Multi-Objective Ensemble Learning." In 2022 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC). IEEE, 2022. http://dx.doi.org/10.1109/icspcc55723.2022.9984412.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Yan, Jie. "Ensemble SVM Regression Based Multi-View Face Detection System." In 2007 IEEE Workshop on Machine Learning for Signal Processing. IEEE, 2007. http://dx.doi.org/10.1109/mlsp.2007.4414300.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Jin, Jing, Yahui Zhao, and Rongyi Cui. "Research on Multi-granularity Ensemble Learning Based on Korean." In CONF-CDS 2021: The 2nd International Conference on Computing and Data Science. New York, NY, USA: ACM, 2021. http://dx.doi.org/10.1145/3448734.3450777.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Jing-Jing Cao, Sam Kwong, Ran Wang, and Ke Li. "AN indicator-based selection multi-objective evolutionary algorithm with preference for multi-class ensemble." In 2014 International Conference on Machine Learning and Cybernetics (ICMLC). IEEE, 2014. http://dx.doi.org/10.1109/icmlc.2014.7009108.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Wang, Peng, Peng Zhang, and Li Guo. "Mining Multi-label Data Streams Using Ensemble-based Active Learning." In Proceedings of the 2012 SIAM International Conference on Data Mining. Philadelphia, PA: Society for Industrial and Applied Mathematics, 2012. http://dx.doi.org/10.1137/1.9781611972825.97.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Tan, Jiajie, and Ning Li. "Ensemble Learning Based Multi-Color Space in Convolutional Neural Network." In 2019 Chinese Control Conference (CCC). IEEE, 2019. http://dx.doi.org/10.23919/chicc.2019.8865681.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Kocyigit, Gokhan, and Yusuf Yaslan. "Dictionary ensemble based multi instance active learning method for image categorization." In 2016 24th Signal Processing and Communication Application Conference (SIU). IEEE, 2016. http://dx.doi.org/10.1109/siu.2016.7495966.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Avila, Nelson Fabian, Von-Wun Soo, Wan-Yu Yu, and Chia-Chi Chu. "Capacity-based service restoration using Multi-Agent technology and ensemble learning." In 2015 18th International Conference on Intelligent System Application to Power Systems (ISAP). IEEE, 2015. http://dx.doi.org/10.1109/isap.2015.7325546.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography