Academic literature on the topic 'AdaBoost (Adaptive Boosting)'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'AdaBoost (Adaptive Boosting).'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "AdaBoost (Adaptive Boosting)"

1

Mendrofa, Rosa Delima, Maria Hosianna Siallagan, Junita Amalia, and Diana Pebrianty Pakpahan. "Credit Risk Analysis With Extreme Gradient Boosting and Adaptive Boosting Algorithm." Journal of Information System,Graphics, Hospitality and Technology 5, no. 1 (2023): 1–7. http://dx.doi.org/10.37823/insight.v5i1.233.

Full text
Abstract:
Credit Risk Analysis digunakan untuk mengenali resiko terhadap pinjaman untuk mencegah penunggakan pembayaran utang. Pemberian uji kelayakan pinjaman dapat di analisis menggunakan model klasifikasi. Untuk menghasilkan model credit risk analysis yang sesuai, penulis mengajukan Algoritma Extreme Gradient Boosting (XGBoost) dan Adaptive Boosting (AdaBoost). Data yang digunakan dalam penelitian ini adalah data pinjaman platform Peer to Peer (P2P) Lending. Penelitian ini menerapkan data preprocessing yang bertujuan untuk menghasilkan data yang lebih baik dan melakukan analisis terhadap data. Analisis dilakukan berdasarkan fitur yang dimiliki oleh peminjam menggunakan algoritma klasifikasi berdasarkan historical data pinjaman peminjam. Fitur yang digunakan seperti jumlah pinjaman yang diajukan, total pinjaman yang ditawarkan, jumlah pembayaran pinjaman, jangka waktu pembayaran, suku bungan pinjaman, jumlah angsuran dan lain lain. Jumlah fitur sebelum dilakukan data reduksi 136 dan setelah direduksi 34 fitur. Fitur tersebut digunakan pada penerapan algoritma XGBoost dan AdaBoost untuk menghasilkan klasifikasi good borrower dan bad borrower. Penulis menggunakan metode evaluasi kurva ROC dan nilai AUC untuk menilai performa dari kedua algoritma. Pada kurva ROC, nilai AUC dari algoritma XGBoost 0,92 dan nilai AUC dari algrithma AdaBoost adalah 0,89. Berdasarkan perbandingan nilai AUC tersebut dapat disimpulkan algoritma XGBoost menghasilkan klasifikasi yang lebih baik untuk model klasifikasi pemberian pinjaman.
APA, Harvard, Vancouver, ISO, and other styles
2

Nayab, Durr e., Rehan Ullah Khan, and Ali Mustafa Qamar. "Performance Augmentation of Base Classifiers Using Adaptive Boosting Framework for Medical Datasets." Applied Computational Intelligence and Soft Computing 2023 (December 22, 2023): 1–10. http://dx.doi.org/10.1155/2023/5542049.

Full text
Abstract:
This paper investigates the performance enhancement of base classifiers within the AdaBoost framework applied to medical datasets. Adaptive boosting (AdaBoost), being an instance of boosting, combines other classifiers to enhance their performance. We conducted a comprehensive experiment to assess the efficacy of twelve base classifiers with the AdaBoost framework, namely, Bayes network, decision stump, ZeroR, decision tree, Naïve Bayes, J-48, voted perceptron, random forest, bagging, random tree, stacking, and AdaBoost itself. The experiments are carried out on five datasets from the medical domain based on various types of cancers, i.e., global cancer map (GCM), lymphoma-I, lymphoma-II, leukaemia, and embryonal tumours. The evaluation focuses on the accuracy, precision, and efficiency of the base classifiers in the AdaBoost framework. The results show that the performance of Naïve Bayes, Bayes network, and voted perceptron is highly improved compared to the rest of the base classifiers, attaining accuracies as high as 94.74%, 97.78%, and 97.78%, respectively. The results also show that in most cases, the base classifiers perform better with AdaBoost compared to their performance, i.e., for voted perceptron, the accuracy is improved up to 13.34%.For bagging, it is improved by up to 7%. This research aims to identify such base classifiers with optimal boosting capabilities within the AdaBoost framework for medical datasets. The significance of these results is that they provide insight into the performance of the base classifiers when used in the boosting framework to enhance the classification performance of classifiers in scenarios where individual classifiers do not perform up to the mark.
APA, Harvard, Vancouver, ISO, and other styles
3

La, Lei, Qiao Guo, Dequan Yang, and Qimin Cao. "Multiclass Boosting with Adaptive Group-BasedkNN and Its Application in Text Categorization." Mathematical Problems in Engineering 2012 (2012): 1–24. http://dx.doi.org/10.1155/2012/793490.

Full text
Abstract:
AdaBoost is an excellent committee-based tool for classification. However, its effectiveness and efficiency in multiclass categorization face the challenges from methods based on support vector machine (SVM), neural networks (NN), naïve Bayes, andk-nearest neighbor (kNN). This paper uses a novel multi-class AdaBoost algorithm to avoid reducing the multi-class classification problem to multiple two-class classification problems. This novel method is more effective. In addition, it keeps the accuracy advantage of existing AdaBoost. An adaptive group-basedkNN method is proposed in this paper to build more accurate weak classifiers and in this way control the number of basis classifiers in an acceptable range. To further enhance the performance, weak classifiers are combined into a strong classifier through a double iterative weighted way and construct an adaptive group-basedkNN boosting algorithm (AGkNN-AdaBoost). We implement AGkNN-AdaBoost in a Chinese text categorization system. Experimental results showed that the classification algorithm proposed in this paper has better performance both in precision and recall than many other text categorization methods including traditional AdaBoost. In addition, the processing speed is significantly enhanced than original AdaBoost and many other classic categorization algorithms.
APA, Harvard, Vancouver, ISO, and other styles
4

Riansyah, Muhammad, Saib Suwilo, and Muhammad Zarlis. "Improved Accuracy In Data Mining Decision Tree Classification Using Adaptive Boosting (Adaboost)." SinkrOn 8, no. 2 (2023): 617–22. http://dx.doi.org/10.33395/sinkron.v8i2.12055.

Full text
Abstract:
The Decision Tree algorithm is a data mining method algorithm that is often applied as a solution to a problem for a classification. The Decision Tree C5.0 algorithm has several weaknesses, including: the C5.0 algorithm and several other decision tree methods are often biased towards modeling whose features have many levels, some problems for the model can occur such as over-fit or under-fit challenges, big changes to decision logic can result in small changes to data training, C5.0 can experience modeling inconvenience, data imbalance causes low accuracy in C5.0 algorithm. The boosting algorithm is an iterative algorithm that gives different weights to the distribution of training data in each iteration. Each iteration of boosting adds weight to examples of misclassification and decreases weight to examples of correct classification, thereby effectively changing the distribution of the training data. One example of a boosting algorithm is adaboost. The purpose of this research is to improve the performance of the Decision Tree C5.0 classification method using adaptive boosting (adaboost) to predict hepatitis disease using the Confusion matrix. Tests that have been carried out with the Confusion Matrix use the Hepatitis dataset in the Decision Tree C5.0 classification which has an accuracy rate of 80.58% with a classification error rate of 19.15%. Whereas in the Decision Tree C5.0 classification Adaboost has a higher accuracy rate of 82.98%, a classification error rate of 17.02%. This difference is caused by the adaboost algorithm, because the adaboost algorithm is able to change a weak classifier into a strong classifier by increasing the weight of the observations, and adaboost is also able to reduce the classifier error rate.
APA, Harvard, Vancouver, ISO, and other styles
5

Zhang, Jiangnan, Kewen Xia, Ziping He, Zhixian Yin, and Sijie Wang. "Semi-Supervised Ensemble Classifier with Improved Sparrow Search Algorithm and Its Application in Pulmonary Nodule Detection." Mathematical Problems in Engineering 2021 (February 18, 2021): 1–18. http://dx.doi.org/10.1155/2021/6622935.

Full text
Abstract:
The Adaptive Boosting (AdaBoost) classifier is a widely used ensemble learning framework, and it can get good classification results on general datasets. However, it is challenging to apply the AdaBoost classifier directly to pulmonary nodule detection of labeled and unlabeled lung CT images since there are still some drawbacks to ensemble learning method. Therefore, to solve the labeled and unlabeled data classification problem, the semi-supervised AdaBoost classifier using an improved sparrow search algorithm (AdaBoost-ISSA-S4VM) was established. Firstly, AdaBoost classifier is used to construct a strong semi-supervised classifier using several weak classifiers S4VM (AdaBoost-S4VM). Next, in order to solve the accuracy problem of AdaBoost-S4VM, sparrow search algorithm (SSA) is introduced in the AdaBoost classifier and S4VM. Then, sine cosine algorithm and new labor cooperation structure are adopted to increase the global optimal solution and convergence performance of sparrow search algorithm, respectively. Furthermore, based on the improved sparrow search algorithm and adaptive boosting classifier, the AdaBoost-S4VM classifier is improved. Finally, the effective improved AdaBoost-ISSA-S4VM classification model was developed for actual pulmonary nodule detection based on the publicly available LIDC-IDRI database. The experimental results have proved that the established AdaBoost-ISSA-S4VM classification model has good performance on labeled and unlabeled lung CT images.
APA, Harvard, Vancouver, ISO, and other styles
6

Prianti, Ade Irma, Rukun Santoso, and Arief Rachman Hakim. "PERBANDINGAN METODE K-NEAREST NEIGHBOR DAN ADAPTIVE BOOSTING PADA KASUS KLASIFIKASI MULTI KELAS." Jurnal Gaussian 9, no. 3 (2020): 346–54. http://dx.doi.org/10.14710/j.gauss.v9i3.28924.

Full text
Abstract:
The company's financial health provides an indication of company’s performance that is useful for knowing the company's position in industrial area. The company's performance needs to be predicted to knowing the company's progress. K-Nearest Neighbor (KNN) and Adaptive Boosting (AdaBoost) are classification methods that can be used to predict company's performance. KNN classifies data based on the proximity of the data distance while AdaBoost works with the concept of giving more weight to observations that include weak learners. The purpose of this study is to compare the KNN and AdaBoost methods to find out better methods for predicting company’s performance in Indonesia. The dependent variable used in this study is the company's performance which is classified into four classes, namely unhealthy, less healthy, healthy, and very healthy. The independent variables used consist of seven financial ratios namely ROA, ROE, WCTA, TATO, DER, LDAR, and ROI. The data used are financial ratio data from 575 companies listed on the Indonesia Stock Exchange in 2019. The results of this study indicate that the prediction of company’s performance in Indonesia should use the AdaBoost method because it has a classification accuracy of 0,84522 which is greater than the KNN method’s accuracy of 0,82087. Keywords: company’s performance, classification, KNN and AdaBoost, classification accuracy.
APA, Harvard, Vancouver, ISO, and other styles
7

Akazue, Maureen, Anthonia Onovughe, Omede Edith, and John Paul A.C. Hampo. "Use of Adaptive Boosting Algorithm to Estimate User's Trust in the Utilization of Virtual Assistant Systems." International Journal of Innovative Science and Research Technology 8, no. 1 (2023): 502–7. https://doi.org/10.5281/zenodo.7568675.

Full text
Abstract:
User trust in technology is an essential factor for the usage of a system or machine. AI enabled technologies such as virtual digital assistants simplify a lot of process for humans starting from simple search to a more complex action like house automation and completion of some transitions notably Amazon’s Alexa. Can human actually trust these AI enabled technologies? Hence, this research applied adaptive boosting ensemble learning approach to predict users trust in virtual assistants. A technology trust dataset was obtained from figshare.com and engineered before training the adaptive boosting (AdaBoost) algorithm to learn the trends and pattern. The result of the study showed that AdaBoost had an accuracy of 94.31% for the testing set.
APA, Harvard, Vancouver, ISO, and other styles
8

Gamal, Heba, Nour Eldin Ismail, M. R. M. Rizk, Mohamed E. Khedr, and Moustafa H. Aly. "A Coherent Performance for Noncoherent Wireless Systems Using AdaBoost Technique." Applied Sciences 9, no. 2 (2019): 256. http://dx.doi.org/10.3390/app9020256.

Full text
Abstract:
Boosting is a machine learning approach built upon the idea of producing a highly precise prediction rule by combining many relatively weak and imprecise rules. The Adaptive Boosting (AdaBoost) algorithm was the first practical boosting algorithm. It remains one of the most broadly used and studied, with applications in many fields. In this paper, the AdaBoost algorithm is utilized to improve the bit error rate (BER) of different modulation techniques. By feeding the noisy received signal into the AdaBoost algorithm, it is able to recover the transmitted data from the noisy signal. Consequently, it reconstructs the constellation diagram of the modulation technique. This is done by removing the noise that affects and changes the signal space of the data. As a result, AdaBoost shows an improvement in the BER of coherently detected binary phase shift keying (BPSK) and quadrature phase shift keying (QPSK). The AdaBoost is next used to improve the BER of the noncoherent detection of the used modulation techniques. The improvement appears in the form of better results of the noncoherent simulated BER in comparison to that of the theoretical noncoherent BER. Therefore, the AdaBoost algorithm is able to achieve a coherent performance for the noncoherent system. The AdaBoost is simulated for several techniques in additive white Gaussian noise (AWGN) and Rayleigh fading channels so, as to verify the improving effect of the AdaBoost algorithm.
APA, Harvard, Vancouver, ISO, and other styles
9

Jang, Seok-Woo, and Sang-Hong Lee. "Harmful Content Detection Based on Cascaded Adaptive Boosting." Journal of Sensors 2018 (October 21, 2018): 1–12. http://dx.doi.org/10.1155/2018/7497243.

Full text
Abstract:
Recently, it has become very easy to acquire various types of image contents through mobile devices with high-performance visual sensors. However, harmful image contents such as nude pictures and videos are also distributed and spread easily. Therefore, various methods for effectively detecting and filtering such image contents are being introduced continuously. In this paper, we propose a new approach to robustly detect the human navel area, which is an element representing the harmfulness of the image, using Haar-like features and a cascaded AdaBoost algorithm. In the proposed method, the nipple area of a human is detected first using the color information from the input image and the candidate navel regions are detected using positional information relative to the detected nipple area. Nonnavel areas are then removed from the candidate navel regions and only the actual navel areas are robustly detected through filtering using the Haar-like feature and the cascaded AdaBoost algorithm. The experimental results show that the proposed method extracts nipple and navel areas more precisely than the conventional method. The proposed navel area detection algorithm is expected to be used effectively in various applications related to the detection of harmful contents.
APA, Harvard, Vancouver, ISO, and other styles
10

Irma Prianti, Ade. "Pebandingan Metode K-Nearest Neighbor dan Adaptive Boosting pada Kasus Klasifikasi Multi Kelas." J Statistika: Jurnal Ilmiah Teori dan Aplikasi Statistika 13, no. 1 (2020): 39–47. http://dx.doi.org/10.36456/jstat.vol13.no1.a3269.

Full text
Abstract:
Kesehatan keuangan perusahaan memberikan suatu indikasi kinerja perusahaan yang berguna untuk mengetahui posisi perusahaan dalam area industri. Kinerja perusahaan perlu diprediksi untuk mengetahui perkembangan perusahaan. K-Nearest Neighbor (KNN) dan Adaptive Boosting (AdaBoost) merupakan metode klasifikasi yang dapat digunakan untuk memprediksi kinerja perusahaan. KNN mengklasifikasikan data berdasarkan kedekatan jarak data sedangkan AdaBoost bekerja dengan konsep memberi bobot lebih pada amatan yang termasuk weak learner. Tujuan dari penelitian ini adalah membandingkan metode KNN dan AdaBoost untuk mengetahui metode yang lebih baik dalam memprediksi kinerja perusahaan di Indonesia. Variabel dependen yang digunakan dalam penelitian ini adalah kinerja perusahaan yang digolongkan ke dalam empat kelas yaitu tidak sehat, kurang sehat, sehat, dan sehat sekali. Variabel independen yang digunakan terdiri atas tujuh rasio keuangan yaitu ROA, ROE, WCTA, TATO, DER, LDAR, dan ROI. Data yang digunakan yaitu data rasio keuangan dari 575 perusahaan yang tercatat di Bursa Efek Indonesia tahun 2019. Hasil penelitian ini menunjukkan bahwa prediksi kinerja perusahaan di Indonesia sebaiknya menggunakan metode AdaBoost karena memiliki akurasi klasifikasi sebesar 0,84522 yang lebih besar dibandingkan akurasi metode KNN sebesar 0,82087
APA, Harvard, Vancouver, ISO, and other styles
More sources

Book chapters on the topic "AdaBoost (Adaptive Boosting)"

1

Shi, Heng, Belkacem Chikhaoui, and Shengrui Wang. "Tree-Based Models for Pain Detection from Biomedical Signals." In Lecture Notes in Computer Science. Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-031-09593-1_14.

Full text
Abstract:
AbstractFor medical treatments, pain is often measured by self-report. However, the current subjective pain assessment highly depends on the patient’s response and is therefore unreliable. In this paper, we propose a physiological-signals-based objective pain recognition method that can extract new features, which have never been discovered in pain detection, from electrodermal activity (EDA) and electrocardiogram (ECG) signals. To discriminate the absence and presence of pain, we establish four classification tasks and build four tree-based classifiers, including Random Forest, Adaptive Boosting (AdaBoost), eXtreme Gradient Boosting (XGBoost), and TabNet. The comparative experiments demonstrate that our method using the EDA and ECG features yields accurate classification results. Furthermore, the TabNet achieves a large accuracy improvement using our ECG features and a classification accuracy of 94.51% using the features selected from the fusion of the two signals.
APA, Harvard, Vancouver, ISO, and other styles
2

El bakrawy, Lamiaa M., and Abeer S. Desuky. "A Hybrid Classification Algorithm and Its Application on Four Real-World Data Sets." In Advanced Bioinspiration Methods for Healthcare Standards, Policies, and Reform. IGI Global, 2022. http://dx.doi.org/10.4018/978-1-6684-5656-9.ch006.

Full text
Abstract:
The aim of this chapter is to propose a hybrid classification algorithm based on particle swarm optimization (PSO) to enhance the generalization performance of the adaptive boosting (AdaBoost) algorithm. AdaBoost enhances any given machine learning algorithm performance by producing some weak classifiers which requires more time and memory and may not give the best classification accuracy. For this purpose, PSO is proposed as a post optimization procedure for the resulted weak classifiers and removes the redundant classifiers. The experiments were conducted on the basis of ionosphere data set, thoracic surgery data set, blood transfusion service center data set (btsc) and Statlog (Australian credit approval) data set. The experimental results show that a given boosted classifier with post optimization based on PSO improves the classification accuracy for all used data. Also, the experiments show that the proposed algorithm outperforms other techniques with best generalization.
APA, Harvard, Vancouver, ISO, and other styles
3

Caceres Hernandez, Danilo, Laksono Kurnianggoro, Alexander Filonenko, and Kang-Hyun Jo. "Obstacle Classification Based on Laser Scanner for Intelligent Vehicle Systems." In Advances in Computational Intelligence and Robotics. IGI Global, 2020. http://dx.doi.org/10.4018/978-1-5225-9924-1.ch010.

Full text
Abstract:
In the field of advanced driver-assistance and autonomous vehicle systems, understanding the surrounding vehicles plays a vital role to ensure a robust and safe navigation. To solve detection and classification problem, an obstacle classification strategy based on laser sensor is presented. Objects are classified according the geometry, distance range, reflectance, and disorder of each of the detected object. In order to define the best number of features that allows the algorithm to classify these objects, a feature analysis is performed. To do this, the set of features were divided into four groups based on the characteristic, distance, reflectance, and the entropy of the object. Finally, the classification task is performed using the support vector machines (SVM) and adaptive boosting (AdaBoost) algorithms. The evaluation indicates that the method proposes a feasible solution for intelligent vehicle applications, achieving a detection rate of 87.96% at 48.32 ms for the SVM and 98.19% at 79.18ms for the AdaBoost.
APA, Harvard, Vancouver, ISO, and other styles
4

Ramalingam, Renugadevi. "An Innovative Investigation on Predicting Forest Fire Using Machine Learning Approach." In AI and IoT for Proactive Disaster Management. IGI Global, 2024. http://dx.doi.org/10.4018/979-8-3693-3896-4.ch004.

Full text
Abstract:
Predicting forest fire occurrences can bolster early detection capabilities and improve early warning systems and responses. Currently, forest and grassland fire prevention and suppression efforts in China face significant hurdles due to the complex interplay of natural and societal factors. While existing models for predicting forest fire occurrences typically consider factors like vegetation, topography, weather conditions, and human activities, the moisture content of forest fuels is a critical aspect closely linked to fire occurrences. Additionally, it introduces forest fuel-related factors, including vegetation canopy water content and evapotranspiration from the top of the vegetation canopy, to construct a comprehensive database for predicting forest fire occurrences. Furthermore, the study develops a forest fire occurrence prediction model using machine learning techniques such as the random forest model (RF), gradient boosting decision tree model (GBDT), and adaptive augmentation model (AdaBoost).
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "AdaBoost (Adaptive Boosting)"

1

Okai, M. I., O. Ogolo, P. Nzerem, and K. S. Ibrahim. "Application of Boosting Machine Learning for Mud Loss Prediction During Drilling Operations." In SPE Nigeria Annual International Conference and Exhibition. SPE, 2024. http://dx.doi.org/10.2118/221583-ms.

Full text
Abstract:
Abstract Lost circulation during drilling operations is a persistent challenge in the oil and gas industry, leading to significant financial losses and increased non-productive time. The common use of lost circulation materials (LCMs) in drilling fluids helps mitigate mud loss only to an extent. However, predicting the extent of mud loss before drilling specific formations would greatly benefit engineers. This study aims to predict mud loss using advanced boosting machine learning frameworks, addressing the need for more accurate forecasting tools. We evaluated three ensemble boosting algorithms—Adaptive Boosting (AdaBoost), Light Gradient Boosting Machine (LightGBM), and Extreme Gradient Boosting (XGBoost)—and compared them to Random Forest, a baseline bagging algorithm. Utilizing a dataset of over 7,000 data points with 27 features from drilling operations in Well MXY at the Utah FORGE field, we found that XGBoost and Random Forest were the most accurate models, with R2 scores of 0.935 and 0.934, respectively. These results indicate that while XGBoost is the top-performing framework, Random Forest remains a robust and reliable method for predicting lost circulation, providing valuable insights for drilling engineers.
APA, Harvard, Vancouver, ISO, and other styles
2

Donat, William, Kihoon Choi, Woosun An, Satnam Singh, and Krishna Pattipati. "Data Visualization, Data Reduction and Classifier Fusion for Intelligent Fault Detection and Diagnosis in Gas Turbine Engines." In ASME Turbo Expo 2007: Power for Land, Sea, and Air. ASMEDC, 2007. http://dx.doi.org/10.1115/gt2007-28343.

Full text
Abstract:
In this paper, we investigate four key issues associated with data-driven approaches for fault classification using the Pratt and Whitney commercial dual-spool turbofan engine data as a test case. The four issues considered here include: (1) Can we characterize, a priori, the difficulty of fault classification via self-organizing maps? (2) Do data reduction techniques improve fault classification performance and enable the implementation of data-driven classification techniques in memory-constrained digital electronic control units (DECUs)?, (3) When does adaptive boosting, an incremental fusion method that successively combines moderately inaccurate classifiers into accurate ones, help improve classification performance?, and (4) How to synthesize classifier fusion architectures to improve the overall diagnostic accuracy? The classifiers studied in this paper are the support vector machine (SVM), probabilistic neural network (PNN), k-nearest neighbor (KNN), principal component analysis (PCA), Gaussian mixture models (GMM), and a physics-based single fault isolator (SFI). As these algorithms operate on large volumes of data and are generally computationally expensive, we reduce the dataset using the multi-way partial least squares (MPLS) method. This has the added benefits of improved diagnostic accuracy and smaller memory requirements. The performance of the moderately inaccurate classifiers is improved using adaptive boosting (AdaBoost). These results are compared to the results of the classifiers alone, as well as different fusion architectures. We show that fusion reduces the variability in diagnostic accuracy, and is most useful when combining moderately inaccurate classifiers.
APA, Harvard, Vancouver, ISO, and other styles
3

Jagtap, Shilpa, J. L. Mudegaonkar, Sanjay Patil, and Dinesh Bhoyar. "A Novel Approach for Diagnosis of Diabetes Using Iris Image Processing Technique and Evaluation Parameters." In National Conference on Relevance of Engineering and Science for Environment and Society. AIJR Publisher, 2021. http://dx.doi.org/10.21467/proceedings.118.37.

Full text
Abstract:
This paper presented here deals with study of identification and verification approach of Diabetes based on human iris pattern. In the pre-processing of this work, region of interest according to color (ROI) concept is used for iris localization, Dougman's rubber sheet model is used for normalization and Circular Hough Transform can be used for pupil and boundary detection. To extract features, Gabor Filter, Histogram of Oriented Gradients, five level decomposition of wavelet transforms likeHaar, db2, db4, bior 2.2, bior6.8 waveletscan be used. Binary coding scheme binaries’ the feature vector coefficients and classifier like hamming distance, Support Vector Machine (SVM), Adaptive Boosting (AdaBoost), Neural Networks (NN), Random Forest (RF) and Linear Discriminative Analysis (LDA) with shrinkage parametercan be used for template matching. Performance parameters such as Computational time, Hamming distance variation, False Acceptance Rate (FAR), False Rejection Rate (FRR), Accuracy, and Match ratio can be calculated for the comparison purpose.
APA, Harvard, Vancouver, ISO, and other styles
4

Khan, Abdul Muqtadir, Abdullah BinZiad, and Abdullah Al Subaii. "Boosting Algorithm Choice in Predictive Machine Learning Models for Fracturing Applications." In SPE/IATMI Asia Pacific Oil & Gas Conference and Exhibition. SPE, 2021. http://dx.doi.org/10.2118/205642-ms.

Full text
Abstract:
Abstract With the advancement in machine learning (ML) applications, some recent research has been conducted to optimize fracturing treatments. There are a variety of models available using various objective functions for optimization and different mathematical techniques. There is a need to extend the ML techniques to optimize the choice of algorithm. For fracturing treatment design, the literature for comparative algorithm performance is sparse. The research predominantly shows that compared to the most commonly used regressors and classifiers, some sort of boosting technique consistently outperforms on model testing and prediction accuracy. A database was constructed for a heterogeneous reservoir. Four widely used boosting algorithms were used on the database to predict the design only from the output of a short injection/falloff test. Feature importance analysis was done on eight output parameters from the falloff analysis, and six were finalized for the model construction. The outputs selected for prediction were fracturing fluid efficiency, proppant mass, maximum proppant concentration, and injection rate. Extreme gradient boost (XGBoost), categorical boost (CatBoost), adaptive boost (AdaBoost), and light gradient boosting machine (LGBM) were the algorithms finalized for the comparative study. The sensitivity was done for a different number of classes (four, five, and six) to establish a balance between accuracy and prediction granularity. The results showed that the best algorithm choice was between XGBoost and CatBoost for the predicted parameters under certain model construction conditions. The accuracy for all outputs for the holdout sets varied between 80 and 92%, showing robust significance for a wider utilization of these models. Data science has contributed to various oil and gas industry domains and has tremendous applications in the stimulation domain. The research and review conducted in this paper add a valuable resource for the user to build digital databases and use the appropriate algorithm without much trial and error. Implementing this model reduced the complexity of the proppant fracturing treatment redesign process, enhanced operational efficiency, and reduced fracture damage by eliminating minifrac steps with crosslinked gel.
APA, Harvard, Vancouver, ISO, and other styles
5

Al-Mudhafar, Watheq J., and David A. Wood. "Tree-Based Ensemble Algorithms for Lithofacies Classification and Permeability Prediction in Heterogeneous Carbonate Reservoirs." In Offshore Technology Conference. OTC, 2022. http://dx.doi.org/10.4043/31780-ms.

Full text
Abstract:
Abstract Rock facies are typically identified either by core analysis to provide visually interpreted lithofacies, or determined indirectly based on suites of recorded well-log data, thereby generating electrofacies interpretations. Since the lithofacies cannot be obtained for all reservoir intervals, drilled section and/or wells, it is commonly essential to model the discrete lithofacies as a function of well-log data (electrofacies) to predict the poorly sampled or non-cored intervals. The process is called predictive lithofacies classification. In this study, measured discrete lithofacies distributions (based on core data) are comparatively modeled with well-log data using two tree-based ensemble algorithms: extreme gradient boosting (XGBoost) and adaptive boosting (AdaBoost) configured as classifiers. The predicted lithofacies are then combined with recorded well-log data for analysis by an XGBoost regression model to predict permeability. The input well-log variables are log porosity, gamma ray, water saturation, neutron porosity, deep resistivity, and bulk density. The data are derived from the Mishrif carbonate reservoir in a giant southern Iraqi oil field. For efficient lithofacies classification and permeability modelling, random sub-sampling cross-validation was applied to the well-log dataset to generate two subsets: training subset for model tuning; and testing subset for prediction of data points unseen during training of the model. Confusion matrices and the total correct percentage (TCP) of predictions are used to measure the prediction performance of each algorithm to identify the most realistic lithofacies classification. The TCPs for XGBoost and AdaBoost classifiers for the training subset were 98% and 100%, respectively. However, the TCPs achieved for the testing subsets were 97%, and 96%, respectively. The mismatch between the measured and predicted permeability from the XGBoost regressor was determined using root mean square error. The XGBoost model provides accurate lithofacies classification and permeability predictions of the cored data. The XGBoost model is therefore considered suitable for providing reliable predictions of lithofacies and permeability for the non-cored intervals of the same well and for non-cored wells in the studied reservoir. The workflow for lithofacies and permeability prediction was fully implemented and visualized using R open-source codes.
APA, Harvard, Vancouver, ISO, and other styles
6

Al-Mudhafar, Watheq J., and David A. Wood. "Tree-Based Ensemble Algorithms for Lithofacies Classification and Permeability Prediction in Heterogeneous Carbonate Reservoirs." In Offshore Technology Conference. OTC, 2022. http://dx.doi.org/10.4043/31780-ms.

Full text
Abstract:
Abstract Rock facies are typically identified either by core analysis to provide visually interpreted lithofacies, or determined indirectly based on suites of recorded well-log data, thereby generating electrofacies interpretations. Since the lithofacies cannot be obtained for all reservoir intervals, drilled section and/or wells, it is commonly essential to model the discrete lithofacies as a function of well-log data (electrofacies) to predict the poorly sampled or non-cored intervals. The process is called predictive lithofacies classification. In this study, measured discrete lithofacies distributions (based on core data) are comparatively modeled with well-log data using two tree-based ensemble algorithms: extreme gradient boosting (XGBoost) and adaptive boosting (AdaBoost) configured as classifiers. The predicted lithofacies are then combined with recorded well-log data for analysis by an XGBoost regression model to predict permeability. The input well-log variables are log porosity, gamma ray, water saturation, neutron porosity, deep resistivity, and bulk density. The data are derived from the Mishrif carbonate reservoir in a giant southern Iraqi oil field. For efficient lithofacies classification and permeability modelling, random sub-sampling cross-validation was applied to the well-log dataset to generate two subsets: training subset for model tuning; and testing subset for prediction of data points unseen during training of the model. Confusion matrices and the total correct percentage (TCP) of predictions are used to measure the prediction performance of each algorithm to identify the most realistic lithofacies classification. The TCPs for XGBoost and AdaBoost classifiers for the training subset were 98% and 100%, respectively. However, the TCPs achieved for the testing subsets were 97%, and 96%, respectively. The mismatch between the measured and predicted permeability from the XGBoost regressor was determined using root mean square error. The XGBoost model provides accurate lithofacies classification and permeability predictions of the cored data. The XGBoost model is therefore considered suitable for providing reliable predictions of lithofacies and permeability for the non-cored intervals of the same well and for non-cored wells in the studied reservoir. The workflow for lithofacies and permeability prediction was fully implemented and visualized using R open-source codes.
APA, Harvard, Vancouver, ISO, and other styles
7

Tackie-Otoo, Bennet Nii, Joshua Nsiah Turkson, Mohamed Mahmoud, Arshad Raza, Shirish Patil, and Victor Darkwah-Owusu. "Comparative Analysis of Ensemble Learning, Evolutionary Algorithm, and Molecular Dynamics Simulation for Enhanced Aqueous H2/Cushion Gases Interfacial Tension Prediction: Implications on Underground H2 Storage." In GOTECH. SPE, 2025. https://doi.org/10.2118/224624-ms.

Full text
Abstract:
Abstract There is a strong push for decarbonization in the Global North and South to combat climate change and avert its dire consequences including rising sea levels, severe weather events, biodiversity loss, and threats to food and water security. Hydrogen (H2) emerges as a sustainable energy carrier in this regard. Underground hydrogen storage (UHS) presents significant potential but requires a thorough understanding of H2 behavior in porous media. One of the crucial parameters is interfacial tension (IFT) which influences capillary entry pressure, H2 column height, and storage capacity. Accurate IFT measurement is therefore imperative but current laboratory methods are resource-intensive. To address the limitation, the current study employed computational techniques to estimate the IFT of aqueous H2/cushion gases systems using minimum resources. The novelty of this investigation lies in the comparative analysis of diverse techniques including ensemble learning, evolutionary algorithm, and molecular dynamics simulation. Specifically, this study employed Random Forest and Adaptive Boosting (AdaBoost) to predict the IFT of aqueous H2/cushion gases systems. We utilized a comprehensive database of 2400 experimental data points encompassing pressure (0.50–45.20 MPa), temperature (298–449 K), brine salinity (0–3.42 mol/kg), and gas mole fractions (0–100 mol%). Numerical statistical error measures and multiple data visualizations were used to evaluate the performance of the developed models. The best smart model was also compared to an evolutionary algorithm and molecular dynamics simulation to ascertain robustness and generalizability. Subsequently, sensitivity analysis and feature importance were performed to comprehend the impact of the individual input parameters on the IFT. Finally, the implication of model-predicted IFT on UHS was extensively investigated. The developed paradigms demonstrated remarkable predictive prowess, yielding a coefficient of determination>0.96, average absolute deviation (AAD)<1.50 mN/m, and root mean square error (RMSE)<2.1 mN/m, with AdaBoost emerging as the upper-echelon paradigm. Additionally, AdaBoost exhibited predictive dominance over genetic programming and molecular-dynamics-based correlations developed by other scholars. Sensitivity analysis identified carbon dioxide mole fraction as the main factor influencing IFT. Analysis also indicate that higher concentrations of the cushion gas, particularly CO2 significantly reduces IFT and this could compromise structural hydrogen trapping and lead to H2 loss. AdaBoost also provided precise results for capillary entry pressure (11.9–12.1 MPa), H2 storage height (1177–1217 m), and storage capacity (690–2226 kg/m2) of a Saudi basaltic formation comparable to experimental findings (2-3% deviation), affirming its potential for practical applications. AdaBoost can also be utilized in conjunction with Monte Carlo simulation for a swift and accurate assessment of the uncertainty associated with the influencing factors of UHS. The paradigms offer a promising approach to estimating IFT using minimal resources. Additionally, the findings facilitate UHS optimization, reduce H2 loss, and improve storage capacity.
APA, Harvard, Vancouver, ISO, and other styles
8

Al-Sahlanee, Dhuha T., Raed H. Allawi, Watheq J. Al-Mudhafar, and Changqing Yao. "Ensemble Machine Learning for Data-Driven Predictive Analytics of Drilling Rate of Penetration (ROP) Modeling: A Case Study in a Southern Iraqi Oil Field." In SPE Western Regional Meeting. SPE, 2023. http://dx.doi.org/10.2118/213043-ms.

Full text
Abstract:
Abstract Modeling the drill bit Rate of Penetration (ROP) is crucial for optimizing drilling operations as maximum ROP causes fast drilling, reflecting efficient rig performance and productivity. In this paper, four Ensemble machine learning (ML) algorithms were adopted to reconstruct ROP predictive models: Random Forest (RF), Gradient Boosting (GB), Extreme Gradient Boost (XGB), and Adaptive Boosting (AdaBoost). The research was implemented on well data for the entire stratigraphy column in a giant Southern Iraqi oil field. The drilling operations in the oil field pass through 19 formations (including 4 oil-bearing reservoirs) from Dibdibba to Zubair in a total depth of approximately 3200 m. From the stratigraphic column, various lithology types exist, such as carbonate and clastic with distinct thicknesses that range from (40-440) m. The ROP predictive models were built given 14 operating parameters: Total Vertical Depth (TVD), Weight on Bit (WOB), Rotation per Minute (RPM), Torque, Total RPM, flow rate, Standpipe Pressure (SPP), effective density, bit size, D exponent, Gamma Ray (GR), density, neutron, and caliper, and the discrete lithology distribution. For ROP modeling and validation, a dataset that combines information from three development wells was collected, randomly subsampled, and then subdivided into 85% for training and 15% for validation and testing. The root means square prediction error (RMSE) and coefficient of correlation (R-sq) were used as statistical mismatch quantification tools between the measured and predicted ROP given the test subset. Except for Adaboost, all the other three ML approaches have given acceptable accurate ROP predictions with good matching between the ROP to the measured and predicted for the testing subset in addition to the prediction for each well across the entire depth. This integrated modeling workflow with cross-validation of combining three wells together has resulted in more accurate prediction than using one well as a reference for prediction. In the ROP optimization, determining the optimal set of the 14 operational parameters leads to the fastest penetration rate and most economic drilling. The presented workflow is not only predicting the proper penetration rate but also optimizing the drilling parameters and reducing the drilling cost of future wells. Additionally, the resulting ROP ML-predictive models can be implemented for the prediction of the drilling rate of penetration in other areas of this oil field and also other nearby fields of the similar stratigraphic columns.
APA, Harvard, Vancouver, ISO, and other styles
9

Khan, Mohammad Rasheed, Zeeshan Tariq, Muhammad Ali, and Mobeen Murtaza. "Predicting Interfacial Tension in CO2/Brine Systems: A Data-Driven Approach and Its Implications for Carbon Geostorage." In International Petroleum Technology Conference. IPTC, 2024. http://dx.doi.org/10.2523/iptc-23568-ms.

Full text
Abstract:
Abstract CO2 Interfacial Tension (IFT) and the reservoir rock-fluid interfacial interactions are critical parameters for successful CO2 geological sequestration, where the success relies significantly on the rock-CO2-brine interactions. IFT behaviors during storage dictate the CO2/brine distribution at pore scale and the residual/structural trapping potentials of storage/caprocks. Experimental assessment of CO2-Brine IFT as a function of pressure, temperature, and readily available organic contaminations on rock surfaces is arduous because of high CO2 reactivity and embrittlement damages. Data-driven machine learning (ML) modeling of CO2-brine IFT are less strenuous and more precise. They can be conducted at geo-storage conditions that are complex and hazardous to attain in the laboratory. In this study, we have applied three different machine learning techniques, including Random Forest (RF), XGBoost (XGB), and Adaptive Gradient Boosting (AGB), to predict the interfacial tension of the CO2 in brine system. The performance of the ML models was assessed through various assessment tests, such as cross-plots, average absolute percentage error (AAPE), root mean square error (RMSE), and coefficient of determination (R2). The outcomes of the predictions indicated that the XGB outperformed the RF, and AdaBoost. The XGB yielded remarkably low error rates. With optimal settings, the output was predicted with 97% accuracy. The proposed methodology can minimize the experimental cost of measuring rheological parameters and serve as a quick assessment tool.
APA, Harvard, Vancouver, ISO, and other styles
10

Jalo, Hoor, Andrei Borg, Elsa Thoreström, et al. "Early Characterization of Stroke Using Video Analysis and Machine Learning." In AHFE 2023 Hawaii Edition. AHFE International, 2023. http://dx.doi.org/10.54941/ahfe1004359.

Full text
Abstract:
Stroke is one of the leading causes of death and disability worldwide and requires an immediate attention as the longer the patient is left untreated, the more sever its outcomes are. Enhancing access to optimal treatment and reducing mortality rates require improving the accuracy of stroke characterization methods in prehospital settings. This study explores how video analysis and machine learning (ML) can be leveraged to identify stroke symptoms on the National Institute of Health Stroke Scale (NIHSS), with the goal of facilitating the prehospital management of patients with suspected stroke. A total of 888 videos were captured from the research group members, who mimicked stroke symptoms including facial palsy, leg and arm paresis, ataxia and dysarthria, following the criteria of the NIHSS. Multiple algorithms, utilized in earlier studies, were examined to predict these symptoms, and their performance was assessed using accuracy, sensitivity and specificity. The best method for detecting facial palsy was found using Histogram of Oriented Gradients (HOG) features in conjunction with Adaptive Boosting (AdaBoost), achieving an accuracy, sensitivity and specificity values of 97.8%, 98.0% and 97.0%, respectively. The identification of arm paresis reached 100% on all metrics using a combination of MediaPipe and SVM. For leg paresis, all algorithms had poor detection rates. The outcome for ataxia for both limbs varied. Google Cloud Speech-to-Text was used to detect dysarthria and reached 100% on all evaluation metrics. These findings suggest that video analysis and ML have the potential to assist early stroke diagnosis, but further research is needed to validate this.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!