Academic literature on the topic 'Min-max normalization'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Min-max normalization.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Min-max normalization"

1

Sinsomboonthong, Saichon. "Performance Comparison of New Adjusted Min-Max with Decimal Scaling and Statistical Column Normalization Methods for Artificial Neural Network Classification." International Journal of Mathematics and Mathematical Sciences 2022 (April 22, 2022): 1–9. http://dx.doi.org/10.1155/2022/3584406.

Full text
Abstract:
In this research, the normalization performance of the proposed adjusted min-max methods was compared to the normalization performance of statistical column, decimal scaling, adjusted decimal scaling, and min-max methods, in terms of accuracy and mean square error of the final classification outcomes. The evaluation process employed an artificial neural network classification on a large variety of widely used datasets. The best method was min-max normalization, providing 84.0187% average ranking of accuracy and 0.1097 average ranking of mean square error across all six datasets. However, the proposed adjusted-2 min-max normalization achieved a higher accuracy and a lower mean square error than min-max normalization on each of the following datasets: white wine quality, Pima Indians diabetes, vertical column, and Indian liver disease datasets. For example, the proposed adjusted-2 min-max normalization on white wine quality dataset achieved 100% accuracy and 0.00000282 mean square error. To conclude, for some classification applications on one of these specific datasets, the proposed adjusted-2 min-max normalization should be used over the other tested normalization methods because it performed better.
APA, Harvard, Vancouver, ISO, and other styles
2

Firmansyah, Muhammad Raihan. "Stroke Classification Comparison with KNN through Standardization and Normalization Techniques." Advance Sustainable Science, Engineering and Technology 6, no. 1 (2024): 02401012. http://dx.doi.org/10.26877/asset.v6i1.17685.

Full text
Abstract:
This study explores the impact of z-score standardization and min-max normalization on K-Nearest Neighbors (KNN) classification for strokes. Focused on managing diverse scales in health attributes within the stroke dataset, the research aims to improve classification model accuracy and reliability. Preprocessing involves z-score standardization, min-max normalization, and no data scaling. The KNN model is trained and evaluated using various methods. Results reveal comparable performance between z-score standardization and min-max normalization, with slight variations across data split ratios. Demonstrating the importance of data scaling, both z-score and min-max achieve 95.07% accuracy. Notably, normalization averages a higher accuracy (94.25%) than standardization (94.21%), highlighting the critical role of data scaling for robust machine learning performance and informed health decisions.
APA, Harvard, Vancouver, ISO, and other styles
3

Anusas-Amornkul, Tanapat, and Naphat Bussabong. "Normalization Technique and Weight Adjustment Analysis for Keystroke Vector Dissimilarity Authentication." WSEAS TRANSACTIONS ON SYSTEMS 23 (September 23, 2024): 206–14. http://dx.doi.org/10.37394/23202.2024.23.23.

Full text
Abstract:
A keystroke dynamics authentication uses keystroke rhythm for each user on a keyboard to verify a real user. The idea is that each user has a unique keystroke rhythm such that it can be determined the identity of a user. To verify a user, a keystroke vector dissimilarity technique was proposed to use keystroke features as a vector and calculate a weight using SoftMax+1 to overcome the Euclidean distance problem. However, the weight has yet to be analyzed in detail. Therefore, this paper aims to find a normalization technique and a weight adjustment to enhance the accuracy of the keystroke vector dissimilarity technique. The normalization techniques and activation functions analyzed in this study are Euclidean norm, Mean normalization, Min-max normalization, Z-score normalization, SoftMax function, and ReLU function. The weight adjustment varies from w+1000 to 1000-w. The results show that the Mean and Min-max normalizations with 10-w as a weight gave the same results at 96.97% accuracy and 3.03% error, which are better than the previous work.
APA, Harvard, Vancouver, ISO, and other styles
4

Allorerung, Petronilia Palinggik, Angdy Erna, Muhammad Bagussahrir, and Samsu Alam. "Analisis Performa Normalisasi Data untuk Klasifikasi K-Nearest Neighbor pada Dataset Penyakit." JISKA (Jurnal Informatika Sunan Kalijaga) 9, no. 3 (2024): 178–91. http://dx.doi.org/10.14421/jiska.2024.9.3.178-191.

Full text
Abstract:
This study investigates four normalization methods (Min-Max, Z-Score, Decimal Scaling, MaxAbs) across prostate, kidney, and heart disease datasets for K-Nearest Neighbor (K-NN) classification. Imbalanced feature scales can hinder K-NN performance, making normalization crucial. Results show that Decimal Scaling achieves 90.00% accuracy in prostate cancer, while Min-Max and Z-Score yield 97.50% in kidney disease. MaxAbs performs well with 96.25% accuracy in kidney disease. In heart disease, Min-Max and MaxAbs attain accuracies of 82.93% and 81.95%, respectively. These findings suggest Decimal Scaling suits datasets with few instances, limited features, and normal distribution. Min-Max and MaxAbs are better for datasets with numerous instances and non-normal distribution. Z-Score fits datasets with a wide range of feature numbers and near-normal distribution. This study aids in selecting the appropriate normalization method based on dataset characteristics to enhance K-NN classification accuracy in disease diagnosis. The experiments involve datasets with different attributes, continuous and categorical data, and binary classification. Data conditions such as the number of instances, the number of features, and data distribution affect the performance of normalization and classification.
APA, Harvard, Vancouver, ISO, and other styles
5

Muhammad Ali, Peshawa J. "Investigating the Impact of Min-Max Data Normalization on the Regression Performance of K-Nearest Neighbor with Different Similarity Measurements." ARO-THE SCIENTIFIC JOURNAL OF KOYA UNIVERSITY 10, no. 1 (2022): 85–91. http://dx.doi.org/10.14500/aro.10955.

Full text
Abstract:
K-nearest neighbor (KNN) is a lazy supervised learning algorithm, which depends on computing the similarity between the target and the closest neighbor(s). On the other hand, min-max normalization has been reported as a useful method for eliminating the impact of inconsistent ranges among attributes on the efficiency of some machine learning models. The impact of min-max normalization on the performance of KNN models is still not clear, and it needs more investigation. Therefore, this research examines the impacts of the min-max normalization method on the regression performance of KNN models utilizing eight different similarity measures, which are City block, Euclidean, Chebychev, Cosine, Correlation, Hamming, Jaccard, and Mahalanobis. Five benchmark datasets have been used to test the accuracy of the KNN models with the original dataset and the normalized dataset. Mean squared error (MSE) has been utilized as a performance indicator to compare the results. It’s been concluded that the impact of min-max normalization on the KNN models utilizing City block, Euclidean, Chebychev, Cosine, and Correlation depends on the nature of the dataset itself, therefore, testing models on both original and normalized datasets are recommended. The performance of KNN models utilizing Hamming, Jaccard, and Mahalanobis makes no difference by adopting min-max normalization because of their ratio nature, and dataset covariance involvement in the similarity calculations. Results showed that Mahalanobis outperformed the other seven similarity measures. This research is better than its peers in terms of reliability, and quality because it depended on testing different datasets from different application fields.
APA, Harvard, Vancouver, ISO, and other styles
6

Wahyusari, Retno, Sunardi Sunardi, and Abdul Fadlil. "Comparison of Machine Learning Methods for Predicting Electrical Energy Consumption." Aviation Electronics, Information Technology, Telecommunications, Electricals, and Controls (AVITEC) 7, no. 1 (2025): 11. https://doi.org/10.28989/avitec.v7i1.2722.

Full text
Abstract:
This research investigates how to accurately predict electrical energy consumption to address growing global energy demands. The study employs three Machine Learning (ML) models: k-Nearest Neighbors (KNN), Random Forest (RF), and CatBoost. To enhance prediction accuracy, the researchers included a data pre-processing step using min-max normalization. The analysis utilized a dataset containing 52,416 records of power consumption from Tetouan City. The dataset was divided into training and testing sets using different ratios (90:10, 80:20, 50:50) to evaluate model performance. Root Mean Square Error (RMSE) and Mean Absolute Percentage Error (MAPE) were used to assess prediction accuracy. Min-max normalization significantly improved KNN's performance (reduced RMSE and MAPE). RF achieved similar accuracy with and without normalization. CatBoost also demonstrated stable performance regardless of normalization. Data pre-processing, specifically min-max normalization, is crucial for improving the accuracy of distance-based algorithms like KNN. Decision tree-based algorithms like RF and CatBoost are less sensitive to data normalization. These findings emphasize the importance of selecting appropriate pre-processing techniques to optimize energy consumption prediction models, which can contribute to better energy management strategies.
APA, Harvard, Vancouver, ISO, and other styles
7

Prasetyowati, Sri Arttini Dwi, Munaf Ismail, and Badieah Badieah. "Implementation of Least Mean Square Adaptive Algorithm on Covid-19 Prediction." JUITA: Jurnal Informatika 10, no. 1 (2022): 139. http://dx.doi.org/10.30595/juita.v10i1.11963.

Full text
Abstract:
This study used Corona Virus Disease-19 (Covid-19) data in Indonesia from June to August 2021, consisting of data on people who were infected or positive Covid-19, recovered from Covid-19, and passed away from Covid-19. The data were processed using the adaptive LMS algorithm directly without pre-processing cause calculation errors, because covid-19 data was not balanced. Z-score and min-max normalization were chosen as pre-processing methods. After that, the prediction process can be carried out using the LMS adaptive method. The analysis was done by observing the error prediction that occurred every month per case. The results showed that data pre-processing using min-max normalization was better than with Z-score normalization because the error prediction for pre-processing using min-max and z-score were 18% and 47%, respectively.
APA, Harvard, Vancouver, ISO, and other styles
8

Ambarwari, Agus, Qadhli Jafar Adrian, and Yeni Herdiyeni. "Analysis of the Effect of Data Scaling on the Performance of the Machine Learning Algorithm for Plant Identification." Jurnal RESTI (Rekayasa Sistem dan Teknologi Informasi) 4, no. 1 (2020): 117–22. http://dx.doi.org/10.29207/resti.v4i1.1517.

Full text
Abstract:
Data scaling has an important role in preprocessing data that has an impact on the performance of machine learning algorithms. This study aims to analyze the effect of min-max normalization techniques and standardization (zero-mean normalization) on the performance of machine learning algorithms. The stages carried out in this study included data normalization on the data of leaf venation features. The results of the normalized dataset, then tested to four machine learning algorithms include KNN, Naïve Bayesian, ANN, SVM with RBF kernels and linear kernels. The analysis was carried out on the results of model evaluations using 10-fold cross-validation, and validation using test data. The results obtained show that Naïve Bayesian has the most stable performance against the use of min-max normalization techniques as well as standardization. The KNN algorithm is quite stable compared to SVM and ANN. However, the combination of the min-max normalization technique with SVM that uses the RBF kernel can provide the best performance results. On the other hand, SVM with a linear kernel, the best performance is obtained when applying standardization techniques (zero-mean normalization). While the ANN algorithm, it is necessary to do a number of trials to find out the best data normalization techniques that match the algorithm.
APA, Harvard, Vancouver, ISO, and other styles
9

Ahmed, Haval A., Peshawa J. Muhammad Ali, Abdulbasit K. Faeq, and Saman M. Abdullah. "An Investigation on Disparity Responds of Machine Learning Algorithms to Data Normalization Method." ARO-THE SCIENTIFIC JOURNAL OF KOYA UNIVERSITY 10, no. 2 (2022): 29–37. http://dx.doi.org/10.14500/aro.10970.

Full text
Abstract:
Data normalization can be useful in eliminating the effect of inconsistent ranges in some machine learning (ML) techniques and in speeding up the optimization process in others. Many studies apply different methods of data normalization with an aim to reduce or eliminate the impact of data variance on the accuracy rate of ML-based models. However, the significance of this impact aligning with the mathematical concept of the ML algorithms still needs more investigation and tests. To identify that, this work proposes an investigation methodology involving three different ML algorithms, which are support vector machine (SVM), artificial neural network (ANN), and Euclidean-based K-nearest neighbor (E-KNN). Throughout this work, five different datasets have been utilized, and each has been taken from different application fields with different statistical properties. Although there are many data normalization methods available, this work focuses on the min-max method, because it actively eliminates the effect of inconsistent ranges of the datasets. Moreover, other factors that are challenging the process of min-max normalization, such as including or excluding outliers or the least significant feature, have also been considered in this work. The finding of this work shows that each ML technique responds differently to the min-max normalization. The performance of SVM models has been improved, while no significant improvement happened to the performance of ANN models. It is been concluded that the performance of E-KNN models may improve or degrade with the min-max normalization, and it depends on the statistical properties of the dataset.
APA, Harvard, Vancouver, ISO, and other styles
10

Gde Agung Brahmana Suryanegara, Adiwijaya, and Mahendra Dwifebri Purbolaksono. "Peningkatan Hasil Klasifikasi pada Algoritma Random Forest untuk Deteksi Pasien Penderita Diabetes Menggunakan Metode Normalisasi." Jurnal RESTI (Rekayasa Sistem dan Teknologi Informasi) 5, no. 1 (2021): 114–22. http://dx.doi.org/10.29207/resti.v5i1.2880.

Full text
Abstract:
Diabetes is a disease caused by high blood sugar in the body or beyond normal limits. Diabetics in Indonesia have experienced a significant increase, Basic Health Research states that diabetics in Indonesia were 6.9% to 8.5% increased from 2013 to 2018 with an estimated number of sufferers more than 16 million people. Therefore, it is necessary to have a technology that can detect diabetes with good performance, accurate level of analysis, so that diabetes can be treated early to reduce the number of sufferers, disabilities, and deaths. The different scale values for each attribute in Gula Karya Medika’s data can complicate the classification process, for this reason the researcher uses two data normalization methods, namely min-max normalization, z-score normalization, and a method without data normalization with Random Forest (RF) as a classification method. Random Forest (RF) as a classification method has been tested in several previous studies. Moreover, this method is able to produce good performance with high accuracy. Based on the research results, the best accuracy is model 1 (Min-max normalization-RF) of 95.45%, followed by model 2 (Z-score normalization-RF) of 95%, and model 3 (without data normalization-RF) of 92%. From these results, it can be concluded that model 1 (Min-max normalization-RF) is better than the other two data normalization models and is able to increase the performance of classification Random Forest by 95.45%.
APA, Harvard, Vancouver, ISO, and other styles
More sources

Book chapters on the topic "Min-max normalization"

1

Kiran, Ajmeera, and D. Vasumathi. "Data Mining: Min–Max Normalization Based Data Perturbation Technique for Privacy Preservation." In Proceedings of the Third International Conference on Computational Intelligence and Informatics. Springer Singapore, 2020. http://dx.doi.org/10.1007/978-981-15-1480-7_66.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Ramalakshmi, K., S. P. Santhoshkumar, L. Krishna Kumari, et al. "Improved Brain Tumor Segmentation Using Min-Max Normalization in a U-Net Architecture." In Integrative Machine Learning and Optimization Algorithms for Disease Prediction. IGI Global Scientific Publishing, 2025. https://doi.org/10.4018/979-8-3373-1087-9.ch010.

Full text
Abstract:
The procedure of manually identifying brain tumors from MRI scans is difficult and time-consuming. It takes a lot of time and work, which increases the possibility of mistakes. Brain tumor segmentation needs to be accurate and consistent in order to be used for cancer identification, treatment scheduling, and outcome estimation. Effective brain tumor segmentation is vital for early brain tumor detection is necessaryfor improving patient prognosis and treatment alternatives. Even with the increasing usage of for brain imaging and the development of AI detection techniques, creating a reliable and effective model to identify and classify cancers from MRI images is still a difficult task. In order to tackle this issue, we suggested an improved UNet based segmentation method for brain tumor partitioning.For brain tumor partitioning, our newly created deep learning-based model performs better than previously developed pre-trained models. The outcomes show that, a dice coefficient of 88.7% and an accuracy of 98.7%, our segmentation model had the best performance.
APA, Harvard, Vancouver, ISO, and other styles
3

Lopez, Kyra Mikaela M., and Ma Sheila A. Magboo. "A Clinical Decision Support Tool to Detect Invasive Ductal Carcinoma in Histopathological Images Using Support Vector Machines, Naïve-Bayes, and K-Nearest Neighbor Classifiers." In Machine Learning and Artificial Intelligence. IOS Press, 2020. http://dx.doi.org/10.3233/faia200765.

Full text
Abstract:
This study aims to describe a model that will apply image processing and traditional machine learning techniques specifically Support Vector Machines, Naïve-Bayes, and k-Nearest Neighbors to identify whether or not a given breast histopathological image has Invasive Ductal Carcinoma (IDC). The dataset consisted of 54,811 breast cancer image patches of size 50px x 50px, consisting of 39,148 IDC negative and 15,663 IDC positive. Feature extraction was accomplished using Oriented FAST and Rotated BRIEF (ORB) descriptors. Feature scaling was performed using Min-Max Normalization while K-Means Clustering on the ORB descriptors was used to generate the visual codebook. Automatic hyperparameter tuning using Grid Search Cross Validation was implemented although it can also accept user supplied hyperparameter values for SVM, Naïve Bayes, and K-NN models should the user want to do experimentation. Aside from computing for accuracy, the AUPRC and MCC metrics were used to address the dataset imbalance. The results showed that SVM has the best overall performance, obtaining accuracy = 0.7490, AUPRC = 0.5536, and MCC = 0.2924.
APA, Harvard, Vancouver, ISO, and other styles
4

Kaushik, Rekha, Pritam Goyal, and Atharv Pandey. "NeuroVoice: Leveraging Neural Networks for Precise Gender Classification in Audio." In Computational Intelligence and Machine Learning. Soft Computing Research Society, 2024. https://doi.org/10.56155/978-81-975670-5-6-5.

Full text
Abstract:
In this paper, the model for gender recognition is developed with voice samples using various machine learning algorithms and acoustic parameters. It is divided at the beginning into the training and test data of the dataset. There then follows a number of key steps and techniques as part of the process that improves the performance of this model. The paper focuses on a holistic approach toward gender classification from audio data through various techniques of data preprocessing, augmentation, feature scaling, model development, and their performance evaluation. First, it encodes the class label (male/female) into a numerical format through label encoding. Then, it extracts critical features like MFCC, Chroma Features, Spectral Contrast, and Pitch to extract the most essential characteristics of the audio. Data augmentation with SMOTE avoids bias in the dataset by creating artificial samples. Features are scaled with Min-Max Scaling to enhance model convergence and performance. Several advanced neural network architectures like MLP with Batch Normalization and Dropout techniques have also been considered. Stratified K-Fold Cross-Validation ensures robustness and avoidance of bias in the evaluation. In this study, each model has an evaluation including performance metrics, such as accuracy, precision, recall, F1-score, confusion matrix, and classification report. Remarkably, this model has an accuracy of 99.7% against the test dataset, improving both in total accuracy and robustness. The findings could be of importance in telecommunication, human-computer interaction, and security systems where accurate gender recognition from voice is required.
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Min-max normalization"

1

Waema, Damaris, Waweru Mwangi, and Petronilla Muriithi. "A Min-Max Based Data Normalization and Maximum Pooling Approach for Improved Maize Leaf Disease Detection." In 2025 IST-Africa Conference (IST-Africa). IEEE, 2025. https://doi.org/10.23919/ist-africa67297.2025.11060470.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Patel, Chetan, Aarsh Pandey, Rajesh Wadhvani, and Deepali Patil. "Forecasting Nonstationary Wind Data Using Adaptive Min-Max Normalization." In 2022 1st International Conference on Sustainable Technology for Power and Energy Systems (STPES). IEEE, 2022. http://dx.doi.org/10.1109/stpes54845.2022.10006473.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Gajera, Vatsal, Shubham, Rishabh Gupta, and Prasanta K. Jana. "An effective Multi-Objective task scheduling algorithm using Min-Max normalization in cloud computing." In 2016 2nd International Conference on Applied and Theoretical Computing and Communication Technology (iCATccT). IEEE, 2016. http://dx.doi.org/10.1109/icatcct.2016.7912111.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Herwanto, Heru Wahyu, Anik Nur Handayani, Aji Prasetya Wibawa, Katya Lindi Chandrika, and Kohei Arai. "Comparison of Min-Max, Z-Score and Decimal Scaling Normalization for Zoning Feature Extraction on Javanese Character Recognition." In 2021 7th International Conference on Electrical, Electronics and Information Engineering (ICEEIE). IEEE, 2021. http://dx.doi.org/10.1109/iceeie52663.2021.9616665.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Sholeh, Muhammad, and Erna Kumalasari Nurnawati. "Comparison of Z-score, min-max, and no normalization methods using support vector machine algorithm to predict student’s timely graduation." In THE 3RD INTERNATIONAL CONFERENCE ON NATURAL SCIENCES, MATHEMATICS, APPLICATIONS, RESEARCH, AND TECHNOLOGY (ICON-SMART2022): Mathematical Physics and Biotechnology for Education, Energy Efficiency, and Marine Industries. AIP Publishing, 2024. http://dx.doi.org/10.1063/5.0202505.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Assiri, Mohammed. "Enhancing Android Security Through Artificial Intelligence: A Hyperparameter-Tuned Deep Learning Approach for Robust Software Vulnerability Detection." In 13th International Conference on Human Interaction & Emerging Technologies: Artificial Intelligence & Future Applications. AHFE International, 2025. https://doi.org/10.54941/ahfe1005919.

Full text
Abstract:
Detecting software vulnerabilities is essential for cybersecurity, particularly in Android systems, which are widely used and vulnerable due to their open-source nature. Conventional signature-based malware detection methods are inadequate against sophisticated and evolving threats. This paper introduces a Hyperparameter-Tuned Deep Learning Approach for Robust Software Vulnerability Detection (HPTDLA-RSVD) aimed at enhancing Android security through an optimized deep learning model. The HPTDLA-RSVD methodology encompasses min-max data normalization, feature selection using the Ant Lion Optimizer (ALO), classification via a Deep Belief Network (DBN), and hyperparameter optimization with the Slime Mould Algorithm (SMA). Experimental evaluations on a benchmark dataset reveal that HPTDLA-RSVD surpasses existing techniques across multiple performance metrics, confirming its efficacy in identifying and mitigating software vulnerabilities on Android platforms.
APA, Harvard, Vancouver, ISO, and other styles
7

Ramanayake, R. M. D. T., and Chethika Abeynayake. "A COMPARATIVE STUDY OF CRITICAL SUCCESS CRITERIA ON SUSTAINABLE HOUSING; A CASE OF - LOW INCOME HOUSING, SRI LANKA." In Beyond sustainability reflections across spaces. Faculty of Architecture Research Unit, 2021. http://dx.doi.org/10.31705/faru.2021.1.

Full text
Abstract:
Sustainable housing is a popular topic with regard to the SDG, sustainable communities and Sustainable cities. Although different researches have come up with regard to different CSC of specific contexts there are very limited studies on CSC on Sustainable low-income housing. This research aims to compare the CSC on Sustainable low-income Housing in designing stage in Sri Lankan Context. 18 CSC were derived from comprehensive literature review and re-examined through the 27 professionals and ranked from community on three locations. Relative Importance Index- RII, Min Max Normalization and Gap analyses were employed in the ranking process of Critical Success Criteria. The highest importance has been ranked with Efficiency use of water and energy, Users Satisfaction and Quality of Housing while least importance is ranked with Maintainability, Public Consultation and community participation and cater for Disables and by Literature, Experts and Community respectively. Anyway, Public Consultation and Community Participation, newly derived CSC which is highly ranked among community is to be concentrated among the professionals for the attention and applications in practices. The findings of the research would support to the designers, architectures, planners specialized in this field to ensure the successful delivery of sustainable housing.
APA, Harvard, Vancouver, ISO, and other styles
8

Kapuge, A. B. A. K. V. S., J. M. S. J. Bandara, and N. Jayasooriya. "Criteria for Assessing the Effectiveness of a Non Real Time Coordinated Cluster of Signalized Intersections." In 5th International Conference on Advances in Highway Engineering & Transportation Systems 2024, edited by H. R. Pasindu. Transport Engineering Division, Department of Civil Engineering, University of Moratuwa, 2024. https://doi.org/10.31705/icahets.2024.2.

Full text
Abstract:
Given the limited resources available for installing advanced signal controllers, many researchers and professionals believe that well-designed coordinated fixed-time signal control, combined with well-defined corridor coordination, is a cost-effective option. In this study, two closely spaced intersections were considered for analysis. It was identified that a subsystem consists of spatial elements (the link road that connects two signalized intersections) and temporal elements (the relative offset between the signals of the two intersections). Throughput, travel time, delay, queue length, and flow efficiency were considered basic performance measurement parameters. The traffic flow analysis was conducted using the Vissim model, with calibration parameters adjusted to suit Sri Lankan conditions based on previous research. Twenty demand models were evaluated, with volume-to-capacity (v/c) ratios ranging from 0.2 to 0.9 across the two intersections. The traffic flow was analyzed under three scenarios: Scenario 1 consisted of 60% through traffic, 20% left-turn traffic, and 20% right-turn traffic. In Scenario 2, the traffic distribution was 60% through traffic, 30% left-turn traffic, and 10% right-turn traffic. Scenario 3 featured 60% through traffic, 10% left-turn traffic, and 30% right-turn traffic. The Vissim model's link road geometry was kept at 500 meters, and both junctions were identical. The east-west direction was treated as the coordinated direction with three lanes, while the north-south cross streets had two lanes. Base saturation flows and offset optimization were calculated using Akcelik’s (1981) method, and cycle timing was determined based on the Webster formula. The study extended its analysis to compare unsynchronized (Case 1) and synchronized (Case 2) models. Initially, travel time was plotted against throughput. Based on the Elbow method, three clusters were identified as the optimal cluster number and then K-means clustering was performed. Generally, Cluster 1 had low flow and moderate traffic, Cluster 2 had moderate to high flow but moderate travel time, and Cluster 3 had moderate flow and high travel time. When the mean values for each cluster were compared in the unsynchronized and synchronized models, signal synchronization improved all key metrics in each cluster, except for queue length on the link road in Cluster 3, and throughput and flow efficiency in Cluster 2. Based on the Independent-Samples T-Test p-value, statistically significant improvements were found in travel time, with an 8.74% improvement in Cluster 2 and a 14.34% improvement in Cluster 3. Delay improvements were 29% and 30% in Clusters 2 and 3, respectively. However, there was a 35.17% increase in queue length on the link road, suggesting that while flow efficiency and delays improved, congestion remained an issue. As a further performance check, the Min-Max Normalization and Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS) methods were used. Both methods ranked Cluster 3 as the highest performer, despite the increase in queue length. Finally, considering the Level of Service (LOS) criteria from the Highway Capacity Manual, Cluster 3 showed an improvement from LOS E and D to LOS C, whereas Cluster 2 showed an improvement from LOS C and B to LOS B and Cluster 1 remained at LOS B and A. The most notable result from this research is the identification of Cluster 3 as the best-performing cluster across multiple scenarios, while Cluster 2 provided moderate improvement and Cluster 1 showed no significant improvement. Therefore, this study suggests that to gain the benefits of synchronization, the intersections considered should initially fall within the range of Cluster 3 or Cluster 2. Further research is required to explore how intersection geometry and coordination direction affect synchronization.
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Min-max normalization"

1

Tayeb, Shahab. Taming the Data in the Internet of Vehicles. Mineta Transportation Institute, 2022. http://dx.doi.org/10.31979/mti.2022.2014.

Full text
Abstract:
As an emerging field, the Internet of Vehicles (IoV) has a myriad of security vulnerabilities that must be addressed to protect system integrity. To stay ahead of novel attacks, cybersecurity professionals are developing new software and systems using machine learning techniques. Neural network architectures improve such systems, including Intrusion Detection System (IDSs), by implementing anomaly detection, which differentiates benign data packets from malicious ones. For an IDS to best predict anomalies, the model is trained on data that is typically pre-processed through normalization and feature selection/reduction. These pre-processing techniques play an important role in training a neural network to optimize its performance. This research studies the impact of applying normalization techniques as a pre-processing step to learning, as used by the IDSs. The impacts of pre-processing techniques play an important role in training neural networks to optimize its performance. This report proposes a Deep Neural Network (DNN) model with two hidden layers for IDS architecture and compares two commonly used normalization pre-processing techniques. Our findings are evaluated using accuracy, Area Under Curve (AUC), Receiver Operator Characteristic (ROC), F-1 Score, and loss. The experimentations demonstrate that Z-Score outperforms no-normalization and the use of Min-Max normalization.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!