Literatura científica selecionada sobre o tema "Hyperparameter search"

Crie uma referência precisa em APA, MLA, Chicago, Harvard, e outros estilos

Selecione um tipo de fonte:

Consulte a lista de atuais artigos, livros, teses, anais de congressos e outras fontes científicas relevantes para o tema "Hyperparameter search".

Ao lado de cada fonte na lista de referências, há um botão "Adicionar à bibliografia". Clique e geraremos automaticamente a citação bibliográfica do trabalho escolhido no estilo de citação de que você precisa: APA, MLA, Harvard, Chicago, Vancouver, etc.

Você também pode baixar o texto completo da publicação científica em formato .pdf e ler o resumo do trabalho online se estiver presente nos metadados.

Artigos de revistas sobre o assunto "Hyperparameter search":

1

Florea, Adrian-Catalin, e Razvan Andonie. "Weighted Random Search for Hyperparameter Optimization". International Journal of Computers Communications & Control 14, n.º 2 (14 de abril de 2019): 154–69. http://dx.doi.org/10.15837/ijccc.2019.2.3514.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
We introduce an improved version of Random Search (RS), used here for hyperparameter optimization of machine learning algorithms. Unlike the standard RS, which generates for each trial new values for all hyperparameters, we generate new values for each hyperparameter with a probability of change. The intuition behind our approach is that a value that already triggered a good result is a good candidate for the next step, and should be tested in new combinations of hyperparameter values. Within the same computational budget, our method yields better results than the standard RS. Our theoretical results prove this statement. We test our method on a variation of one of the most commonly used objective function for this class of problems (the Grievank function) and for the hyperparameter optimization of a deep learning CNN architecture. Our results can be generalized to any optimization problem dened on a discrete domain.
2

Ghawi, Raji, e Jürgen Pfeffer. "Efficient Hyperparameter Tuning with Grid Search for Text Categorization using kNN Approach with BM25 Similarity". Open Computer Science 9, n.º 1 (8 de agosto de 2019): 160–80. http://dx.doi.org/10.1515/comp-2019-0011.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
AbstractIn machine learning, hyperparameter tuning is the problem of choosing a set of optimal hyperparameters for a learning algorithm. Several approaches have been widely adopted for hyperparameter tuning, which is typically a time consuming process. We propose an efficient technique to speed up the process of hyperparameter tuning with Grid Search. We applied this technique on text categorization using kNN algorithm with BM25 similarity, where three hyperparameters need to be tuned. Our experiments show that our proposed technique is at least an order of magnitude faster than conventional tuning.
3

Yang, Eun-Suk, Jong Dae Kim, Chan-Young Park, Hye-Jeong Song e Yu-Seop Kim. "Hyperparameter tuning for hidden unit conditional random fields". Engineering Computations 34, n.º 6 (7 de agosto de 2017): 2054–62. http://dx.doi.org/10.1108/ec-11-2015-0350.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Purpose In this paper, the problem of a nonlinear model – specifically the hidden unit conditional random fields (HUCRFs) model, which has binary stochastic hidden units between the data and the labels – exhibiting unstable performance depending on the hyperparameter under consideration. Design/methodology/approach There are three main optimization search methods for hyperparameter tuning: manual search, grid search and random search. This study shows that HUCRFs’ unstable performance depends on the hyperparameter values used and its performance is based on tuning that draws on grid and random searches. All experiments conducted used the n-gram features – specifically, unigram, bigram, and trigram. Findings Naturally, selecting a list of hyperparameter values based on a researchers’ experience to find a set in which the best performance is exhibited is better than finding it from a probability distribution. Realistically, however, it is impossible to calculate using the parameters in all combinations. The present research indicates that the random search method has a better performance compared with the grid search method while requiring shorter computation time and a reduced cost. Originality/value In this paper, the issues affecting the performance of HUCRF, a nonlinear model with performance that varies depending on the hyperparameters, but performs better than CRF, has been examined.
4

Wen, Long, Xingchen Ye e Liang Gao. "A new automatic machine learning based hyperparameter optimization for workpiece quality prediction". Measurement and Control 53, n.º 7-8 (21 de julho de 2020): 1088–98. http://dx.doi.org/10.1177/0020294020932347.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Workpiece quality prediction is very important in modern manufacturing industry. However, traditional machine learning methods are very sensitive to their hyperparameters, making the tuning of the machine learning methods essential to improve the prediction performance. Hyperparameter optimization (HPO) approaches are applied attempting to tune hyperparameters, such as grid search and random search. However, the hyperparameters space for workpiece quality prediction model is high dimension and it consists with continuous, combinational and conditional types of hyperparameters, which is difficult to be tuned. In this article, a new automatic machine learning based HPO, named adaptive Tree Pazen Estimator (ATPE), is proposed for workpiece quality prediction in high dimension. In the proposed method, it can iteratively search the best combination of hyperparameters in the automatic way. During the warm-up process for ATPE, it can adaptively adjust the hyperparameter interval to guide the search. The proposed ATPE is tested on sparse stack autoencoder based MNIST and XGBoost based WorkpieceQuality dataset, and the results show that ATPE provides the state-of-the-art performances in high-dimensional space and can search the hyperparameters in reasonable range by comparing with Tree Pazen Estimator, annealing, and random search, showing its potential in the field of workpiece quality prediction.
5

Hinz, Tobias, Nicolás Navarro-Guerrero, Sven Magg e Stefan Wermter. "Speeding up the Hyperparameter Optimization of Deep Convolutional Neural Networks". International Journal of Computational Intelligence and Applications 17, n.º 02 (junho de 2018): 1850008. http://dx.doi.org/10.1142/s1469026818500086.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Most learning algorithms require the practitioner to manually set the values of many hyperparameters before the learning process can begin. However, with modern algorithms, the evaluation of a given hyperparameter setting can take a considerable amount of time and the search space is often very high-dimensional. We suggest using a lower-dimensional representation of the original data to quickly identify promising areas in the hyperparameter space. This information can then be used to initialize the optimization algorithm for the original, higher-dimensional data. We compare this approach with the standard procedure of optimizing the hyperparameters only on the original input. We perform experiments with various state-of-the-art hyperparameter optimization algorithms such as random search, the tree of parzen estimators (TPEs), sequential model-based algorithm configuration (SMAC), and a genetic algorithm (GA). Our experiments indicate that it is possible to speed up the optimization process by using lower-dimensional data representations at the beginning, while increasing the dimensionality of the input later in the optimization process. This is independent of the underlying optimization procedure, making the approach promising for many existing hyperparameter optimization algorithms.
6

Yang, Zeshi, e Zhiqi Yin. "Efficient Hyperparameter Optimization for Physics-based Character Animation". Proceedings of the ACM on Computer Graphics and Interactive Techniques 4, n.º 1 (26 de abril de 2021): 1–19. http://dx.doi.org/10.1145/3451254.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Physics-based character animation has seen significant advances in recent years with the adoption of Deep Reinforcement Learning (DRL). However, DRL-based learning methods are usually computationally expensive and their performance crucially depends on the choice of hyperparameters. Tuning hyperparameters for these methods often requires repetitive training of control policies, which is even more computationally prohibitive. In this work, we propose a novel Curriculum-based Multi-Fidelity Bayesian Optimization framework (CMFBO) for efficient hyperparameter optimization of DRL-based character control systems. Using curriculum-based task difficulty as fidelity criterion, our method improves searching efficiency by gradually pruning search space through evaluation on easier motor skill tasks. We evaluate our method on two physics-based character control tasks: character morphology optimization and hyperparameter tuning of DeepMimic. Our algorithm significantly outperforms state-of-the-art hyperparameter optimization methods applicable for physics-based character animation. In particular, we show that hyperparameters optimized through our algorithm result in at least 5x efficiency gain comparing to author-released settings in DeepMimic.
7

Han, Junjie, Cedric Gondro e Juan Steibel. "98 Using differential evolution to improve predictive accuracy of deep learning models applied to pig production data". Journal of Animal Science 98, Supplement_3 (2 de novembro de 2020): 27. http://dx.doi.org/10.1093/jas/skaa054.048.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Abstract Deep learning (DL) is being used for prediction in precision livestock farming and in genomic prediction. However, optimizing hyperparameters in DL models is critical for their predictive performance. Grid search is the traditional approach to select hyperparameters in DL, but it requires exhaustive search over the parameter space. We propose hyperparameter selection using differential evolution (DE), which is a heuristic algorithm that does not require exhaustive search. The goal of this study was to design and apply DE to optimize hyperparameters of DL models for genomic prediction and image analysis in pig production systems. One dataset consisted of 910 pigs genotyped with 28,916 SNP markers to predict their post-mortem meat pH. Another dataset consisted of 1,334 images of pigs eating inside a single-spaced feeder classified as: “single pig” or “multiple pigs.” The accuracy of genomic prediction was defined as the correlation between the predicted pH and the observed pH. The image classification prediction accuracy was the proportion of correctly classified images. For genomic prediction, a multilayer perceptron (MLP) was optimized. For image classification, MLP and convolutional neural networks (CNN) were optimized. For genomic prediction, the initial hyperparameter set resulted in an accuracy of 0.032 and for image classification, the initial accuracy was between 0.72 and 0.76. After optimization using DE, the genomic prediction accuracy was 0.3688 compared to 0.334 using GBLUP. The top selected models included one layer, 60 neurons, sigmoid activation and L2 penalty = 0.3. The accuracy of image classification after optimization was between 0.89 and 0.92. Selected models included three layers, adamax optimizer and relu or elu activation for the MLP, and one layer, 64 filters and 5×5 filter size for the CNN. DE can adapt the hyperparameter selection to each problem, dataset and model, and it significantly increased prediction accuracy with minimal user input.
8

Tsai, Chun-Wei, e Zhi-Yan Fang. "An Effective Hyperparameter Optimization Algorithm for DNN to Predict Passengers at a Metro Station". ACM Transactions on Internet Technology 21, n.º 2 (30 de março de 2021): 1–24. http://dx.doi.org/10.1145/3410156.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
As one of the public transportation systems, metro is certainly an indispensable part in urban areas of a metropolis today. Several successful results have shown that deep learning technologies might provide an effective way to predict the number of passengers at a metro station. However, most information systems based on deep learning technologies are usually designed and tuned manually by using domain knowledge and trial-and-error; thus, how to find out a set of suitable hyperparameters for a deep neural network (DNN) has become a critical research issue. To deal with the problem of hyperparameter setting for a DNN in solving the prediction of passengers at a metro station, a novel metaheuristic algorithm called search economics for hyperparameter optimization is presented to improve the accuracy rate of such a prediction system. The basic idea of the proposed algorithm is to divide the solution space into a set of subspaces first and then assign a different number of search agents to each subspace based on the “potential of each subspace.” The potential is estimated based on the objective values of the searched solutions, the objective values of the probe solutions, and the computation time invested in each subspace. The proposed method is compared with Bayesian, random forest, support vector regression, DNN, and DNN with different hyperparameter search algorithms, namely, grid search, simulated annealing, and particle swarm optimization. The simulation results using the data provided by the government of Taipei city, Taiwan, indicate that the proposed method outperforms all the other forecasting methods compared in this article in terms of the mean absolute percentage error.
9

Contreras, Pablo, Johanna Orellana-Alvear, Paul Muñoz, Jörg Bendix e Rolando Célleri. "Influence of Random Forest Hyperparameterization on Short-Term Runoff Forecasting in an Andean Mountain Catchment". Atmosphere 12, n.º 2 (10 de fevereiro de 2021): 238. http://dx.doi.org/10.3390/atmos12020238.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
The Random Forest (RF) algorithm, a decision-tree-based technique, has become a promising approach for applications addressing runoff forecasting in remote areas. This machine learning approach can overcome the limitations of scarce spatio-temporal data and physical parameters needed for process-based hydrological models. However, the influence of RF hyperparameters is still uncertain and needs to be explored. Therefore, the aim of this study is to analyze the sensitivity of RF runoff forecasting models of varying lead time to the hyperparameters of the algorithm. For this, models were trained by using (a) default and (b) extensive hyperparameter combinations through a grid-search approach that allow reaching the optimal set. Model performances were assessed based on the R2, %Bias, and RMSE metrics. We found that: (i) The most influencing hyperparameter is the number of trees in the forest, however the combination of the depth of the tree and the number of features hyperparameters produced the highest variability-instability on the models. (ii) Hyperparameter optimization significantly improved model performance for higher lead times (12- and 24-h). For instance, the performance of the 12-h forecasting model under default RF hyperparameters improved to R2 = 0.41 after optimization (gain of 0.17). However, for short lead times (4-h) there was no significant model improvement (0.69 < R2 < 0.70). (iii) There is a range of values for each hyperparameter in which the performance of the model is not significantly affected but remains close to the optimal. Thus, a compromise between hyperparameter interactions (i.e., their values) can produce similar high model performances. Model improvements after optimization can be explained from a hydrological point of view, the generalization ability for lead times larger than the concentration time of the catchment tend to rely more on hyperparameterization than in what they can learn from the input data. This insight can help in the development of operational early warning systems.
10

Jervis, Michael, Mingliang Liu e Robert Smith. "Deep learning network optimization and hyperparameter tuning for seismic lithofacies classification". Leading Edge 40, n.º 7 (julho de 2021): 514–23. http://dx.doi.org/10.1190/tle40070514.1.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Deep learning is increasingly being applied in many aspects of seismic processing and interpretation. Here, we look at a deep convolutional neural network approach to multiclass seismic lithofacies characterization using well logs and seismic data. In particular, we focus on network performance and hyperparameter tuning. Several hyperparameter tuning approaches are compared, including true and directed random search methods such as very fast simulated annealing and Bayesian hyperparameter optimization. The results show that improvements in predictive capability are possible by using automatic optimization compared with manual parameter selection. In addition to evaluating the prediction accuracy's sensitivity to hyperparameters, we test various types of data representations. The choice of input seismic data can significantly impact the overall accuracy and computation speed of the optimized networks for the classification challenge under consideration. This is validated on a 3D synthetic seismic lithofacies example with acoustic and lithologic properties based on real well data and structure from an onshore oil field.

Teses / dissertações sobre o assunto "Hyperparameter search":

1

Wang, Jiexin. "Policy Hyperparameter Exploration for Behavioral Learning of Smartphone Robots". 京都大学 (Kyoto University), 2017. http://hdl.handle.net/2433/225744.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
2

Gabere, Musa Nur. "Prediction of antimicrobial peptides using hyperparameter optimized support vector machines". Thesis, University of the Western Cape, 2011. http://etd.uwc.ac.za/index.php?module=etd&action=viewtitle&id=gen8Srv25Nme4_7345_1330684697.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:

Antimicrobial peptides (AMPs) play a key role in the innate immune response. They can be ubiquitously found in a wide range of eukaryotes including mammals, amphibians, insects, plants, and protozoa. In lower organisms, AMPs function merely as antibiotics by permeabilizing cell membranes and lysing invading microbes. Prediction of antimicrobial peptides is important because experimental methods used in characterizing AMPs are costly, time consuming and resource intensive and identification of AMPs in insects can serve as a template for the design of novel antibiotic. In order to fulfil this, firstly, data on antimicrobial peptides is extracted from UniProt, manually curated and stored into a centralized database called dragon antimicrobial peptide database (DAMPD). Secondly, based on the curated data, models to predict antimicrobial peptides are created using support vector machine with optimized hyperparameters. In particular, global optimization methods such as grid search, pattern search and derivative-free methods are utilised to optimize the SVM hyperparameters. These models are useful in characterizing unknown antimicrobial peptides. Finally, a webserver is created that will be used to predict antimicrobial peptides in haemotophagous insects such as Glossina morsitan and Anopheles gambiae.

3

Lundh, Felix, e Oscar Barta. "Hyperparameters relationship to the test accuracy of a convolutional neural network". Thesis, Högskolan i Skövde, Institutionen för informationsteknologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-19846.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Machine learning for image classification is a hot topic and it is increasing in popularity. Therefore the aim of this study is to provide a better understanding of convolutional neural network hyperparameters by comparing the test accuracy of convolutional neural network models with different hyperparameter value configurations. The focus of this study is to see whether there is an influence in the learning process depending on which hyperparameter values were used. For conducting the experiments convolutional neural network models were developed using the programming language Python utilizing the library Keras. The dataset used for this study iscifar-10, it includes 60000 colour images of 10 categories ranging from man-made objects to different animal species. Grid search is used for instantiating models with varying learning rate and momentum, width and depth values. Learning rate is only tested combined with momentum and width is only tested combined with depth. Activation functions, convolutional layers and batch size are tested individually. Grid search is compared against Bayesian optimization to see which technique will find the most optimized learning rate and momentum values. Results illustrate that the impact different hyperparameters have on the overall test accuracy varies. Learning rate and momentum affects the test accuracy greatly, however suboptimal values for learning rate and momentum can decrease the test accuracy severely. Activation function, width and depth, convolutional layer and batch size have a lesser impact on test accuracy. Regarding Bayesian optimization compared to grid search, results show that Bayesian optimization will not necessarily find more optimal hyperparameter values.
4

Myrberger, Axel, e Essen Benjamin Von. "Classifying True and Fake Telecommunication Signals With Deep Learning". Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-297675.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
This project aimed to classified artificiality gener-ated,fake, and authentic,true, telecommunication signals, basedupon their frequency response, using methods from deep learn-ing. Another goal was to accomplish this with the least amountof dimension of data possible. The datasets used contained of anequal amount of measured, provided by Ericsson, and generated,by a WINNER II implementation in Matlab, frequency responses.It was determined that a normalized version of the absolute valueof the complex frequency response was enough information for afeedforward network to do a sufficient classification. To improvethe accuracy of the network we did a hyperparameter search,which allowed us to reach an accuracy of 90 percent on our testdataset. The results show that it is possible for neural networksto differentiate between true and fake telecommunication signalsbased on their frequency response, even if it is hard for a humanto tell the difference.
Målet med det här projektet var att klassificera artificiellt genererade signaler, falska, och riktiga, sanna, telekommunikation signaler med hjälp av signalernas frekvens- svar med djup inlärningsmetoder, deep learning. Ett annat mål med projektet var att klassificera signalerna med minsta möjliga antalet dimensioner av datan. Datasetet som användes bestod av till hälften av uppmät data som Ericsson har tillhandahållit, och till hälften av generad data ifrån en WINNER II modell implementerad i Matlab. En slutsats som kunde dras är att en normaliserad version av beloppet av det komplexa frekvenssvaret innehöll tillräckligt med information för att träna ett feedforward nätverk till att uppnå en hög klassificeringssäkerhet. För att vidare öka tillförlitligheten av nätverket gjordes en hyperparametersökning, detta ökade tillförligheten till 90 procent för testdataseten. Resultaten visar att det är möjligt för neurala nätverk att skilja mellan sanna och falska telekommunikations- signaler baserat på deras frekvenssvar, även om det är svårt för människor att skilja signalerna åt.
Kandidatexjobb i elektroteknik 2020, KTH, Stockholm
5

Stynsberg, John. "Incorporating Scene Depth in Discriminative Correlation Filters for Visual Tracking". Thesis, Linköpings universitet, Datorseende, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-153110.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Visual tracking is a computer vision problem where the task is to follow a targetthrough a video sequence. Tracking has many important real-world applications in several fields such as autonomous vehicles and robot-vision. Since visual tracking does not assume any prior knowledge about the target, it faces different challenges such occlusion, appearance change, background clutter and scale change. In this thesis we try to improve the capabilities of tracking frameworks using discriminative correlation filters by incorporating scene depth information. We utilize scene depth information on three main levels. First, we use raw depth information to segment the target from its surroundings enabling occlusion detection and scale estimation. Second, we investigate different visual features calculated from depth data to decide which features are good at encoding geometric information available solely in depth data. Third, we investigate handling missing data in the depth maps using a modified version of the normalized convolution framework. Finally, we introduce a novel approach for parameter search using genetic algorithms to find the best hyperparameters for our tracking framework. Experiments show that depth data can be used to estimate scale changes and handle occlusions. In addition, visual features calculated from depth are more representative if they were combined with color features. It is also shown that utilizing normalized convolution improves the overall performance in some cases. Lastly, the usage of genetic algorithms for hyperparameter search leads to accuracy gains as well as some insights on the performance of different components within the framework.

Capítulos de livros sobre o assunto "Hyperparameter search":

1

Wistuba, Martin, Nicolas Schilling e Lars Schmidt-Thieme. "Hyperparameter Search Space Pruning – A New Component for Sequential Model-Based Hyperparameter Optimization". In Machine Learning and Knowledge Discovery in Databases, 104–19. Cham: Springer International Publishing, 2015. http://dx.doi.org/10.1007/978-3-319-23525-7_7.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
2

Plate, Tony. "Controlling the hyperparameter search in MacKay’s Bayesian neural network framework". In Lecture Notes in Computer Science, 93–112. Berlin, Heidelberg: Springer Berlin Heidelberg, 1998. http://dx.doi.org/10.1007/3-540-49430-8_5.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
3

Plate, Tony. "Controlling the Hyperparameter Search in MacKay’s Bayesian Neural Network Framework". In Lecture Notes in Computer Science, 91–110. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-35289-8_7.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
4

Ordozgoiti, Bruno, e Lluís A. Belanche Muñoz. "Off-the-Grid: Fast and Effective Hyperparameter Search for Kernel Clustering". In Machine Learning and Knowledge Discovery in Databases, 399–415. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-67661-2_24.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
5

Florea, Adrian Cătălin, e Răzvan Andonie. "A Dynamic Early Stopping Criterion for Random Search in SVM Hyperparameter Optimization". In IFIP Advances in Information and Communication Technology, 168–80. Cham: Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-319-92007-8_15.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
6

Khamitov, Kamil, Nina Popova, Yuri Konkov e Tony Castillo. "Tuning ANNs Hyperparameters and Neural Architecture Search Using HPC". In Communications in Computer and Information Science, 536–48. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-64616-5_46.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
7

Jääsaari, Elias, Ville Hyvönen e Teemu Roos. "Efficient Autotuning of Hyperparameters in Approximate Nearest Neighbor Search". In Advances in Knowledge Discovery and Data Mining, 590–602. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-16145-3_46.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
8

Hinaut, Xavier, e Nathan Trouvain. "Which Hype for My New Task? Hints and Random Search for Echo State Networks Hyperparameters". In Lecture Notes in Computer Science, 83–97. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-86383-8_7.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
9

Lopez, Kyra Mikaela M., e Ma Sheila A. Magboo. "A Clinical Decision Support Tool to Detect Invasive Ductal Carcinoma in Histopathological Images Using Support Vector Machines, Naïve-Bayes, and K-Nearest Neighbor Classifiers". In Machine Learning and Artificial Intelligence. IOS Press, 2020. http://dx.doi.org/10.3233/faia200765.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
This study aims to describe a model that will apply image processing and traditional machine learning techniques specifically Support Vector Machines, Naïve-Bayes, and k-Nearest Neighbors to identify whether or not a given breast histopathological image has Invasive Ductal Carcinoma (IDC). The dataset consisted of 54,811 breast cancer image patches of size 50px x 50px, consisting of 39,148 IDC negative and 15,663 IDC positive. Feature extraction was accomplished using Oriented FAST and Rotated BRIEF (ORB) descriptors. Feature scaling was performed using Min-Max Normalization while K-Means Clustering on the ORB descriptors was used to generate the visual codebook. Automatic hyperparameter tuning using Grid Search Cross Validation was implemented although it can also accept user supplied hyperparameter values for SVM, Naïve Bayes, and K-NN models should the user want to do experimentation. Aside from computing for accuracy, the AUPRC and MCC metrics were used to address the dataset imbalance. The results showed that SVM has the best overall performance, obtaining accuracy = 0.7490, AUPRC = 0.5536, and MCC = 0.2924.
10

Larsen, Kai R., e Daniel S. Becker. "Why Use Automated Machine Learning?" In Automated Machine Learning for Business, 1–22. Oxford University Press, 2021. http://dx.doi.org/10.1093/oso/9780190941659.003.0001.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
Resumo:
Machine learning is involved in search, translation, detecting depression, likelihood of college dropout, finding lost children, and to sell all kinds of products. While barely beyond its inception, the current machine learning revolution will affect people and organizations no less than the Industrial Revolution’s effect on weavers and many other skilled laborers. Machine learning will automate hundreds of millions of jobs that were considered too complex for machines ever to take over even a decade ago, including driving, flying, painting, programming, and customer service, as well as many of the jobs previously reserved for humans in the fields of finance, marketing, operations, accounting, and human resources. This section explains how automated machine learning addresses exploratory data analysis, feature engineering, algorithm selection, hyperparameter tuning, and model diagnostics. The section covers the eight criteria considered essential for AutoML to have significant impact: accuracy, productivity, ease of use, understanding and learning, resource availability, process transparency, generalization, and recommended actions.

Trabalhos de conferências sobre o assunto "Hyperparameter search":

1

Balaprakash, Prasanna, Michael Salim, Thomas D. Uram, Venkat Vishwanath e Stefan M. Wild. "DeepHyper: Asynchronous Hyperparameter Search for Deep Neural Networks". In 2018 IEEE 25th International Conference on High Performance Computing (HiPC). IEEE, 2018. http://dx.doi.org/10.1109/hipc.2018.00014.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
2

Lopez-Ramos, Luis M., e Baltasar Beferull-Lozano. "Online Hyperparameter Search Interleaved with Proximal Parameter Updates". In 2020 28th European Signal Processing Conference (EUSIPCO). IEEE, 2021. http://dx.doi.org/10.23919/eusipco47968.2020.9287537.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
3

Buratti, Benedetto J., e Eli Upfal. "Ordalia: Deep Learning Hyperparameter Search via Generalization Error Bounds Extrapolation". In 2019 IEEE International Conference on Big Data (Big Data). IEEE, 2019. http://dx.doi.org/10.1109/bigdata47090.2019.9006144.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
4

Cho, Minsu, e Chinmay Hegde. "Reducing the Search Space for Hyperparameter Optimization Using Group Sparsity". In ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019. http://dx.doi.org/10.1109/icassp.2019.8682434.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
5

Li, Zhenzhen, Lianwen Jin, Chunlin Yang e Zhuoyao Zhong. "Hyperparameter search for deep convolutional neural network using effect factors". In 2015 IEEE China Summit and International Conference on Signal and Information Processing (ChinaSIP). IEEE, 2015. http://dx.doi.org/10.1109/chinasip.2015.7230511.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
6

Sanchez, Odnan Ref, Matteo Repetto, Alessandro Carrega e Raffaele Bolla. "Evaluating ML-based DDoS Detection with Grid Search Hyperparameter Optimization". In 2021 IEEE 7th International Conference on Network Softwarization (NetSoft). IEEE, 2021. http://dx.doi.org/10.1109/netsoft51509.2021.9492633.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
7

Zhang, Michael, Chandra Krintz, Markus Mock e Rich Wolski. "Seneca: Fast and Low Cost Hyperparameter Search for Machine Learning Models". In 2019 IEEE 12th International Conference on Cloud Computing (CLOUD). IEEE, 2019. http://dx.doi.org/10.1109/cloud.2019.00071.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
8

Wendt, Alexander, Marco Wuschnig e Martin Lechner. "Speeding up Common Hyperparameter Optimization Methods by a Two-Phase-Search". In IECON 2020 - 46th Annual Conference of the IEEE Industrial Electronics Society. IEEE, 2020. http://dx.doi.org/10.1109/iecon43393.2020.9254801.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
9

Shekar, B. H., e Guesh Dagnew. "Grid Search-Based Hyperparameter Tuning and Classification of Microarray Cancer Data". In 2019 Second International Conference on Advanced Computational and Communication Paradigms (ICACCP). IEEE, 2019. http://dx.doi.org/10.1109/icaccp.2019.8882943.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.
10

Alibrahim, Hussain, e Simone A. Ludwig. "Hyperparameter Optimization: Comparing Genetic Algorithm against Grid Search and Bayesian Optimization". In 2021 IEEE Congress on Evolutionary Computation (CEC). IEEE, 2021. http://dx.doi.org/10.1109/cec45853.2021.9504761.

Texto completo da fonte
Estilos ABNT, Harvard, Vancouver, APA, etc.

Vá para a bibliografia