Are any conclusion about optimal hyperparameters of optmization algorithm for use on stratigraphic forward models?

Answer from top 10 papers

The provided papers do not directly address the optimization of hyperparameters for stratigraphic forward models. Instead, they focus on hyperparameter optimization in various other domains such as healthcare, neural networks, energy methods, machine learning algorithms, reinforcement learning, and software quality prediction. Each paper proposes or evaluates different methods for hyperparameter tuning, including swarm intelligence algorithms (Ashraf et al., 2021), bilevel optimization (Mackay et al., 2019), Bayesian optimization (Belete & Huchaiah, 2021), genetic algorithms (Probst et al., 2018), and other optimization strategies (Chadha et al., 2022; Dernoncourt & Lee, 2016; Ibrahim et al., 2023; Malhotra & Cherukuri, 2024; Raji et al., 2022; Wojciuk et al., 2024).
While these studies offer insights into hyperparameter optimization, they do not provide specific conclusions for stratigraphic forward models, which are computational models used in geology to simulate sediment deposition and erosion over geological time scales. The principles and methods discussed in the papers could potentially be adapted to the context of stratigraphic forward models, but this would require further research and experimentation specific to the field of geology and stratigraphy.
In summary, the papers reviewed do not yield direct conclusions regarding the optimal hyperparameters of optimization algorithms for use on stratigraphic forward models. However, the optimization techniques and findings presented could inform future research aimed at determining the best hyperparameter settings for such models, acknowledging the need for domain-specific adaptations and validations (Ashraf et al., 2021; Belete & Huchaiah, 2021; Chadha et al., 2022; Dernoncourt & Lee, 2016; Ibrahim et al., 2023; Mackay et al., 2019; Malhotra & Cherukuri, 2024; Probst et al., 2018; Raji et al., 2022; Wojciuk et al., 2024).

Source Papers

Optimizing hyperparameters of deep reinforcement learning for autonomous driving based on whale optimization algorithm.

Deep Reinforcement Learning (DRL) enables agents to make decisions based on a well-designed reward function that suites a particular environment without any prior knowledge related to a given environment. The adaptation of hyperparameters has a great impact on the overall learning process and the learning processing times. Hyperparameters should be accurately estimated while training DRL algorithms, which is one of the key challenges that we attempt to address. This paper employs a swarm-based optimization algorithm, namely the Whale Optimization Algorithm (WOA), for optimizing the hyperparameters of the Deep Deterministic Policy Gradient (DDPG) algorithm to achieve the optimum control strategy in an autonomous driving control problem. DDPG is capable of handling complex environments, which contain continuous spaces for actions. To evaluate the proposed algorithm, the Open Racing Car Simulator (TORCS), a realistic autonomous driving simulation environment, was chosen to its ease of design and implementation. Using TORCS, the DDPG agent with optimized hyperparameters was compared with a DDPG agent with reference hyperparameters. The experimental results showed that the DDPG’s hyperparameters optimization leads to maximizing the total rewards, along with testing episodes and maintaining a stable driving policy.

Open Access
Self-Tuning Networks: Bilevel Optimization of Hyperparameters using Structured Best-Response Functions

Hyperparameter optimization can be formulated as a bilevel optimization problem, where the optimal parameters on the training set depend on the hyperparameters. We aim to adapt regularization hyperparameters for neural networks by fitting compact approximations to the best-response function, which maps hyperparameters to optimal weights and biases. We show how to construct scalable best-response approximations for neural networks by modeling the best-response as a single network whose hidden units are gated conditionally on the regularizer. We justify this approximation by showing the exact best-response for a shallow linear network with L2-regularized Jacobian can be represented by a similar gating mechanism. We fit this model using a gradient-based hyperparameter optimization algorithm which alternates between approximating the best-response around the current hyperparameters and optimizing the hyperparameters using the approximate best-response function. Unlike other gradient-based approaches, we do not require differentiating the training loss with respect to the hyperparameters, allowing us to tune discrete hyperparameters, data augmentation hyperparameters, and dropout probabilities. Because the hyperparameters are adapted online, our approach discovers hyperparameter schedules that can outperform fixed hyperparameter values. Empirically, our approach outperforms competing hyperparameter optimization methods on large-scale deep learning problems. We call our networks, which update their own hyperparameters online during training, Self-Tuning Networks (STNs).

Grid search in hyperparameter optimization of machine learning models for prediction of HIV/AIDS test results

In this work, we propose hyperparameters optimization using grid search to optimize the parameters of eight existing models and apply the best parameters to predict the outcomes of HIV tests from the Ethiopian Demographic and Health Survey (EDHS), HIV/AIDS dataset. The core challenge of this work is to find the right or optimum parameter values that generate the optimal model and uncertain training computing costs and test predictive models using various values of hyperparameters. To overcome these challenges, we explore the effects of hyperparameters optimizations by applying a proposed grid search hyperparameter optimization (GSHPO) on the considered models to robust the prediction power. An extensive number of experiments are conducted to affirm the feasibility of our proposed methods. These experiments are done in two separate phases. In the first phase, we test our method with the selected models before hyperparameter optimization is applying (using the default parameters). The second phase of the experiment is done after the hyperparameter optimization is applying (using GSHPO). During the experiment, the 10-fold cross validation technique is used to solve the bias of the models. The proposed system helped to tune the hyperparameters using the grid search approach to the prediction algorithms. Several standard metrics are used to assess the method's efficiency, like accuracy, precision, recall, f1-score, AUC-ROC, MAE, RMSE, R2 and confusion matrix to compare results of each experiments. The results obtained by after applying 10-fold cross validation techniques and the proposed GSHPO are promising. Our findings suggest that the hyper-parameters of tuning models have a statistically important positive impact on the models' prediction accuracy.

Optimizing Hyperparameters and Architecture of Deep Energy Method

The deep energy method (DEM) employs the principle of minimum potential energy to train neural network models to predict displacement at a state of equilibrium under given boundary conditions. The accuracy of the model is contingent upon choosing appropriate hyperparameters. The hyperparameters have traditionally been chosen based on literature or through manual iterations. The displacements predicted using hyperparameters suggested in the literature do not ensure the minimum potential energy of the system. Additionally, they do not necessarily generalize to different load cases. Selecting hyperparameters through manual trial and error and grid search algorithms can be highly time-consuming. We propose a systematic approach using the Bayesian optimization algorithms and random search to identify optimal values for these parameters. Seven hyperparameters are optimized to obtain the minimum potential energy of the system under compression, tension, and bending loads cases. In addition to Bayesian optimization, Fourier feature mapping is also introduced to improve accuracy. The models trained using optimal hyperparameters and Fourier feature mapping could accurately predict deflections compared to finite element analysis for linear elastic materials. The deflections obtained for tension and compression load cases are found to be more sensitive to values of hyperparameters compared to bending. The approach can be easily extended to 3D and other material models.

Open Access
A systematic review of hyperparameter tuning techniques for software quality prediction models

BACKGROUND: Software quality prediction models play a crucial role in identifying vulnerable software components during early stages of development, and thereby optimizing the resource allocation and enhancing the overall software quality. While various classification algorithms have been employed for developing these prediction models, most studies have relied on default hyperparameter settings, leading to significant variability in model performance. Tuning the hyperparameters of classification algorithms can enhance the predictive capability of quality models by identifying optimal settings for improved accuracy and effectiveness. METHOD: This systematic review examines studies that have utilized hyperparameter tuning techniques to develop prediction models in software quality domain. The review focused on diverse areas such as defect prediction, maintenance estimation, change impact prediction, reliability prediction, and effort estimation, as these domains demonstrate the wide applicability of common learning algorithms. RESULTS: This review identified 31 primary studies on hyperparameter tuning for software quality prediction models. The results demonstrate that tuning the parameters of classification algorithms enhances the performance of prediction models. Additionally, the study found that certain classification algorithms exhibit high sensitivity to their parameter settings, achieving optimal performance when tuned appropriately. Conversely, certain classification algorithms exhibit low sensitivity to their parameter settings, making tuning unnecessary in such instances. CONCLUSION: Based on the findings of this review, the study conclude that the predictive capability of software quality prediction models can be significantly improved by tuning their hyperparameters. To facilitate effective hyperparameter tuning, we provide practical guidelines derived from the insights obtained through this study.

Improving classification accuracy of fine-tuned CNN models: Impact of hyperparameter optimization

The immense popularity of convolutional neural network (CNN) models has sparked a growing interest in optimizing their hyperparameters. Discovering the ideal values for hyperparameters to achieve optimal CNN training is a complex and time-consuming task, often requiring repetitive numerical experiments. As a result, significant attention is currently being devoted to developing methods aimed at tailoring hyperparameters for specific CNN models and classification tasks. While existing optimization methods often yield favorable image classification results, they do not provide guidance on which hyperparameters are worth optimizing, the appropriate value ranges for those hyperparameters, or whether it is reasonable to use a subset of training data for the optimization process. This work is focused on the optimization of hyperparameters during transfer learning, with the goal of investigating how different optimization methods and hyperparameter selections impact the performance of fine-tuned models. In our experiments, we assessed the importance of various hyperparameters and identified the ranges within which optimal CNN training can be achieved. Additionally, we compared four hyperparameter optimization methods—grid search, random search, Bayesian optimization, and the Asynchronous Successive Halving Algorithm (ASHA). We also explored the feasibility of fine-tuning hyperparameters using a subset of the training data. By optimizing the hyperparameters, we observed an improvement in CNN classification accuracy of up to 6%. Furthermore, we found that achieving a balance in class distribution within the subset of data used for parameter optimization is crucial in establishing the optimal set of hyperparameters for CNN training. The results we obtained demonstrate that hyperparameter optimization is highly dependent on the specific task and dataset at hand.

Open Access
Simple Deterministic Selection-Based Genetic Algorithm for Hyperparameter Tuning of Machine Learning Models

Hyperparameter tuning is a critical function necessary for the effective deployment of most machine learning (ML) algorithms. It is used to find the optimal hyperparameter settings of an ML algorithm in order to improve its overall output performance. To this effect, several optimization strategies have been studied for fine-tuning the hyperparameters of many ML algorithms, especially in the absence of model-specific information. However, because most ML training procedures need a significant amount of computational time and memory, it is frequently necessary to build an optimization technique that converges within a small number of fitness evaluations. As a result, a simple deterministic selection genetic algorithm (SDSGA) is proposed in this article. The SDSGA was realized by ensuring that both chromosomes and their accompanying fitness values in the original genetic algorithm are selected in an elitist-like way. We assessed the SDSGA over a variety of mathematical test functions. It was then used to optimize the hyperparameters of two well-known machine learning models, namely, the convolutional neural network (CNN) and the random forest (RF) algorithm, with application on the MNIST and UCI classification datasets. The SDSGA’s efficiency was compared to that of the Bayesian Optimization (BO) and three other popular metaheuristic optimization algorithms (MOAs), namely, the genetic algorithm (GA), particle swarm optimization (PSO) and biogeography-based optimization (BBO) algorithms. The results obtained reveal that the SDSGA performed better than the other MOAs in solving 11 of the 17 known benchmark functions considered in our study. While optimizing the hyperparameters of the two ML models, it performed marginally better in terms of accuracy than the other methods while taking less time to compute.

Open Access