Constant optimization in symbolic regression is an important task addressed by several researchers. It has been demonstrated that continuous optimization techniques are adequate to find good values for the constants by minimizing the prediction error. In this paper, we evaluate several continuous optimization methods that can be used to perform constant optimization in symbolic regression.
In this paper we investigate how to efficiently apply Approximate-Karush–Kuhn–Tucker proximity measures as stopping criteria for optimization algorithms that do not generate approximations to Lagrange multipliers. We prove that the KKT error measurement tends to zero when approaching a solution and we develop a simple model to compute the KKT error measure requiring only the
In the last decades, a number of novel meta-heuristics and hybrid algorithms have been proposed to solve a great variety of optimization problems. Among these, constrained optimization problems are considered of particular interest in applications from many different domains. The presence of multiple constraints can make optimization problems particularly hard to solve, thus imposing the
Several constrained and unconstrained optimization problems have been adequately solved over the years thanks to advances in the metaheuristics area. In the last decades, different metaheuristics have been proposed employing new ideas, and hybrid algorithms that improve the original metaheuristics have been developed. One of the most successfully employed metaheuristics is the Differential Evolution. In
Protein structures prediction (PSP) is a computationally complex problem. Simplified models of the protein molecule (such as the HP Model) and the use of evolutionary algorithms (EAs) are among the most investigated techniques for PSP. However, the evaluation of a structure represented by the HP model considers only the number of hydrophobic contacts, which doesn’t