From 3ae8d0403bec740ed3ffd27a633a8e5dd5ac3a4d Mon Sep 17 00:00:00 2001 From: rolo Date: Fri, 10 Apr 2026 10:36:56 +0200 Subject: [PATCH] Fix error in epf logic --- src/popt/loop/optimize.py | 2 +- src/popt/update_schemes/enopt.py | 4 +++- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/src/popt/loop/optimize.py b/src/popt/loop/optimize.py index 55d171d9..44938f5d 100644 --- a/src/popt/loop/optimize.py +++ b/src/popt/loop/optimize.py @@ -198,7 +198,7 @@ def run_loop(self): previous_state = self.xk self.epf['r'] *= self.epf['r_factor'] # increase penalty factor self.obj_func_tol *= self.epf['tol_factor'] # decrease tolerance - self.obj_func_values = self.fun(self.xk, **self.epf) + self.obj_func_values = self.fun(self.xk, epf=self.epf) self.iteration = 0 self.epf_iteration += 1 optimize_result = ot.get_optimize_result(self) diff --git a/src/popt/update_schemes/enopt.py b/src/popt/update_schemes/enopt.py index e3fa869e..cae55043 100644 --- a/src/popt/update_schemes/enopt.py +++ b/src/popt/update_schemes/enopt.py @@ -97,7 +97,9 @@ def __set__variable(var_name=None, defalut=None): # Set other optimization parameters self.obj_func_tol = __set__variable('tol', 1e-6) - self.alpha = __set__variable('alpha', 0.1) + self.alpha = __set__variable('step_size', 0.1) + self.alpha = __set__variable('alpha', 0.1) # accept either 'step_size' or 'alpha' for step size, + # with 'alpha' as the default self.alpha_cov = __set__variable('alpha_cov', 0.001) self.beta = __set__variable('beta', 0.0) # this is stored in the optimizer class self.nesterov = __set__variable('nesterov', False) # use Nesterov acceleration if value is true