diff --git a/src/selection.jl b/src/selection.jl index d42f638..7f93015 100644 --- a/src/selection.jl +++ b/src/selection.jl @@ -57,7 +57,7 @@ There are 7 optimizers available: ``` """ function select_optimizers( - nlp::AbstractNLPModel{T, S}, + nlp::AbstractNLPModel{T, S}; verbose = 1, highest_derivative_available::Integer = 2, ) where {T, S} @@ -151,9 +151,9 @@ function select_optimizers( return select end -function select_optimizers(model::JuMP.Model, args...; kwargs...) +function select_optimizers(model::JuMP.Model; kwargs...) nlp = MathOptNLPModel(model) - return select_optimizers(nlp, args...; kwargs...) + return select_optimizers(nlp; kwargs...) end """Checker whether optimizers are Generic only""" diff --git a/src/solve.jl b/src/solve.jl index 818a9da..4235dd3 100644 --- a/src/solve.jl +++ b/src/solve.jl @@ -4,7 +4,11 @@ function minimize( highest_derivative_available::Integer = 2, kwargs..., ) - select = select_optimizers(nlp, verbose, highest_derivative_available) + select = select_optimizers( + nlp; + verbose = verbose, + highest_derivative_available = highest_derivative_available, + ) (verbose ≥ 1) && println("Solve using $(first(select).name):") solver = first(select) return minimize(Val(Symbol(solver.name)), nlp; verbose = verbose, kwargs...) @@ -16,7 +20,11 @@ function minimize( highest_derivative_available::Integer = 2, kwargs..., ) - select = select_optimizers(nlp, verbose, highest_derivative_available) + select = select_optimizers( + nlp; + verbose = verbose, + highest_derivative_available = highest_derivative_available, + ) nls_select = select[select.specialized_nls, :] solver = if !isempty(nls_select) first(nls_select)