-
Couldn't load subscription status.
- Fork 58
Description
Ipopt v1.8.0 adds the Ipopt._VectorNonlinearOracle set. The purpose of this issue is to explain our motivation for adding the set, and what we plan to do about it in the future.
Current status and future plans
Ipopt._VectorNonlinearOracle is experimental. You are encouraged to try it out and report back experiences (see usage instructions below). It is marked as experimental (it begins with an underscore) because we plan to make a change to the API no later than September 30, 2025 (six months after adding it).
We are considering three options:
- Rename the set to
Ipopt.VectorNonlinearFunctionand publicly support it as part of the v1.X API of Ipopt.jl - Move the set to MathOptInterface as part of the public API for MathOptInterface.jl, and add support for it in other packages like KNITRO.jl and NLopt.jl
- Remove the set in favor of a different approach to vector-valued user defined functions.
If the set is useful for a small number of users who use only Ipopt.jl, and no one shows wider interest, we will choose (1).
If the set is useful for a number of different groups, including by people who want to use it with different NLP solvers, and the API proves to be a good design, we will choose (2).
If issues arise, we reserve the right to delete the set entirely and pursue a different approach in (3).
Thus, if you use this set please comment below with feedback of how you are using it and whether we could improve it.
Compatibility
The _VectorNonlinearOracle feature is experimental. It relies on a private API feature of Ipopt.jl that will change in a future release. If you use this feature, you must pin the version of Ipopt.jl in your Project.toml to ensure that future updates to Ipopt.jl do not break your existing code.
A known good version of Ipopt.jl is v1.8.0. Pin the version using:
[compat]
Ipopt = "=1.8.0"
Definition
Ipopt._VectorNonlinearOracle is defined as:
_VectorNonlinearOracle(;
dimension::Int,
l::Vector{Float64},
u::Vector{Float64},
eval_f::Function,
jacobian_structure::Vector{Tuple{Int,Int}},
eval_jacobian::Function,
hessian_lagrangian_structure::Vector{Tuple{Int,Int}} = Tuple{Int,Int}[],
eval_hessian_lagrangian::Union{Nothing,Function} = nothing,
) <: MOI.AbstractVectorSetwhich represents the set:
where f is defined by the vectors l and u, and the callback oracles eval_f, eval_jacobian, and eval_hessian_lagrangian.
Function
The eval_f function must have the signature
eval_f(ret::AbstractVector, x::AbstractVector)::Nothingwhich fills ret.
Jacobian
The eval_jacobian function must have the signature
eval_jacobian(ret::AbstractVector, x::AbstractVector)::Nothingwhich fills the sparse Jacobian ret.
The one-indexed sparsity structure must be provided in the jacobian_structure argument.
Hessian
The eval_hessian_lagrangian function is optional.
If eval_hessian_lagrangian === nothing, Ipopt will use a Hessian approximation instead of the exact Hessian.
If eval_hessian_lagrangian is a function, it must have the signature
eval_hessian_lagrangian(
ret::AbstractVector,
x::AbstractVector,
μ::AbstractVector,
)::Nothingwhich fills the sparse Hessian of the Lagrangian ret.
The one-indexed sparsity structure must be provided in the hessian_lagrangian_structure argument.
Use with JuMP
This section provides some usage examples for JuMP.
Example: f(x) <= 1
A simplest example is to represent a constraint x^2 + y^2 <= 1 in the problem:
This can be achieved as follows
julia> using JuMP, Ipopt
julia> set = Ipopt._VectorNonlinearOracle(;
dimension = 2,
l = [-Inf],
u = [1.0],
eval_f = (ret, x) -> (ret[1] = x[1]^2 + x[2]^2),
jacobian_structure = [(1, 1), (1, 2)],
eval_jacobian = (ret, x) -> ret .= 2.0 .* x,
hessian_lagrangian_structure = [(1, 1), (2, 2)],
eval_hessian_lagrangian = (ret, x, u) -> ret .= 2.0 .* u[1],
);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, 0 <= x <= 1)
@variable(model, 0 <= y <= 1)
@objective(model, Max, x + y)
@constraint(model, [x, y] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
value(x), value(y), 1 / sqrt(2)
end
(0.707106783471979, 0.707106783471979, 0.7071067811865475)Example: y = F(x)
To model
We can modify our previous example to:
which is implemented as follows:
julia> using JuMP, Ipopt
julia> set = Ipopt._VectorNonlinearOracle(;
dimension = 3,
l = [0.0],
u = [0.0],
eval_f = (ret, x) -> (ret[1] = x[1]^2 + x[2]^2 - x[3]),
jacobian_structure = [(1, 1), (1, 2), (1, 3)],
eval_jacobian = (ret, x) -> begin
ret[1:2] .= 2.0 .* x[1:2]
ret[3] = -1.0
end,
hessian_lagrangian_structure = [(1, 1), (2, 2)],
eval_hessian_lagrangian = (ret, x, u) -> ret .= 2.0 .* u[1],
);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, 0 <= x <= 1)
@variable(model, 0 <= y <= 1)
@variable(model, z <= 1)
@objective(model, Max, x + y)
# x^2 + y^2 - z == 0
@constraint(model, [x, y, z] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
value(x), value(y), value(z), 1 / sqrt(2)
end
(0.707106783471979, 0.707106783471979, 1.0000000064625545, 0.7071067811865475)Example: multiple rows
A more involved example is:
which is implemented as follows:
julia> set = Ipopt._VectorNonlinearOracle(;
dimension = 5,
l = [0.0, 0.0],
u = [0.0, 0.0],
eval_f = (ret, x) -> begin
ret[1] = x[1]^2 - x[4]
ret[2] = x[2]^2 + x[3]^3 - x[5]
return
end,
jacobian_structure = [(1, 1), (2, 2), (2, 3), (1, 4), (2, 5)],
eval_jacobian = (ret, x) -> begin
ret[1] = 2 * x[1]
ret[2] = 2 * x[2]
ret[3] = 3 * x[3]^2
ret[4] = -1.0
ret[5] = -1.0
return
end,
hessian_lagrangian_structure = [(1, 1), (2, 2), (3, 3)],
eval_hessian_lagrangian = (ret, x, u) -> begin
ret[1] = 2 * u[1]
ret[2] = 2 * u[2]
ret[3] = 6 * x[3] * u[2]
return
end,
);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, x[i in 1:3] == i)
@variable(model, y[1:2])
@constraint(model, [x; y] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
value.(x), value.(y)
end
([1.0, 2.0, 3.0], [1.0, 31.0])Example: no Hessian
The Hessian evaluator is optional:
julia> using JuMP, Ipopt
julia> set = Ipopt._VectorNonlinearOracle(;
dimension = 2,
l = [-Inf],
u = [1.0],
eval_f = (ret, x) -> (ret[1] = x[1]^2 + x[2]^2),
jacobian_structure = [(1, 1), (1, 2)],
eval_jacobian = (ret, x) -> ret .= 2.0 .* x,
);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, 0 <= x <= 1)
@variable(model, 0 <= y <= 1)
@objective(model, Max, x + y)
@constraint(model, [x, y] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
value(x), value(y), 1 / sqrt(2)
end
(0.7071067847123265, 0.7071067847123265, 0.7071067811865475)Example: vector-valued user-defined function and ForwardDiff
Constructing the Jacobian and Hessian is tedious and error prone. Here is an example that uses ForwardDiff.jl:
julia> using JuMP, Ipopt, ForwardDiff
julia> function build_set_from_f(f; input_dimension, output_dimension)
input_indices = 1:input_dimension
output_indices = input_dimension.+(1:output_dimension)
return Ipopt._VectorNonlinearOracle(;
dimension = input_dimension + output_dimension,
l = zeros(output_dimension),
u = zeros(output_dimension),
# eval_f := f(x) - y
eval_f = (ret, args) -> begin
ret .= f(args[input_indices]) .- args[output_indices]
end,
# eval_jacobian := [∇f(x) -I]
jacobian_structure = vcat(
[(i, j) for j in 1:input_dimension for i in 1:output_dimension],
[(j, input_dimension + j) for j in 1:output_dimension],
),
eval_jacobian = (ret, args) -> begin
ret .= vcat(
vec(ForwardDiff.jacobian(f, args[input_indices])),
fill(-1.0, output_dimension),
)
end,
# eval_hessian_lagrangian := ∑ uᵢ ∇²fᵢ(x) := ∇²(u'f(x))
hessian_lagrangian_structure = [
(i, j) for j in 1:input_dimension for i in 1:input_dimension if j >= i
],
eval_hessian_lagrangian = (ret, args, u) -> begin
H = ForwardDiff.hessian(x -> u' * f(x), args[input_indices])
k = 0
for j in 1:input_dimension
for i in 1:j
k += 1
ret[k] = H[i, j]
end
end
end,
)
end
build_set_from_f (generic function with 1 method)
julia> f(x) = [x[1]^2, x[2]^2 + x[3]^3]
f (generic function with 1 method)
julia> set = build_set_from_f(f; input_dimension = 3, output_dimension = 2);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, x[i in 1:3] == i)
@variable(model, y[1:2])
@constraint(model, [x; y] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
f(value.(x)), value.(y)
end
([1.0, 31.0], [1.0, 31.0])