# KSComp¶

KSComp provides a way to aggregate many constraints into a single constraint. This is usually done for performance reasons, in particular, to reduce the calculation time needed for the total derivatives of your model. The KSComp implements the Kreisselmeier-Steinhauser Function to aggregate constraint vector input “g” into a single scalar output ‘KS’.

By default, the constraint vector “g” is assumed be of the form where g<=0 satisfies the constraints, but other forms can be specified using the “upper” and “lower_flag” options.

The output “KS” should be constrained with an upper-bound of zero to satisfy the aggregated constraint. By default, it is left to the user to provide this constraint. However, setting option “add_constraint” to True will cause the KSComp to automatically add a constraint to the optimization.

## KSComp Options¶

Option

Default

Acceptable Values

Acceptable Types

Description

False

[True, False]

[‘bool’]

If True, add a constraint on the resulting output of the KSComp. If False, the user will be expected to add a constraint explicitly.

None

N/A

[‘int’, ‘float’]

distributed

False

[True, False]

[‘bool’]

True if the component has variables that are distributed across multiple processes.

lower_flag

False

[True, False]

[‘bool’]

Set to True to reverse sign of input constraints.

parallel_deriv_color

None

N/A

[‘str’]

If specified, this design var will be grouped for parallel derivative calculations with other variables sharing the same parallel_deriv_color.

ref

None

N/A

[‘int’, ‘float’]

Unit reference for constraint, if added, default is one.

ref0

None

N/A

[‘int’, ‘float’]

Zero-reference for constraint, if added, default is zero.

rho

50.0

N/A

N/A

Constraint Aggregation Factor.

scaler

None

N/A

[‘int’, ‘float’]

Scaler for constraint, if added, default is one.

units

None

N/A

[‘str’]

Units to be assigned to all variables in this component. Default is None, which means variables are unitless.

upper

0.0

N/A

N/A

Upper bound for constraint, default is zero.

vec_size

1

N/A

[‘int’]

The number of rows to independently aggregate.

width

1

N/A

[‘int’]

Width of constraint vector.

## KSComp Constructor¶

The call signature for the KSComp constructor is:

KSComp.__init__(**kwargs)[source]

Initialize the KS component.

Parameters
**kwargsdict of keyword arguments

Keyword arguments that will be mapped into the Component options.

## KSComp Example¶

The following example is perhaps the simplest possible. It shows a component that represents a constraint of width two. We would like to aggregate the values of this constraint vector into a single scalar value using the KSComp.

import numpy as np

import openmdao.api as om

prob = om.Problem()
model = prob.model

x=np.zeros((2, )),
y=np.zeros((2, ))), promotes_inputs=['x'])

model.connect('comp.y', 'ks.g')

prob.setup()
prob.set_val('x', np.array([5.0, 4.0]))
prob.run_model()

print(prob.get_val('ks.KS'))

[[15.]]


A more practical example that uses the KSComp can be found in the beam optimization example.

You can also independently aggregate multiple rows of an output as separate constraints by declaring the vec_size argument:

import numpy as np

import openmdao.api as om

prob = om.Problem()
model = prob.model

x=np.zeros((2, 2)),
y=np.zeros((2, 2))), promotes_inputs=['x'])

model.connect('comp.y', 'ks.g')

prob.setup()
prob.set_val('x', np.array([[5.0, 4.0], [10.0, 8.0]]))
prob.run_model()

print(prob.get_val('ks.KS'))

[[15.]
[30.]]


## KSComp Option Examples¶

Normally, the input constraint vector is assumed to be of the form g<=0 is satisfied. If you would like to set a different upper bound for the constraint, you can declare it in the “upper” option in the options dictionary.

In the following example, we specify a new upper bound of 16 for the constraint vector. Note that the KS output is still satisfied if it is less than zero.

upper

import numpy as np

import openmdao.api as om

prob = om.Problem()
model = prob.model

x=np.zeros((2, )),
y=np.zeros((2, ))), promotes_inputs=['x'])

model.connect('comp.y', 'ks.g')

model.ks.options['upper'] = 16.0
prob.setup()
prob.set_val('x', np.array([5.0, 4.0]))
prob.run_model()

print(prob['ks.KS'])

[[-1.]]

Normally, the input constraint vector is satisfied if it is negative and violated if it is positive. You can reverse this behavior by setting the “lower_flag” option to True. In the following example, we turn on the “lower_flag” so that positive values of the input constraint are considered satisfied. Note that the KS output is still satisfied if it is less than zero.

lower_flag

import numpy as np

import openmdao.api as om

prob = om.Problem()
model = prob.model

x=np.zeros((2, )),
y=np.zeros((2, ))), promotes_inputs=['x'])

model.connect('comp.y', 'ks.g')

model.ks.options['lower_flag'] = True
prob.setup()
prob.set_val('x', np.array([5.0, 4.0]))
prob.run_model()

print(prob.get_val('ks.KS'))

[[-12.]]

Typically, the KSComp is used to provide a constraint which aggregates many values into a single scalar constraint. Consider the following simple example, where we seek to maximize the peak of a parabola but also keep the peak of the parabola below a certain threshold value. Clearly, the solution here is to have the peak of the parabola lie on the peak constraint.

Note the resulting value of the offset “k” is not exactly 4.0 as we might expect. The KS function provides a differentiable constraint aggregation, but the resulting scalar constraint is slightly conservative.

import numpy as np
import openmdao.api as om
import matplotlib.pyplot as plt

n = 50
prob = om.Problem()
model = prob.model

prob.driver = om.ScipyOptimizeDriver()

model.add_subsystem('comp', om.ExecComp('y = -3.0*x**2 + k',
x=np.zeros((n, )),
y=np.zeros((n, )),
k=0.0), promotes_inputs=['x', 'k'])

model.connect('comp.y', 'ks.g')

prob.setup()
prob.set_val('x', np.linspace(-np.pi/2, np.pi/2, n))
prob.set_val('k', 5.)

prob.run_driver()

print(max(prob.get_val('comp.y')) <= 4.0)

fig, ax = plt.subplots()

x = prob.get_val('x')
y = prob.get_val('comp.y')

ax.plot(x, y, 'r.')
ax.plot(x, 4.0*np.ones_like(x), 'k--')
ax.set_xlabel('x')
ax.set_ylabel('y')
ax.grid(True)
ax.text(-0.25, 0, f"k = {prob.get_val('k')[0]:6.3f}")

plt.show()


units

Finally, note that you can pass a units option to the KSComp that will define units on its input and output variables. There is only one unit, shared between both inputs and outputs.

import openmdao.api as om
from openmdao.utils.units import convert_units

n = 10

model = om.Group()


[2.7432]