"""Define the ExecComp class, a component that evaluates an expression."""
import re
import time
from itertools import product
from contextlib import contextmanager
import numpy as np
from numpy import ndarray, imag
from openmdao.core.system import _DEFAULT_COLORING_META
from openmdao.utils.coloring import _ColSparsityJac, _compute_coloring
from openmdao.core.constants import INT_DTYPE
from openmdao.core.explicitcomponent import ExplicitComponent
from openmdao.utils.units import valid_units
from openmdao.utils import cs_safe
from openmdao.utils.om_warnings import issue_warning, DerivativesWarning, SetupWarning
from openmdao.utils.array_utils import get_random_arr
# regex to check for variable names.
VAR_RGX = re.compile(r'([.]*[_a-zA-Z]\w*[ ]*\(?)')
# Names of metadata entries allowed for ExecComp variables.
_allowed_meta = {'value', 'val', 'shape', 'units', 'res_units', 'desc',
'ref', 'ref0', 'res_ref', 'lower', 'upper', 'src_indices',
'flat_src_indices', 'tags', 'shape_by_conn', 'copy_shape', 'compute_shape',
'constant'}
# Names that are not allowed for input or output variables (keywords for options)
_disallowed_names = {'has_diag_partials', 'units', 'shape', 'shape_by_conn', 'run_root_only',
'constant', 'do_coloring'}
[docs]def check_option(option, value):
"""
Check option for validity.
Parameters
----------
option : str
The name of the option.
value : any
The value of the option.
Raises
------
ValueError
"""
if option == 'units' and value is not None and not valid_units(value):
raise ValueError("The units '%s' are invalid." % value)
[docs]def array_idx_iter(shape):
"""
Return an iterator over the indices into a n-dimensional array.
Parameters
----------
shape : tuple
Shape of the array.
Yields
------
int
"""
for p in product(*[range(s) for s in shape]):
yield p
[docs]class ExecComp(ExplicitComponent):
"""
A component defined by an expression string.
Parameters
----------
exprs : str, tuple of str or list of str
An assignment statement or iter of them. These express how the
outputs are calculated based on the inputs. In addition to
standard Python operators, a subset of numpy and scipy functions
is supported.
**kwargs : dict of named args
Initial values of variables can be set by setting a named
arg with the var name. If the value is a dict it is assumed
to contain metadata. To set the initial value in addition to
other metadata, assign the initial value to the 'val' entry
of the dict.
Attributes
----------
_kwargs : dict of named args
Initial values of variables.
_exprs : list
List of expressions.
_codes : list
List of code objects.
_exprs_info : list
List of tuples containing output and inputs for each expression.
complex_stepsize : double
Step size used for complex step which is used for derivatives.
_manual_decl_partials : bool
If True, at least one partial has been declared by the user.
_requires_fd : dict
Contains a mapping of 'of' variables to a tuple of the form (wrts, functs) for those
'of' variables that require finite difference to be used to compute their derivatives.
_constants : dict of dicts
Constants defined in the expressions. The key is the name of the constant and the value
is a dict of metadata.
_coloring_declared : bool
If True, coloring has been declared manually.
_inarray : ndarray or None
If using internal CS, this is a complex array containing input values.
_outarray : ndarray or None
If using internal CS, this is a complex array containing output values.
_indict : dict or None
If using internal CS, this maps input variable views in _inarray to input names.
_viewdict : dict or None
If using internal CS, this maps input, output, and constant names to their corresponding
views/values.
"""
[docs] def __init__(self, exprs=[], **kwargs):
r"""
Create a <Component> using only an expression string.
Given a list of assignment statements, this component creates
input and output variables at construction time. All variables
appearing on the left-hand side of an assignment are outputs,
and the rest are inputs. Each variable is assumed to be of
type float unless the initial value for that variable is supplied
in \*\*kwargs. Derivatives are calculated using complex step.
The following functions are available for use in expressions:
========================= ====================================
Function Description
========================= ====================================
abs(x) Absolute value of x
acos(x) Inverse cosine of x
acosh(x) Inverse hyperbolic cosine of x
arange(start, stop, step) Array creation
arccos(x) Inverse cosine of x
arccosh(x) Inverse hyperbolic cosine of x
arcsin(x) Inverse sine of x
arcsinh(x) Inverse hyperbolic sine of x
arctan(x) Inverse tangent of x
arctan2(y, x) 4-quadrant arctangent function of y and x
asin(x) Inverse sine of x
asinh(x) Inverse hyperbolic sine of x
atan(x) Inverse tangent of x
cos(x) Cosine of x
cosh(x) Hyperbolic cosine of x
dot(x, y) Dot product of x and y
e Euler's number
erf(x) Error function
erfc(x) Complementary error function
exp(x) Exponential function
expm1(x) exp(x) - 1
fmax(x, y) Element-wise maximum of x and y
fmin(x, y) Element-wise minimum of x and y
inner(x, y) Inner product of arrays x and y
isinf(x) Element-wise detection of np.inf
isnan(x) Element-wise detection of np.nan
kron(x, y) Kronecker product of arrays x and y
linspace(x, y, N) Numpy linear spaced array creation
log(x) Natural logarithm of x
log10(x) Base-10 logarithm of x
log1p(x) log(1+x)
matmul(x, y) Matrix multiplication of x and y
maximum(x, y) Element-wise maximum of x and y
minimum(x, y) Element-wise minimum of x and y
ones(N) Create an array of ones
outer(x, y) Outer product of x and y
pi Pi
power(x, y) Element-wise x**y
prod(x) The product of all elements in x
sin(x) Sine of x
sinh(x) Hyperbolic sine of x
sum(x) The sum of all elements in x
tan(x) Tangent of x
tanh(x) Hyperbolic tangent of x
tensordot(x, y) Tensor dot product of x and y
zeros(N) Create an array of zeros
========================= ====================================
Notes
-----
If a variable has an initial value that is anything other than 1.0,
either because it has a different type than float or just because its
initial value is != 1.0, you must use a keyword arg
to set the initial value. For example, let's say we have an
ExecComp that takes an array 'x' as input and outputs a float variable
'y' which is the sum of the entries in 'x'.
.. code-block:: python
import numpy
import openmdao.api as om
excomp = om.ExecComp('y=sum(x)', x=numpy.ones(10, dtype=float))
In this example, 'y' would be assumed to be the default type of float
and would be given the default initial value of 1.0, while 'x' would be
initialized with a size 10 float array of ones.
If you want to assign certain metadata for 'x' in addition to its
initial value, you can do it as follows:
.. code-block:: python
excomp = ExecComp('y=sum(x)',
x={'val': numpy.ones(10, dtype=float),
'units': 'ft'})
"""
options = {}
for name in _disallowed_names:
if name in kwargs:
options[name] = kwargs.pop(name)
super().__init__(**options)
# change default coloring values
self._coloring_info.method = 'cs'
self._coloring_info.num_full_jacs = 2
# if complex step is used for derivatives, this is the stepsize
self.complex_stepsize = 1.e-40
if isinstance(exprs, str):
exprs = [exprs]
self._exprs = exprs[:]
self._exprs_info = []
self._codes = []
self._kwargs = kwargs
self._manual_decl_partials = False
self._no_check_partials = True
self._constants = {}
self._coloring_declared = False
self._inarray = None
self._outarray = None
self._indict = None
self._viewdict = None
[docs] def initialize(self):
"""
Declare options.
"""
self.options.declare('has_diag_partials', types=bool, default=False,
desc='If True, treat all array/array partials as diagonal if both '
'arrays have size > 1. All arrays with size > 1 must have the '
'same flattened size or an exception will be raised.')
self.options.declare('units', types=str, allow_none=True, default=None,
desc='Units to be assigned to all variables in this component. '
'Default is None, which means units may be provided for variables'
' individually.',
check_valid=check_option)
self.options.declare('shape', types=(int, tuple, list), allow_none=True, default=None,
desc='Shape to be assigned to all variables in this component. '
'Default is None, which means shape may be provided for variables'
' individually.')
self.options.declare('shape_by_conn', types=bool, default=False,
desc='If True, shape all inputs and outputs based on their '
'connection. Default is False.')
self.options.declare('do_coloring', types=bool, default=True,
desc='If True (the default), compute the partial jacobian '
'coloring for this component.')
self.options.undeclare("distributed")
[docs] @classmethod
def register(cls, name, callable_obj, complex_safe):
"""
Register a callable to be usable within ExecComp expressions.
Parameters
----------
name : str
Name of the callable.
callable_obj : callable
The callable.
complex_safe : bool
If True, the given callable works correctly with complex numbers.
"""
global _expr_dict, _not_complex_safe
if not callable(callable_obj):
raise TypeError(f"{cls.__name__}: '{name}' passed to register() of type "
f"'{type(callable_obj).__name__}' is not callable.")
if name in _expr_dict:
raise NameError(f"{cls.__name__}: '{name}' has already been registered.")
if name in _disallowed_names:
raise NameError(f"{cls.__name__}: cannot register name '{name}' because "
"it's a reserved keyword.")
if '.' in name:
raise NameError(f"{cls.__name__}: cannot register name '{name}' because "
"it contains '.'.")
_expr_dict[name] = callable_obj
if not complex_safe:
_not_complex_safe.add(name)
[docs] def setup(self):
"""
Set up variable name and metadata lists.
"""
if self._exprs:
self._setup_expressions()
def _setup_expressions(self):
"""
Set up the expressions.
This is called during setup_procs and after each call to "add_expr" from configure.
"""
global _not_complex_safe
exprs = self._exprs
kwargs = self._kwargs
shape = self.options['shape']
shape_by_conn = self.options['shape_by_conn']
if shape is not None and shape_by_conn:
raise RuntimeError(f"{self.msginfo}: Can't set both shape and shape_by_conn.")
self._exprs_info = exprs_info = []
outs = set()
allvars = set()
constants = set()
self._requires_fd = {}
for expr in exprs:
lhs, _, rhs = expr.partition('=')
onames = self._parse_for_out_vars(lhs)
vnames, fnames = self._parse_for_names(rhs)
constants.update([n for n, val in kwargs.items()
if isinstance(val, dict) and 'constant' in val and val['constant']])
# remove constants
vnames = vnames.difference(constants)
allvars.update(vnames)
outs.update(onames)
if onames.intersection(allvars):
# we have a used-before-calculated output
violators = sorted([n for n in onames if n in allvars])
raise RuntimeError(f"{self.msginfo}: Outputs {violators} are used before "
"being calculated, so this ExecComp is not a valid explicit "
"component.")
exprs_info.append((onames, vnames, fnames))
if _not_complex_safe.intersection(fnames):
for o in onames:
self._requires_fd[o] = (vnames, fnames)
allvars.update(outs)
if self._requires_fd:
inps = []
for out, (rhsvars, funcs) in self._requires_fd.items():
iset = rhsvars.difference(outs)
self._requires_fd[out] = (iset, funcs)
inps.extend(iset)
self._no_check_partials = False
self.set_check_partial_options(wrt=inps, method='fd')
kwargs2 = {}
init_vals = {}
units = self.options['units']
# make sure all kwargs are legit
for varname, val in kwargs.items():
if varname not in allvars and varname not in constants:
msg = f"{self.msginfo}: arg '{varname}' in call to ExecComp() " \
f"does not refer to any variable in the expressions {exprs}"
if varname in ('promotes', 'promotes_inputs', 'promotes_outputs'):
msg += ". Did you intend to promote variables in the 'add_subsystem' call?"
raise RuntimeError(msg)
if isinstance(val, dict):
dct = val
vval = dct.get('val')
vshape = dct.get('shape')
vshape_by_conn = dct.get('shape_by_conn')
vcopy_shape = dct.get('copy_shape')
vcompute_shape = dct.get('compute_shape')
vconstant = dct.get('constant')
vunits = dct.get('units')
if vconstant:
if vval is None:
raise RuntimeError(f"{self.msginfo}: arg '{varname}' in call to ExecComp() "
"is a constant but no value is given")
for ignored_meta in ['units', 'shape']:
if ignored_meta in dct:
issue_warning(f"arg '{varname}' in call to ExecComp() "
f"is a constant. The {ignored_meta} will be ignored",
prefix=self.msginfo, category=SetupWarning)
self._constants[varname] = vval
continue # TODO should still do some checking here!
diff = set(dct.keys()) - _allowed_meta
if diff:
raise RuntimeError("%s: the following metadata names were not "
"recognized for variable '%s': %s" %
(self.msginfo, varname, sorted(diff)))
kwargs2[varname] = dct.copy()
if units is not None:
if vunits is not None and vunits != units:
raise RuntimeError("%s: units of '%s' have been specified for "
"variable '%s', but units of '%s' have been "
"specified for the entire component." %
(self.msginfo, vunits, varname, units))
else:
kwargs2[varname]['units'] = units
if shape is not None:
if vshape is not None and vshape != shape:
raise RuntimeError("%s: shape of %s has been specified for "
"variable '%s', but shape of %s has been "
"specified for the entire component." %
(self.msginfo, vshape, varname, shape))
elif vval is not None and np.atleast_1d(vval).shape != shape:
raise RuntimeError("%s: value of shape %s has been specified for "
"variable '%s', but shape of %s has been "
"specified for the entire component." %
(self.msginfo, np.atleast_1d(vval).shape,
varname, shape))
else:
init_vals[varname] = np.ones(shape)
if vval is not None:
init_vals[varname] = vval
del kwargs2[varname]['val']
if vshape_by_conn or vcopy_shape or vcompute_shape:
if vshape is not None or vval is not None:
raise RuntimeError(f"{self.msginfo}: Can't set 'shape' or 'val' for "
f"variable '{varname}' along with 'copy_shape', "
"compute_shape, or 'shape_by_conn'.")
if vshape is not None:
if varname not in init_vals:
init_vals[varname] = np.ones(vshape)
elif np.atleast_1d(init_vals[varname]).shape != vshape:
raise RuntimeError("%s: shape of %s has been specified for variable "
"'%s', but a value of shape %s has been provided." %
(self.msginfo, str(vshape), varname,
str(np.atleast_1d(init_vals[varname]).shape)))
del kwargs2[varname]['shape']
else:
init_vals[varname] = val
if self._static_mode:
var_rel2meta = self._static_var_rel2meta
else:
var_rel2meta = self._var_rel2meta
for var in sorted(allvars):
meta = kwargs2.get(var, {
'units': units,
'shape': shape,
'shape_by_conn': shape_by_conn})
# if user supplied an initial value, use it, otherwise set to 1.0
if var in init_vals:
val = init_vals[var]
else:
val = 1.0
if var in var_rel2meta:
# Input/Output already exists, but we may be setting defaults for the first time.
# Note that there is only one submitted dictionary of defaults.
current_meta = var_rel2meta[var]
for kname, kvalue in meta.items():
if kvalue is not None:
current_meta[kname] = kvalue
new_val = kwargs[var].get('val')
if new_val is not None:
# val is normally ensured to be a numpy array in add_input/add_output,
# do the same here...
current_meta['val'] = np.atleast_1d(new_val)
else:
# new input and/or output.
if var in outs:
current_meta = self.add_output(var, val, **meta)
else:
if 'constant' in meta:
meta.pop('constant', None)
current_meta = self.add_input(var, val, **meta)
if var not in init_vals:
init_vals[var] = current_meta['val']
self._codes = self._compile_exprs(self._exprs)
[docs] def add_expr(self, expr, **kwargs):
"""
Add an expression to the ExecComp.
Parameters
----------
expr : str
An assignment statement that expresses how the outputs are calculated based on the
inputs. In addition to standard Python operators, a subset of numpy and scipy
functions is supported.
**kwargs : dict of named args
Initial values of variables can be set by setting a named arg with the var name. If
the value is a dict it is assumed to contain metadata. To set the initial value in
addition to other metadata, assign the initial value to the 'val' entry of the dict.
Do not include for inputs whose default kwargs have been declared on previous
expressions.
"""
if not isinstance(expr, str):
typ = type(expr).__name__
msg = f"Argument 'expr' must be of type 'str', but type '{typ}' was found."
raise TypeError(msg)
self._exprs.append(expr)
for name in kwargs:
if name in self._kwargs:
raise NameError(f"Defaults for '{name}' have already been defined in a previous "
"expression.")
self._kwargs.update(kwargs)
if not self._static_mode:
self._setup_expressions()
def _compile_exprs(self, exprs):
compiled = []
outputs = set()
for i, expr in enumerate(exprs):
# Quick dupe check.
lhs_name = expr.partition('=')[0].strip()
if lhs_name in outputs:
# Can't add two equations with the same output.
raise RuntimeError(f"{self.msginfo}: The output '{lhs_name}' has already been "
"defined by an expression.")
else:
outputs.add(lhs_name)
try:
compiled.append(compile(expr, expr, 'exec'))
except Exception:
raise RuntimeError("%s: failed to compile expression '%s'." %
(self.msginfo, exprs[i]))
return compiled
def _parse_for_out_vars(self, s):
vnames = set([x.strip() for x in re.findall(VAR_RGX, s)
if not x.endswith('(') and not x.startswith('.')])
for v in vnames:
if v in _expr_dict:
raise NameError("%s: cannot assign to variable '%s' "
"because it's already defined as an internal "
"function or constant." % (self.msginfo, v))
return vnames
def _parse_for_names(self, s):
names = [x.strip() for x in re.findall(VAR_RGX, s) if not x.startswith('.')]
vnames = set()
for n in names:
if n.endswith('('):
continue
vnames.add(n)
fnames = [n[:-1] for n in names if n[-1] == '(']
to_remove = []
for v in vnames:
if v in _disallowed_names:
raise NameError("%s: cannot use variable name '%s' because "
"it's a reserved keyword." % (self.msginfo, v))
if v in _expr_dict:
expvar = _expr_dict[v]
if callable(expvar):
raise NameError("%s: cannot use '%s' as a variable because "
"it's already defined as an internal "
"function or constant." % (self.msginfo, v))
else:
to_remove.append(v)
for f in fnames:
if f not in _expr_dict:
raise NameError(f"{self.msginfo}: can't use '{f}' as a function because "
"it hasn't been registered.")
return vnames.difference(to_remove), fnames
def __getstate__(self):
"""
Return state as a dict.
Returns
-------
dict
State to get.
"""
state = self.__dict__.copy()
del state['_codes']
return state
def __setstate__(self, state):
"""
Restore state from `state`.
Parameters
----------
state : dict
State to restore.
"""
self.__dict__.update(state)
self._codes = self._compile_exprs(self._exprs)
[docs] def declare_partials(self, *args, **kwargs):
"""
Declare information about this component's subjacobians.
Parameters
----------
*args : list
Positional args to be passed to base class version of declare_partials.
**kwargs : dict
Keyword args to be passed to base class version of declare_partials.
Returns
-------
dict
Metadata dict for the specified partial(s).
"""
if 'method' not in kwargs or kwargs['method'] not in ('cs', 'fd'):
raise RuntimeError(f"{self.msginfo}: declare_partials must be called with method='cs' "
"or method='fd'.")
if self.options['has_diag_partials']:
raise RuntimeError(f"{self.msginfo}: declare_partials cannot be called manually if "
"has_diag_partials has been set.")
self._manual_decl_partials = True
return super().declare_partials(*args, **kwargs)
def _get_coloring(self):
"""
Get the Coloring for this system.
If necessary, load the Coloring from a file or dynamically generate it.
Returns
-------
Coloring or None
Coloring object, possible loaded from a file or dynamically generated, or None
"""
if self.options['do_coloring']:
return super()._get_coloring()
def _setup_partials(self):
"""
Check that all partials are declared.
"""
has_diag_partials = self.options['has_diag_partials']
if not self._manual_decl_partials:
if self.options['do_coloring'] and not has_diag_partials:
rank = self.comm.rank
sizes = self._var_sizes
if not self._has_distrib_vars and (sum(sizes['input'][rank]) > 1 and
sum(sizes['output'][rank]) > 1):
if not self._coloring_declared:
super().declare_coloring(wrt=('*', ), method='cs', show_summary=False)
self._coloring_info.dynamic = True
self._manual_decl_partials = False # this gets reset in declare_partials
self._declared_partials_patterns = {}
else:
self.options['do_coloring'] = False
self._coloring_info.dynamic = False
meta = self._var_rel2meta
decl_partials = super().declare_partials
for outs, vs, _ in self._exprs_info:
ins = sorted(set(vs).difference(outs))
for out in sorted(outs):
for inp in ins:
if has_diag_partials:
ival = meta[inp]['val']
oval = meta[out]['val']
iarray = isinstance(ival, ndarray) and ival.size > 1
if iarray and isinstance(oval, ndarray) and oval.size > 1:
if oval.size != ival.size:
raise RuntimeError(
"%s: has_diag_partials is True but partial(%s, %s) "
"is not square (shape=(%d, %d))." %
(self.msginfo, out, inp, oval.size, ival.size))
# partial will be declared as diagonal
inds = np.arange(oval.size, dtype=INT_DTYPE)
else:
inds = None
decl_partials(of=out, wrt=inp, rows=inds, cols=inds)
else:
decl_partials(of=out, wrt=inp)
super()._setup_partials()
if self._manual_decl_partials:
undeclared = []
for outs, vs, _ in self._exprs_info:
ins = sorted(set(vs).difference(outs))
for out in sorted(outs):
out = '.'.join((self.pathname, out))
for inp in ins:
inp = '.'.join((self.pathname, inp))
if (out, inp) not in self._subjacs_info:
undeclared.append((out, inp))
if undeclared:
idx = len(self.pathname) + 1
undeclared = ', '.join([' wrt '.join((f"'{of[idx:]}'", f"'{wrt[idx:]}'"))
for of, wrt in undeclared])
issue_warning(f"The following partial derivatives have not been "
f"declared so they are assumed to be zero: [{undeclared}].",
prefix=self.msginfo, category=DerivativesWarning)
def _setup_vectors(self, root_vectors):
"""
Compute all vectors for all vec names.
Parameters
----------
root_vectors : dict of dict of Vector
Root vectors: first key is 'input', 'output', or 'residual'; second key is vec_name.
"""
super()._setup_vectors(root_vectors)
if not self._use_derivatives:
self._manual_decl_partials = True # prevents attempts to use _viewdict in compute
self._iodict = _IODict(self._outputs, self._inputs, self._constants)
self._relcopy = False
if not self._manual_decl_partials:
if self._force_alloc_complex:
# we can use the internal Vector complex arrays
# set complex_step_mode so we'll get the full complex array
self._inputs.set_complex_step_mode(True)
self._outputs.set_complex_step_mode(True)
self._indict = self._inputs._get_local_views()
outdict = self._outputs._get_local_views()
self._inarray = self._inputs.asarray(copy=False)
self._outarray = self._outputs.asarray(copy=False)
self._inputs.set_complex_step_mode(False)
self._outputs.set_complex_step_mode(False)
else:
# we make our own complex 'copy' of the Vector arrays
self._inarray = np.zeros(len(self._inputs), dtype=complex)
self._outarray = np.zeros(len(self._outputs), dtype=complex)
self._indict = self._inputs._get_local_views(self._inarray)
outdict = self._outputs._get_local_views(self._outarray)
self._relcopy = True
# combine lookup dicts for faster exec calls
viewdict = self._indict.copy()
viewdict.update(outdict)
viewdict.update(self._constants)
self._viewdict = _ViewDict(viewdict)
[docs] def compute(self, inputs, outputs):
"""
Execute this component's assignment statements.
Parameters
----------
inputs : `Vector`
`Vector` containing inputs.
outputs : `Vector`
`Vector` containing outputs.
"""
if not self._manual_decl_partials:
if self._relcopy:
self._inarray[:] = self._inputs.asarray(copy=False)
self._exec()
outs = outputs.asarray(copy=False)
if outs.dtype == self._outarray.dtype:
outs[:] = self._outarray
else:
outs[:] = self._outarray.real
else:
self._exec()
return
if self._iodict._inputs is not inputs:
self._iodict = _IODict(outputs, inputs, self._constants)
for i, expr in enumerate(self._codes):
try:
# inputs, outputs, and _constants are vectors
exec(expr, _expr_dict, self._iodict) # nosec:
# limited to _expr_dict
except Exception as err:
raise RuntimeError(f"{self.msginfo}: Error occurred evaluating '{self._exprs[i]}':"
f"\n{err}")
def _linearize(self, jac=None, sub_do_ln=False):
"""
Compute jacobian / factorization. The model is assumed to be in a scaled state.
Parameters
----------
jac : Jacobian or None
Ignored.
sub_do_ln : bool
Flag indicating if the children should call linearize on their linear solvers.
"""
if self._requires_fd:
if 'fd' in self._approx_schemes:
fdins = {wrt.rsplit('.', 1)[1] for wrt in self._approx_schemes['fd']._wrt_meta}
else:
fdins = set()
for _, (inps, funcs) in self._requires_fd.items():
diff = inps.difference(fdins)
if diff:
raise RuntimeError(f"{self.msginfo}: expression contains functions "
f"{sorted(funcs)} that are not complex safe. To fix this, "
f"call declare_partials('*', {sorted(diff)}, method='fd') "
f"on this component prior to setup.")
self._requires_fd = False # only need to do this check the first time around
super()._linearize(jac, sub_do_ln)
[docs] def declare_coloring(self,
wrt=_DEFAULT_COLORING_META['wrt_patterns'],
method=_DEFAULT_COLORING_META['method'],
form=None,
step=None,
per_instance=_DEFAULT_COLORING_META['per_instance'],
num_full_jacs=_DEFAULT_COLORING_META['num_full_jacs'],
tol=_DEFAULT_COLORING_META['tol'],
orders=_DEFAULT_COLORING_META['orders'],
perturb_size=_DEFAULT_COLORING_META['perturb_size'],
min_improve_pct=_DEFAULT_COLORING_META['min_improve_pct'],
show_summary=_DEFAULT_COLORING_META['show_summary'],
show_sparsity=_DEFAULT_COLORING_META['show_sparsity']):
"""
Set options for deriv coloring of a set of wrt vars matching the given pattern(s).
Parameters
----------
wrt : str or list of str
The name or names of the variables that derivatives are taken with respect to.
This can contain input names, output names, or glob patterns.
method : str
Method used to compute derivative: "fd" for finite difference, "cs" for complex step.
form : str
Finite difference form, can be "forward", "central", or "backward". Leave
undeclared to keep unchanged from previous or default value.
step : float
Step size for finite difference. Leave undeclared to keep unchanged from previous
or default value.
per_instance : bool
If True, a separate coloring will be generated for each instance of a given class.
Otherwise, only one coloring for a given class will be generated and all instances
of that class will use it.
num_full_jacs : int
Number of times to repeat partial jacobian computation when computing sparsity.
tol : float
Tolerance used to determine if an array entry is nonzero during sparsity determination.
orders : int
Number of orders above and below the tolerance to check during the tolerance sweep.
perturb_size : float
Size of input/output perturbation during generation of sparsity.
min_improve_pct : float
If coloring does not improve (decrease) the number of solves more than the given
percentage, coloring will not be used.
show_summary : bool
If True, display summary information after generating coloring.
show_sparsity : bool
If True, display sparsity with coloring info after generating coloring.
"""
super().declare_coloring(wrt, method, form, step, per_instance, num_full_jacs,
tol, orders, perturb_size, min_improve_pct,
show_summary, show_sparsity)
self._coloring_declared = True
self._manual_decl_partials = True
def _exec(self):
for i, expr in enumerate(self._codes):
try:
exec(expr, _expr_dict, self._viewdict) # nosec:
except Exception as err:
raise RuntimeError(f"{self.msginfo}: Error occurred evaluating "
f"'{self._exprs[i]}':\n{err}")
def _compute_coloring(self, recurse=False, **overrides):
"""
Compute a coloring of the partial jacobian.
This assumes that the current System is in a proper state for computing derivatives.
Parameters
----------
recurse : bool
Ignored.
**overrides : dict
Any args that will override either default coloring settings or coloring settings
resulting from an earlier call to declare_coloring.
Returns
-------
list of Coloring
The computed colorings.
"""
if self._manual_decl_partials:
# use framework approx coloring
return super()._compute_coloring(**overrides)
info = self._coloring_info
info.update(overrides)
if not self._coloring_declared and info['method'] is None:
info['method'] = 'cs'
if info['method'] != 'cs':
raise RuntimeError(f"{self.msginfo}: 'method' for coloring must be 'cs' if partials "
"and/or coloring are not declared manually using declare_partials "
"or declare_coloring.")
if info.coloring is None and info.static is None:
info['dynamic'] = True
# match everything
info['wrt_matches'] = None
sparsity_start_time = time.perf_counter()
step = self.complex_stepsize * 1j
inv_stepsize = 1.0 / self.complex_stepsize
inarr = self._inarray
oarr = self._outarray
if self.options['has_diag_partials']:
# we should never get here
raise NotImplementedError("has_diag_partials not supported with coloring yet")
# compute perturbations
starting_inputs = self._inputs.asarray(copy=not self._relcopy)
in_offsets = starting_inputs.copy()
in_offsets[in_offsets == 0.0] = 1.0
in_offsets *= info['perturb_size']
# use special sparse jacobian to collect sparsity info
jac = _ColSparsityJac(self, info)
for i in range(info['num_full_jacs']):
inarr[:] = starting_inputs + in_offsets * get_random_arr(in_offsets.size, self.comm)
for i in range(inarr.size):
inarr[i] += step
self._exec()
jac.set_col(self, i, imag(oarr * inv_stepsize))
inarr[i] -= step
if not self._relcopy:
self._inputs.set_val(starting_inputs)
sparsity, sp_info = jac.get_sparsity(self)
sparsity_time = time.perf_counter() - sparsity_start_time
coloring = _compute_coloring(sparsity, 'fwd')
if not self._finalize_coloring(coloring, info, sp_info, sparsity_time):
return [None]
# compute mapping of col index to wrt varname
self._col_idx2name = idxnames = [None] * len(self._inputs)
plen = len(self.pathname) + 1
for name, slc in self._inputs.get_slice_dict().items():
name = name[plen:]
for i in range(slc.start, slc.stop):
idxnames[i] = name
# get slice dicts using relative name keys
self._out_slices = {n[plen:]: slc for n, slc in self._outputs.get_slice_dict().items()}
self._in_slices = {n[plen:]: slc for n, slc in self._inputs.get_slice_dict().items()}
return [coloring]
def _compute_colored_partials(self, partials):
"""
Use complex step method with coloring to update the given Jacobian.
Parameters
----------
partials : `Jacobian`
Contains sub-jacobians.
"""
step = self.complex_stepsize * 1j
inv_stepsize = 1.0 / self.complex_stepsize
inarr = self._inarray
oarr = self._outarray
out_names = self._var_rel_names['output']
inarr[:] = self._inputs.asarray(copy=False)
scratch = np.zeros(oarr.size)
idx2name = self._col_idx2name
out_slices = self._out_slices
in_slices = self._in_slices
for icols, nzrowlists in self._coloring_info.coloring.color_nonzero_iter('fwd'):
# set a complex input value
inarr[icols] += step
# solve with complex input value
self._exec()
imag_oar = imag(oarr * inv_stepsize)
for icol, rows in zip(icols, nzrowlists):
scratch[rows] = imag_oar[rows]
input_name = idx2name[icol]
loc_i = icol - in_slices[input_name].start
for u in out_names:
key = (u, input_name)
if key in partials:
# set the column in the Jacobian entry
part = scratch[out_slices[u]]
partials[key][:, loc_i] = part
part[:] = 0.
# restore old input value
inarr[icols] -= step
[docs] def compute_partials(self, inputs, partials):
"""
Use complex step method to update the given Jacobian.
Parameters
----------
inputs : Vector or dict
Vector containing parameters (p).
partials : `Jacobian`
Contains sub-jacobians.
"""
if self._manual_decl_partials:
return
if self.under_complex_step:
raise RuntimeError(f"{self.msginfo}: Can't compute complex step partials when higher "
"level system is using complex step unless you manually call "
"declare_partials and/or declare_coloring on this ExecComp.")
if self._coloring_info.coloring is not None:
self._compute_colored_partials(partials)
return
step = self.complex_stepsize * 1j
out_names = self._var_rel_names['output']
inv_stepsize = 1.0 / self.complex_stepsize
has_diag_partials = self.options['has_diag_partials']
inarr = self._inarray
indict = self._indict
vdict = self._viewdict
inarr[:] = self._inputs.asarray(copy=False)
for inp, ival in indict.items():
psize = ival.size
if has_diag_partials or psize == 1:
# set a complex inpup value
ival += step
# solve with complex input value
self._exec()
for u in out_names:
if (u, inp) in partials:
partials[u, inp] = imag(vdict[u] * inv_stepsize).flat
# restore old input value
ival -= step
else:
for i, idx in enumerate(array_idx_iter(ival.shape)):
# set a complex input value
ival[idx] += step
# solve with complex input value
self._exec()
for u in out_names:
if (u, inp) in partials:
# set the column in the Jacobian entry
partials[u, inp][:, i] = imag(vdict[u] * inv_stepsize).flat
# restore old input value
ival[idx] -= step
class _ViewDict(object):
def __init__(self, dct):
self.dct = dct
def __getitem__(self, name):
return self.dct[name]
def __setitem__(self, name, value):
try:
self.dct[name][:] = value
except ValueError:
# see if value fits if size 1 dimensions are removed
sqz = np.squeeze(value)
if np.squeeze(self.dct[name]).shape == sqz.shape:
self.dct[name][:] = sqz
else:
raise
# need __contains__ here else we get weird KeyErrors in certain situations when evaluating
# the compiled expressions. Non-compiled expressions evaluate just fine, but after compilation,
# and only in rare circumstances (like running under om trace), KeyErrors for 0, 1, ...
# are mysteriously generated.
def __contains__(self, name):
return name in self.dct
class _IODict(object):
"""
A dict wrapper that contains 2 different dicts.
Items are first looked for in the outputs
and then the inputs.
Attributes
----------
_inputs : dict-like
The inputs object to be wrapped.
_outputs : dict-like
The outputs object to be wrapped.
_constants : dict-like
The constants object to be wrapped.
"""
def __init__(self, outputs, inputs, constants):
"""
Create the dict wrapper.
Parameters
----------
outputs : dict-like
The outputs object to be wrapped.
inputs : dict-like
The inputs object to be wrapped.
constants : dict-like
The constants object to be wrapped.
"""
self._outputs = outputs
self._inputs = inputs
self._constants = constants
def __getitem__(self, name):
try:
return self._inputs[name]
except KeyError:
try:
return self._outputs[name]
except KeyError:
return self._constants[name]
def __setitem__(self, name, value):
try:
self._outputs[name][:] = value
except ValueError:
# see if value fits if size 1 dimensions are removed
sqz = np.squeeze(value)
if np.squeeze(self._outputs[name]).shape == sqz.shape:
self._outputs[name][:] = sqz
else:
raise
def __contains__(self, name):
return name in self._inputs or name in self._outputs or name in self._constants
def _import_functs(mod, dct, names=None):
"""
Map attributes attrs from the given module into the given dict.
Parameters
----------
mod : object
Module to check.
dct : dict
Dictionary that will contain the mapping
names : iter of str, optional
If supplied, only map attrs that match the given names
"""
if names is None:
names = dir(mod)
for name in names:
if isinstance(name, tuple):
name, alias = name
else:
alias = name
if not name[0] == '_':
dct[name] = getattr(mod, name)
dct[alias] = dct[name]
_expr_dict = {} # this dict will act as the local scope when we eval our expressions
_not_complex_safe = set() # this is the set of registered functions that are not complex safe
_import_functs(np, _expr_dict,
names=['arange', 'ones', 'zeros', 'linspace', # Array creation
'e', 'pi', # Constants
'isinf', 'isnan', # Logic
'log', 'log10', 'log1p', 'power', # Math operations
'exp', 'expm1', 'fmax', 'min', 'max', 'diff',
'fmin', 'maximum', 'minimum',
'sum', 'dot', 'prod', # Reductions
'tensordot', 'matmul', # Linear algebra
'outer', 'inner', 'kron',
'sin', 'cos', 'tan', ('arcsin', 'asin'), # Trig
('arccos', 'acos'), ('arctan', 'atan'),
'sinh', 'cosh', 'tanh', ('arcsinh', 'asinh'), # Hyperbolic trig
('arccosh', 'acosh')])
# if scipy is available, add some functions
try:
import scipy.special
except ImportError:
pass
else:
_import_functs(scipy.special, _expr_dict, names=['erf', 'erfc'])
# put any functions that need custom complex-safe versions here
_expr_dict['abs'] = cs_safe.abs
_expr_dict['arctan2'] = cs_safe.arctan2
class _NumpyMsg(object):
"""
A class that will raise an error if an attempt is made to access any attribute/function.
"""
def __init__(self, namespace):
"""
Construct the _NumpyMsg object.
Parameters
----------
namespace : str
The numpy namespace (e.g. 'numpy' or 'np).
"""
self.namespace = namespace
def __getattr__(self, name):
"""
Attempt to access an attribute/function.
Parameters
----------
name : str
The name of the attribute/function.
Raises
------
RuntimeError
When an attempt is made to access any attribute/function.
"""
raise RuntimeError('\n'.join([
" ExecComp supports a subset of numpy functions directly, without the '%s' prefix.",
" '%s' is %ssupported (See the documentation)."
]) % (self.namespace, name, '' if name in _expr_dict else 'not '))
_expr_dict['np'] = _NumpyMsg('np')
_expr_dict['numpy'] = _NumpyMsg('numpy')
@contextmanager
def _temporary_expr_dict():
"""
During a test, it's useful to be able to save and restore the _expr_dict.
"""
global _expr_dict, _not_complex_safe
save = (_expr_dict.copy(), _not_complex_safe.copy())
try:
yield
finally:
_expr_dict, _not_complex_safe = save