Second-Order Adjoint Sensitivity Analysis Methodology for Computing Exactly Response Sensitivities to Uncertain Parameters and Boundaries of Linear Systems: Mathematical Framework ()
1. Introduction
The earliest use of adjoint operators for computing exactly and efficiently the first-order sensitivities of responses of a large-scale linear system comprising many parameters has appeared in the report by Wigner [2] on his work on the “nuclear pile” (i.e., nuclear reactor) using the linear neutron transport or diffusion equations. Cacuci [2] [3] conceived the rigorous 1st-order adjoint sensitivity analysis methodology for generic large-scale nonlinear systems involving operator responses, comprising functional-type responses as particular cases. Representative developments of the adjoint sensitivity analysis methodology were reviewed in the books by Cacuci [4] and Cacuci, Ionescu-Bujor and Navon [5]. The first-order adjoint sensitivity analysis methodology was extended by Cacuci [6] [7] [8] to the “Second-Order Adjoint Sensitivity Analysis Methodology (2nd-ASAM)” for linear and nonlinear systems, to enable the exact and efficient computation of all of the second-order sensitivities (i.e., functional derivatives) of model responses to parameters. However, all of the above works, as well as the vast majority of the application of the first- and second-order adjoint sensitivity analysis methodology considered that the phase-space location of the physical model’s boundary is perfectly well known, and did not investigate the consequences of imprecisely known (i.e., uncertain) model boundaries.
Komata [9] seems to have been the first author to use an adjoint equation (in this case; the adjoint neutron diffusion equation) for investigating the effects of perturbations in boundary conditions for simple idealized geometries in neutron diffusion theory. First-order perturbation theory in conjunction with adjoint operators was subsequently used [10] - [15], either formally or in conjunction with variational formulations, to obtain approximate first-order sensitivities to boundary parameters of responses that were linear functionals (or ratios thereof) of the neutron flux in the context of linear neutron diffusion or neutron transport problems. Furthermore, the works referenced in the foregoing have considered the computation of response sensitivities to model parameters separately (rather than simultaneously) from the computations of response sensitivities to imprecisely known domain boundary parameters. Also, none of the above works considered responses that are simultaneously functions of the forward and adjoint state functions.
Noteworthy, Cacuci [16] [17] has presented methodologies for the exact computation of 1st- and 2nd-order sensitivities of responses that are functionals of both the forward and the adjoint fluxes in a multiplying nuclear system with source. However, these works [16] [17] have specifically considered the neutron transport equation (rather than generic linear problems) within precisely known domains, without considering response sensitivities to uncertain domain parameters.
The aim of this work is to present the novel “Second-Order Comprehensive Adjoint Sensitivity Analysis Methodology (2nd-CASAM),” which has the following features that generalize and extend all previously published works on this topic:
2) The system response considered within the 2nd-CASAM framework is an operator-valued response that depends on both the forward and adjoint state-functions. Functional-valued responses are subsumed as particular cases.
2) The 2nd-CASAM framework enables the efficient and exact computation of the 1st- and 2nd-order response sensitivities not only to uncertain model parameters but also to uncertain domain boundaries of generic linear systems.
This work is structured as follows: Section 2 presents the mathematical framework of the generic linear physical system comprising imprecisely known parameters and boundaries. Section 3 presents the mathematical framework of the 2nd-CASAM. Section 4 offers concluding remarks. An accompanying work [18] presents an illustrative application of the general 2nd-CASAM methodology to a paradigm model of generic particle/radiation transmission and/or evolution which has fundamental applications in many fields, including nuclear engineering (e.g., radiation detection, chemical reprocessing of spent reactor fuel, radioactive decay, etc.).
2. Mathematical Modeling of Response-Coupled Forward and Adjoint Linear Systems Comprising Uncertain Responses, Parameters and Boundaries
For the mathematical derivations to be presented in this work, the vectors are column vectors unless explicitly stated otherwise. Bold letters will be used to denote matrices and vectors. Transposition will be indicated by a dagger
superscript. A linear physical system is generally represented by means of
coupled operator equations of the form
in which the operators
act linearly on the state functions
. In matrix form, a linear system can be written as follows:
. (1)
The quantities appearing in the Equation (1) are defined as follows:
2)
denotes a
-dimensional column vector whose components are the physical system’s imprecisely known parameters, which are subject to uncertainties;
, where
denotes a
-dimensional subset of the set of real scalars. The vector
is considered to include any imprecisely known model parameters that may enter into defining the system’s boundary in the phase-se space of independent variables. The symbol “
” will be used to denote “is defined as” or “is by definition equal to.”
2)
denotes the
-dimensional phase-space position vector, defined on a phase-space domain denoted as
which is defined as follows:
. The lower boundary-point of an independent variable is denoted as
, where the superscript “
” denotes “lower” (e.g., the inner radius of a sphere or cylinder, the lower range of an energy-variable, etc.), while the corresponding upper boundary-point is denoted as
, where the superscript “u” denotes “upper” (e.g., the outer radius of a sphere or cylinder, the upper range of an energy-variable, etc.). A typical example of “boundaries that depend on imprecisely known parameters” is provided by the boundary conditions needed for models based on diffusion theory, in which the respective “flux and/or current conditions” for the “boundaries facing vacuum” are imposed on the “extrapolated boundary” of the respective spatial domain. As is well known, the “extrapolated boundary” depends not only on the imprecisely known physical dimensions of the problem’s domain, but also on the medium’s microscopic transport cross sections and atomic number densities. For subsequent derivations, it is convenient to define the “vectors of endpoints”
and
. The boundary of
, which will be denoted as
, comprises all of the endpoints
and
of the respective intervals on which the components of
are defined, i.e.,
.
3)
denotes a
-dimensional column vector whose components represent the system’s dependent variables (also called “state functions”);
4)
denotes a
-dimensional column vector defined on a vector space denoted as
, whose elements represent inhomogeneous source terms that depend either linearly or nonlinearly on
; the components of
may be operators, rather than just functions, acting on
and
;
5)
denotes a
-component column vector whose components are operators (including differential, difference, integral, distributions, and/or infinite matrices) acting linearly on
and nonlinearly on
.
6) All of the equalities in this work are considered to hold in the weak (“distributional”) sense, since the right-sides (“sources”) of Equation (1) and of other various equations to be derived in this work may contain distributions (“generalized functions/functionals”), particularly Dirac-distributions and derivatives and/or integrals thereof.
When differential operators appear in Equation (1), a corresponding set of boundary and/or initial conditions, which are essential to define the domain of
, must also be given. Since this work considers only systems that are linear in the state function
, the accompanying boundary and/or initial conditions must also be linear in
, which means that they can be represented in operator form as follows:
(2)
In Equation (2), the operator
is a matrix comprising, as components, operators that act linearly on
and nonlinearly on
; the quantity
denotes the total number of boundary and initial conditions. The operator
is a
-dimensional vector comprising components that are operators acting, in general, nonlinearly on
.
The nominal solution of Equations (1) and (2) is denoted as
, and is obtained by solving these equations at the nominal parameter values
. In other words, the vectors
and
satisfy the following equations:
(3)
(4)
Equations (3) and (4) represent the “base-case” or nominal state of the physical system. Throughout this work, the superscript “0” will be used to denote “nominal” or “expected” values.
The vector-valued function
is considered to be the unique nontrivial solution of the physical problem described by Equations (1) and (2). The linear operator
is considered in this work to admit an adjoint operator, which will be denoted as
. The specific form of the adjoint operator
depends upon the specific inner product and associated Hilbert space chosen for defining it. In view of Equations (1) and (2), the adjoint operator
is most often defined in a Hilbert space (which will henceforth be denoted as
) endowed with the following inner product, denoted as
, between two the vectors
and
, each having square integrable components and the same structure as the
-dimensional column vector
:
(5)
In Equation (5), the “product” notation
compactly denotes the respective multiple integrals, while the dot indicates the “scalar product of two vectors” defined as follows:
. (6)
Formally, the inner product introduced in Equation (5) defines a self-dual Hilbert space, which shall be denoted as
. It is important to note that the inner product defined in Equation (5) is continuous in
, holding at any value particular value of
in the neighborhood of the nominal parameter values
, including at
.
The formal adjoint operator
is defined through the following relationship, which holds for an arbitrary vector
:
(7)
In Equation (7), the formal adjoint operator
is the
matrix
(8)
comprising elements
obtained by transposing the formal adjoints of the operators
, while
denotes the associated bilinear form evaluated on
. The adjoint state function
is the solution of an adjoint system which can be written as follows:
(9)
where the source
for the adjoint equation is usually related to the system response under consideration. The domain of
is determined by selecting appropriate adjoint boundary and/or initial conditions, represented here in operator form as follows:
(10)
In Equation (10), the subscript “A” indicates “adjoint,” and the letter “B” indicates “boundary and/or initial conditions.” The nominal value of the adjoint state function
will be denoted as
, and is obtained by solving the adjoint system at the nominal parameter values
:
(11)
(12)
The system response (i.e., result of interest) associated with the problem modeled by Equations (1) and (2) is typically a real-valued nonlinear operator (function) of the system’s forward and adjoint state-functions (i.e., dependent variables) and parameters, denoted as
, which can be generally represented in the form
(13)
where
is a suitably differentiable function with respect to its arguments.
3. The Second-Order Comprehensive Adjoint Sensitivity Analysis Methodology (2nd-CASAM)
As has been already mentioned in the foregoing, the model and boundary parameters are considered to be imprecisely known quantities. Their true values may differ from their nominal values by quantities denoted as
, where
. Since the forward and the adjoint state functions
and
are related to the model and boundary parameters
through Equations (1), (2), (9) and (10), it follows that variations
in the model and boundary parameter will cause corresponding variations
around the nominal value
of the forward state function, as well as variations
around the nominal value
of the adjoint state function. In turn, the variations in the model’s parameters, boundaries and state functions will induce variations in the system’s response. For the derivations to follow, it is convenient to define the following vectors in the combined phase-space of state-functions, model and boundary parameters: 1) the vector
, comprising as components the state functions and parameters; 2) the vector
of nominal values of the state functions and parameters; 3) the vector
of variations
,
, and
in the state functions and parameters around their respective nominal values. The 2nd-CASAM formalism underlying the computation of the 1st-order sensitivities of the response
is presented in Section 3.2, while the mathematical methodology for computing the 2nd-order sensitivities is presented in Section 3.2.
3.1. Derivation of the 1st-Level Adjoint Sensitivity System (1st-LASS) for Computing Exactly and Efficiently the 1st-Order Response Sensitivities to Model and Boundary Parameters
As shown by Cacuci [2] [3], in front of the sentence, the most general definition of the 1st-order total sensitivity of the operator-valued model response
to variations
around the nominal values
, respectively, is provided by the first-order Gateaux-variation
of
, which is defined as follows:
(14)
for
, where
denotes the field of real scalars, and all (i.e., arbitrary) vectors
in a neighborhood
around
. The G-differential
is an operator defined on the same domain as
, has the same range as
, and provides the total 1st-order sensitivity of
with respect to variations in the model’s parameters and state functions. The G-differential
satisfies the relation
, with
.
The existence of the G-variation
does not guarantee its numerical computability. Numerical methods most often require that
be linear in
in a neighborhood
around
. The necessary and sufficient conditions for the G-variation
of a nonlinear operator
to be linear in
in a neighborhood
around
, and thus admit a total first-order G-derivative, are as follows:
1)
must satisfy a weak Lipschitz condition at
; (15)
2)
must satisfy the following condition for
:
(16)
Numerical methods (e.g., Newton’s method and variants thereof) for solving Equations (1) and (2) also require the existence of the 1st-order G-derivatives of original model equations, in which case the components of the operators which appear in these equations must also satisfy the conditions described in Equations (15) and (16). Of course, if the first-order G-derivatives of the system’s response do not exist, the computation of higher-order response sensitivities (G-derivatives) would be moot. Therefore, the conditions described in Equations (15) and (16) will henceforth be considered to be satisfied by the operators underlying the physical system, in which case the partial G-derivatives of
at
with respect to
,
and
exist. These derivatives are row vectors defined as
,
, and
, respectively, so that the first-order G-variation
can be written in the form
(17)
where
(18)
(19)
The notation
indicates that the quantity within the braces is to be evaluated at
. The quantity
in Equation (18) is called the “direct-effect term” because it can be computed after the base-case values
become available. On the other hand, the “indirect-effect term”
defined in Equation (19) can be quantified only after having determined the variations
and
in terms of the variations
. The first-order relationship between the variations
and
is determined by determining the G-differentials of Equations (1) and (2). Thus, applying the definition of the G-differential to Equations (1) and (2) yields the following equations:
(20)
(21)
where the superscript “(1)” indicates “1st-Level” and the letter “B” indicates “boundary and/or initial conditions,” and where the following definitions were used:
(22)
(23)
The partial G-derivatives appearing in Equation (22) are matrices defined as follows:
(24)
(25)
The partial G-derivatives that appear in Equation (23) are also matrices, having structures and components similar to those defined in Equations (24) and (25).
The first-order relationship between the variations
and
is determined by taking the G-differentials of Equations (9) and (10), which yields the following system of equations:
(26)
(27)
where the superscript “(1)” indicates “1st-Level,” the letter “A” indicates “adjoint,” the letter “B” indicates “boundary and/or initial conditions,” and where the following definitions were used:
(28)
(29)
The system of equations comprising Equations (20), (21), (26) and (27) are called the “1st-Level Forward Sensitivity System” (1st-LFSS). In principle, the 1st-LFSS could to be solved for each possible vector of parameter variations
in a neighborhood around
to obtain
and
as functions of
. In turn, the functions
and
thus obtained could be used in Equation (17) to compute the total response sensitivity
. Computing the (total) response sensitivity
by using the
-dependent solutions
and
of the 1st-LFSS is called the Forward Sensitivity Analysis Method/Procedure (FSAM). From the standpoint of computational costs and effort, the FSAM would require
large-scale computations to solve the 1st-LFSS. Therefore, the FSAM is advantageous to employ only if, in the problem under consideration, the number
of model and boundary parameters is considerably less than the number of responses of interest. This is rarely the case in practice, however, since most problems of practical interest are characterized by many model parameters and comparatively few responses. In such situations, it is not economical to employ the FSAM since it becomes prohibitively expensive to solve repeatedly the
-dependent 1st-LFSS in order to determine the variations
and
for all possible variations
.
In most practical situations, the number of model parameters exceeds significantly the number of responses of interest. In such cases, the exact computation of the 1st-order response sensitivities can be performed most efficiently by using the 1st-order adjoint sensitivity analysis methodology introduced by Cacuci [2] [3] and extended to 2nd-order by Cacuci [4] [5] [6], which expresses the indirect effect defined in Equation (35) in terms of a suitably defined 1st-level adjoint function, thus eliminating the need to compute the variations
and
by solving the 1st-LFSS. As has been first shown by Cacuci [3], the adjoint sensitivity analysis methodology cannot be applied directly to an operator-valued response, but only to responses that are functionals (i.e., scalar-valued operators) of the state functions. For this purpose, the operator-valued response
is expressed through its generalized Fourier (spectral) expansion
(30)
where the Fourier (spectral) coefficients
are functionals of the parameters, the forward and adjoint state variables, defined as follows:
(31)
and where the quantities
are the basis functions (e.g., orthogonal polynomials, trigonometric functions, etc.) chosen for the domain
. Although the expansion/summation in Equation (30) comprises in principle infinitely many terms, only a finite number, denoted as NF, of terms will be computed in practice. It is therefore important to choose the basis-functions
so as to minimize the total number of terms NF that would be needed to reproduce the response
within an a priori chosen error criterion. In particular, if the model response is ab initio simply a functional (as opposed to a function or operator) of the forward and/or adjoint functions
,
, then the expansion in Equation (31) would contain just the single
term. The nominal value of the functional defined in Equation (31) is denoted as follows:
(32)
The 1st-order G-differential of the response
defined in Equation (30) is
(33)
where the functional
has the following expression:
(34)
with the “indirect-effect term”
defined as
(35)
and the “direct-effect term”
defined as
(36)
The appearance of the variations
and
in the indirect-effect term
defined in Equation (35) can be eliminated by expressing the right-side of Equation (35) in terms of adjoint functions that are the solutions of a 2nd-Level Adjoint Sensitivity System (2nd-LASS) which is constructed following the steps outlined by Cacuci [4] [5] [6] :
1) Introduce a Hilbert space, denoted as
, having as elements two-component vectors of the form
, with square integrable components having the same structure as the
-dimensional vectors
and/or
. The inner product between two vectors
and
will be denoted as
and is defined as follows:
(37)
In particular, the inner product defined in Equation (37) also holds at the nominal parameter values
.
2) Using Equation (37), construct the inner product of Equations (20) and (26) with a vector
to obtain the following relation:
(38)
The relation expressed by Equation (38) is continuous in the neighborhood of the nominal values
.
3) Using the definition of the adjoint operator
provided in Equation (7) makes it possible to recast the left-side of Equation (38) into the following form:
(39)
where
denotes the associated bilinear form evaluated on
.
4) Require the right-side of Equation (35) and the first term on the right-side of Equation (39) to represent the same functional, which is accomplished by requiring that
(40)
which implies that the following system of equations is to be satisfied in the weak (distributional) sense:
(41)
(42)
Since the right-sides of Equations (41) and (42) depend on the index f, it follows that the functions
and
will also depend on the index f, which means that they will be distinct for each value of
, a fact that has been emphasized by showing the index f in the arguments of these functions.
5) Determine the adjoint boundary and/or initial conditions for the functions
and
, represented in operator form as
(43)
where the superscript “(1)” indicates “1st-Level.” The boundary and/or initial conditions represented by Equation (43) are obtained by requiring that they satisfy the following two criteria:
a) They must be independent of unknown values of
,
and
.
b) The substitution of the boundary and/or initial conditions represented by Equations (21), (27) and (43) into the expression of the bilinear concomitant
, must cause all terms containing unknown values of
and
in this bilinear concomitant to vanish. The selection of the boundary conditions for the functions
and
, based on the two criteria mentioned above, eliminates the appearance of the unknown values of the variations
and
, and consequently reduces the bilinear concomitant
to a residual quantity that contains boundary terms involving only known values of
, the index f and the nominal parameter values
. This residual quantity will be denoted by
, which is linear in
and can be therefore written in the form
(44)
In general, the residual quantity
does not automatically vanish, although it may do so in particular instances. By considering extensions of
and
,
could be forced to vanish, if necessary.
6) The system comprising Equations (41)-(43) will be called the First-Level Adjoint Sensitivity System (1st-LASS) and the functions
and
will be called the first-level adjoint sensitivity functions. It is important to note that the 1st-LASS is independent of the variations
in the model and boundary parameters. It is noteworthy that the function
satisfies an “adjoint-type” operator equation, i.e., Equation (41), while the function
satisfies “forward-type” operator equation, i.e., Equation (42). The operator equations underlying the 1st-LASS are to be solved using the nominal values
of the model’s parameters and state functions. By continuity, however, Equations (41)-(43) are valid not only at the nominal values
, but also within a neighborhood
of
.
7) Using the right-side of Equation (38) to replace the left-side of Equation (39), replacing the right-sides of Equation (40) into the right-side of Equation (39), and recalling that the boundary conditions in Equations (21), (27), and (43) have reduced the bilinear concomitant
to the residual quantity
, yields the following form for Equation (39):
(45)
8) The second term on the right-side of Equation (45) is actually the indirect-effect term defined in Equation (35), so that Equation (45) can be re-written as follows:
(46)
As the right-side of Equation (46) indicates, all unknown values of have been eliminated from the expression of the indirect effect term,
by having replaced the variations
and
with functions that depend on the 1st-level adjoint functions
and
. The identity
in Equation (46) explicitly indicates that the indirect-effect term now depends on the 1st-level adjoint functions
and
rather than on the unknown values of the variations
and
.
9) Inserting the results provided in Equations (22), (28) and (44) into Equation (46) yields the following expression for the indirect-effect term:
(47)
Adding the results obtained in Equation (47) to the direct-effect term expressed by Equation (36) yields the following expression for the total 1st-order sensitivity
of the spectral expansion coefficient
:
(48)
where the quantity
denotes the 1st-order partial sensitivities (G-derivatives) of the functional
with respect to the model/boundary parameters
, evaluated at the nominal values of the 1st-level adjoint functions and model and boundary parameters, and is given by the following expression, for each
:
(49)
In particular, contributions to the sensitivities
with respect to the specific parameters that characterize the imprecisely known model’s domain boundary can arise from the second term of the right-side of Equation (49), i.e., the boundary term
, and from the last two terms on the right-side of Equation (49) involving the boundary end-points
and
.
3.2. Derivation of the 2nd-Level Adjoint Sensitivity System (2nd-LASS) for Computing Exactly and Efficiently the 2nd-Order Response Sensitivities to Model and Boundary Parameters
The 2nd-order total sensitivity, denoted as
, of the functional
is provided by the 2nd-order G-differential of
to variations
in the model/boundary parameters and state functions around the respective nominal values
. The 2nd-order total sensitivity
is defined inductively as “the total first-order G-variation (or 1st-order total G-differential) of the 1st-order total response variation (or 1st-order G-differential),” as follows:
(50)
for
, and all (i.e., arbitrary) vectors of variations
around
. The total 2nd-order sensitivity
is bilinear in the vectors of variations
, and will exists if and only if the 1st-order partial sensitivities
satisfy the conditions expressed in Equations (15) and (16) in a neighborhood
of
. In order to proceed, it will be assumed that this is indeed the case. For the derivations to follow, it is convenient to introduce the vectors
, comprising as components the vectors of model parameters, forward and adjoint state functions, as well as the 1st-level adjoint functions. The nominal value of the vector
will be denoted as
, and the variations around
will be denoted as
. Using the definition provided in Equation (50), the 1st-order G-differential
of the functional
defined in Equation (49) can be computed, for each
, in the usual manner, namely:
(51)
where
denotes the “direct-effect term”, while
denotes the “indirect-effect term,” which are defined as follows:
(52)
(53)
For notational simplicity, the arguments of the functional
have been omitted in Equation (53). The index f, has been kept in the argument of the direct and indirect-effect terms in order to serve as a reminder that the 2nd-order sensitivities
are distinct for each
and each
. Since the direct-effect term,
, is independent of: 1) variations
and
in the forward and adjoint state functions; and 2) variations
and
in the 1st-level adjoint functions, itcan be computed immediately at this stage, without needing any additional large-scale computations. On the other hand, the “indirect-effect term”,
, cannot be computed at this stage because the variations
are not available. The variations
and
could be determined by solving the 1st-LFSS, which is computationally expensive. Furthermore, the variations
and
are the solutions of the system of equations obtained by taking 1st-order G-differentials of the 1st-LASS, i.e., Equations (41)-(43), which yields the following equations:
(54)
(55)
(56)
Equations (54)-(56) together with the 1st-LFSS constitute the 2nd-Level Forward Sensitivity System (2nd-LFSS). For further developments, it is convenient to write Equations (20), (26), (54) and (55) in the following matrix form:
(57)
where
(58)
(59)
Since the source-terms of the 2nd-LFSS depend on the parameter variations
, the 2nd-LFSS is computationally expensive to solve. To avoid the need for solving the 2nd-LFSS, the indirect-effect term defined in Equation (53) will be expressed in terms of a 2nd-Level Adjoint Sensitivity System (2nd-LASS), which will be constructed by following the general principles introduced by Cacuci [4] [5] [6], comprising the following sequence of steps:
1) Define a Hilbert space, denoted as
, having vector-valued elements of the form the
, with components that are
-dimensional vectors of the form
with square-integrable components
. The inner-product in
, denoted as
, of two vector-valued functions
and
, is defined as follows:
(60)
In particular, the inner product defined in Equation (60) also holds at the nominal parameter values
. Using the definition provided in Equation (60), construct the inner product of a vector
with Equation (57) to obtain the following relation:
(61)
The relation expressed by Equation (60) is continuous in the neighborhood of the nominal values
.
2) Using the definition of the adjoint operator in the Hilbert space
, the left-side of Equation (61) is recast as follows:
(62)
where
denotes the bilinear concomitant defined on the phase-space boundary
, and where
denotes the formal adjoint of the operator
, having components
defined as the transposed formal adjoints of the components of
, i.e.,
(63)
3) The first term on the right-side of Equation (62) is now required to represent the indirect-effect term defined in Equation (53), which is accomplished by requiring that the following relation be satisfied:
(64)
4) The boundary conditions given in Equations (56) are now implemented in Equation (62), thereby reducing by half the number of unknown boundary-values in the bilinear concomitant
. The boundary conditions for the functions
are now chosen so as to eliminate the remaining unknown boundary-values of the functions
,
,
and
in
while ensuring that Equation (64) is well posed. The boundary conditions thus chosen for the adjoint functions
and
can be represented in operator form as follows:
(65)
In most cases, the above choice of boundary conditions for the 1st-level adjoint function
will cause the bilinear concomitant
in Equation (62) to vanish. Even when it does not vanish, however, this bilinear concomitant will be reduced to a residual quantity, which will be denoted as
, and which will contain only known values of its arguments.
5) Replacing the right-side of Equation (61) by the left-side of Equation (62), replacing the right-side of Equation (64) in the first term on the right-side of Equation (62), and recalling that the boundary conditions in Equations (56) and (65) have reduced the bilinear concomitant
to the residual quantity
, yields the following form for Equation (62):
(66)
6) The second term on the right-side of Equation (66) is actually the indirect-effect term,
, defined in Equation (53), so that Equation (66) can be re-written in the following detailed form:
(67)
As the right-side of Equation (67) indicates, all unknown values of the variations
have been eliminated from the expression of the indirect effect term,
, by having replaced these unknown variations by quantities that depend on the functions
.
The system of equations comprising Equations (64) and (65) will be called the 2nd-Level Adjoint Sensitivity System (2nd-LASS) for the 2nd-level adjoint functions
,
,
. The explicit form of Equation (64) is
(68)
(69)
(70)
(71)
It is important to note that the 2nd-LASS is independent of the variations
in the model and boundary parameters. The 2nd-level adjoint functions
and
are the solutions of equations involving the forward operator
, while the 2nd-level adjoint functions
and
are the solutions of equations involving the adjoint operator
. Thus, the 2nd-LASS is solved successively by using two “forward” and two “adjoint” computations, for each of the imprecisely known scalar model parameters. The operator equations underlying the 1st-LASS are to be solved using the nominal values
of the model’s parameters and state functions. By continuity, however, Equations (68)-(71) are valid not only at the nominal parameter values
, but also within a neighborhood
of
.
Replacing the expression obtained in Equation (67) for the indirect-effect term
together with the expression in Equation (52) for the direct-effect term
into Equation (51) yields the following expression for the total 2nd-order response sensitivity
:
(72)
where
,
, denotes the 2nd-order partial sensitivity of the response to the model parameters and is given by the following expression:
(73)
All quantities appearing in Equation (73) are to be computed at the nominal values for the model and boundary parameters, although this requirement has not been indicated explicitly in Equation (73) in order to simplify the notations. In particular, contributions to the 2nd-order response sensitivities involving boundary parameters will arise not only from the last two terms in Equation (73), i.e., from the residual term on the domain’s boundary and from the direct-effect term, but may also arise from other terms involving mixed second-order derivatives with respect to the imprecisely known boundary parameters.
Since the 2nd-LASS is independent of parameter variations
, each of the G-differentials
contains one complete row (running on the index
) of the mixed second-order partial sensitivities,
, which can be computed exactly and efficiently after solving only once the 2nd-Level Adjoint Sensitivity System (2nd-LASS) to obtain the 2nd-level adjoint functions
, for each index
and index
. Thus, for each index
, the exact computation of all of the partial second-order sensitivities,
requires at most
large-scale (adjoint) computations using the 2nd-LASS, rather than
large-scale computations as would be required by re-computations and/or forward methods. It is also important to note that by solving the 2nd-LASS
-times, the “off-diagonal” 2nd-order mixed sensitivities
,
, will be computed twice, in two different ways (i.e., using distinct 2nd-level adjoint functions), thereby providing an independent intrinsic (numerical) verification that the respective sensitivities are computed accurately.
Another important characteristic of using the 2nd-LASS is the flexibility it provides for prioritizing the computation of the 2nd-order sensitivities. The 2nd-order mixed G-derivatives (sensitivities) corresponding to the largest relative 1st-order response sensitivity should be computed first; the second largest relative 1st-order response sensitivity should be considered next, and so on. Computing 2nd-order partial sensitivities that correspond to vanishing 1st-order sensitivities may also be of interest, since vanishing 1st-order sensitivities may indicate critical points of the response in the phase-space of model parameters. Thus, only the 2nd-order partial sensitivities which are deemed important should be computed; the unimportant ones can be deliberately neglected while knowing the error incurred by neglecting them.
It is noteworthy that the solving the equations underlying the 1st-LASS and 2nd-LASS require computational solvers for solving (i.e., “inverting”) either the forward linear operator
or the adjoint linear operator
. Only the various source on the right-sides of the 1st-LASS and 2nd-LASS differ from one another. Therefore, the same computer program and “solvers” can be used for solving the 1st-LASS and the 2nd-LASS as were already used for solving the original system of equations. For this reason, the 2nd-LASS was designated as a “second-level” rather than a “second-order” adjoint sensitivity system, since the 2nd-LASS does not involve any explicit 2nd-order G-derivatives of the operators underlying the original system, but involves the inversion of the same operators that are needed to be inverted for solving the 1st-LASS and the original system of equations.
4. Conclusions
This work has presented the “Second-Order Comprehensive Adjoint Sensitivity Analysis Methodology (2nd-CASAM)” for the efficient and exact computation of 1st- and 2nd-order response sensitivities to imprecisely known parameters and domain boundary for a generic/general model for linear physical systems. The model’s response (i.e., model result of interest) is considered to be a generic function/operator that depends on the model’s forward and adjoint state functions, as well as on the imprecisely known boundary and model parameters. It has been shown that the novel 2nd-CASAM has the following features that generalize and extend all previously published works on this topic:
1) The system response considered within the 2nd-CASAM framework is an operator-valued response that depends on both the forward and adjoint state-functions. Functional-valued responses are subsumed as particular cases.
2) The 2nd-CASAM framework enables the efficient and exact computation of the 1st- and 2nd-order response sensitivities not only to uncertain model parameters but also to uncertain domain boundaries of generic linear systems. In particular, contributions to the first-order response sensitivities with respect to the specific parameters that characterize the imprecisely known model’s domain boundary can arise solely from boundary terms or directly from the response if its definition includes the domain’s boundary. On the other hand, mixed second-order response sensitivities involving boundary parameters may arise from all source terms (of the 2nd-LASS) that involve the imprecisely known boundary parameters.
Companion works [18] [19] present illustrative applications of the 2nd-CASAM to various responses for a paradigm model of generic particle/radiation transmission and/or evolution which has fundamental applications in many fields, including nuclear engineering (e.g., radiation detection, chemical reprocessing of spent reactor fuel, radioactive decay, etc.).