578
Views
5
CrossRef citations to date
0
Altmetric
Article

Generalized formulation of extended cross-section adjustment method based on minimum variance unbiased linear estimation

&
Pages 87-104 | Received 20 Jul 2018, Accepted 25 Sep 2018, Published online: 24 Oct 2018

ABSTRACT

By introducing a new assumption of linear estimation, we derive a new formulation of the extended cross-section adjustment (EA) method, which minimizes the variance of the design target core parameters. The new formulation is derived on the basis of minimum variance unbiased estimation with no use of the assumption of normal distribution. In this formulation, we found that EA has infinitely many solutions as the adjusted cross-section set. The new formulation of EA can represent all the possible solutions minimizing the variance of the design target core parameters and includes a special case identical to the classical Bayesian EA method, which was derived on the basis of the Bayes theorem under the assumption of normal distribution. Moreover, we prove that the special case minimizes not only the variance of the design target core parameters but also the variance of the nuclear data. Meanwhile, we show that the new assumption of linear estimation is consistent with the Kalman filter and demonstrate that we can formulate similarly the extended bias factor method, the conventional cross-section adjustment method, and the regressive cross-section adjustment method with no use of the assumption of normal distribution.

1. Introduction

As a neutronics design method for an innovative nuclear reactor system, the original cross-section adjustment methodology was established in 1970s–1980s [Citation1Citation4]. With regard to the cross-section adjustment methodology, a comprehensive and detailed comparison study has been reported in references [Citation5,Citation6]. This methodology enables us to improve the prediction accuracy of neutronic characteristics by reducing uncertainties in nuclear data. In the cross-section adjustment methodology, the nuclear data are modified with the use of integral experimental quantities, such as measurement data acquired in critical experiments; the modified nuclear data are provided as an adjusted cross-section set for predicting the design target core parameters.

On the other hand, a formulation of the extended bias factor (EB) method [Citation7] was proposed in 2007 as an alternative methodology for improving the prediction accuracy of the design target core parameters by using integral experimental quantities. In EB, a bias factor is obtained by a semifictitious experimental value, which is defined by a linear combination of integral experimental quantities; the design prediction value is corrected by the bias factor.

Subsequently, a formulation of the extended cross-section adjustment (EA) method [Citation8] was proposed in 2012; EA can generate an adjusted cross-section set that reproduces the design prediction values and accuracies equivalent to EB. This formulation has revealed that EB minimizes the variance of the design prediction value of the target core parameter by considering cross-correlations in the analysis method error between the integral experimental quantities and the design target core parameters. Hence, in the case that there are strong cross-correlations in the analysis method error, EB can improve the design prediction accuracy more than the original cross-section adjustment methodology. Thus, the advantage of EB has been integrated into the cross-section adjustment methodology. In addition, the formulation of EA has revealed a difference between EA and EB. That is, the assumption of normal distribution required in the derivation of EA is not used in the derivation of EB. To explain this difference, the cross-section adjustment methods based on minimum variance unbiased estimation (MVUE) [Citation9] were proposed in 2016. In reference [Citation9], it has been shown that the MVUE-based rigorous extended cross-section adjustment (MREA) method, which was derived with no use of the assumption of normal distribution, can yield the design prediction values and accuracies equivalent to EB. Thus, it has been confirmed that the design prediction values and accuracies by EA are consistent with those by EB.

Moreover, the derivation of the MVUE-based cross-section adjustment methods [Citation9] has clarified that the original cross-section adjustment methodology minimizes the variance of the nuclear data. This finding led to a formulation of the regressive cross-section adjustment (RA) method [Citation9], which minimizes the variance of the integral experimental quantities. Although RA is unpractical as a neutronics design method, the formulation of RA is useful when we discuss what is optimized, that is, which variance is minimized, in the cross-section adjustment methodology. Thus, there are three variations of the cross-section adjustment methodology depending on the optimization target. In the present paper, we refer to the original cross-section adjustment methodology that minimizes the variance of the nuclear data as the conventional cross-section adjustment (CA) method. In particular, the original CA that was derived on the basis of the Bayes theorem under the assumption of normal distribution is called the classical Bayesian conventional cross-section adjustment (CBCA) method.

At that point in time, there was an unclear difference in the adjusted cross-section set with regard to CA. That is, the adjusted cross-section set by the MVUE-based rigorous conventional cross-section adjustment (MRCA) method [Citation9], which was derived with no use of the assumption of normal distribution, differs from that by CBCA. In reference [Citation9], it has been explained that a hypothetical mathematical operation, called ‘projection simplification,’ permits us to derive the MVUE-based simplified conventional cross-section adjustment (MSCA) method that can yield the adjusted cross-section set equivalent to CBCA. This explanation, however, is insufficient from the mathematical point of view. To explain this inconsistency in the adjusted cross-section sets between MRCA and CBCA, the dimension-reduced conventional cross-section adjustment (DRCA) method [Citation10] was proposed in 2018. The derivation of DRCA has revealed that the formulation for the adjusted cross-section set equivalent to CBCA can be derived with no use of the assumption of normal distribution by introducing a concept of dimensionality reduction. Thus, the difference in the adjusted cross-section set has been explained mathematically with regard to CA.

As for EA, however, there remains a similar problem in the adjusted cross-section set. That is, the adjusted cross-section set by the MREA method [Citation9], which was derived with no use of the assumption of normal distribution, differs from that by the classical Bayesian extended cross-section adjustment (CBEA) method [Citation8], which was derived on the basis of the Bayes theorem under the assumption of normal distribution. Similarly to MSCA, the MVUE-based simplified extended cross-section adjustment (MSEA) method, which was derived by using the projection simplification, was proposed in reference [Citation9] to explain the difference, but this explanation is also insufficient. Thus, it has not been clearly explained yet why this difference in the adjusted cross-section set is occurred with regard to EA.

In the present paper, with no use of the assumption of normal distribution, we derive a new formulation that can yield the adjusted cross-section set equivalent to CBEA. To derive the new formulation of EA, we focus attention on a fact that the formulation of the Kalman filter [Citation11,Citation12] can be derived with no use of the assumption of normal distribution. For this reason, we make a review on the derivation of the Kalman filter to clarify the relationship with the cross-section adjustment methodology. In accordance with this review, we introduce a new assumption of linear estimation into the derivation of the cross-section adjustment methods based on the MVUE. In the present paper, we refer to this derivation procedure based on MVUE with the use of the new assumption as minimum variance unbiased linear estimation (MVULE).

First, we derive the formulation of CA based on MVULE, namely, the MVULE-based conventional cross-section adjustment (MLCA) method. Next, we demonstrate that the formulation of EB can be also derived by the same procedure. That is, we formulate the MVULE-based extended bias factor (MLEB) method. Finally, we derive the formulation of the MVULE-based extended cross-section adjustment (MLEA) method and the MVULE-based regressive cross-section adjustment (MLRA) method.

In Section 2, we review the cross-section adjustment methods, the extended bias factor method, and the Kalman filter. In Section 3, the derivations of MLCA, MLEB, MLEA, and MLRA are described. In Section 4, we discuss the derived formulation and conclude in Section 5. The acronyms are described in the Abbreviations section at the end of this article.

2. Review

2.1. Review of cross-section adjustment method

2.1.1. Precondition

To review the formulation for the cross-section adjustment methodology, we first define the following numbers of data used in the methodology:

na: the number of nuclear data to be adjusted;

n(1): the number of integral experimental quantities used for adjustment; and

n(2): the number of the design target core parameters to be considered in adjustment of EA.

In accordance with reference [Citation9], MRCA is equivalent to CBCA under the condition of the overdetermined (well-posed) problem, where n(1)na (cf. Section 4.1.2 in reference [Citation9]). For this reason, we review the formulation under the condition of the underdetermined (ill-posed) problem, where n(1)<na. With regard to EAs – including MREA, MSEA, and CBEA, a precondition that n(2)<n(1)(<na) is adopted. As for DRCA, we focus on the special case equivalent to CBCA, namely, DRCA2 [Citation10].

2.1.2. Unified formula

The adjusted cross-section set by the foregoing methods can be expressed in the unified form [Citation9]:

(1) Tx  T0+BxD1Re(1)Rc(1)(T0),(1)

where

(2) Bx{MG(1)T+G(2)+Vm(12)T(x=CBEA,MSEA)G(1)+(G(2)G(1)+)+(G(2)MG(1)T+Vm(12)T)(x=MREA)MG(1)T(x=CBCA,MSCA,DRCA2)G(1)+G(1)MG(1)T(x=MRCA),(2)
(3) DG(1)MG(1)T+Ve+m(1),(3)

and

(4) Ve+m(1)Ve(1)+Vm(1).(4)

Here, we have defined as follows:

  • T0Rna×1: the unadjusted cross-section set;

  • Re(1)Rn(1)×1: the measurement value of the integral experimental quantities;

  • Rc(1)(T)Rn(1)×1: the calculation value of the integral experimental quantities obtained using the cross-section set T;

  • G(1)Rn(1)×na: the sensitivity coefficient matrix of the integral experimental quantities with respect to the nuclear data;

  • G(2)Rn(2)×na: the sensitivity coefficient matrix of the design target core parameters with respect to the nuclear data;

  • MRna×na: the variance–covariance matrix of the unadjusted cross-section set;

  • Ve(1)Rn(1)×n(1): the variance–covariance matrix with respect to the experimental error;

  • Vm(1)Rn(1)×n(1): the variance–covariance matrix with respect to the analysis method error; and

  • Vm(12)Rn(1)×n(2): the cross-correlation matrix between the integral experimental quantities and the design target core parameters with respect to the analysis method error,

where XRm×n means that X is an m×n real matrix. In addition, the superscript of ‘+’ denotes the Moore–Penrose pseudoinverse, and the superscript of ‘T’ represents the matrix transpose. Note that D is the full variance–covariance matrix of the integral experimental quantities and is supposed to be invertible (i.e. nonsingular) since the inverse of D appears in EquationEquation (1). Moreover, it is assumed that G(1) and G(2) have full rank. Then, recalling the precondition that n(2)<n(1)<na, we have

(5) G(i)G(i)+=I(i=1,2),(5)

where I denotes the identity matrix.

In accordance with the formulation of the MVUE-based cross-section adjustment methods (cf. EquationEquations (103), Equation(113), and (112) in reference [Citation9]), we can also write the variance-covariance matrix of the adjusted cross-section set, the variance of the calculation values of the integral experimental quantities reanalyzed using the adjusted cross-section set, and the variance of the prediction values of the design target core parameters in the unified forms:

(6) Mx=M+BxD1BxTMG(1)TD1BxTBxD1G(1)M,(6)
(7) VarRc(1)(Tx)=G(1)MxG(1)T+Vm(1)FxVm(1)Vm(1)FxT,(7)

and

(8) VarRc(2)(Tx)=G(2)MxG(2)T+Vm(2)LxVm(12)Vm(12)TLxT,(8)

respectively, where

(9) Fx  G(1)BxD1,(9)

and

(10) Lx  G(2)BxD1.(10)

Here, we have defined as follows:

  • Rc(2)(T)n(2)×1: the calculation value of the design target core parameters obtained using the cross-section set T; and

  • Vm(2)n(2)×n(2): the variance–covariance matrix of the design target core parameters with respect to the analysis method error.

2.1.3. Detailed definition

In addition, let us define the following symbols:

  • TtRna×1: the (unknown) true cross-section set;

  • Rt(i)Rn(i)×1: the true value of integral experimental quantities (i=1) or of the design target core parameters (i=2);

  • ΔT0Rna×1: the error of the unadjusted cross-section set;

  • Δe(1)Rn(1)×1: the experimental error of the integral experimental quantities; and

  • Δm(i)Rn(i)×1: the analysis method error of the integral experimental quantities (i=1) or of the design target core parameters (i=2).

Then, some symbols can be redefined by

(11) G(i)  dRc(i)dT|T=T0(i=1,2),(11)
(12) ΔT0  T0Tt,(12)
(13) Δe(1)  Re(1)Rt(1),(13)

and

(14) Δm(i)  Rc(i)(Tt)Rt(i)(i=1,2).(14)

Similarly, we can redefine the variance–covariance and cross-correlation matrices by

(15) M  VarT0,(15)
(16) Ve(1)  VarRe(1),(16)
(17) Vm(i)  VarRc(i)(Tt)(i=1,2),(17)

and

(18) Vm(ij)  CovRc(i)(Tt), Rc(j)(Tt)(i=1,2;j=1,2).(18)

Here, we have defined Var and Cov for column vectors X and Y by

(19) VarX  EXEXXEXT,(19)

and

(20) CovX,Y  EXEXYEYT,(20)

respectively, where E denotes the expectation value. Note that when i=j in EquationEquation (18), Vm(ii) is reduced to Vm(i) by definition. Moreover, all of the following variance-covariance matrices are symmetric:

(21) MT=M,(21)
(22) Ve(1)T=Ve(1),(22)

and

(23) Vm(i)T=Vm(i).(23)

On the other hand, a variance–covariance matrix is always nonnegative definite. That is, a variance-covariance matrix is either positive definite or positive semidefinite. Hence, M, Ve(1), and Vm(1) are nonnegative definite. Moreover, G(1)MG(1)T, D, and D1 are symmetric and nonnegative definite in accordance with the following theorems:

  • For an n×n matrix A and an n×m matrix B, if A is (symmetric and) nonnegative definite, BTAB is (symmetric and) nonnegative definite (e.g. Theorem 14.2.9 in reference [Citation13]);

  • For n×n matrices A and B, if both A and B are nonnegative definite, A+B is nonnegative definite (e.g. Lemma 14.2.4 in reference [Citation13]); and

  • For an n×n matrix A, if A is invertible and symmetric, A1 is symmetric (e.g. Equation (2.4) in reference [Citation13]);

  • For an n×n matrix A, if A is invertible and nonnegative definite, A1 is nonnegative definite (e.g. Corollary 14.2.11 in reference [Citation13]).

2.2. Review of extended bias factor method

Given that the design prediction values by CBEA are equal to those by EB [Citation8], the design prediction values by EB can be defined by

(24) REB(2)  Rc(2)(T0)+G(2)MG(1)T+Vm(12)TD1Re(1)Rc(1)(T0).(24)

This definition can be obtained by assuming linearity as it is done in EquationEquation (75), which is described later, and substituting EquationEquations (1) and (Equation2) as follows:

(25) Rc(2)(TCBEA)=Rc(2)(T0)+G(2)TCBEAT0=Rc(2)(T0)+G(2)MG(1)T+G(2)+Vm(12)TD1Re(1)Rc(1)(T0)=REB(2),(25)(25)

where EquationEquation (5) is used.

2.3. Review of Kalman filter

Next, we review a derivation of the Kalman filter [Citation14] and examine the relationship with the cross-section adjustment methodology.

2.3.1. Non-linear Kalman filter and its linearization

Let us consider the following equations as a nonlinear discrete-time state model for time series data:

(26) sk+1=f(sk)+wk,(26)
(27) yk=h(sk)+vk,(27)

where

  • skRn×1: the process state vector at time tk;

  • wkRn×1: the system error, which is a white noise contribution to the state vector for time interval (tk+1,tk);

  • ykRm×1: the measurement (observation vector) at time tk;

  • vkRm×1: the measurement error at time tk;

  • f(): the time-independent state transition function; and

  • h(): the time-independent observation function giving the ideal (noiseless) connection between the measurement and the state vector.

It is assumed that an approximate sk can be determined by some means. When the difference between the actual and the approximate state vectors is defined by Δsk, the actual state vector is written by

(28) sk=sk+Δsk.(28)

To linearize the equations of the discrete-time state model, let us approximate the state transition function and the observation function by the Taylor series expansion around sk:

(29) f(sk)f(sk)+fsk|sk=skΔsk,(29)

and

(30) h(sk)h(sk)+hsk|sk=skΔsk,(30)

where the higher-order terms are ignored. Thus, EquationEquation (26) can be written by

(31) sk+1+Δsk+1=f(sk)+ΦkΔsk+wk,(31)

where

(32) Φk  fsk|sk=skRn×n.(32)

Meanwhile, when we need an approximation of the state vector at the next time step, we can adopt a transition of the approximate state vector at the previous time step. That is, sk+1 can be determined such that sk+1=f(sk). Thus, EquationEquation (31) can be rewritten as

(33) Δsk+1=ΦkΔsk+wk.(33)

On the other hand, EquationEquation (27) can be expressed by

(34) ykh(sk)=HkΔsk+vk,(34)

where

(35) Hk  hsk|sk=skRm×n.(35)

To simplify the linearized model described by EquationEquations (33) and (Equation34), we replace the state vector and the observation vector by

(36) xk  Δsk=sksk,(36)

and

(37) zk  ykh(sk),(37)

respectively. Thus, the equations of the discrete-time state model can be rewritten as

(38) xk+1=Φkxk+wk,(38)
(39) zk=Hkxk+vk.(39)

This formulation is known as the linear Kalman filter.

2.3.2. Linear Kalman filter

Next, we review the state model equations of the linear Kalman filter. The covariance matrices of the system noise and the observation noise are defined by

(40) EwkwiT=Qk(i = k)O(i  k),(40)
(41) EvkviT=Rk(i = k)O(i  k),(41)

and

(42) EwkviT=O(forallkandi),(42)

where O denotes the zero matrix. Let us denote the prior estimate by xˆk0, which is the state vector estimated from the information prior to time tk, where the hat () represents an estimate, and the subscript of ‘0’ means an estimate prior to assimilating the measurement at time tk.

Let us assume that the covariance matrix of the prior estimate is known. That is, the estimation error is defined explicitly by

(43) ek0xkxˆk0,(43)

then the covariance matrix of the prior estimate can be written by

(44) Pk0=Eek0ek0T=Exkxˆk0xkxˆk0T.(44)

Next, let us consider improving the prior estimate by using the measurement at time tk. For this purpose, a linear blending of the measurement zk and the prior estimate xˆk0 is adopted in the form:

(45) xˆk=xˆk0+KkzkHkxˆk0,(45)

where

xˆk: the updated estimate of the state vector and

Kk: the linear blending factor.

Meanwhile, the covariance matrix of the updated estimate is expressed by

(46) Pk=EekekT=ExkxˆkxkxˆkT.(46)

Substituting EquationEquation (39) into EquationEquation (45) yields

(47) xˆk=xˆk0+KkHkxk+vkHkxˆk0.(47)

Thus, substituting EquationEquation (47) in EquationEquation (46) and considering that there is no correlation between ek0 and vk, we obtain

(48) Pk=Exkxˆk0KkHkxk+vkHkxˆk0xkxˆk0KkHkxk+vkHkxˆk0T=Eek0KkHkek0+vkek0KkHkek0+vkT=IKkHkEek0ek0TIKkHkT+KkEvkvkTKkT=IKkHkPk0IKkHkT+KkRkKkT=Pk0KkHkPk0Pk0HkTKkT+KkHkPk0HkT+RkKkT.(48)

To derive this equation, we have used that the following equation holds for arbitrary matrices A and B:

(49) ABT=BTAT.(49)

To determine the blending factor that minimizes the trace of the covariance matrix of the updated estimate, we consider the partial derivative:

(50) KktrPk=2KktrPk0HkTKkT+KktrKkHkPk0HkT+RkKkT=2Pk0HkT+2KkHkPk0HkT+Rk,(50)

where tr denotes the matrix trace, which is the sum of the diagonal elements. To derive EquationEquation (50), we have used the following equations for arbitrary matrices A and B:

(51) trA+B=trA+trB,(51)

and

(52) trAT=trA.(52)

Moreover, we have used the following derivative formulas for arbitrary constant matrices A, B, C:

(53) XtrA+trB=XtrA+XtrB,(53)
(54) XtrAXTB=BA,(54)

and

(55) XtrAXBXTC=ATCTXBT+CAXB.(55)

To determine the blending factor that minimizes the trace of Pk, we set EquationEquation (50) equal to zero and solve for Kk. The result is

(56) Kk=Pk0HkTHkPk0HkT+Rk1.(56)

This equation is called the Kalman gain. Substituting this equation into EquationEquation (45), we obtain

(57) xˆk=xˆk0+Pk0HkTHkPk0HkT+Rk1zkHkxˆk0.(57)

This equation enables us to estimate the state vector at time tk.

2.3.3. Relationship with cross-section adjustment method

As defined by EquationEquations (1) – (Equation3), the adjusted cross-section set by CBCA is written by

(58) TCBCA=T0+MG(1)TG(1)MG(1)T+Ve+m(1)1Re(1)Rc(1)(T0).(58)

Although this equation is similar to EquationEquation (57), it differs from EquationEquation (57) in that it deals with the nonlinearity of Rc(1)(). To eliminate this difference, let us consider the following correspondence in light of the derivation procedure for linearizing the Kalman filter:

(59) xˆk=sˆkskTCBCAT0,(59)
(60) xˆk0=sˆk0skT0T0 (=O),(60)
(61) zk=ykh(sk)Re(1)Rc(1)(T0),(61)
(62) Pk0M,(62)
(63) HkG(1),(63)

and

(64) RkVe+m(1).(64)

Thus, we can understand that EquationEquation (58) is identical to EquationEquation (57).

Moreover, in light of EquationEquation (64), we can find another correspondence:

(65) vkΔe(1)+Δm(1).(65)

From this correspondence, we can understand that the cross-section adjustment method is a linearized model of the nonlinear Kalman filter described by the state model equations:

(66) T1=f(T0)+w0,(66)
(67) Re(1)+Δe(1)=Rc(1)(T0)+Δm(1),(67)

where T1 represents the cross-section set at time t1. Although EquationEquation (66) is inessential for the cross-section adjustment methodology, the state model described by EquationEquation (67) is reasonable as the observation of C/E (calculation/experiment) values for the integral experimental quantities. On the other hand, since the linearization of the nonlinear Kalman filter is known as the extended Kalman filter (EKF), we can understand that the cross-section adjustment methodology is a kind of EKF.

2.4. Discussion on assumption of linear estimation

As described before, the Kalman gain is determined by minimizing the variance with no use of the assumption of normal distribution. This procedure is similar to determining the linear combination factor F in the derivation of the MVUE-based cross-section adjustment methods. Hence, we can find yet another correspondence:

(68) KkF.(68)

Substituting the foregoing correspondence, that is, EquationEquations (59) – (Equation61), and (Equation68), into EquationEquation (45), we obtain

(69) TˆT0=FRe(1)Rc(1)(T0).(69)

This equation differs from the following equation that represents the assumption of linear estimation used in the derivation of the MVUE-based cross-section adjustment methods [Citation9]:

(70) Rˆ(1)Rc(1)(T0)=FRe(1)Rc(1)(T0).(70)

In light of the review of the Kalman filter, it is expected that we can formulate a new cross-section adjustment method if we adopt EquationEquation (69) as the assumption of linear estimation.

On the other hand, we can use another equation as the assumption of linear estimation. That is, if we consider the design target core parameters as the state vector, the assumption of linear estimation can be written by

(71) Rˆ(2)Rc(2)(T0)=FRe(1)Rc(1)(T0).(71)

Since this equation estimates directly the design target core parameters, it is expected that a new formulation corresponding to EB can be derived if EquationEquation (71) is used as the assumption of linear estimation.

In the following, we derive MLCA with the use of EquationEquation (69) as the assumption of linear estimation. Subsequently, MLEB is derived by using EquationEquation (71). Finally, we derive MLEA and MLRA with EquationEquation (69).

3. Derivation

3.1. Assumptions

In the present paper, we adopt the same assumptions used in the derivation of the MVUE-based cross-section adjustment methods [Citation9] except for the assumption of linear estimation. However, since these assumptions are essential for the formulation, we summarize them in the following.

3.1.1. Assumptions on number of data

Unlike the foregoing review of the cross-section adjustment methods, we employ no particular assumption (precondition) regarding the relationship between the number of integral experimental quantities and the number of nuclear data:

(72) n(1)><na.(72)

That is, it is unnecessary to distinguish whether the problem is overdetermined or underdetermined in the present formulation. With regard to EA, it is assumed that the number of design target core parameters is smaller than the number of nuclear data:

(73) n(2)<na.(73)

In addition, we assume that G(2) has full rank:

(74) rankG(2)=n(2).(74)

Note that G(2) is a wide matrix and has full row rank in light of EquationEquations (73) and (Equation74). By contrast, at this moment, we pay no attention to whether G(1) is of full rank.

3.1.2. Assumption of linearity

Let us assume that the variation of two calculation values obtained by different cross-section sets T and T can be evaluated by the first-order approximation in the form:

(75) Rc(1)(T)Rc(1)(T)G(1)(TT).(75)

Then, the difference between the experimental value and the calculation value of integral experimental quantities can be expressed by

(76) Re(1)Rc(1)(T0)=Re(1)Rt(1)Rc(1)(Tt)Rt(1)Rc(1)(T0)Rc(1)(Tt)Δe(1)Δm(1)G(1)ΔT0,(76)

where EquationEquations (Equation12)–(Equation14) are used.

3.1.3. Assumption of unbiased estimation

Let us assume that the unbiased estimation is valid all for the nuclear data included in the unadjusted cross-section set, the experimental value and the calculation value of the integral experimental quantities, and the calculation value of the design target core parameters. That is, we assume that the expectation value is equal to the true value:

(77) ET0=Tt,(77)
(78) ERe(1)=Rt(1),(78)

and

(79) ERc(i)(Tt)=Rt(i)(i=1,2).(79)

Then, in light of EquationEquations (19) and (Equation20), the foregoing covariance and cross-correlation matrices can be rewritten as

(80) M=EΔT0ΔT0T,(80)
(81) Ve(1)=EΔe(1)Δe(1)T,(81)
(82) Vm(i)=EΔm(i)Δm(i)T(i=1,2),(82)

and

(83) Vm(ij)=EΔm(i)Δm(j)T(i=1,2;j=1,2).(83)

3.1.4. Assumption of error independence

Moreover, we assume that the experimental error, the analysis method error, and the cross-section-induced errors with respect to the integral experimental quantities and the design target core parameters are independent of each other:

(84) EΔxiΔxjT=O(i  j),(84)

where

(85) ΔxiΔe(1)(i=1)Δm(1),Δm(2)(i=2)G(1)ΔT0(i=3)G(2)ΔT0(i=4).(85)

3.2. Derivation of conventional cross-section adjustment method

Let us derive MLCA by using EquationEquation (69) as the assumption of linear estimation.

3.2.1. Derivation of adjusted cross-section set

In light of EquationEquation (12), the estimate of the cross-section set can be written by

(86) Tˆ=Tt+(TˆT0)+ΔT0.(86)

Substituting EquationEquations (69) and (Equation76) in EquationEquation (86), we have

(87) Tˆ=Tt+FΔe(1)Δm(1)G(1)ΔT0+ΔT0=Tt+IFG(1)ΔT0+FΔe(1)FΔm(1).(87)

Thus, the variance of Tˆ can be expressed by

(88) VarTˆ=ETˆTtTˆTtT=IFG(1)EΔT0ΔT0TIFG(1)T+FEΔe(1)Δe(1)TF+FEΔm(1)Δm(1)TF=M+FG(1)MG(1)TFTMG(1)TFTFG(1)M+FVe(1)FT+FVm(1)FT=M+FG(1)MG(1)T+Ve+m(1)FTMG(1)TFTMG(1)TFTT(88)

where EquationEquations (Equation84), (Equation80), (Equation81), (Equation82) and (Equation4) are used. To determine the linear combination factor F that minimizes the variance, let us consider the partial derivative of the trace of this variance with respect to F:

(89) FtrVarTˆ=FtrFG(1)MG(1)T+Ve+m(1)FT2FtrMG(1)TFT=2FG(1)MG(1)T+Ve+m(1)2MG(1)T(89)

where EquationEquations (51)–(Equation55) are used. It is possible to determine the linear combination factor matrix F that minimizes the variance by setting EquationEquation (89) equal to zero. Let FMLCA denote the minimizing factor of F, then we have

(90) FMLCAG(1)MG(1)T+Ve+m(1)=MG(1)T.(90)

Premultiplying both sides of this equation by the inverse of D defined by EquationEquation (3), we obtain

(91) FMLCA=MG(1)TD1(91)

Substituting FMLCA into F in EquationEquation (69) and defining TMLCA in this case as Tˆ, we have

(92) TMLCA=T0+MG(1)TD1Re(1)Rc(1)(T0).(92)

This equation is identical to that for the adjusted cross-section set by CBCA:

(93) TMLCA=TCBCA.(93)

In summary, if EquationEquation (69) is valid as the assumption of linear estimation, we can derive the formula identical to CBCA by using neither the projection simplification nor the concept of dimensionality reduction.

3.2.2. Derivation of variances

Given that EquationEquations (6)–(Equation8) can be derived with no use of the assumption of linear estimation (cf. Sections 4.2–4.4 in reference [Citation9]), we can adopt these equations also in MLCA. Thus, we can understand that the prediction accuracies by MLCA are also equivalent to CBCA.

3.3. Derivation of extended bias factor method

Next, let us derive MLEB by using EquationEquation (71) as the assumption of linear estimation.

3.3.1. Derivation of estimate of design target core parameters

The calculation values of the design target core parameters obtained using the unadjusted cross-section set can be written by

(94) Rc(2)(T0)=Rt(2)+Rc(2)(T0)Rc(2)(Tt)+Rc(2)(Tt)Rt(2)Rt(2)+G(2)(T0Tt)+Δm(2)=Rt(2)+G(2)ΔT0+Δm(2),(94)

where EquationEquations (75), (Equation14), and (Equation12) are used. Substituting EquationEquations (94) and (Equation76) in EquationEquation (71), we have

(95) Rˆ(2)=Rt(2)+G(2)ΔT0+Δm(2)+FΔe(1)Δm(1)G(1)ΔT0=Rt(2)+G(2)FG(1)ΔT0+FΔe(1)FΔm(1)+Δm(2).(95)(95)

Thus, the variance of Rˆ(2) can be written by

(96) VarRˆ(2)=ERˆ(2)Rt(2)Rˆ(2)Rt(2)T=G(2)FG(1)EΔT0ΔT0TG(2)FG(1)T+FEΔe(1)Δe(1)TFT+FEΔm(1)Δm(1)TFT+EΔm(2)Δm(2)TFEΔm(1)Δm(2)TEΔm(2)Δm(1)TFT=G(2)MG(2)T+FG(1)MG(1)TFTG(2)MG(1)TFTFG(1)MG(2)T+FVe(1)FT+FVm(1)FT+Vm(2)FVm(12)Vm(12)TFT=G(2)MG(2)T+FG(1)MG(1)T+Ve+m(1)FTG(2)MG(1)TFTG(2)MG(1)TFTT+Vm(2)Vm(12)TFTTVm(12)TFT.(96)

where EquationEquationEquations (84), Equation(80)–(Equation83), and (4) are used. To determine F that minimizes the variance, let us consider the partial derivative:

(97) FtrVarRˆ(2)=FtrFG(1)MG(1)T+Ve+m(1)FT2FtrG(2)MG(1)TFT2FtrVm(12)TFT=2FG(1)MG(1)T+Ve+m(1)2G(2)MG(1)T2Vm(12)T,(97)

where EquationEquations (51)Equation(55) are used. We can determine the linear combination factor matrix that minimizes the variance by setting EquationEquation (97) equal to zero. Let FMLEB denote the minimizing factor, then we have

(98) FMLEBG(1)MG(1)T+Ve+m(1)=G(2)MG(1)T+Vm(12)T.(98)

Premultiplying both sides of this equation by D1 yields

(99) FMLEB=G(2)MG(1)T+Vm(12)TD1.(99)

Substituting FMLEB of this equation into F of EquationEquation (71) and defining Rˆ(2) in this case by REB(2), we have

(100) RMLEB(2)=Rc(2)(T0)+G(2)MG(1)T+Vm(12)TD1Re(1)Rc(1)(T0).(100)

3.3.2. Comparison with original extended bias factor method

From a comparison between EquationEquations (100) and (Equation24), we obtain

(101) RMLEB(2)=REB(2).(101)

Thus, we can interpret that EB is a kind of the linear Kalman filter in which the design target core parameters are estimated directly as the state vector.

3.4. Derivation of extended cross-section adjustment method

Finally, let us derive MLEA by using EquationEquation (69) as the assumption of linear estimation.

3.4.1. Minimization of variance of design target core parameters

Similarly to EquationEquation (94), the calculation values of the design target core parameters obtained using the estimated cross-section set can be written by

(102) Rc(2)(Tˆ)=Rt(2)+Rc(2)(Tˆ)Rc(2)(Tt)+Rc(2)(Tt)Rt(2)Rt(2)+G(2)(TˆTt)+Δm(2)=Rt(2)+G(2)(TˆT0)+G(2)ΔT0+Δm(2).(102)

Substituting EquationEquations (69) and (Equation76) in EquationEquation (102), we have

(103) Rc(2)(Tˆ)=Rt(2)+G(2)FΔe(1)Δm(1)G(1)ΔT0+G(2)ΔT0+Δm(2)=Rt(2)+G(2)G(2)FG(1)ΔT0+G(2)FΔe(1)G(2)FΔm(1)+Δm(2).(103)

Thus, the variance of Rc(2)(Tˆ) can be written by

(104) VarRc(2)(Tˆ)=ERc(2)(Tˆ)Rt(2)Rc(2)(Tˆ)Rt(2)T=G(2)G(2)FG(1)EΔT0ΔT0TG(2)G(2)FG(1)T+G(2)FEΔe(1)Δe(1)TG(2)FT+G(2)FEΔm(1)Δm(1)TG(2)FT+EΔm(2)Δm(2)TG(2)FEΔm(1)Δm(2)TEΔm(2)Δm(1)TG(2)FT=G(2)MG(2)T+G(2)FG(1)MG(1)TFTG(2)TG(2)MG(1)TFTG(2)TG(2)FG(1)MG(2)T+G(2)FVe(1)FTG(2)T+G(2)FVm(1)FTG(2)T+Vm(2)G(2)FVm(12)Vm(12)TFTG(2)T=G(2)MG(2)T+G(2)FG(1)MG(1)T+Ve+m(1)FTG(2)TG(2)MG(1)TFTG(2)T  G(2)MG(1)TFTG(2)TT+Vm(2)Vm(12)TFTG(2)TTVm(12)TFTG(2)T.(104)

where EquationEquations (84), (Equation80)– (Equation83), and (Equation4) are used. To determine F that minimizes the variance, let us consider the partial derivative:

(105) FtrVarRc(2)(Tˆ)=FtrG(2)FG(1)MG(1)T+Ve+m(1)FTG(2)T2FtrG(2)MG(1)TFTG(2)T2FtrVm(12)TFTG(2)T   =2G(2)TG(2)FG(1)MG(1)T+Ve+m(1)2G(2)TG(2)MG(1)T2G(2)TVm(12)T.(105)

where EquationEquations (51)Equation–(55) are used. By setting EquationEquation (105) equal to zero, we can determine the linear combination factor matrix that minimizes the variance. Let FMLEA denote the minimizing factor, then we have

(106) G(2)TG(2)FMLEAG(1)MG(1)T+Ve+m(1)=G(2)TG(2)MG(1)T+Vm(12)T.(106)

Premultiplying both sides of this equation by D1 yields

(107) G(2)TG(2)FMLEA=G(2)TG(2)MG(1)T+Vm(12)TD1.(107)

Note that G(2)TG(2) is a singular matrix in light of EquationEquation (73). Therefore, EquationEquation (107) has infinitely many solutions of FMLEA. In other words, we cannot determine uniquely the adjusted cross-section set for MLEA. Although a minimum norm solution was adopted in the derivation of MREA and MSEA to determine a unique solution by using the Moore–Penrose pseudoinverse (cf. EquationEquation (64) in reference [Citation9]), this treatment remains open to discussion. In the present derivation, we treat explicitly these numerous solutions.

3.4.2. Derivation of numerous solutions of adjusted cross-section set

To express the solutions for EquationEquation (107), we introduce a generalized inverse for a singular matrix, which includes a rectangular matrix. Note that the generalized inverse of a matrix is not unique. The Moore–Penrose pseudoinverse is a special case of the generalized inverse, and is determined uniquely. By definition, for the generalized inverse of an arbitrary matrix A, we have

(108) AAA=A,(108)

where the superscript of ‘’ denotes the generalized inverse. Furthermore, the generalized inverse of a matrix of full row rank is a right inverse, even if it is not the Moore–Penrose pseudoinverse (e.g. Lemma 9.2.8 in reference [Citation13]). That is, for a matrix of full row rank B, we always have

(109) BB=I.(109)

In light of EquationEquations (73) and (Equation74), G(2) has full row rank. Hence, G(2) is the right inverse of G(2):

(110) G(2)G(2)=I.(110)

Thus, EquationEquation (107) can be rewritten as

(111) G(2)TG(2)FMLEA=G(2)TG(2)MG(1)T+G0(2)Vm(12)TD1,(111)

where G0(2) is a generalized inverse of G(2). To indicate that G0(2) is a particular one of the numerous generalized inverses of G(2), we have defined the new symbol of G0(2).

To simplify EquationEquation (111), let us define

(112) BMLEA0  MG(1)T+G0(2)Vm(12)T,(112)

then EquationEquation (111) can be rewritten as

(113) G(2)TG(2)FMLEAG(2)BMLEA0D1=O.(113)

Premultiplying both sides of this equation by FMLEABMLEA0D1T, we obtain

(114) G(2)FMLEAG(2)BMLEA0D1TG(2)FMLEAG(2)BMLEA0D1=O,(114)

where EquationEquation (49) is used. In general, for an arbitrary matrix A, we have A=O if and only if ATA=O (e.g. Corollary 5.3.2 in reference [Citation13]). Hence, EquationEquation (114) means that EquationEquation (111) is equivalent to the equation:

(115) G(2)FMLEA=G(2)BMLEA0D1.(115)

Meanwhile, all the solutions for AX=B can be expressed by X=X0+Z, where X0 is an arbitrary particular solution for AX=B and Z is all the solutions for AZ=O (e.g. Theorem 11.2.3 in reference [Citation13]). Moreover, all the solutions for AZ=O can be expressed by Z=IAAY, where Y is an arbitrary matrix (e.g. Theorem 11.2.1 in reference [Citation13]).

Let us adopt BMLEA0D1 as a particular solution of FMLEA for EquationEquation (115), then we obtain the following expression as the all possible solutions for EquationEquation (115):

(116) FMLEA=BMLEA0D1+IG1(2)G(2)Y,(116)

where the superscript of ‘’ indicates that there are numerous solutions; G1(2) is a generalized inverse of G(2); and YRna×n(1) is an arbitrary matrix. Note that the generalized inverse G1(2) may differ from the previously defined G0(2). Hence, we have defined the new symbol of G1(2).

Substituting EquationEquation (116) in EquationEquation (69), we obtain the following equation that represents all the solutions of the adjusted cross-section set in MLEA:

(117) TMLEA=T0+BMLEA0+IG1(2)G(2)YDD1Re(1)Rc(1)(T0).(117)

3.4.3. Derivation of variances

Similarly to MLCA, since EquationEquations (6)– (Equation8) can be derived with no use of the assumption of linear estimation (cf. Sections 4.24.4 in reference [Citation9]), we can adopt these equations in MLEA:

(118) MMLEA=M+BMLEAD1BMLEATMG(1)TD1BMLEATBMLEAD1G(1)M,(118)
(119) VarRc(1)(TMLEA)=G(1)MMLEAG(1)T+Vm(1)FMLEAVm(1)Vm(1)FMLEAT,(119)

and

(120) VarRc(2)(TMLEA)=G(2)MMLEAG(2)T+Vm(2)LMLEAVm(12)Vm(12)TLMLEAT,(120)

where

(121) LMLEA  G(2)BMLEAD1.(121)

We can use these equations in the application of MLEA to nuclear reactor design. For instance, we can evaluate the variances (i.e. accuracies) of the design target core parameters by using EquationEquation (120).

3.5. Derivation of regressive cross-section adjustment method

Although it is known that RA is unpractical as the neutronics design method, we derive MLRA since it is meaningful for understanding the difference of the optimization target in comparison with CA and EA. By following the derivation procedure similar to MLEA, we can derive the formulation of MLRA while taking into account the difference of the optimization target. That is, by replacing the superscript of ‘(2)’ in the derivation of MLEA with the superscript of ‘(1)’, we can obtain the formula for the adjusted cross-section set in MLRA:

(122) TMLRA=T0+BMLCA0+IG1(1)G(1)YDD1Re(1)Rc(1)(T0),(122)

where

(123) BMLCA0  MG(1)T+G0(1)Vm(1),(123)

and G0(1) and G1(1) are particular ones of the generalized inverses of G(1). Similarly to MLEA, the following assumptions are required additionally to derive MLRA:

(124) n(1)<na,(124)

and

(125) rankG(1)=n(1).(125)

Meanwhile, to derive EquationEquation (122), we have used that Vm(11)=Vm(1) holds in light of EquationEquations (82) and (Equation83).

3.6. Summary of derived results by unified formula

In accordance with EquationEquations (1) and (Equation2), Bx for MLCA can be defined by

(126) Bx  MG(1)T(x=MLCA),(126)

and Bx for MLEA and MLRA can be defined by

(127) Bx MG(1)T+G0(2)Vm(12)T+IG1(2)G(2)YD(x=MLEA)MG(1)T+G0(1)Vm(1)+IG1(1)G(1)YD(x=MLRA),(127)

where

YRna×n(1): an arbitrary matrix; and

G0(i),G1(i)Rn(i)×na(i=1,2): particular ones of the generalized inverses of G(i) – it is allowed to determine G0(i) and G1(i) independently.

4. Discussion

4.1. Consistency of MLEA with EB

Let us confirm that MLEA reproduces the design prediction values by EB. Since the design prediction values are obtained using the adjusted cross-section set, we can confirm that the prediction values by MLEA are equal to EquationEquation (24):

(128) Rc(2)(TMLEA)=Rc(2)(T0)+G(2)(TMLEAT0)=Rc(2)(T0)+G(2)(BMLEA0+IG1(2)G(2)YD)D1Re(1)Rc(1)(T0)=Rc(2)(T0)+G(2)BMLEA0D1Re(1)Rc(1)(T0)=REB(2),(128)(128)

where EquationEquations (75), (Equation1), (Equation127), (Equation112), and (Equation108) are used.

4.2. Quasi-solutions of CA and compensation effect

As an analogy from the foregoing discussion, we can deduce that all the adjusted cross-section sets defined by the following equation can reproduce the calculation values of the integral experimental quantities of CBCA:

(129) TquasiCA=T0+MG(1)T+IG(1)G(1)YDD1Re(1)Rc(1)(T0).(129)

That is, for all the adjusted cross-section sets expressed by this equation, we have

(130) Rc(1)(TquasiCA)=Rc(1)(T0)+G(1)TquasiCAT0=Rc(1)(T0)+G(1)MG(1)TD1Re(1)Rc(1)(T0)=Rc(1)(T0)+G(1)TCBCAT0=Rc(1)(TCBCA),(130)

where EquationEquations (75), (Equation129), (Equation108), (Equation1), and (Equation2) are used. Note that, however, these adjusted cross-section sets are not always the optimal solution that minimizes the variance of the nuclear data.

Thus, we can understand that there are infinitely many solutions of the adjusted cross-section set for solely reproducing the calculation values of the integral experimental quantities. This fact is known as a ‘compensation effect’ [Citation15] in the use of integral experimental benchmarks in nuclear data evaluations.

4.3. Practical solution of MLEA

With regard to MLEA, we have obtained the mathematical formulation representing all the solutions of the adjusted cross-section set that minimize the variance of the design target core parameters. Next, let us consider which solution should be employed in practice. To determine uniquely the solution, we introduce an additional constraint condition. From a practical point of view, we adopt the constraint condition of minimizing the variance of the nuclear data.

4.3.1. Conjecture

Let us make a conjecture that the variance of the adjusted cross-section set in MLEA (cf. EquationEquation (127)) is minimized when we adopt the case where Y=O (n.b. the term of G1(2) vanishes) and G0(2)=G(2)+.

Hence, we make the following definitions:

(131) BMLEAmin  BMLEA|Y=O,G0(2)=G(2)+,(131)
(132) TMLEAmin  T0+BMLEAminD1Re(1)Rc(1)(T0),(132)

and

(133) MMLEAmin  VarTMLEAmin.(133)

Note that BMLEAmin is identical to Bx for CBEA:

(134) BMLEAmin=MG(1)T+G(2)+Vm(12)T=BCBEA.(134)

In the following, we prove that the trace of MMLEAmin has the minimum value among the traces of MMLEA:

(135) trMMLEA  trMMLEAmin.(135)

4.3.2. Proof

The trace of EquationEquation (6), which represents the covariance matrix of the adjusted cross-section set, can be expressed by

(136) trMx=trM+trBxD1BxT2trMG(1)TD1BxT,(136)

where EquationEquations (51) and (Equation52) are used. Thus, the difference between the left-hand side and the right-hand side of EquationEquation (135) can be written by

(137) trMMLEAtrMMLEAmin=trBMLEAD1BMLEAT2trMG(1)TD1BMLEATtrBMLEAminD1BMLEAminT+2trMG(1)TD1BMLEAminT.(137)

To simplify this equation, let us define

(138) P  IG1(2)G(2),(138)

then BMLEA defined by EquationEquation (127) can be rewritten as

(139) BMLEA=BMLEA0+PYD.(139)

Note that P is an idempotent matrix:

(140) P2=P.(140)

In addition, with respect to P, we have

(141) G(2)P=PG(2)=O.(141)

To derive EquationEquations (140) and (Equation141), we have used EquationEquation (108).

Substituting EquationEquation (139) in EquationEquation (137), we obtain

(142) trMMLEAtrMMLEAmin=trBMLEA0+PYDD1BMLEA0+PYDT2trMG(1)TD1BMLEA0+PYDTtrBMLEAminD1BMLEAminT+2trMG(1)TD1BMLEAminT=trPYDYTPT+trBMLEA0D1BMLEA0T+2trBMLEA0YTPT2trMG(1)TD1BMLEA0T2trMG(1)TYTPTtrBMLEAminD1BMLEAminT+2trMG(1)TD1BMLEAminT,(142)

where EquationEquations (52) and (Equation49) are used. Substituting EquationEquations (112) and (Equation134) into EquationEquation (142), we obtain

(143) trMMLEAtrMMLEAmin=trPYDYTPT+trMG(1)T+G0(2)Vm(12)TD1MG(1)T+G0(2)Vm(12)TT+2trMG(1)T+G0(2)Vm(12)TYTPT2trMG(1)TD1MG(1)T+G0(2)Vm(12)TT2trMG(1)TYTPTtrMG(1)T+G(2)+Vm(12)TD1MG(1)T+G(2)+Vm(12)TT+2trMG(1)TD1MG(1)T+G(2)+Vm(12)TT=trPYDYTPT+2trG0(2)Vm(12)TYTPT+trG0(2)Vm(12)TD1Vm(12)G0(2)TtrG(2)+Vm(12)TD1Vm(12)G(2)+T.(143)

Note that EquationEquation (143) is considered as a function of Y, G0(2), and P.

Meanwhile, if a specific generalized inverse of an m×n matrix A is given by A, all the generalized inverses of A can be expressed by

(144) A=A+ZAAZAA,(144)

where Z is an n×m arbitrary matrix (e.g. Theorem 9.2.7 in reference [Citation13]). Since the Moore–Penrose pseudoinverse is a special case of the generalized inverse, we may adopt G(2)+ as the specific generalized inverse in the above equation. Hence, by using an arbitrary matrix Z0Rna×n(2), we can express all the generalized inverses in the form:

G0(2)=G(2)++Z0G(2)+G(2)Z0G(2)G(2)+
(145) =G(2)++QZ0,(145)

where

(146) Q  IG(2)+G(2).(146)

To derive EquationEquation (145), we have used that G(2)G(2)+=I in light of EquationEquation (110). Note that Q is symmetric and idempotent:

(147) QT=Q=Q2.(147)

In addition, with regard to Q, we have

(148) G(2)Q=QG(2)+=O.(148)

Similarly to G0(2), by using an arbitrary matrix Z1Rna×n(2), all the generalized inverses can be expressed by

(149) G1(2)=G(2)++QZ1.(149)

Thus, EquationEquation (138) can be rewritten as

(150) P=IG(2)++QZ1G(2)  =QIZ1G(2),(150)

where EquationEquation (146) is used. Note that G0(2) is a function of Z0 as shown in EquationEquation (145), and that P is a function of Z1 as shown in EquationEquation (150). Hence, we can express EquationEquation (143) as a function of arbitrary matrices Y, Z0, and Z1 by substituting EquationEquations (145) and (Equation150) in EquationEquation (143).

First, substituting EquationEquation (145) in EquationEquation (143), we obtain

(151) trMMLEAtrMMLEAmin=trPYDYTPT+2trG(2)++QZ0Vm(12)TYTPT+tr(G(2)++QZ0Vm(12)TD1Vm(12)G(2)++QZ0T)trG(2)+Vm(12)TD1Vm(12)G(2)+T=trPYDYTPT+2trG(2)++QZ0Vm(12)TYTPT+trQZ0Vm(12)TD1Vm(12)Z0TQ+2trG(2)+Vm(12)TD1Vm(12)Z0TQ.(151)

In light of EquationEquation (148), the fourth term of the most right-hand side of EquationEquation (151) vanishes:

(152) 2trG(2)+Vm(12)TD1Vm(12)Z0TQ=trQG(2)+Vm(12)TD1Vm(12)Z0T=0.(152)

To derive this equation, we have used the following equation for arbitrary matrices A, B, and C:

(153) trABC=trCAB=trBCA.(153)

Second, let us substitute EquationEquation (150) in EquationEquation (151), then the second term of the most right-hand side of EquationEquation (151) can be written by

(154) 2trG(2)++QZ0Vm(12)TYTPT=2trPTG(2)++QZ0Vm(12)TYT=2trIZ1G(2)TQG(2)++QZ0Vm(12)TYT=2trIZ1G(2)TQZ0Vm(12)TYT=2trQZ0Vm(12)TYTIZ1G(2)T,(154)

where EquationEquations (153), (Equation49), (Equation148), and (Equation147) are used.

Finally, EquationEquation (151) is expressed as a function of Y, Z0, and Z1:

(155) trMMLEAtrMMLEAmin=trQIZ1G(2)YDYTIZ1G(2)TQ+2trQZ0Vm(12)TYTIZ1G(2)T+trQZ0Vm(12)TD1Vm(12)Z0TQ.(155)

Note that the right-hand side of this equation is in the quadratic form of Y, Z0, and Z1, and that it can be factorized. That is, we can rewrite EquationEquation (155) as

(156) trMMLEAtrMMLEAmin=trQIZ1G(2)YD+QZ0Vm(12)TD1QIZ1G(2)YD+QZ0Vm(12)TT.(156)

By expanding the right-hand side of EquationEquation (156), we can confirm that it is equal to the right-hand side of EquationEquation (155). As mentioned in Section 2.1.3, for an n×n matrix A and an n×m matrix B, if A is symmetric and nonnegative definite, BTAB is a symmetric nonnegative definite matrix. Moreover, for an arbitrary symmetric nonnegative definite matrix C, trC0 holds (e.g. Corollary 14.7.3 in reference [Citation13]). Recalling that D1 is symmetric and nonnegative definite, we find that EquationEquation (135) holds for arbitrary matrices Y, Z0, and Z1. Q.E.D.

4.3.3. Numerical verification

The correctness of EquationEquation (156) derived in the proof was verified by numerical calculations using a simplified case (n(1)=5, n(2)=3, na=36). In this numerical verification, random numbers were generated to determine the arbitrary matrices Y, Z0, and Z1; and the right-hand side of EquationEquation (156) was evaluated. Subsequently, G0(2) and G1(2) were determined by using EquationEquations (145) and (Equation149); and the left-hand side of EquationEquation (156) was obtained by using EquationEquation (6). Finally, we confirmed that the left-hand side of EquationEquation (156) agreed well with the right-hand side of EquationEquation (156).

Moreover, and show another calculation result of the difference of traces, namely trMMLEAtrMMLEAmin, for the simplified case. In this calculation, we defined by Y=yJ and Z0=Z1=zJ, where y and z are scalar variables, and J represents the all-ones matrix, in which all elements are unity. Thus, by using y and z, we changed Y and Z0(=Z1) around the zero matrix (i.e. around y=z=0). From these figure, we can see that the difference of traces becomes zero when Y=O and Z0=O (i.e. G0(2)=G(2)+). That is, trMMLEAmin has the minimum value among trMMLEA.

Figure 1. Difference of the traces of MMLEA and MMLEAmin as a function of Y(=yJ) and Z0(=Z1=zJ) within the ranges of 0y2 and 0z4.

Figure 1. Difference of the traces of MMLEA∗ and MMLEAmin as a function of Y(=yJ) and Z0(=Z1=zJ) within the ranges of 0≤y≤2 and 0≤z≤4.

Figure 2. Difference of the traces of MMLEA and MMLEAmin as a function of Y(=yJ) and Z0(=Z1=zJ) within the ranges of 2y0 and 4z0.

Figure 2. Difference of the traces of MMLEA∗ and MMLEAmin as a function of Y(=yJ) and Z0(=Z1=zJ) within the ranges of −2≤y≤0 and −4≤z≤0.

4.4. Role of assumption of normal distribution

With regard to MLCA, it was found that the formulation equivalent to CBCA can be derived by introducing the assumption of linear estimation expressed by EquationEquation (69). Meanwhile, the same formula can be derived by using the other assumptions, that is, MSCA and DRCA. For this reason, we examine these assumptions necessary for deriving the formulation equivalent to CBCA. summarizes the assumptions used in the derivation of CAs – including CBCA, MRCA, DRCA, MLCA, and MRCA.

Note that the assumption of normal distribution includes an assumption that the mean value is equal to the true value. Meanwhile, the mean value of the normal distribution is equal to the expectation value. Hence, we can understand that the assumption of normal distribution includes the assumption of unbiased estimation. Moreover, it should be noted that, when we assume the linear estimation of EquationEquation (70), it is necessary to introduce the assumption of minimum norm solution:

(157) Tˆ=T0+G(1)+FRe(1)Rc(1)(T0),(157)

Table 1. Assumptions used in the derivation of the conventional cross-section adjustment methods

where FRn(1)×n(1) is another linear combination factor (cf. EquationEquation (40) in reference [Citation9]). Note that the dimension of F differs from that of F used in the present derivation. We can see that both EquationEquations (157) and (Equation69) include a concept of dimensionality expansion in the case that na>n(1): the former expands the dimension by G(1)+Rna×n(1), and the latter by FRna×n(1). Thus, we can deduce that either the projection simplification or the dimensionality reduction is required to compensate the dimensionality expansion introduced by the assumption of minimum norm solution.

Moreover, from this table, it is seen that the assumption of normal distribution adopted in CBCA is equivalent to a set of the assumptions of:

  • unbiased estimation, linear estimation of EquationEquation (70), minimum norm solution, and projection simplification.

  • unbiased estimation, linear estimation of EquationEquation (70), minimum norm solution, and dimensionality reduction.

  • unbiased estimation and linear estimation of EquationEquation (69).

In summary, when we apply CBCA under the condition of the underdetermined problem, the assumption of normal distribution plays an important role to expand properly the information of the integral experimental quantities to the adjusted cross-section set.

4.5. Comparison with derivation based on Bayes theorem

From the derivations of MLCA, MLEA, MLRA, and MLEB, we can deduce that the assumptions of linear estimation expressed by EquationEquations (69)–(Equation71) correspond to conditional probabilities maximized in derivations based on the Bayes theorem:

(158) P(Tˆ|Re(1)),(158)
(159) P(Rˆ(1)|Re(1)),(159)

and

(160) P(Rˆ(2)|Re(1)),(160)

respectively, where P(A|B) denotes the probability distribution of A under the condition that B is given. In fact, EquationEquation (158) is used explicitly in the derivation of CBCA [Citation1,Citation4], and EquationEquation (160) is utilized in the derivation of CBEA [Citation8] in the modified form:

(161) P(Rc(2)(Tˆ)|Re(1)).(161)

Besides CBEA, a lot of new techniques based on the Bayes theorem have been recently proposed (e.g. [Citation16Citation18]). Hence, a detailed comparison with these techniques would be a future task. The foregoing correspondence between the linear estimations and the conditional probabilities, however, will be useful for the comparison of our methodologies with the other techniques based on the Bayes theorem.

5. Conclusions

In comparison with the classical Bayesian conventional cross-section adjustment (CBCA) method, the derivation of the Kalman filter was reviewed. This review revealed that CBCA is equivalent to a linearized model of the nonlinear Kalman filter, while the assumption of linear estimation used in the cross-section adjustment methodology differs from that in the Kalman filter. For this reason, by introducing the assumption used in the Kalman filter, we formulated the conventional, extended, and regressive cross-section adjustment methods (CA, EA, and RA) with no use of the assumption of normal distribution; these methods are called the conventional, extended, and regressive cross-section adjustment methods based on minimum variance unbiased linear estimation (MLCA, MLEA, and MLRA), respectively. Consequently, we found that MLCA has the formulation equivalent to CBCA. From these findings, we discussed the role of the assumption of normal distribution adopted in CBCA. Moreover, we demonstrated that the extended bias factor (EB) method can be formulated with the same procedure; this method is referred to as MLEB. The derivation of MLEB is simple, and would be helpful to understand the difference between EB and EA.

With regard to MLEA, we derived a generalized formulation that can express all the numerous solutions minimizing the variance of the design target core parameters. The formulation of MLEA includes a special case identical to the conventional Bayesian extended cross-section adjustment (CBEA) method; when we adopt the special case, the variance of adjusted cross-section set is also minimized. Thus, we can interpret that CBEA minimizes not only the variance of the design target core parameters but also the variance of the nuclear data. For this reason, we recommend to use the special case of MLEA, which is equivalent to CBEA, from a practical point of view.

On the other hand, in our review, we showed that CBCA is a type of the Kalman filter. Since many enhanced techniques based on the Kalman filter have been proposed in the other fields, we expect that an application of these techniques lead to further improvement in the cross-section adjustment methodology. Moreover, MLEA itself is an enhancement of the Kalman filter. Therefore, MLEA would be applicable in various fields other than nuclear reactor design.

Abbreviations

  1. derivation procedures

    CB : classical Bayesian (inference)

    MVUE : minimum variance unbiased estimation

    MVULE : minimum variance unbiased linear estimation

  2. variations of cross-section adjustment methodology

    CA : conventional cross-section adjustment method

    EA : extended cross-section adjustment method

    RA : regressive cross-section adjustment method

  3. CAs

    CBCA : classical Bayesian conventional cross-section adjustment method

    MRCA : MVUE-based rigorous conventional cross-section adjustment method

    MSCA : MVUE-based simplified conventional cross-section adjustment method

    MLCA : MVULE-based conventional cross-section adjustment method

    DRCA : dimension-reduced conventional cross-section adjustment method

  4. EAs

    CBEA : classical Bayesian extended cross-section adjustment method

    MREA : MVUE-based rigorous extended cross-section adjustment method

    MSEA : MVUE-based simplified extended cross-section adjustment method

    MLEA : MVULE-based extended cross-section adjustment method

  5. RA

    MLRA : MVULE-based regressive cross-section adjustment method

  6. extended bias factor methods

    EB : (original) extended bias factor method

    MLEB : MVULE-based extended bias factor method

Acknowledgments

The authors express their deep gratitude to the coauthors of references [8–10], since the present study is due to discussion with them through the series of studies. The authors are also grateful to Dr S. Takeda of Osaka University for useful comments on the present study. K. Yokoyama wishes to extend his special thanks to Mr M. Ishikawa of JAEA for great encouragement and helpful discussion based on his expertise.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

  • Dragt JB. Statistical considerations on techniques for adjustment of Differential cross sections with measured integral parameters. In: STEK, The fast-thermal coupled facility of RCN at Petten. RCN-122. Petten: Reactor Centrum Nederland; 1970. p. 85–105.
  • Gandini A, Petilli M, Salvatores M. Nuclear data and integral measurement correlation for fast reactors. Statistical formulation and analysis of methods. The Consistent Approach. International Symposium of Physics of Fast Reactors, 1973 October 16-19. Tokyo, Japan; Vol. 1. p. 612–628.
  • Dragt JB, Dekker JWM, Guppelaar H, et al. Methods of adjustment and error evaluation of neutron capture cross sections; Application to fission product nuclides. Nucl Sci Eng. 1977;62:117–129.
  • Takeda T, Yoshimura A, Kamei T. Prediction uncertainty evaluation methods of core performance parameters in large liquid-metal fast breeder reactors. Nucl Sci Eng. 1989;103:157–165.
  • Salvatores M, Palmiotti G, Aliberti G, et al. Methods and issues for the combined use of integral experiments and covariance data: results of a NEA International Collaborative Study. Nuclear Data Sheets. 2014;118:38–71.
  • de Saint-Jean C, Dupont E, Ishikawa M, et al. Assessment of existing nuclear data adjustment methodologies. Paris: OECD/NEA; 2011. NEA/NSC/WPEC/DOC(2010)429.
  • Kugo T, Mori T, Takeda T. Theoretical study on new bias factor methods to effectively use critical experiments for improvement of prediction accuracy of neutronic characteristics. J Nucl Sci Technol. 2007;44(12):1509–1517.
  • Yokoyama K, Ishikawa M, Kugo T. Extended cross-section adjustment method to improve the prediction accuracy of core parameters. J Nucl Sci Technol. 2012;49(12):1165–1174.
  • Yokoyama K, Yamamoto A. Cross-section adjustment methods based on minimum variance unbiased estimation. J Nucl Sci Technol. 2016;53(10):1622–1638.
  • Yokoyama K, Yamamoto A, Kitada T. Dimension-reduced cross-section adjustment methods based on minimum variance unbiased estimation. J Nucl Sci Technol. 2018;55(3):319–334.
  • Kalman RE, New A. Approach to linear filtering and prediction problems. Trans ASME J Basic Eng. 1960;82:35–45.
  • Kalman RE, Bucy RS. New results in linear filtering and prediction therory. Trans ASME J Basic Eng. 1961;83:95–108.
  • Harville DA. Matrix algebra from a statistician’s perspective. 1st ed. New York (NY): Springer; 1997.
  • Brown RG, Hwang PYC. Introduction to random signals and applied Kalman filtering. 4th ed. Hoboken (NJ): John Wiley & Sons, Inc; 2012.
  • Palmiotti G, Salvatores M, Yokoyama K, et al. Methods and approaches to provide feedback from nuclear and covariance data adjustment for improvement of nuclear data files. Paris: OECD/NEA; 2017. NEA/NSC/WPEC/DOC(2016)6.
  • Cacuci DG, Ionescu-Bujor M. Best-estimate model Calibration and Prediction Through Experimental Data Assimilation – I: mathematical Framework. Nucl Sci Eng. 2010;165:18–44.
  • Hoefer A, Buss O, Hennebach M, et al. MOCABA: A general Monte Carlo-Bayes pro- cedure for improved predictions of integral functions of nuclear data. Ann Nucl Energy. 2015;77:514–521.
  • Endo T, Yamamoto A, Watanabe T. Bias factor method using random sampling technique. J Nucl Sci Technol. 2016;53:1494–1501.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.