NMNR-Unit-1 â Eigenvalue Problems and Solution Strategies â Inner Iterations. 1/60 ..... Therefore, with respect to
Lecture Notes for the Course on NUMERICAL MODELS FOR NUCLEAR REACTORS Prof. WALTER AMBROSINI University of Pisa, Italy
Unit 1 – Eigenvalue Problems with Neutron Diffusion and Solution Strategies: Inner Iterations
NOTICE: These notes were prepared by Prof. Ambrosini mainly on the basis of the material adopted by Prof. Bruno Montagnini for the lectures he held up to years ago, when he left to Prof. Ambrosini the charge of the course held for the Degree in Nuclear Engineering at the University of Pisa. This material is freely distributed to Course attendees or to anyone else requesting it. It has not the worth of a textbook and it is not intended to be an official publication. It was conceived as the notes that the teacher himself would take of his own lectures in the paradoxical case he could be both teacher and student at the same time (sometimes space and time stretch and fold in strange ways). It is also used as slides to be projected during lectures to assure a minimum of uniform, constant quality lecturing, regardless of the teacher’s good and bad days. As such, the material contains reference to classical textbooks and material whose direct reading is warmly recommended to students for a more accurate understanding. In the attempt to make these notes as original as feasible and reasonable, considering their purely educational purpose, most of the material has been completely re-interpreted in the teacher’s own view and personal preferences about notation. In this effort, errors in details may have been introduced which will be promptly corrected in further versions after discovery. Requests of clarification, suggestions, complaints or even sharp judgements in relation to this material can be directly addressed to Prof. Ambrosini at the e-mail address:
[email protected]
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
1/60
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
2
EIGENVALUE (OR CRITICALITY) PROBLEMS • The static neutronic design of nuclear reactor core involves different aspects related to: ♦ the calculation of the k eff and of the flux spatial distribution in different configurations ♦ the prediction of the evolution of reactor isotopical composition with the increase of burn-up: the so-called “life” or “depletion” calculations • Eigenvalue problems, which can be expressed also in forms other than the simple calculation of k eff (e.g., identifying the critical absorption cross section, liquid boron concentration or the critical reactor size) are quite relevant in the definition of reactor core parameters as the lattice pitch, the level of enrichment and control rod characteristics • In heterogeneous reactors eigenvalue problems require the numerical solution of neutron diffusion (or more seldom transport) equations dealing with detailed geometry and isotopic composition data • Both space and energy discretization is required in this step
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
3
CRITICALITY CONDITIONS FOR A PLANE HOMOGENEOUS SLAB Reminder of the analytic solution One-group differential problem
νΣ f d 2φ D φ =0 − Σ aφ + dx 2 k φ (0) = φ (a ) = 0
0
a
x
The following positions can be adopted:
d 2φ dx
2
=
Σ a −ν Σ f k D
µ = µ (k ) =
φ = µφ
Σ a −ν Σ f k D
The general solution of the above equation is
φ ( x ) = C1e
µx
+ C2e −
µx
Boundary conditions are then imposed
φ (0) = C1 + C 2 = 0
φ (a ) = C1e
µa
+ C2e −
µa
=0
This homogeneous algebraic linear system of equations has a nontrivial solution only if
1 det µ a e
1 − µa = 0 e
i.e., when
⇒
e− µ a − e µ a = 0 ⇒ e2 2 µ a = ln(1) + i 2nπ = i 2nπ
µa
=1 (n = 0,±1,±2, ...)
In fact, it can be checked that it is
e ln (1)+ i 2 nπ = e 0 [cos(2nπ ) + i sen(2nπ )] = 1
(n = 0,±1,±2, ...)
Therefore, excluding n = 0 leading to the trivial solution , it must be:
µn = −
n 2π 2 a
2
(n = 1,2, ...)
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
4
1 Normalised Harmonic Mode
0.8 0.6 0.4 0.2 0 -0.2 n=1 n=2 n=3
-0.4 -0.6 -0.8 -1 0
0.2
0.4
0.6
0.8
1
Normalised Coordinate
Since in the definition of µ only k is variable, it is:
µn =
Σa − ν Σ f kn D
n2π2
=−
⇒
a2
νΣ f
kn =
Σa + D
nπ a
φ n ( x ) = C sen
As well known, it is
(n = 1,2, ...)
n2π2 a2
x
(n = 1,2, ...)
Considering the different harmonic modes it can be observed that • the eigenvalues k are such as it is k1 > k2 > ... • the fundamental mode is the only one being characterised by φ1 (x ) > 0 for 0 ≤ x ≤ a ; • on the other hand, we know from Reactor Physics that it is the one on which the neutron flux settles in steady-state conditions and it is 2
2
B = B1 =
π2
d 2φ
φ ( x ) = φ1 ( x )
a2
dx 2
= − B 2φ
• therefore, for steady-state problems, considering that the layer of multiplying material is homogeneous it can be shown that k1 =
ν Σf Σ a + DB 2
a
=
(Σ
ν Σ f ∫ φdx 0
a
+ DB 2
)∫ φdx a
0
a
=
∫ ν Σ f φ dx 0
a
∫0
a
Σ aφ dx + ∫ DB 2φ dx 0
a
∫0ν Σ f φ dx
= a
a
d 2φ
∫0 Σ aφ dx −∫0 D dx 2 dx
or k1 =
neutrons produced by fissions = keff total number of neutrons in the previous generation (absorption + leakage)
i.e., k1 is found to be the effective multiplication factor NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
5
CRITICALITY CONDITIONS FOR A PLANE HOMOGENEOUS SLAB Numerical Solution
x0 = 0
x1
xi
x i +1
xn
x n +1 = a
x
The spatial dicretisation of the problem can be obtained by subdividing the range (0, a ) into segments having uniform length, h , and putting φ i = φ (xi ) Different discretization approaches can be adopted for turning the differential equation into an algebraic one; here we choose
dφ dx
≈
φ i +1 − φ i h
x = xi
(forward finite difference form)
However, we could also assume
dφ dx
≈
φi − φi −1 h
x = xi
(backward finite difference form)
A closer examination, as well as the rigorous application of Taylor expansion, reveals that the two approximations are actually more accurate for expressing the derivatives at the locations x i +1 2 = ( x i + x i +1 ) / 2 and x i −1 2 = ( x i −1 + x i ) / 2 respectively; that is:
dφ dx
≈ x = xi +1 2
φi +1 − φi
dφ dx
h
≈ x = xi −1 2
φi − φi −1 h
In order to obtain an approximation of the second order derivative, a symmetric expression (“centred” in x i ) can be then assumed
d 2φ dx 2
x = xi
1 dφ ≈ h dx
x = xi +1 2
=
1 φ −φ φ −φ i −1 ≈ i +1 i − i h h h x = xi −1 2 φi +1 − 2φi + φi −1
dφ − dx
h2
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
6
Making use of these approximations, the differential equation expressing 1D diffusion in the layer can be reverted to this finite difference approximation form
D
φi +1 − 2φi + φi −1 h
2
νΣ f + − Σ a φi = 0 k
(i = 1, ..., n )
while the boundary conditions take the form
φ0 = 0
φ n+1 = 0
A system of n equations in the n unknowns of the nodal fluxes {φ1 , φ 2 , ..., φ n } is then obtained. Putting
ai ,i −1 = ai ,i +1 =
D h2
ai ,i = −
2D ν Σ f + − Σa 2 k h
a homogeneous linear algebraic system is obtained, having the classical form of a three-point equation
(i = 1, ..., n )
a i ,i −1φ i −1 + a i ,i φ i + a i ,i +1φ i +1 = 0 Writing the system in the form Aφ = 0 it can be noted that: • in the system matrix the only coefficients being different from zero can be found in the main diagonal and the two adjoining ones: the matrix is three-diagonal
× × 0 0 A= 0 0 0 0
× × × 0 0 0 0 0
0 × × × 0 0 0 0
0 0 × × × 0 0 0
0 0 0 × × × 0 0
0 0 0 0 × × × 0
0 0 0 0 0 × × ×
0 0 0 0 0 0 × ×
• the neutron flux in each node is therefore directly related only to the one in the neighbouring nodes, a numerical effect clearly deriving by the discretization of the leakage term (the second order derivative).
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
7
SOLUTION (at least from a purely theoretical standpoint): • The “characteristic equation” det A ( k ) = 0 can be solved by any suitable means, identifying the n single or multiple eigenvalues k1 , k 2 , ..., k n . •
The eigenvectors corresponding to each k i are then identified: ♦ an appropriate non-zero minor of A(ki ) having n − 1 order can be selected; ♦ the corresponding algebraic system of order n − 1 obtained by assigning and arbitrary value to one of the unknowns φ i can be solved by any suitable technique.
Therefore, with respect to the analytic solution, the finite difference discretised equation provides only n eigenvalues, representing approximations of the eigenfunctions of the problem in the continuum space. φ (1)
φ1(1) ≡ φn(1)
φ (n)
φ1( n ) ≡ φn( n )
For instance, in this particular case it can be demonstrated that (we skip the demonstration):
φ (j1) = C sen and
k = k1 =
π hj a
νΣf 2D π h Σ a + 2 1 − cos h a
( j = 1, ..., n ) →
h →0
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
νΣf Σa + D
π2 a2
8
NOTES: • In most neutronic problems, it is just necessary to evaluate the fundamental eigenvalue, k1 , and the corresponding fundamental eigenvector, φ (1) . • Obviously in real life reactors, an accurate multidimensional space and energy description is necessary, aiming at determining reliable estimates of the k eff and of the steady-state neutron flux distribution. • Only in some selected cases, it is necessary to get information about the higher order harmonic modes, as it is the case for Boiling Water Reactors (BWRs) to evaluate the possible occurrence of “out-ofphase” or “regional” oscillations due to the thermal-hydraulic feedback.
Core − wide
Regional
Power Oscillations
Power Oscillations
• In fact, the thermal-hydraulic feedback may make more probable the occurrence of normally “subcritical” modes, which may mostly govern the transient evolution.
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
9
MULTIGROUP EIGENVALUE PROBLEMS A multigroup eigenvalue problem can be expressed making use of the diffusion theory in the form
div Dg (r )grad r φ g (r ) − Σ r , g (r ) φ g (r ) + +
χg
k
( ) ( Σ r φ r ∑ s, g′→g g′ )
g ′< g
G
( ) ( ν Σ r φ r ∑ f ,g′ g′ ) = 0
g ′=1
φ g (r ) + d g
dφ g ( r )
dn
(r ∈V , g = 1, ..., G ) (r ∈ ∂V ,
=0
g = 1, ..., G )
in which k is the parameter giving rise to eigenvalues; it is also
∫ χ(E )dE
χg =
∆E g
i.e., χ g is the fraction of neutrons produced by fission in g-th energy group and the “removal cross section” is defined as
Σ r , g (r ) = Σ a , g (r ) +
∑ Σ s, g →g′ (r )
g ′> g
In order to solve these equations, it can be noted that, defining the total fission neutron source as
ψ(r ) =
G
ν Σ ( r ) φ ( r ∑ f , g′ g′ )
g ′=1
and putting
S g (r ) = χ gψ (r )
the steady-state diffusion equations with G energy groups become 1 div D1 (r )grad r φ1 (r ) − Σ r ,1 (r ) φ1 (r ) + S1 (r ) = 0 k 1 div D2 (r )grad r φ 2 (r ) − Σ r , 2 (r ) φ 2 (r ) + Σ s,1→2 (r ) φ1 (r ) + S 2 (r ) = 0 k 2 1 div D3 (r )grad r φ 3 (r ) − Σ r ,3 (r ) φ 3 (r ) + ∑ Σ s , g ′→3 (r ) φ g ′ (r ) + S 3 (r ) = 0 k g ′=1 etc.. NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
10
It can be noted that:
• the first equation involves only the fission source and the neutron flux in the first energy group; • the subsequent equations involve the fission source, the neutron flux in the group and the slowing down terms from previous (higher energy) groups; • the initial homogeneous problem has been formally turned into a non-homogeneous one. It is therefore possible to envisage a solution procedure based on the following steps:
1. a tentative distribution of the fission source, ψ (r ) , and a tentative value of the multiplication factor, k , are assigned as ψ (0 ) (r ) and k (0 ) respectively; this allows obtaining the group sources S g(0 ) (r ) 2. after an appropriate spatial discretization (see later on), the group equations are solved in sequence:
• the first equation is solved, obtaining an updated estimate of the neutron flux in the first group, φ1(1) (r ) ; • the slowing down source from the first group appearing in the second equation can be therefore evaluated, then solving the equation for φ 2(1) (r ) ; • a similar procedure is adopted for all the higher index group equations, evaluating the appropriate slowing down sources from higher energy groups 3. the fission source can be then updated on the basis of the new fluxes, obtaining ψ (1) (r ) , and the estimate of the eigenvalues is also updated according to a “generational” formulation () ∫ψ ( r )dV 1
k( ) = 1
neutrons generated by fissions at the new iteration = V 1 neutron source adopted at the previous iteration ψ ( 0) ( r )dV ( 0) ∫ k V
then, the sources S g(1) (r ) are calculated again and the processi s repeated from step 2 until convergence. NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
11
The scheme includes two different phases:
• iterations on the fission source, named outer iterations; • the solution of the equation of single group equations, performed as if they were representing mono-energetic decoupled equations having the form
div D(r )grad r φ(r ) − Σ a (r ) φ(r ) + S (r ) = 0
since very often the solution of these equations is performed by using iterative methods, this step is named inner iterations.
The adopted iteration scheme is described in the figure reported in the next page. NOTES:
• The scheme is general enough to be applied making use of any operator describing the neutronic behaviour, including the transport equation. In fact, by changing the diffusion operator with the more accurate transport one, the same “lower triangular” structure of the slowing down matrix Σ s , g ′→ g ( r ) appears, whenever a slowing down process is addressed. This allows to apply the same strategy, whenever a suitable spatial discretization of the transport operator is adopted.
• The lower-triangular structure of the differential scattering matrix is obviously lost when thermalisation problems are dealt with. In fact, when the energy of neutrons becomes comparable with that of target nuclei, “up-scattering” may occur in addition to “down-scattering”. In such cases, the scattering source becomes as “coupled” as the fission source. A similar treatment of these terms can be therefore envisaged.
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
12
Initial estimate of k and ψ
Solution of the steady diffusion equation in each group inner iterations
g =1
inner iterations
g=2
Outer iterations
inner iterations
g =G
Updating k and ψ and group sources
Convergence test
NO
YES END
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
13
INNER ITERATIONS FORM OF THE FINITE DIFFERENCE EQUATIONS • As explained above, a key point in solving eigenvalues problems is the numerical solution of group equations having the general monokinetic non-homogeneous form:
− div D(r )grad r φ(r ) + Σ a (r )φ(r ) = S (r )
• In the light of numerical solution, it is important to highlight the main features of the algebraic linear systems obtained by the discretization of diffusion equations in one or more space dimensions
One-dimensional Problems • In the case of the criticality problem related to a slab containing multiplicating material dealt with above, it was shown that the obtained linear system was homogeneous and had a tri-diagonal matrix; i.e., three-point equations were obtained:
(i = 1, ..., n )
a i ,i −1φ i −1 + a i ,i φ i + a i ,i +1φ i +1 = 0
• It is clearly understood that a similar discretization in the case of a fixed source (non-homogeneous) problem for a uniform property slab
−D
d 2φ dx
2
+ Σ aφ ( x ) = S
φ (0 ) = φ (a ) = 0
provides three-point formulations
−D
φi +1 − 2φi + φi −1 h
2
+ Σ aφ i = S i
φ 0 = φ n+1 = 0
In particolar, putting
ai ,i −1 = ai ,i +1 = −
D h
2
ai ,i =
it is
ai ,i −1φ i −1 + ai ,iφ i + ai ,i +1φ i +1 = S i NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
2D h
2
+ Σa
(i = 1, ..., n ) 14
• However, uniform property problems are scarcely relevant for practical applications; it is by far more interesting to deal with multi-layer problems EXERCISE Let us consider two different ways of providing a discretization of such problems by a “finite volume” technique 1st METHOD: fluxes evaluated at the interfaces between nodes
Vi −
Vi +
Ai−
r0
r1
Ai+
ri −1 ri −1 2 ri
ri +1 2 ri +1
r
rN
• We now assume the presence of material layers having different nuclear properties in 1D plane, cylindrical or spherical geometry • Intervals (“nodes”) are assigned such as each layer contains an integer number of them and each interval has uniform properties • Fluxes are evaluated at the interfaces between two intervals • Each “flux point” is assigned a control volume Vi = Vi − ∪ Vi+ , made by half of the left and half of the right intervals; in general, these two halves have different nuclear properties • Volumes Vi − and Vi+ the surfaces Ai− e Ai+ separating them are defined as follows: ♦ plane geometry: Vi − = ri − ri −1 2 , Vi + = ri +1 2 − ri , Ai− = Ai+ = 1 ♦ cylindrical
geometry:
(
)
Vi− = π ri2 − ri2−1 / 2 ,
(
)
Vi + = π ri2+1 / 2 − ri2 ,
Ai− = 2πri −1 2 , Ai+ = 2πri +1 2
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
15
♦ spherical
geometry:
Vi − =
(
)
4 π ri3 − ri3−1 / 2 , 3
Vi + =
(
)
4 π ri3+1 / 2 − ri3 , 3
Ai− = 4π ri2−1 / 2 , Ai+ = 4π ri2+1 / 2
in which it is ri −1 / 2 = (ri −1 + ri ) / 2 and ri +1 / 2 = (ri + ri +1 ) / 2
• Diffusion equations are therefore integrated throughout Vi :
− ∫ divD grad φ dV + ∫ Σ a φ dV = ∫ S dV Vi
Vi
Vi
♦ the Gauss (divergence) theorem is made use of
∫ divD grad φ dV = Vi
∫
D
Ai+
dφ dφ dφ dA − ∫ D dA = Di+ Ai+ du du du A− i
− Di− Ai− ri +1 / 2
dφ du
ri −1 / 2
♦ a linear distribution of the flux in each interval is assumed for purpose of approximating gradients + + φi +1 − φi − − φi − φi −1 φ = − divD grad dV D A D i i i Ai ∫V − r r ri − ri −1 i + 1 i i ♦ absorption and source terms are finally evaluated
∫ Σ φ dV = (Σ a
)
− − a ,i i
V + Σ +a ,iVi + φi
Vi
∫ S dV = S V
− − i i
+ S i+Vi +
Vi
• In the layers close to the boundaries 3rd kind boundary conditions are assumed
γ 0 φ(r0 ) + δ 0
dφ dr
=0
γ N φ(rN ) + δ N
r0
dφ =0 dr rN
allowing to assume neutron flux zeroing at the physical or the extrapolated boundary by appropriate choices of the γ and δ coefficients • It is easy to check that also in this case three-point formulas are obtained NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
16
ai ,i −1φi −1 + ai ,iφi + ai ,i +1φi +1 = bi
( i = 0,..., N )
where the unknowns are N + 1 since the boundary fluxes are generally non-zero. It is, in fact:
− ∫ divD grad φ dV + ∫ Σ a φ dV = ∫ S dV Vi
− Di+ Ai+
φi +1 − φi ri +1 − ri
Vi
+ Di− Ai−
φi − φi −1 ri − ri −1
(
Vi
)
+ Σ a−,iVi − + Σ a+,iVi + φi = Si−Vi − + Si+Vi +
Di+ Ai+ Di+ Ai+ Di− Ai− Di− Ai− − + φi +1 + φi −1 φi − ri +1 − ri ri − ri −1 ri +1 − ri ri − ri −1
(
)
+ Σ −a ,iVi − + Σ +a ,iVi + φi = Si−Vi − + Si+Vi +
or Di+ Ai+ Di− Ai− Di− Ai− Di+ Ai+ − − + + − φi −1 + + + Σ a ,iVi + Σ a ,iVi φi − φi +1 ri − ri −1 r − r r − r r − r +1 i −1 +1 i i i ii ai ,i −1
ai , i +1
ai , i
= Si−Vi − + Si+Vi + bi
The boundary conditions can be imposed by specializing the above equation, valid for internal nodes, to the case of “boundary” nodes. 2nd METHOD: fluxes evaluated at mid-node locations
Vi Ai−
Ai+
r1 / 2 r1 ri −1 2 ri ri +1 2 NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
rN rN +1 / 2
r 17
• Intervals (“nodes”) are assigned such as each layer contains an integer number of them and each interval has uniform properties • Fluxes are evaluated in the centres of each interval (nodes) • Each node is assigned the volume Vi of the interval • In this case, the volume Vi and the surfaces Ai− e Ai+ are defined by ♦ plane geometry: Vi = ri +1 / 2 − ri −1 2 , Ai− = Ai+ = 1 ♦ cylindrical geometry: Vi = π(ri2+1 / 2 − ri2−1 / 2 ), Ai− = 2πri −1 2 , Ai+ = 2πri +1 2 ♦ spherical
Vi =
geometry:
(
)
4 π ri3+1 / 2 − ri3−1 / 2 , 3
Ai− = 4π ri2−1 / 2 ,
Ai+ = 4π ri2+1 / 2
where ri −1 / 2 = (ri −1 + ri ) / 2 and ri +1 / 2 = (ri + ri +1 ) / 2
• Diffusion equations are again integrated throughout each Vi
− ∫ divD grad φ dV + ∫ Σ a φ dV = ∫ S dV Vi
Vi
Vi
♦ use is made of the divergence theorem
∫ divD grad φ dV = ∫
Ai+
Vi
D
dφ dφ dφ dA − ∫ D dA = Di Ai+ du du du A− i
− Di Ai− ri +1 / 2
dφ du
ri −1 / 2
♦ a linear distribution of the flux is assumed within each semiinterval connecting the node to each interface + ∫ divD grad φ dV = Di Ai Vi
φ i +1 / 2 − φi φ − φi −1/ 2 − Di Ai− i ri +1 / 2 − ri ri − ri −1/ 2
♦ interfacial fluxes thus apperaing are eliminated by assuming the continuity of currents throuh the interfaces; e.g.: − Di
φi +1/ 2 − φi ri +1/ 2 − ri
= − Di +1
φi +1 − φi +1/ 2 ri +1 − ri +1/ 2
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
18
Di
φ i +1 / 2
Di +1 φ i +1 ri +1 / 2 − ri ri +1 − ri +1 / 2 = Di Di +1 + ri +1 / 2 − ri ri +1 − ri +1 / 2 φi +
♦ absorption and source terms are then evaluated
∫ Σ a φ dV = Σ a,iVi
∫ S dV = Si Vi
φi
Vi
Vi
♦ It is then obtained φ −φ φ −φ − Di Ai+ i +1/ 2 i + Di Ai− i i −1/ 2 + Σ a ,iVi φ i = S i Vi ri +1/ 2 − ri ri − ri −1/ 2 that represents again a three-point formula in the nodal fluxes, if it is considered that
φi +1 / 2
Di Di +1 φi + φi +1 ri +1 / 2 − ri ri +1 − ri +1 / 2 = Di Di +1 + ri +1 / 2 − ri ri +1 − ri +1 / 2
φ i −1 / 2
Di −1 Di φ i −1 + φi ri −1 / 2 − ri −1 ri − ri −1 / 2 = Di −1 Di + ri −1 / 2 − ri −1 ri − ri −1 / 2
In fact, let’s check that: Di +1 Di + φ φ i i 1 + φ D A ri +1/ 2 − ri ri +1 − ri +1/ 2 −φ − Di Ai+ i +1/ 2 i = − i − φi Di Di +1 ri +1/ 2 − ri ri +1/ 2 − ri + ri +1/ 2 − ri ri +1 − ri +1/ 2 + i
Di Di Di +1 Di +1 φ + φ − φ + i i +1 i ri +1 − ri +1/ 2 r r r r − − Di Ai+ ri +1/ 2 − ri i +1/ 2 i i +1 i +1/ 2 =− Di Di +1 ri +1/ 2 − ri + r r r r − − i 1 / 2 i i 1 i 1 / 2 + + + Di
Di +1 r −r r −r = − i +1/ 2 i i +1 i +1/ 2 Ai+ (φi +1 − φi ) Di Di +1 + ri +1/ 2 − ri ri +1 − ri +1/ 2
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
19
In order to understand the meaning of the above formulation, let us rewrite it as: + φi +1 / 2 − φi right volume current × surface area = − Di Ai
ri +1/ 2 − ri
Di
Di +1 r −r r −r D (φi +1 − φi ) = − i +1/ 2 i i +1 i +1/ 2 Ai+ (φi +1 − φi ) = − Ai+ Di Di +1 ∆ r i ,i +1 + ri +1/ 2 − ri ri +1 − ri +1/ 2 weighetd averaged value of D / ∆r between nodes i and i +1
___________________________________________ Suggested personal study activity Prove that
Di Ai−
φi − φi −1/ 2 ri − ri −1/ 2
Di −1 Di r −r r −r D = i i −1/ 2 i +1/ 2 i Ai− (φi − φi −1 ) = Ai− (φi − φi −1 ) Di −1 Di ∆r i −1,i + ri − ri −1/ 2 ri +1/ 2 − ri
as it can be obtained from the previous relation by an appropriate change of the indices.
Noting the above we recognise that the steady neutron balance equation in the volume φ −φ φ −φ − Di Ai+ i +1/ 2 i + Di Ai− i i −1/ 2 + Σ a ,iVi φ i = S i Vi ri +1/ 2 − ri ri − ri −1/ 2 can be rewritten as: D D − Ai+ (φi +1 − φi ) + Ai− (φi − φi −1 ) + Σ a ,i Vi φi = S i Vi ∆r i ,i +1 ∆r i −1,i source rate absorption rate leakage rate of neutrons at the right surface leakage rate of neutrons at the left surface
holding for each internal point of the discretization. So, it is: D D ai ,i +1 = − Ai+ ai ,i −1 = − Ai− ∆r i ,i +1 ∆r i −1,i D D ai ,i = Ai+ + Ai− + Σ a ,i Vi ∆r i ,i +1 ∆r i −1,i
bi = Si Vi
and ai ,i −1φi −1 + ai ,iφi + ai ,i +1φi +1 = bi NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
20
In the layers close to the boundaries 3rd kind boundary conditions can be again assumed
γ 0 φ(r0 ) + δ 0
dφ dr
γ N φ(rN ) + δ N
=0 r0
dφ =0 dr rN
thus closing the problem. ____________________________________________________________ Suggested personal study activity Try to express the formulations for different boundary conditions in the first and the last node. Hint Restart from the equation
− Di Ai+
φi +1/ 2 − φi ri +1/ 2 − ri
+ Di Ai−
φi − φi −1/ 2 ri − ri −1/ 2
+ Σ a,iVi φ i = S i Vi
and assume: a) i = 1 and φ1 / 2 = 0 ; what does this condition mean? how does the equation for the first node look like? φ1 − φ1 / 2 = 0 ; what does this condition mean? how does the equation b) i = 1 and r1 − r1 / 2 for the first node look like? φ1 − φ1/ 2 + γφ1/ 2 = 0 ; what does this condition mean? how does the c) i = 1 and δ 1 2 r1 − r1/ 2 equation for the first node look like? Further Hint For the first node the equation is:
− D1 A1+
φ3 / 2 − φ1 r3 / 2 − r1
+ D1 A1−
φ1 − φ1/ 2 r1 − r1/ 2
+ Σ a ,1V1 φ1 = S1 V1
or also (remember the above developments)
φ −φ D − Ai+ (φ2 − φ1 ) + D1 A1− 1 1/ 2 + Σ a ,1V1 φ1 = S V 1 1 r1 − r1/ 2 ∆r 1, 2 So, work on this form. You must end up with a formulation that has the form:
a1,1φ1 + a1,2φ2 = b1 Repeat the above for the last node i = N with N the number of intervals.
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
21
Suggested personal study activity Try showing that in case of Cartesian coordinates, constant diffusion coefficient and uniform spatial discretization the equation
D − Ai+ (φi +1 − φi ) ∆r i ,i +1
D + Ai− (φi − φi −1 ) + Σ a ,i Vi φi = S i Vi ∆r i −1,i source rate absorption rate
leakage rate of neutrons at the right surface leakage rate of neutrons at the left surface
is equivalent to the “finite difference” approximation
−D
φi +1 − 2φi + φi −1 h2
+ Σ a ,iφi = Si
NB: These discretization techniques can be used also for other physical problems (e.g., for heat conduction, electrostatic potential) having an elliptic character
− div k (r ) grad r T (r ) = q ′′′(r ) div ε(r ) grad rV (r ) = 0
As a general conclusion: The finite volume discretization of steady-state diffusion equations in 1D leads to linear algebraic systems having a three-diagonal matrix This conclusion can be extended to some discretization schemes other than finite volumes, e.g., to “coarse-mesh” methods to be described later on
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
22
SOLUTION OF THREE-DIAGONAL MATRIX SYSTEMS • The solution of TDM systems is particularly simple by an algorithm being a special case of the Gauss elimination scheme. • The matrix is represented in the form (e.g., for N = 8 ): v1 u 2 0 0 0 0 0 0
w1 v2
0 w2
0 0
0 0
0 0
0 0
u3 0 0
v3 u4 0
w3 v4 u5
0 w4 v5
0 0 w5
0 0 0
0 0
0 0
0 0
u6 0
v6 u7
w6 v7
0
0
0
0
0
u8
0 φ1 S 1 0 φ 2 S 2 0 φ 3 S 3 0 φ 4 S 4 ⋅ = 0 φ 5 S 5 0 φ 6 S 6 w 7 φ 7 S 7 v 8 φ 8 S 8
putting
v i = a i ,i
u i = a i ,i −1
wi = a i ,i +1
The following two-sweeps technique is adopted known as Thomas’ algorithm (or TDMA=Three-Diagonal Matrix Algorithm) 1. with increasing index (first sweep):
α1 =
w1 v1
β1 =
S1 v1
αi =
wi v i − u i α i −1
βi =
S i − u i β i −1 v i − u i α i −1
(i = 2, ..., N − 1) (i = 2, ..., N )
2. with decreasing index (second sweep):
φN = βN
φ i = β i − α i φ i +1
In particular:
• the first sweep eliminates an unknown in the equations; • the second is needed to solve the lower triangular system thus obtained. NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
23
Multidimensional Problems • Let us consider the quarter reactor lattice represented in the figure
y
• It is discretised through a 2D grid with rectangular (or square) meshes yj • If the reactor is homogeneous, the steady diffusion equation has the form d 2φ d 2φ xi − D 2 + 2 + Σ a φ = S dx dy • By a simple finite-difference equation, it is φ i +1, j − 2φ i , j + φ i −1, j φ i , j +1 − 2φ i , j + φ i , j −1 +Σ φ = S − D + a i, j i, j 2 2 h h x y clearly representing a 5 point formulation
x
• Similarly, for a three-dimensional problem d 2φ d 2φ d 2φ − D 2 + 2 + 2 + Σ a φ = S dy dz dx a 7 point formula is obtained φ i +1, j ,k − 2φ i , j ,k + φ i −1, j ,k − D h x2 +
φ i , j +1,k − 2φ i , j ,k + φ i , j −1,k h y2
+
φ i , j ,k +1 − 2φ i , j ,k + φ i , j ,k −1 + Σ a φ i , j ,k = S i , j ,k 2 hz
• For non-uniform properties, similar techniques as the ones shown for the 1D case can be applied • In fact, though the above formulations are related to the case of uniform properties, it is quite clear that non-uniformity does not change the structure of the involved matrices
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
24
CHARACTERISTICS OF THE OBTAINED LINEAR SYSTEM MATRICES • We just noted that in 1D cases three-diagonal matrices are obtained, to be efficiently dealt with by the Thomas algorithm • In multi-dimensional cases, 5 or seven point formulas are instead obtained; the single-index numbering adopted for the grid nodes determines the structure of the matrix • For instance, in the case of a homogeneous grid with 9 nodes numbered as in the figure, a following banded matrix system is obtained a11 a 21 0 a 41 0 0 0 0 0
a12
0
a14
0
0
0
0
a 22 a 32
a 23 a 33
0 0
a 25 0
0 a 36
0 0
0 0
0 a 52
0 0
a 44 a 54
a 45 a 55
0 a 56
a 47 0
0 a 58
0 0
a 63 0
0 a 74
a 65 0
a 66 0
0 a 77
0 a 78
0 0
0 0
0 0
a 85 0
0 a 96
a87 0
a88 a 98
7
8
9
4
5
6
1
2
3
0 φ1 S 1 0 φ 2 S 2 0 φ 3 S 3 0 φ 4 S 4 0 ⋅ φ 5 = S 5 a 69 φ 6 S 6 0 φ 7 S 7 a89 φ 8 S 8 a 99 φ 9 S 9
where a i ,i = Σ a +
4D h
2
a i, j = −
D h2
(i ≠ j )
In fact, from the notation with two indices φi +1, j − 2φi , j + φi −1, j φi , j +1 − 2φi , j + φi , j −1 + Σ aφ i , j = S i , j − D + 2 2 h h x y reordering with a single index and assuming hx = hy = h as in the figure,
e.g. the equation for the node number 5 becomes φ − 2φ5 + φ4 φ8 − 2φ5 + φ 2 − D 6 + + Σ aφ5 = S 5 h2 h2
NB: We must easily understand shift from multiple index to single index numbering of the unknown variable.
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
25
It can be therefore noted that matrices obtained in 2D and 3D cases are band matrices. Moreover, the following general properties are obtained by most of the discretization procedures: ♦ a i ,i > 0 (i = 1,..., N ) ⇒ the main diagonal is non-zero (1)
♦ ai, j = a j ,i ≤ 0 i ≠ j ⇒ the matrix is symmetric and off-diagonal terms have opposite sign with respect to diagonal ones (i = 1,..., N ) ⇒ the matrix is diagonally dominant ♦ ai ,i > ∑ j ai , j i≠ j
(note that this occurs if Σ a ≠ 0 !!!)
(1) Obviously, a particular sign of the equations is assumed to write the inequalities so we obtained positive diagonal terms. Reversing the sign does not change the meaning of the above conclusions. NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
26
REMINDER ABOUT THE METHODS FOR SOLVING SYSTEMS OF LINEAR ALGEBRAIC EQUATIONS • The Cramer rule represents an inefficient method for solving linear algebraic equation systems, involving too many operations • As known, other methods are therefore adopted, subdivided into two main categories ♦ Direct Methods: they provide the virtually exact solution (only round-off error comes into play) with a finite number of operations ♦ Iterative Methods: they provide the solution as a limit a successive approximations • In neutronic problems, as well as in many other engineering problems, the latter are the most frequently adopted ones: in our case, this is where “internal iterations” get their name from • Let us consider an algebraic system specified in the form: A φ = s
Direct Methods • Gauss or successive elimination method ♦ By this method, equations are linearly combined in order to obtain an equivalent linear system having an upper triangular matrix, whose solution is straightforward a11 a 21 a 31 a 41 a 51
a12
a13
a14
a 22 a 32 a 42 a 52
a 23 a 33 a 43 a 53
a 24 a 34 a 44 a 54
a15 φ1 S1 aˆ11 a 25 φ 2 S 2 0 a 35 ⋅ φ 3 = S 3 → 0 a 45 φ 4 S 4 0 0 a 55 φ 5 S 5
aˆ12 aˆ 22
aˆ13 aˆ 23
aˆ14 aˆ 24
0 0 0
aˆ 33 0 0
aˆ 34 aˆ 44 0
aˆ15 φ1 Sˆ1 aˆ 25 φ 2 Sˆ 2 aˆ 35 ⋅ φ 3 = Sˆ 3 aˆ 45 φ 4 Sˆ 4 aˆ 55 φ 5 Sˆ 5
♦ The solution of the obtained system is immediate, since it is just needed to solve the last (trivial) equation and to back-substitute in the previous ones ♦ The steps to be followed in this respect are:
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
27
1. the first equation is added side by side to each subsequent one, after multiplying it by appropriate factors; in particular, for the r-th equation ( r ≥ 2 ), the factor is − a11 a 21 a 31 a 41 a 51
a12 a 22
a13 a 23
a14 a 24
a 32 a 42 a 52
a 33 a 43 a 53
a 34 a 44 a 54
a15 φ1 S1 a11 0 a 25 φ2 S2 a 35 ⋅ φ 3 = S 3 → 0 a 45 φ 4 S 4 0 0 a 55 φ 5 S 5
a r1 ; it is: a11
a12 a ′22 ′ a 32
a13 ′ a 23 ′ a 33
a14 ′ a 24 ′ a 34
a ′42 ′ a 52
′ a 43 ′ a 53
′ a 44 ′ a 54
a15 φ1 S1 a ′25 φ 2 S 2′ ′ ⋅ φ 3 = S 3′ a 35 a ′45 φ 4 S 4′ ′ φ 5 S 5′ a 55
2. then, the second equation is added side by side to all the subsequent ones ( r ≥ 3 ) after multiplying it by − a11 0 0 0 0
a12 ′ a 22 ′ a 32
a13 a ′23 ′ a 33
a14 a ′24 ′ a 34
′ a 42 ′ a 52
a ′43 ′ a 53
a ′44 ′ a 54
a15 φ1 S1 a11 0 ′ φ 2 S 2′ a 25 ′ ⋅ φ 3 = S 3′ → 0 a 35 ′ φ 4 S 4′ a 45 0 0 ′ φ 5 S 5′ a 55
a12 ′ a 22 0 0 0
a ′r 2 : a ′22
a13 ′ a 23 ′′ a 33
a14 ′ a 24 ′′ a 34
′′ a 43 ′′ a 53
′′ a 44 ′′ a 54
a15 φ1 S1 ′ φ 2 S 2′ a 25 ′′ ⋅ φ 3 = S 3′′ a 35 ′ φ 4 S 4′′ a ′45 ′′ φ 5 S 5′′ a 55
3. a similar procedure is adopted up to the ( N − 1)-th equation, thus obtaining an upper triangular system matrix.
♦ In general, in order to reduce round-off errors, it is possible to change the order of equations in such a way that: a11 ≥ a r1
(r = 2,3,..)
′ ≥ a r′ 2 a 22
(r = 3,4..) etc.
a technique referred to as partial pivoting
♦ As previously mentioned, the Thomas algorithm represents a the particular formulation of the Gauss method for the simpler cases of three-diagonal matrix systems
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
28
• Gauss-Jordan Method ♦ In this case, equations are linearly combined in order to obtain a diagonal matrix system a11 a 21 a31 a41 a51
a12
a13
a14
a22 a32
a23 a33
a24 a34
a42
a43
a44
a52
a53
a54
a15 φ1 S1 aˆ11 0 ˆ a25 φ2 S2 0 a22 a35 ⋅ φ3 = S3 → 0 0 a45 φ4 S4 0 0 0 a55 φ5 S5 0
0
0
0 aˆ33
0
0
0 aˆ44
0
0
0 φ1 Sˆ1 0 φ2 Sˆ2 0 ⋅ φ3 = Sˆ3 0 φ4 Sˆ4 aˆ55 φ5 Sˆ5
♦ The procedure is similar to the one of Gauss method, but each equation is added to “all” the others (both preceding and following ones) after multiplications by appropriate factors ♦ Also in this case “pivoting” techniques can be used • Factorisation Methods ♦ The Gauss method can be considered as a particular case of triangular factorisation methods ♦ In general, by factorisation methods the matrix A is factored in the product of a lower triangular matrix, L , by an upper triangular one, U A = LU
the solution of the given system is then reduced to the solution of two system with triangular matrix (that is immediate) LU φ = s
⇒
Lψ = s
Uφ = ψ
The difficulty of the application of the method is therefore reduced to the step of factorisation, for which different algorithms (e.g., Doolittle, Crout, Cholesky) are available in classical textbooks (2)
(2) See, e.g., A. Quarteroni, R. Sacco, F. Saleri, Numerical Mathematics, Springer, 2007. NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
29
Iterative Methods General Considerations A general iterative method for solving linear algebraic systems has the form φ ( ) = H φ( k
k −1)
( k = 1, 2,...)
+c
where H is termed the iteration matrix and c is an appropriate vector, whose definition characterise the specific numerical scheme. If φ is the exact solution of the system A φ = s , a convergence condition of the numerical scheme will be φ = Hφ+c
By subtracting the two latter relationships it is φ( k ) − φ = H φ( k −1) − φ
( k = 1, 2,...)
k k and, introducing the error vector ε( ) = φ( ) − φ , it is
ε ( k ) = H ε( k −1)
( k = 1, 2,...)
ε ( k ) = H k ε ( 0)
( k = 1, 2,...)
and then Making use of vector and matrix norms (see below), it is ε( k ) ≤ H k ε( 0)
( k = 1, 2,...)
Necessary and sufficient condition for the convergence of the iterative scheme, i.e. in order to be lim ε ( k →∞
k)
=0
for any ε ( 0) is that lim Η k = 0 k →∞
that is sometimes expressed by saying that the iteration matrix must be convergent.
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
30
Definition. A matrix
H ∈ n×n
is said to be convergent if lim Η k = 0 k →∞
Definition. The spectral radius ρ ( H ) of a matrix is the maximum modulus of any of its eigenvalues. In our purposes it is sufficient to remember that: Theorem: A necessary and sufficient condition for a matrix H ∈ n×n to be convergent is that ρ ( H ) < 1 . Therefore, in order to apply with success an iterative method it is necessary and sufficient that the iteration matrix has spectral radius ρ (H) < 1.
• Reminder about norms and spectral radius Definition: A vector norm is an application given two vectors v e w ∈ n and α ∈ , it is
n : n →
such that,
n ( v ) ≥ 0, e n ( v ) = 0 if and only if v = 0 n (α v ) = α n ( v ) n ( v + w) ≤ n ( v) + n (w)
The most common vector norms are: (L∞-norm or also maximum norm) ♦ v ∞ = max vi i
♦ v 1 = ∑ vi
(L1-norm or also absolute norm)
i
♦ v2=v
(L2-norm or also Euclidean norm)
These norms can be formally obtained by putting general formulation
p = ∞, 1, 2
in the
1 p
v
p
p = ∑ vi i
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
31
Given a vector norm, p , an application from
n×n
to
is said to
be a compatible (or induced or subordinate) matrix norm if it satisfies H
p
= sup
Hv
v≠0
v
p
p
It can be shown that the three norms that are compatible with (or are induced from or are subordinate to) the three above considered vector norms are
♦ H
∞
= max ∑ hi ,k i
maximum absolute row sum norm
k
♦ H 1 = max ∑ hi,k k
maximum absolute column sum norm
i
( ) T
♦ H2= ρ H H
spectral norm
These norms are also said natural norms. Some properties of the matrix norms are shortly recalled:
♦ H ≥0
(H
= 0 ⇔ H = 0)
♦ αH = α H ♦ H+J ≤ H + J
H⋅J ≤ H J
Hv ≤ H v
In our purposes it is interesting to remember the following Theorem (by Hirsch): For any of the three natural norms it is ρ (H) ≤ H .
In fact, as a consequence of this theorem, it is: A sufficient condition for a matrix H ∈ n×n to be convergent is that any of its natural norms be less than unity.
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
32
The Jacobi Method To obtain the iterative scheme, the system matrix is decomposed in a diagonal part and an off-diagonal one: =
A a11 a12 a a22 21 a31 a32 an1 an 2
a13
a23 a33
an3
−
D
a1n a11 0 a2 n 0 a22 0 a3n = 0 0 ann 0
0
0 0 − a12 0 −a 0 21 0 − − a31 − a32 ann − an1 − an 2
0 a33
0
E − a13
− a23 0
− an 3
− a1n − a2 n − a3n 0
The linear system is then rewritten as: Dφ = Eφ + s It must be noted that, since in our cases it is always a ii ≠ 0 , it is possible to calculate D−1 , being the diagonal matrix having non-zero elements equal to the reciprocal of the non-zero elements of D . As a consequence, it is: φ = D−1E φ + D−1 s and, putting Η J = D−1E
q = D−1 s
it is φ = HJ φ + q
It can be recognised that it is bii = 0 and 0 a21 − a 22 Η J = a31 − a33 − an1 ann
−
−
−
a1n a11 a2 n − a22 a − 3n a33 0
a12 a11
−
a13 a11
−
0
−
a23 a22
a32 a33
0
an 3 ann
an 2 ann
−
When A is diagonally dominant, it is therefore: n ΗJ ∞ < 1 ⇒ ⇒ ∑ H J ,ij < 1 (i = 1,..., n)
ρ (ΗJ ) < 1
j =1
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
33
It is then straightforward defining the iterative process: φ(
φ( ) = q 1
m +1)
= H J φ( ) + q m
By these definitions, it is: φ( ) = q 1
φ( ) = Η J q + q 2
φ( ) = Η J 2 q + Η J q + q 3
φ(
m +1)
= Η J m q + Η J m −1 q + + Η J q + q
For the above von Neumann series it can be recognised that −1 m +1 lim φ( ) = ( I − Η J ) q = φ where φ = Η J φ + q m →∞
A simple proof of the above can be found in similarity with the treatment of scalar geometrical series considering that
( I − ΗJ ) φ( m+1) = ( I − ΗJ ) Η J m q + Η J m−1 q + + Η J q + q Manipulating the RHS it can be easily found that
( I − ΗJ ) ΗJ m q + Η J m−1 q + + ΗJ q + q = Η J m q + Η J m−1 q + + Η J q + q − Η J m+1 q + Η J m q + + Η J 2 q + qΗ J = q − Η J m+1 q = ( I − Η J m+1 )
and then
( I − Η J ) φ( m+1) = ( I − Η J m+1 ) q
or φ( m+1) = ( I − Η J )
−1
(I − Η ) q m +1
J
Since Η J is a convergent matrix, it follows lim φ(
m →∞
m +1)
−1
= (I − ΗJ ) q
implying that lim φ(
m →∞
m +1)
−1
= ( I − Η J ) q = φ = exact solution
since only for the exact solution it may be: φ = HJ φ + q NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
34
NOTE: The above arguments, together with the previously mentioned results on the convergence of general iterative schemes for the solution of linear algebraic systems, are enough for stating that the Jacobi method converges for diagonally dominant system matrices. This is also enough to state that the Jacobi method converges for the linear systems obtained by the common space discretisations of the diffusion equation. However, since we accepted most of the results without a rigorous proof, in order to better understand the underlying principles, the treatment of a special case will be easy and instructive. Case in which HJ is symmetric Let us consider the case in which HJ is a real and symmetric matrix. This is e.g. the case of the uniform property reactor with a uniform spatial discretisation. This case is particularly simple, since it is recalled that for any real symmetric matrix (lets’ call it A) an orthogonal matrix U always exists (i.e., a matrix whose transpose is equal to the inverse matrix, UT = U −1 ) such that AU = U diag {λ1 , λ2 ,..., λn }
As a consequence, it can be inferred that:
• the columns of U are eigenvectors of A, corresponding to the eigenvalues λ i ; • these eigenvectors represent an orthonormal basis of
n .
Such an orthonormal basis will be identified in the following as ϕ1 , ϕ 2 , ..., ϕ n
With the above assumptions, the following can be shown. Theorem 1. Necessary and sufficient condition for convergence of the Jacobi method is that ρ ( H J ) < 1 Dem.: Defining the error vector as ε ( m ) = φ( m ) − φ
it is (see above) NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
35
ε ( m +1) = H J m ε(1)
If H J is symmetric, its eigenvalues generate all Therefore
n .
n
ε = ∑ ch(1) ϕ h (1)
h =1
It follows that n n n ε ( m +1) = H J m ∑ ch(1) ϕ h = ∑ ch(1) H J m ϕ h = ∑ ch(1) λhm ϕ h h=1 h =1 h =1
(1) ε( m+1) = 0 if and only if For any arbitrary set of c h it is therefore mlim →∞ λ h < 1 for any h, i.e., if and only if ρ ( H J ) < 1 .
NOTE: Speaking in gross intuitive terms, the above means that the requirement ρ ( H J ) < 1 fulfils the request that H J is an operator that contracts any vector to which it is applied; in particular it contracts the error vector. Summarising, it is: The strong (or strict) diagonal dominance of A is a sufficient condition for the convergence of the Jacobi method. It can be also shown that: The weak diagonal dominance and the irreducibility of the matrix A are sufficient conditions for the Jacobi method convergence Before commenting this result, it is recalled that: Def.: A matrix is said weakly diagonally dominant if
∑ j ≠i
least for a single value of r it is
∑ j ≠r
j
j
ai , j ≤ ai ,i , but at
ar , j < ar , r .
Def.: A n×n matrix is said to be reducible if and only if a permutation matrix P exists (i.e., a matrix that can be obtained by the identity matrix NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
36
by permutation of the columns) such that PT AP is block triangular (or block diagonal as a special case). Def.: An irreducible matrix is not reducible. NOTE: Let us comment this results on an intuitive physico-numerical basis. A reducible matrix is such that: ˆ A 11 ˆ A 21 T ˆ P AP = A 31 ˆ A k1
0
ˆ A 32
0 ˆ A
0 ˆ A
22
ˆ A k2
33
ˆ A k3
0 0 0 ˆ A kk
Block triangular
ˆ A 11 0 T P AP = 0 0
0 ˆ A
22
0 0
0
0 ˆ A
0
33
0 0 0 ˆ A kk
Block diagonal
The condition of irreducibility is also expressed in terms of the directed graph of the matrix is strictly connected. The graph is a geometrical figure (that in our purposes can be well represented by the discretisation lattice), where edges or arcs join the point Pr to point Ps if a rs ≠ 0 . A graph is strictly connected if for any couple Pi and Pj of points there exists an oriented 7 8 9 path leading from Pi to Pj. For instance, in the figure it is possible to join P1 to P6 through the arcs P1→P2, P2→P5 and P5→P6 since it is a12 ≠ 0 , a 25 ≠ 0 1 2 3 and a 56 ≠ 0 . Considering the system matrices obtained by the discretization of diffusion problems it can be understood that this is true for any other couple of points, since the discretised leakage terms always join a point to 2, 4 or 6 other points making a strictly connected chain of paths. Therefore, the graph is strictly coupled and the matrix irreducible. 4
5
6
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
37
Unfolding the mathematical jargon for our purposes Let us now unfold the numerical jargon, to draw meaningful conclusions in our specific case: • the operation PT AP has the result of a renumbering of the discretization nodes; • whenever this renumbering would lead to a block-triangular or blockdiagonal (for a symmetric case) system matrix, regions of the reactor would be “decoupled” from each other, because the neutron flux in them would be independently determined; • this situation does not normally occur, except in cases in which infinitely absorbing curtains (or empty spaces with zero flux) purposely separate reactor regions; • therefore, except for these extreme cases, the system matrices involved in the discretization of the diffusion equation are irreducible; • the strict diagonal dominance of the system matrix is assured only in cases in which absorption cross sections are non-zero; e.g., it was shown that for a simple 2D problem with uniform properties it is a i ,i = Σ a +
4D
h2
a i, j = −
D h2
(i ≠ j )
• whenever Σ a vanishes somewhere in the reactor, the strict diagonal dominance of the system matrix is lost; • from a physical standpoint, nodes in which absorption is zero would have an unbounded flux in the presence of a source, unless leakage towards neighbouring nodes will take place; • in this light, the role of irreducibility is to restore the possibility to calculate a limited neutron flux in poorly absorbing regions. It is interesting to consider that the concept of “nuclear reactor coupling” is frequently used to mean that reactor regions behave as a single system, with a strict interplay. When a reactor is said to be “weakly coupled” it is often meant that the “migration length” is relatively small with respect to reactor size. In such a case, though different reactor regions do anyway interact among each other, this interaction may be weak enough to make them behave somehow separately: in fact, the probability of neutrons generated in a region to have influence on the behaviour of other regions may be small, though never zero. NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
38
The above demonstrates how much physical and numerical aspects interact in determining the conditions of convergence for the Jacobi method. The simple idea behind the Jacobi method Beyond the formalism that is necessary to convince about convergence properties, it should be noted that the simple idea behind the Jacobi scheme is to solve a given linear system of equations a1,1 x1 + a1, 2 x2 + … + a1, n xn = b1 a x + a x + … + a x = b 2,1 1 2, 2 2 2, n n 2 … an ,1 x1 + an, 2 x2 + … + an , n xn = bn
by the “strange” solution procedure
( (
) )
x1( m +1) = b1 − a1, 2 x2( m ) … − a1, n xn ( m ) a1,1 ( m +1) ( m) ( m) = b2 − a2,1x1 … − a1, n xn a2 , 2 x2 … xn ( m +1) = bn − an ,1 x1( m ) … − an , n −1 xn −1( m ) an, n
(
)
in which only the i-th unknown is retained at the left hand side and all the others are assigned a tentative value (or a previous iteration one) at the right hand side. The fact that such a rough procedure may lead to convergence for a vast class of system matrices must be regarded as a consequence of the fact that the iteration matrix, having a spectral radius smaller than unity, contacts the error made with the initial guess in further iterations. Practical implementation of the Jacobi method Advantage is taken from the limited number of nodes involved in discretised diffusion equations (5 in 2D and 7 in 3D) and the nodes are numbered with multiple subscripts (one for each coordinate). It is thus possible to define the coefficients of the equation related to N 7
8
9
3,3
1,3
2,3
3,2 3,1
W 4
5
6
1,2
2,2
1
2
3
1,1
2,1
E S
the a given node, with reference to the neighbouring points NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
39
E.g., in two dimensions it is
−
D ( ) (φi, j +1 − 2φi, j + φi, j −1 ) + Σ a φi, j = S i, j φ − 2 φ + φ − i + 1 , j i , j i − 1 , j 2 2
D hx
hy
that becomes
φi , j = Wi , jφi −1, j + Ei , jφi +1, j + Si , jφi , j −1 + Ni , jφi , j +1 + qi , j where D hy2 Si , j = Ni , j = 2D 2D Σa + 2 + 2 hx hy
D hx2 Wi , j = Ei , j = 2D 2D Σa + 2 + 2 hx hy
S i, j
qi, j = Σa +
2D h x2
+
2D h y2
The iterative scheme has therefore the form φi(,mj +1) = Wi , jφi(−m1,) j + Ei , jφi(+m1,) j + Si , jφi(,mj −)1 + Ni , jφi(,mj +)1 + qi , j
In the 3D case other two additional coefficients would appear (e.g., Ui , j e Di , j ). Convergence rate a) If the matrix HJ has a single eigenvalue having maximum modulus, λ 1 , and its eigenvectors form a complete basis, it is: ε
( m +1)
m n (1) (1) λh = H J ε = ∑ ch λ ϕ h = λ c1 ϕ1 + ∑ ch m ϕ h λ1 h =1 h= 2 m
n
(1)
(1)
m h
m 1
therefore ( m+1)
εj
→ C λ1m = C ρ m ( H J )
b) If B is symmetric (uniform reactor with uniform discrettzation) it can be shown that the eigenvalues are patterned in pairs ± λ i and there are two eigenvalues having maximum modulus. Even, in this case it is anyway NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
40
ε(
m +1)
2
(
= ε(
m +1)
, ε(
m +1)
)
n n 1 n 1 1 = ∑ ch( ) λhm ϕ h , ∑ ch( ) λhm ϕ h = ∑ ch( ) λhm h =1 h =1 h =1
therefore (we use the shorthand notation ρ ( H J ) = ρH ε ( m +1)
2
J
2
)
m 2 λ 2 2 h = ρ H2 mJ ∑ ch(1) + ∑ ch(1) → ρ H2 mJ ∑ ch(1) 2m ρHJ λh < ρ H J λh = ρ H J λh = ρHJ
2
or ε(
m +1)
→ C ρ HmJ
Since after a convenient number of iterations the error decrease at each iteration is measured by the spectral radius, the asymptotic rate of convergence can be therefore defined as: R∞ = − ln ρ ( H J )
The number of iterations to be performed for reducing the error by a factor σ < 1 can be estimated assuming that at each iteration the error is reduced by a factor ρ ( B ) ρ m (HJ ) < σ
⇒
m>
m ln ρ ( H J ) < ln σ
⇒
ln σ ln σ =− ln ρ ( H J ) R∞
In the purpose of a rapid convergence, it is therefore desirable that ρ ( H J ) 1 φ(
m +1)
m m +1 m = φ( ) + ω φ( ) − φ( ) GS
where φ( m +1) − φ( m ) = L φ( m +1) + ( U − I ) φ( m ) + q GS
Therefore φ(
m +1)
m m +1 m = φ( ) + ω L φ( ) + ( U − I ) φ( ) + q
or φ(
m +1)
= ω L φ(
m +1)
+ (1 − ω ) I + ω U φ( ) + ωq m
Writing this vector relationship by components, it is (m+1)
φi
i −1
= ω ∑ li , jφ j
(m+1)
+ (1 − ω )φi
(m )
n
+ ω ∑ ui , jφ (j m ) + ωqi
j =1
j =i +1
from which it can be checked that i −1
φi(m +1) = φi(m ) + ω ∑ li , jφ j(m +1) + j =1
n
∑u
i, j
φ j(m ) + qi − φi(m ) = φi(m ) + ω [φi(m +1) − φi(m ) ]GS
j =i +1
Using the two subscripts notation for a 2D case, it is φi(,mj +1) = (1 − ω ) φi(,mj ) + ω Wi , jφi(−m1,+j1) + ω Si , jφi(,mj −+11) + ω Ei , jφi(+m1,) j + ω Ni , jφi(,mj +)1 + ω qi , j
[
= φi(,mj ) + ω φi(,mj +1) − φi(,mj )
]
GS
The numbering of nodes, which is irrelevant in the Jacobi method, is anyway important for the Gauss-Seidel and the SOR ones
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
44
For both the Gauss-Seidel and the SOR method it is possible to adopt a “checkerboard” variant; two steps are performed at each iteration: 1)
fluxes are updated in staggered nodes on the basis of old fluxes only updated flux 19
20
21
22
23
24
13
14
15
16
17
18
7
8
9
10
11
12
1
2
3
4
5
6
old flux
φi(,mj +1) = Wi , jφi(−m1,) j + Ei , jφi(+m1,) j + Si , jφi(,mj −)1 + Ni , jφi(,mj +)1 + qi , j 2)
then, fluxes are updated in the remaining nodes on the basis of updated fluxes only updated flux 19
20
21
22
23
24
13
14
15
16
17
18
7
8
9
10
11
12
1
2
3
4
5
6
old flux
φi(,mj +1) = Wi , jφi(−m1,+j1) + Ei , jφi(+m1,+j1) + Si , jφi(,mj −+11) + Ni , jφi(,mj ++11) + qi , j A greater symmetry is adopted avoiding the generation of spurious asymmetric components in the error
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
45
The iteration matrix in SOR is found considering that φ( m +1) = ( I − ω L ) (1 − ω ) I + ω U φ( m ) + ω ( I − ωL ) q −1
−1
It can be found that the minimum of the spectral radius of the iteration matrix in SOR is found when: ω = ωopt =
2
(where 1 ≤ ω opt ≤ 2 )
1 + 1 − ρ 2 ( HJ )
In order to calculate ωopt it is necessary to know ρ ( H J ) ; therefore, it is advisable to perform a number of iterations with the Jacobi method to estimate its asymptotical convergence rate. For a symmetric H J , it is n
q = ∑ qh ϕ h
ϕh ⋅ ϕ j = δ h, j
with
h =1
then n 2 n n 2 H J mq = ( H J mq, H J mq ) = ∑ qh λhm ϕ h , ∑ qh λhm ϕ h = ∑ qh λh2 m h =1 h =1 h=1 n
=
∑q
2 h
ρ
λh = ρ B
n
+
2m HJ
∑q
2 h
λ
2m h
=ρ
λh < ρ B
2m HJ
n 2 ∑ qh + λh = ρ B
n
∑
qh
λh < ρ B
2
λ2hm ρ H2 m J
As a consequence:
lim
m →∞
HJ
m +1
q
H J mq
2 2
2m+2 n n 2 2 λh ∑ qh + ∑ qh ρ 2 m + 2 λh < ρ B HJ 2 λh = ρ B = lim ρ B n 2m n m →∞ ∑ q 2 + ∑ q 2 λh h λh = ρ B h ρ H2 mJ λh < ρ B
2 = ρHJ
It is:
(ρ SOR )ω =ω
opt
=
2 1 + 1 − ρ H2 J
−1
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
46
Line OverRelaxation Method (LOR)
♦ Since the three-diagonal matrices are efficiently dealt with by the Thomas’ algorithm, it is tried to make use of the three-point structure existing in each direction in 2D and 3D problems 7 8 9 ♦ For instance, in the case of the 9 node grid in the figure, the obtained banded matrix can be interpreted as a block matrix whose diagonal blocks are tridiagonal a11 a12 a a 21 22 0 a32 a41 0 0 a52 0 0 0 0 0 0 0 0
0 a23
a14 0
0 a25
0 0
0 0
0 0
a33 0 0
0 a44 a54
0 a45 a55
a36 0 a56
0 a47 0
0 0 a58
a63 0 0
0 a74 0
a65 0 a85
a66 0 0
0 a77 a87
0 a78 a88
0
0
0
a96
0
a98
0 φ1 S1 0 φ 2 S 2 0 φ3 S 3 0 φ 4 S 4 0 ⋅ φ5 = S 5 a69 φ6 S 6 0 φ 7 S 7 a89 φ8 S8 a99 φ9 S9
ˆ A 11 ˆ → A 21 0
ˆ A 12 ˆ A
22
ˆ A 32
4
5
6
1
2
3
0 φˆ1 sˆ1 ˆ ⋅ φˆ = sˆ A 23 2 2 ˆ sˆ ˆ A 33 φ3 3
It should be noted that blocks Aˆ i ,i contain the coefficients of fluxes on a same line and, therefore, they are three-diagonal as in 1D cases On the other hand, blocks Aˆ i ,i−1 e Aˆ i ,i+1 contain the coefficients of the fluxes in the lines below and above the one in consideration The idea is then pursued to write the system in the form ˆ φˆ + A i ,i −1 i −1
ˆ A i ,i
ˆ φˆ = sˆ φˆi + A i ,i +1 i +1 i
threediagonal
applying a Gauss-Seidel-like strategy line to be calculated line already calculated line not yet calculated ( m +1) ( m +1) ˆ ˆ ˆ ˆ ˆ ˆ Ai ,i φi = sˆ i − Ai ,i −1φi −1 − Ai ,i +1φi(+m1) tridiagonal
Introducing also an overrelaxation, the LOR scheme is obtained m +1 ˆ ˆ φˆ ( m +1) + (1 − ω ) A ˆ φˆ ( m ) − ω A ˆ φˆ ( m) A φˆi( ) = ωsˆ i − ω A i ,i i ,i −1 i −1 i ,i i i ,i +1 i +1
tri diagonal
It is therefore possible to apply the direct Thomas algorithm on each line covering “line-by-line” the whole domain
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
47
updated flux 19
20
21
22
23
24
13
14
15
16
17
18
7
8
9
10
11
12
1
2
3
4
5
6
old flux
♦ With reference to the more relevant example in the above figure, adopting the two subscript notation, it is: Wi , jφi(−1, j ) + Oi , jφi(, j ) + Ei , jφi(+1, j ) = −ω Si , jφi(, j −1 ) − ω Ni , jφi(, j +)1 + ω qi , j m +1
m +1
m +1
m +1
m
three point structure
{
+ (1 - ω ) Wi , jφi(−m1,) j + Oi , jφi(,mj ) + Ei , jφi(+m1,) j
}
where the right hand side exhibit a clear three-point structure
♦ Though the technique can be applied preferentially in a single direction, it is more convenient to alternate the directions, pointing out the three-point structure alternatively in each one of them Si , jφi(, j −1 ) + Oi , jφi(, j ) + Ni , jφi(, j +1 ) = −ω Wi , jφi(−1, j ) − ω Ei , jφi(+1,) j + ω qi , j m +1
m +1
m +1
m +1
m
three point structure
{
+ (1 - ω ) Si , jφi(−m1,) j + Oi , jφi(,mj ) + Ni , jφi(+m1,) j
}
♦ The following ADI method makes use of a similar technique
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
48
Alternating Direction Implicit Method (ADI)
♦ The method is due to Peaceman and Rachford (1955) and is based on the decomposition of matrix A in the form A=
H
+
+
V
Σ
" horizontal " coefficients
" vertical " coefficients
coefficients
of the discretized Laplacian
of the discretized Laplacian
related to absorption
For a uniform property case, it is D 2D D D 2D D − 2 φi −1, j + 2 φi , j − 2 φi +1, j − 2 φi , j −1 + 2 φi , j − 2 φi , j +1 + Σ a ,ijφi , j = Si , j hx hx hx hy hy hy
↓ Hφ
↓ Vφ
♦ For purpose of illustration, it can be argued that the above splitting of matrix a is used to firstly obtain equivalent forms of the linear system a first form is the following: Aφ = s ⇒
[ H + V + Σ] φ = s
1 1 H + Σφ = s −V + Σφ 2 2
or
⇒
⇒
1 1 H + 2 Σ + V + 2 Σ φ = s
1 1 H + 2 Σ + ωm I φ = s − V + 2 Σ − ωm I φ
1 1 H + 2 Σ + ωm I φ = ωm I − V − 2 Σ φ + s
a second form is trivially obtained by exchanging the roles of H and V in the former one: 1 1 V + 2 Σ + ωm I φ = ωm I − H − 2 Σ φ + s
♦ In the above, ωm is a forcing parameter that could be thought of as the introduction of some degree of inertia in the iterations; this aspect will appear more clearly by treating the transient version of the algorithm; here this will become a key aspect for achieving fast convergence in some cases ♦ Taking profit of the three-point structure in the two directions, semi-iterations are performed adopting both forms:
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
49
1 1 ( m) ( m +1 2) H + Σ + ω I = ω I − V − Σ φ + s φ m m 2 2 three- point structure
1 1 ( m +1 2) ( m +1) V + Σ + ω I = ω I − H − Σ φ +s φ m m 2 2 three- point structure
♦ Since the matrices at the left hand side have a three-point structure (they are not both three-diagonal anyway), profit can be taken of the Thomas’ algorithm to evaluate “implicitly” each row or column by “alternating” the sweeping directions: hence the name of the method ♦ It is interesting to rewrite the two semi-iterations in single step form from the first semi-iteration it is φ
( m +1 2 )
−1
−1
1 1 1 = H + Σ + ωm I ωm I − V − Σ φ( m ) + H + Σ + ωm I s 2 2 2
m+1 substituting into the second and solving for φ( ) it is
φ
( m +1)
−1
−1
1 1 1 1 = V + Σ + ωm I ωm I − H − Σ H + Σ + ωm I ωm I − V − Σ φ( m ) 2 2 2 2 + source terms
♦ The resulting iteration matrix is, therefore −1
J ωm
−1
1 1 1 1 = V + Σ + ωm I ωm I − H − Σ H + Σ + ωm I ωm I − V − Σ 2 2 2 2
It can be shown that ρ ( Jω
m
) 0
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
50
♦ The demonstration of this result is based on the following steps: 1 2
1 2
it is shown that H + Σ and V + Σ are symmetric and positive definite (SPD), i.e., such that φT Aφ > 0, ∀φ ≠ 0 ; as such, all their eigenvalues are real and positive; (3) it is then shown that if ωm > 0 , any eigenvalue of Jω , has a magnitude less than unity. m
♦ Let us skip the demonstration and focus to understand why the ADI method is particularly successful for separable and nearly separable problems: it is first recalled that for separable problems defined by the partial differential equation ∂ 2φ ∂ 2φ + 2 + Bl2,kφ = 0 2 ∂x ∂y
the general solution has the form φ = ϕl ,k ( x, y ) ∼ X l ( x ) Yk ( y ) the eigenfunctions
ϕl , k ( x, y )
are therefore such that
1 d 2 X l 1 d 2Yk + + Bl2,k = 0 2 2 X l dx Yk dy
and thanks to the separability it is 1 d 2 Xl + Bl2 = 0 X l dx 2
1 d 2Yk + Bk2 = 0 Yk dy 2
a direct consequence of the above is that simultaneously the equations 2 1 ∂ ϕl , k + Bl2 = 0 ϕl ,k ∂x 2
ϕl , k ( x, y )
satisfy
2 1 ∂ ϕl , k + Bk2 = 0 ϕl ,k ∂y 2
i.e., ϕl ,k ( x, y ) is at the same time an eigenfunction of both the above eigenvalue equations, but with different eigenvalues;
(3) From the relationship φ
T
Aφ > 0, ∀φ ≠ 0 , assuming ϕ as an eigenvalue of A , it follows ϕT Aϕ > 0
⇒ ϕ λϕ > 0 ⇒ λϕ ϕ > 0 ⇒ λ > 0 . Moreover, it is recalled that the eigenvalues of a symmetric matrix are real as it is for any hermitian matrix. T
T
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
51
these properties are retained also by the discretised form of the equations, so that the matrices H and V have the same 1 2
1 2
eigenvectors, as it is of H + Σ and V + Σ , though they have different eigenvalues 1 ( k ,l ) ( k ,l ) H + Σ ϕ = λk ϕ 2
( k = 1,..., nx )
1 ( k ,l ) ( k ,l ) V + Σ ϕ = µl ϕ 2
( l = 1,..., n ) y
♦ Considering the well known properties of functions of matrices, it can be understood that the spectral radius of matrix Jω is given by (4) (5) m
ρ ( Jω
m
) = max ((ω k ,l
ωm − λk )(ωm − µl ) 0
where λ k > 0 e µ l > 0 as they are eigenvector of the symmetric 1 2
1 2
positive defined real matrices H + Σ and V + Σ . ♦ In fact, the matrices ωm I ± H + Σ and ωm I ± V + Σ have still 2 2
1
( k ,l )
ϕ
1
as eigenvectors and the eigenvalues are ω m ± λ k and
ωm ± µl ♦ As a consequence, also J ω has ϕ( ) as eigenvectors with the mentioned spectral radius. ♦ In theory, for separable cases it is possible to define a very efficient iterative procedure: k ,l
m
(4) It is recalled that, given a function f ( z ) of a complex variable z having a MacLaurin series expansion ∞
f ( z ) = ∑ an z n that converges for n=0
∞
z < R , the series f ( A ) = ∑ an A n converges if the eigenvalues of the matrix A n=0
have modulus less than R. The function is then called “well defined” (see e.g., R. Bronson, Matrix Operations, McGraw-Hill, 1989). As a consequence of the above, it can be easily verified that the matrix f ( A ) has the same eigenvectors of A and its eigenvalues are f ( λi ) where λi are the eigenvalues of A . (5) Moreover, if two matrices A and B have the same eigenvectors with eigenvalues λi and µi respectively, also AB has the same eigenvectors ϕ i with eigenvalues λi µi . In fact, it is easily seen that ABϕ i = Aµi ϕ i = λi µi ϕ i . NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
52
if p is the number of different eigenvalues λ k e µ l , p different values of ωm , can be used in order to minimize the spectral radius of the overall iteration matrix
J = J J ...J J ω p ω p −1 ω 2 ω1 ωm tot that has as a spectral radius the quantity
ρ (J ω ) = max m tot k ,l
(ω1 − λ k )(ω1 − µ l ) (ω p − λ k )(ω p − µ l ) ... (ω1 + λ k )(ω1 + µ l ) (ω p + λ k )(ω p + µ l )
putting at each one of the p steps ω m equal to a given λ k or
µ l , it would be obviously ρ(J ωm ) = 0 tot
since determining the eigenvalues λ k e µ l would require a considerable amount of computational work, it is instead accepted to choose the ω m ’s with the criterion
min
ω1 , ω2 , ...,ω p
(ω1 − x )(ω2 − x )...(ω p − x ) x∈(α,β ) (ω1 + x )(ω2 + x )...(ω p + x ) max
where (α, β) contains the eigenvalues.
♦ For non-separable problems, there is no rigorous theory for selecting the optimum forcing parameters; therefore: ADI has optimum performances for separable cases (6) but it might be not very efficient on some non-separable cases for nearly-separable cases (most of the addressed conditions) it is possible to evaluate the optimum values of ω m referring to the separable case that can be considered closest to the addressed one, obtaining generally good performances; formulations defining the ω m ’s are reported in relevant textbooks on this basis.
(6) It is remarked that in this context, by “separable” and “non-separable” problems it is simply meant that the reference partial differential equation problems are amenable or not to the solution by the technique of variable separation. NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
53
Gradient Methods
♦ Considering the system Ax=s
with A real and symmetric, we will identify with φ its exact solution.
♦ It is interesting to consider different quantitative definitions of the error: T T E1 ( x ) = ( φ − x ) ⋅ ( φ − x ) = ε ⋅ ε T E2 ( x ) = ( s − A x ) ⋅ ( s − A x ) = r ⋅ r T
T E3 ( x ) = ( φ − x ) ( s − A x ) = ε Aε T
While the first and second definitions provide positive quantities (being scalar products of a vector by itself), the third one represents a quadratic form, having A as associated matrix, and can be greater or less than zero.
♦ Recalling that, if A is symmetric and positive definite (SPD) the quadratic form is by definition positive (in fact, it is εT Aε > 0 for any non-zero ε ), we will assume that this is the case, taking E3 ( x ) as a measure of the error. ♦ It is therefore E3 ( x ) ≥ 0
and
E3 ( x ) = 0 ⇒ x = φ
meaning that to find the system solution we should minimise E3 ( x )
♦ In this purpose, geometrical arguments can be set up by considering that the locus of points at constant error
x
φ
φ
E3 ( x ) = c o s t.
represents a hyper-ellipsoid in n .
♦ In fact, the canonical form of the quadratic polynomial can be found to represent an hyper-ellipsoid by making use of the transform NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
54
−ε = x − φ = U z
where U is the orthogonal matrix ( UT U = I o UT = U −1 ) such that UT AU = diag {λ1 , λ2 ,..., λn } = Λ
(it always exists for symmetric matrices). The direction of the axes in the new reference system obtained by the transform are the eigenvectors of A, being an othonormal basis of n . It is: E3 ( z ) = εT A ε = ( U z ) A U z = zT UT AU z = zT Λ z T
or E3 (z ) = λ1 z12 + λ 2 z 22 + ... + λ n z n2 = c o s t.
Since the eigenvalues of any SPD matrix are positive the above clearly represents a hyper-ellipsoid having axes a i ∝ c o s t. λ i , which are measures the error. The basic idea in the gradient methods is therefore to search for better approximations of the solutions moving to hyper-ellipsoid with smaller and smaller axes. Method of the “steepest descent” The above idea is translated into mathematical formalism by the relationship m +1 m m m x( ) = x( ) + α ( )p ( ) where x( ) represents the old approximation (on the old hyper( m) ellipsoid), p represents a vector identifying the direction of ( m) movement and α ( m) measures the distance along p necessary to m
m+1 reach x( )
It must be understood that ( m) o the choice of p must be effective enough in order to move inwards the hyper-ellipsoid surface towards its centre
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
55
( m)
o computing p as a diametral direction would be too costly from the computational point of view: simpler recipes must be therefore identified. In the “steepest descent” method, the adopted recipe is to move inwards from the hyper-ellipsoid along the vector normal to the tangent plane, i.e. along the “gradient” of E3 (hence, the name) Therefore, in this case it is: p( ) = vector normal to the surface in x( m) m
α (m ) = scalar chosen as to minimise
(
)
E3 x( m+1)
In the canonical reference frame, it is conveniently assumed p′( m) = −Λ z ( m) since the gradient is immediately found to be ∂E3 = 2λi zi ∂zi
( )
grad z E3 z ( m ) ∝ Λ z ( m)
⇒
However, since it is necessary to operate in the original reference frame it is needed to back-transform this result p(
m)
= U p′(
m)
= UΛ z (
(
)
m)
= UUT AU z (
m)
= − A ε(
m)
= A x( m) − φ = A x( m) − s = −r ( m)
(m )
♦ α is then chosen on the basis of the mentioned minimum criterion, thus obtaining m m ∂E3 r ( )T r ( ) ( m) α = − ( m)T ( m) =0 ⇒ ( m) r Ar ∂α ♦ The final iterative scheme is therefore x
( m +1)
=x
( m)
r ( )T r ( ) ( m ) + ( m )T ( m ) r r Ar m
m
x( 0 ) = x0
♦ It should be noted that: if z i(m ) = 0 ⇒ p i′ (m ) = 0 , i.e., whenever a component of the solution is exact, this advantage is preserved since the subsequent approximations will keep on the zi = 0 plane; NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
56
if the hyper-ellispoid is a solid of revolution along some axis (i.e., it is round, λ i = λ j , i ≠ j ), convergence is quicker; as a limiting case, if the hyper-ellipsoid is an hyper-sphere, convergence is achieved in a single iteration.
♦ An example related to n=2 is shown in the figure below:
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
57
Coniugate Gradient Method
♦ It should be considered a direct method since, in principle, it allows to obtain the solution in n iterations (n is the dimension of the system matrix) ♦ Actually, as for any other direct method, round-off errors may accumulate, reducing the method to a semi-iterative one. ♦ The adopted formulations are, in this case x(
m +1)
= x( ) + α ( ) p (
p(
m +1)
= r(
m +1)
= r ( ) − α ( ) A p(
r(
m
m
m +1)
α
+ β ( )p(
m
m
m
(m ) ( 0) ( 0) and β where p = r and α ( m)
m)
p ( m )T r ( m ) = ( m )T p A p( m)
(m )
m) m)
are given by β
( m)
r ( m +1)T A p ( m ) = − ( m )T p A p( m)
♦ As it can be noted, these definitions, whose demonstration we omit, though different from the ones of the steepest descent method, are anyway simple; their meaning is the following: vectors p( ) constitute a set of A-orthogonal vectors (or Aconjugate, hence the name of the method): m
p ( m +1)T A p ( k ) = 0
( k = 0,1,..., m )
with such definitions, the residuals r ( m) , instead, turn out to be orthogonal among each other: r ( m +1)T r ( k ) = 0
( k = 0,1,..., m )
Since the n -th residual should be orthogonal to the others, r ( 0) , r (1) , …, r ( n −1) , it should be zero. r(n) = 0 ⇒
x( n ) = φ = exact solution
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
58
In the case n=2 the relation of A-conjugation between two directions defines the relation existing between the direction of two parallel tangent lines to an ellipsis and its diameter passing through the points of tangency: this illustrates the effectiveness of the method in finding the exact solution (i.e., the centre).
Modern versions of the conjugate gradient method can be applied to matrices the do not require to be SPD; the bi-conjugate gradient method is one such algorithm.
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
59
Sources and Suggested Readings In English language R. S. Varga, Matrix Iterative Analysis, Prentice Hall, 1962. A. Quarteroni, R. Sacco, F. Saleri, Numerical Mathematics, Springer, 2007. R. Bronson, Matrix Operations, McGraw-Hill, 1989. In Italian language G. Ghelardoni, P. Marzulli, Argomenti di Analisi Numerica, ETS Università, Pisa, 1980. G. Gambolati, Elementi di Calcolo Numerico, Edizioni Libreria Cortina, Padova. Web references The Matrix Reference Manual, Mike Brookes, Imperial College, London, UK http://www.ee.ic.ac.uk/hp/staff/dmb/matrix/intro.html#Intro
NMNR-Unit-1 – Eigenvalue Problems and Solution Strategies – Inner Iterations
60