Search for robust algebraic preconditioners Miroslav Tma Institute - - PowerPoint PPT Presentation

search for robust algebraic preconditioners
SMART_READER_LITE
LIVE PREVIEW

Search for robust algebraic preconditioners Miroslav Tma Institute - - PowerPoint PPT Presentation

Search for robust algebraic preconditioners Miroslav Tma Institute of Computer Science Academy of Sciences of the Czech Republic tuma@cs.cas.cz based on joint work with Michele Benzi, Rafael Bru, Jurjen Duintjer Tebbens, Jos Marn, Jos


slide-1
SLIDE 1

Search for robust algebraic preconditioners

Miroslav Tůma

Institute of Computer Science Academy of Sciences of the Czech Republic tuma@cs.cas.cz

based on joint work with Michele Benzi, Rafael Bru, Jurjen Duintjer Tebbens, José Marín, José Mas, Miroslav Rozložník, Jennifer Scott et al.

Emory University November 2, 2009, Atlanta

1 / 44

slide-2
SLIDE 2

Motivation: I.

Solving large, sparse systems of linear algebraic equations

Ax = b

2 / 44

slide-3
SLIDE 3

Motivation: I.

Solving large, sparse systems of linear algebraic equations

Ax = b

Contemporary decompositional interpretation of the Gaussian elimination (GE): Householder at the end of the latest 50’s.

2 / 44

slide-4
SLIDE 4

Motivation: I.

Solving large, sparse systems of linear algebraic equations

Ax = b

Contemporary decompositional interpretation of the Gaussian elimination (GE): Householder at the end of the latest 50’s. Both different and similar role of GE in the two basic solving approaches: Direct methods and iterative methods Case of our interest: Relaxed GE (incomplete decompositions of various kinds).

2 / 44

slide-5
SLIDE 5

Motivation: II.

Incomplete decompositions and their implementation. GE: We need sparsity (in the input matrix, elimination graphs’ estimates, intermediate data) and the speed of the whole computation.

3 / 44

slide-6
SLIDE 6

Motivation: II.

Incomplete decompositions and their implementation. GE: We need sparsity (in the input matrix, elimination graphs’ estimates, intermediate data) and the speed of the whole computation. The sparsity does not seem to be particularly critical when considering plain incomplete decompositions (ID). But, fast implementations of contemporary ID may cause problems.

3 / 44

slide-7
SLIDE 7

Motivation: II.

Incomplete decompositions and their implementation. GE: We need sparsity (in the input matrix, elimination graphs’ estimates, intermediate data) and the speed of the whole computation. The sparsity does not seem to be particularly critical when considering plain incomplete decompositions (ID). But, fast implementations of contemporary ID may cause problems. Fortunately, some data structures originally developed for direct methods (and not used there anymore) can be successfully used. Fast implementations of sophisticated GE modifications are possible

3 / 44

slide-8
SLIDE 8

Motivation: III.

Incomplete decompositions and robustness Robustness of ID jointly with the iterative method is what really critical is.

4 / 44

slide-9
SLIDE 9

Motivation: III.

Incomplete decompositions and robustness Robustness of ID jointly with the iterative method is what really critical is. Partial robustness: in its evaluation (breakdown-free property).

◮ May be based on relaxing accuracy of decomposition (decomposing a

different matrix)

◮ Or, may promote density of the decomposition (restricting the

incompleteness (numerically or structurally))

Stability of ID: important in combination with iterative methods.

4 / 44

slide-10
SLIDE 10

Motivation: III.

Incomplete decompositions and robustness Robustness of ID jointly with the iterative method is what really critical is. Partial robustness: in its evaluation (breakdown-free property).

◮ May be based on relaxing accuracy of decomposition (decomposing a

different matrix)

◮ Or, may promote density of the decomposition (restricting the

incompleteness (numerically or structurally))

Stability of ID: important in combination with iterative methods. Is is to possible to guarantee more robustness for decompositions by relating them to GE?

4 / 44

slide-11
SLIDE 11

Motivation: III.

Incomplete decompositions and robustness Robustness of ID jointly with the iterative method is what really critical is. Partial robustness: in its evaluation (breakdown-free property).

◮ May be based on relaxing accuracy of decomposition (decomposing a

different matrix)

◮ Or, may promote density of the decomposition (restricting the

incompleteness (numerically or structurally))

Stability of ID: important in combination with iterative methods. Is is to possible to guarantee more robustness for decompositions by relating them to GE? In the other words, how far are we from GE-aware decompositions?

4 / 44

slide-12
SLIDE 12

Motivation: IV.

ID affects the iterative method via its inverse.

500 1000 1500 2000 500 1000 1500 2000 nz = 13151 500 1000 1500 2000 500 1000 1500 2000 nz = 126945

matrix ADD20 rather precise inverse (2 its BiCGStab)

5 / 44

slide-13
SLIDE 13

Motivation: IV.

500 1000 1500 2000 500 1000 1500 2000 nz = 13151 500 1000 1500 2000 500 1000 1500 2000 nz = 61505

matrix ADD20 less precise inverse

6 / 44

slide-14
SLIDE 14

Motivation: IV.

500 1000 1500 2000 500 1000 1500 2000 nz = 13151 500 1000 1500 2000 500 1000 1500 2000 nz = 9752

matrix ADD20 even less precise inverse

7 / 44

slide-15
SLIDE 15

Motivation: IV.

500 1000 1500 2000 500 1000 1500 2000 nz = 13151 500 1000 1500 2000 500 1000 1500 2000 nz = 7525

matrix ADD20 rough inverse

8 / 44

slide-16
SLIDE 16

Motivation: IV.

500 1000 1500 2000 500 1000 1500 2000 nz = 13151 500 1000 1500 2000 500 1000 1500 2000 nz = 3499

matrix ADD20 very rough inverse

9 / 44

slide-17
SLIDE 17

Motivation: IV.

500 1000 1500 2000 500 1000 1500 2000 nz = 13151 500 1000 1500 2000 500 1000 1500 2000 nz = 3260

matrix ADD20 ILU decomposition (similar size as the "very rough inverse")

10 / 44

slide-18
SLIDE 18

Motivation: IV.

500 1000 1500 2000 500 1000 1500 2000 nz = 13151 500 1000 1500 2000 500 1000 1500 2000 nz = 3260

matrix ADD20 inverted ILU decomposition

11 / 44

slide-19
SLIDE 19

Motivation: V.

Concluded motivation Consulting / employing matrix inverse may provide useful information

12 / 44

slide-20
SLIDE 20

Motivation: V.

Concluded motivation Consulting / employing matrix inverse may provide useful information Two extreme cases of incomplete decompositions:

◮ approximate inverse decompositions (direct ID) ◮ direct incomplete decompositions (inverse ID) 12 / 44

slide-21
SLIDE 21

Motivation: V.

Concluded motivation Consulting / employing matrix inverse may provide useful information Two extreme cases of incomplete decompositions:

◮ approximate inverse decompositions (direct ID) ◮ direct incomplete decompositions (inverse ID)

Our tools: joint treatment of both direct and inverse decompositions.

12 / 44

slide-22
SLIDE 22

Motivation: V.

Concluded motivation Consulting / employing matrix inverse may provide useful information Two extreme cases of incomplete decompositions:

◮ approximate inverse decompositions (direct ID) ◮ direct incomplete decompositions (inverse ID)

Our tools: joint treatment of both direct and inverse decompositions. Is this a way to GE-aware decompositions?

12 / 44

slide-23
SLIDE 23

Motivation: V.

Concluded motivation Consulting / employing matrix inverse may provide useful information Two extreme cases of incomplete decompositions:

◮ approximate inverse decompositions (direct ID) ◮ direct incomplete decompositions (inverse ID)

Our tools: joint treatment of both direct and inverse decompositions. Is this a way to GE-aware decompositions? What we do not discuss here? Modifications of the basic algorithm (basic diagonal modifications, general diagonal compensations with respect to some matvecs etc.) a priori diagonal changes matrix pre/post processings embedding into a more general (e.g. multilevel) scheme. Analysis of the described schemes

12 / 44

slide-24
SLIDE 24

Summarizing our starting points and goals

Starting points Approximate inverse decompositions (Kolotilina, Yeremin, 1993; Benzi, Meyer, T., 1996; Benzi, T., 1998 etc.)

13 / 44

slide-25
SLIDE 25

Summarizing our starting points and goals

Starting points Approximate inverse decompositions (Kolotilina, Yeremin, 1993; Benzi, Meyer, T., 1996; Benzi, T., 1998 etc.) Successful use of parts of factorized matrix inverse used in inverse-based incomplete decompositions (Bollhöfer, Saad, 2002; Bollhöfer, 2003)

13 / 44

slide-26
SLIDE 26

Summarizing our starting points and goals

Starting points Approximate inverse decompositions (Kolotilina, Yeremin, 1993; Benzi, Meyer, T., 1996; Benzi, T., 1998 etc.) Successful use of parts of factorized matrix inverse used in inverse-based incomplete decompositions (Bollhöfer, Saad, 2002; Bollhöfer, 2003) A particular goal: Combined use of direct and inverse incomplete decompositions

13 / 44

slide-27
SLIDE 27

Summarizing our starting points and goals

Starting points Approximate inverse decompositions (Kolotilina, Yeremin, 1993; Benzi, Meyer, T., 1996; Benzi, T., 1998 etc.) Successful use of parts of factorized matrix inverse used in inverse-based incomplete decompositions (Bollhöfer, Saad, 2002; Bollhöfer, 2003) A particular goal: Combined use of direct and inverse incomplete decompositions One of the tools: generalized biconjugation formula

13 / 44

slide-28
SLIDE 28

Summarizing our starting points and goals

Starting points Approximate inverse decompositions (Kolotilina, Yeremin, 1993; Benzi, Meyer, T., 1996; Benzi, T., 1998 etc.) Successful use of parts of factorized matrix inverse used in inverse-based incomplete decompositions (Bollhöfer, Saad, 2002; Bollhöfer, 2003) A particular goal: Combined use of direct and inverse incomplete decompositions One of the tools: generalized biconjugation formula Here we try to get inside GE, not to study/defend a synthetic approach.

13 / 44

slide-29
SLIDE 29

Outline

1

Limits of standard algebraic approaches

2 Standard biconjugation and matrix inverses 3 Direct-inverse decompositions 4 A flavor of applications different from preconditioning 5 Conclusions

14 / 44

slide-30
SLIDE 30

Outline

1

Limits of standard algebraic approaches

2 Standard biconjugation and matrix inverses 3 Direct-inverse decompositions 4 A flavor of applications different from preconditioning 5 Conclusions

15 / 44

slide-31
SLIDE 31

Limits of ID: BCSSTK38, n = 8032, nz = 181, 746

ID: Limitations in predictability and efficiency

10 10

1

10

2

10

3

10

4

10

8

10

10

10

12

10

14

10

16

10

18

10

20

10

22

10

24

10

26

norm of the residual matrix / iterations bandwidth residual matrix norm iterations

Generally no clear dependence on the error size, pattern etc. This is a very common kind of behavior

16 / 44

slide-32
SLIDE 32

Overcoming some limits: individual level preassignments

Experiments: Kohn-Sham equation, n=250500

Effects individual level preassignments

20 40 60 80 100 120 140 160 1 2 3 4 5 6 7 8 x 10

6

preconditioner size number of iterations standard setting new setting 1 new setting 2 new setting 3

17 / 44

slide-33
SLIDE 33

Overcoming some limits: individual level preassignments

Experiments: Kohn-Sham equation, n=250500

Effects individual level preassignments

20 40 60 80 100 120 140 160 1 2 3 4 5 6 7 8 x 10

6

preconditioner size number of iterations standard setting new setting 1 new setting 2 new setting 3

see the talk of Jennifer Scott at SIAM ALA 2009, Monterey

17 / 44

slide-34
SLIDE 34

Outline

1

Limits of standard algebraic approaches

2 Standard biconjugation and matrix inverses 3 Direct-inverse decompositions 4 A flavor of applications different from preconditioning 5 Conclusions

18 / 44

slide-35
SLIDE 35

Generalized Gram-Schmidt (GGS)

Generalized Gram-Schmidt: basics of SPD case Orthogonalize columns of I using the inner product , A We get (instead of A = QDR with R unit upper triangular):

I = ZU

◮ U is unit upper triangular, as usual (U = LT for A = LLT ). ◮ Z is orthogonal in , A

ZT AZ = D (Biconjugate decomposition)

◮ But: Z is unit upper triangular as well (Z = L−T for A = LLT )

Easy to reveal decomposed matrix inverse:

A−1 = ZD−1ZT,

19 / 44

slide-36
SLIDE 36

Generalized Gram-Schmidt: II.

Resulting direct and inverse ID may be practical in the incomplete case

I = ZDU

A ≈ LLT , U ≈ LT , Z ≈ L−1 Origins: more papers in 40’s and early 50’s (Escalator method by Morris (1946), Vector method by Purcell (1952), Fox, Huskey, Wilkinson (1948)). The sparse incomplete method can be implemented: AINV (Benzi, Meyer, T., 1996; Benzi, T., 1998) Computational procedures to compute sparse incomplete U in this way: RIF (Benzi, T., 2003) As we will see, both Z and U can be computed breakdown-free, but this is not all that we may want.

20 / 44

slide-37
SLIDE 37

Generalized Gram-Schmidt: III.

Generalized Gram-Schmidt: the (SPD) algorithm I = ZU ≡ [z1, . . . , zn] (uij)i,j for i=1, n for j=1, i-1 with nonzero uij = eT

j Azi(j)

zi(j) = zi(j−1) − zj(j−1) eT

j Azi(j−1)

eT

j Azj(j−1)

end j end i Forcing partial robustness: different formulas which are the same in exact arithmetic: the breakdown-free variant SAINV But: in order to get U we must get Z: direct factor is obtained via the inverse factor

21 / 44

slide-38
SLIDE 38

Generalized Gram-Schmidt: IV.

Generalized Gram-Schmidt I = ZU: the data dependence graphically

Z L done not used used done U

T =

22 / 44

slide-39
SLIDE 39

Generalized Gram-Schmidt: IV.

Generalized Gram-Schmidt I = ZU: the data dependence graphically

Z L done not used used done U

T =

One way transfer of information

22 / 44

slide-40
SLIDE 40

Summarization the two general problems

Two resulting general problems

23 / 44

slide-41
SLIDE 41

Summarization the two general problems

Two resulting general problems

1

Is there a practical scheme of decomposition that would have an arbitrary transfer of information between direct and inverse factors?

23 / 44

slide-42
SLIDE 42

Summarization the two general problems

Two resulting general problems

1

Is there a practical scheme of decomposition that would have an arbitrary transfer of information between direct and inverse factors?

2

Very good behavior of SAINV observed also in the incomplete case with respect to the non-stabilized algorithm cannot be explained just by breakdown-free property

23 / 44

slide-43
SLIDE 43

Summarization the two general problems

Two resulting general problems

1

Is there a practical scheme of decomposition that would have an arbitrary transfer of information between direct and inverse factors?

2

Very good behavior of SAINV observed also in the incomplete case with respect to the non-stabilized algorithm cannot be explained just by breakdown-free property What is behind the clearly superior performance of the stabilized decomposition with respect to its standard form? Is it possible to get similar enhancement for direct decompositions?

23 / 44

slide-44
SLIDE 44

Summarization the two general problems

Two resulting general problems

1

Is there a practical scheme of decomposition that would have an arbitrary transfer of information between direct and inverse factors?

2

Very good behavior of SAINV observed also in the incomplete case with respect to the non-stabilized algorithm cannot be explained just by breakdown-free property What is behind the clearly superior performance of the stabilized decomposition with respect to its standard form? Is it possible to get similar enhancement for direct decompositions? We have (some) answers for both of these problems

  • 1. Arbitrary direct-inverse decompositions
  • 2. Transforming the problem via projections.

23 / 44

slide-45
SLIDE 45

Summarization the two general problems

Two resulting general problems

1

Is there a practical scheme of decomposition that would have an arbitrary transfer of information between direct and inverse factors?

2

Very good behavior of SAINV observed also in the incomplete case with respect to the non-stabilized algorithm cannot be explained just by breakdown-free property What is behind the clearly superior performance of the stabilized decomposition with respect to its standard form? Is it possible to get similar enhancement for direct decompositions? We have (some) answers for both of these problems

  • 1. Arbitrary direct-inverse decompositions
  • 2. Transforming the problem via projections.

Of course, it remains a lot to do to improve GE-based decompositions from inside.

23 / 44

slide-46
SLIDE 46

Outline

1

Limits of standard algebraic approaches

2 Standard biconjugation and matrix inverses 3 Direct-inverse decompositions 4 A flavor of applications different from preconditioning 5 Conclusions

24 / 44

slide-47
SLIDE 47

New shifted biconjugation

Note: general nonsymmetric formulation is used here A−1 = ZZT ← − A−1 = ZD−1W T

25 / 44

slide-48
SLIDE 48

New shifted biconjugation

Note: general nonsymmetric formulation is used here A−1 = ZZT ← − A−1 = ZD−1W T Nonsymmetric recursions: z(j)

i

= z(j−1)

i

− z(j−1)

j

ajz(j−1)

i

ajz(j−1)

j

, w(j)

i

= w(j−1)

i

− w(j−1)

j

aT

j w(j−1) i

aT

j w(j−1) j

25 / 44

slide-49
SLIDE 49

New shifted biconjugation

Note: general nonsymmetric formulation is used here A−1 = ZZT ← − A−1 = ZD−1W T Nonsymmetric recursions: z(j)

i

= z(j−1)

i

− z(j−1)

j

ajz(j−1)

i

ajz(j−1)

j

, w(j)

i

= w(j−1)

i

− w(j−1)

j

aT

j w(j−1) i

aT

j w(j−1) j

⇓ s−1I − A−1 = ZD−1V T

25 / 44

slide-50
SLIDE 50

New shifted biconjugation

Note: general nonsymmetric formulation is used here A−1 = ZZT ← − A−1 = ZD−1W T Nonsymmetric recursions: z(j)

i

= z(j−1)

i

− z(j−1)

j

ajz(j−1)

i

ajz(j−1)

j

, w(j)

i

= w(j−1)

i

− w(j−1)

j

aT

j w(j−1) i

aT

j w(j−1) j

⇓ s−1I − A−1 = ZD−1V T Analogical recursions: zi = sei −

i−1

  • j=1

vT

j ei

dj zj , vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj, Z and D are the same in both recursions

25 / 44

slide-51
SLIDE 51

More on the new biconjugation

The (s−1I − A−1)−1) biconjugation introduced by Bru, Cerdán, Marín, Mas, 2003. The incomplete algorithm was proposed as an approximate inverse preconditioner. (factor Z)

26 / 44

slide-52
SLIDE 52

More on the new biconjugation

The (s−1I − A−1)−1) biconjugation introduced by Bru, Cerdán, Marín, Mas, 2003. The incomplete algorithm was proposed as an approximate inverse preconditioner. (factor Z) It was shown that this new biconjugation can be used to get a direct decomposition (factor U) as well, Bru, Marín, Mas, T., 2008. s−1I − A−1 = ZD−1V T and A = LDU and Z = U−1 s−1I − U−1D−1L−1 = U−1D−1V T s−1I = U−1D−1(L−1 + V T ) upper triangular ր

տ lower triangular

26 / 44

slide-53
SLIDE 53

More on the new biconjugation: II.

Pictorially V =

           

... −sL−T ... UT D ...

           

, diag(V ) = D − sI. (1)

27 / 44

slide-54
SLIDE 54

More on the new biconjugation: II.

Pictorially V =

           

... −sL−T ... UT D ...

           

, diag(V ) = D − sI. (1) V obtained by a simple recursion for its columns

27 / 44

slide-55
SLIDE 55

More on the new biconjugation: II.

Pictorially V =

           

... −sL−T ... UT D ...

           

, diag(V ) = D − sI. (1) V obtained by a simple recursion for its columns The new recursions provide scaled U and L−1 at the same time! Dropping can interconnect their computation.

27 / 44

slide-56
SLIDE 56

New biconjugation in the SPD case

Note that s−1I − A−1 = ZD−1V T , V = LD − sL−T, Z = L−T vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj,

28 / 44

slide-57
SLIDE 57

New biconjugation in the SPD case

Note that s−1I − A−1 = ZD−1V T , V = LD − sL−T, Z = L−T vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj, We do not need to compute Z at all!

28 / 44

slide-58
SLIDE 58

New biconjugation in the SPD case

Note that s−1I − A−1 = ZD−1V T , V = LD − sL−T, Z = L−T vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj, We do not need to compute Z at all! This is correct strictly mathematically, but computationally?

28 / 44

slide-59
SLIDE 59

New biconjugation in the SPD case

Note that s−1I − A−1 = ZD−1V T , V = LD − sL−T, Z = L−T vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj, We do not need to compute Z at all! This is correct strictly mathematically, but computationally? Still the inverse factor influences the direct factor. L−1 − → L

28 / 44

slide-60
SLIDE 60

New biconjugation in the SPD case

Note that s−1I − A−1 = ZD−1V T , V = LD − sL−T, Z = L−T vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj, We do not need to compute Z at all! This is correct strictly mathematically, but computationally? Still the inverse factor influences the direct factor. L−1 − → L But, dropping can interconnect computation of both L and L−1.

28 / 44

slide-61
SLIDE 61

New biconjugation in the SPD case

Note that s−1I − A−1 = ZD−1V T , V = LD − sL−T, Z = L−T vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj, We do not need to compute Z at all! This is correct strictly mathematically, but computationally? Still the inverse factor influences the direct factor. L−1 − → L But, dropping can interconnect computation of both L and L−1. We drop L using sizes of entries in L−1 and vice versa: balanced incomplete factorization, Bru, Mas, Marín, T. 2008. Is is the best strategy we can do?

28 / 44

slide-62
SLIDE 62

Balanced incomplete factorization (BIF) experiments

SPD experiments: I.

Example: matrix PWTK, n=217,918, nnz=5,926,171

29 / 44

slide-63
SLIDE 63

Balanced incomplete factorization (BIF) experiments

SPD experiments: I.

Example: matrix PWTK, n=217,918, nnz=5,926,171

1 2 3 4 5 6 x 10

6

5 10 15 20 25 time to compute the preconditioner (in seconds) size of the preconditioner (in the number of nonzeros) RIF BIF

29 / 44

slide-64
SLIDE 64

Balanced incomplete factorization (BIF) experiments: II.

Of course: not only pros; cons as well Taking approximate inverses into account, dropping must be always

  • strong. Prefiltration of entries of A is a must.

30 / 44

slide-65
SLIDE 65

Balanced incomplete factorization (BIF) experiments: II.

Of course: not only pros; cons as well Taking approximate inverses into account, dropping must be always

  • strong. Prefiltration of entries of A is a must.

We used the inverse-based dropping rules based on Saad, Bollhöfer, 2002, but dropping should be further investigated. It seems that sometimes any rules influence entries of the factors nonuniformly. Also, our dropping often forces skipping a lot of updates in the

  • decomposition. Is this really the right way to go?

30 / 44

slide-66
SLIDE 66

Balanced incomplete factorization (BIF) experiments: II.

Of course: not only pros; cons as well Taking approximate inverses into account, dropping must be always

  • strong. Prefiltration of entries of A is a must.

We used the inverse-based dropping rules based on Saad, Bollhöfer, 2002, but dropping should be further investigated. It seems that sometimes any rules influence entries of the factors nonuniformly. Also, our dropping often forces skipping a lot of updates in the

  • decomposition. Is this really the right way to go?

The convergence curve is often rather flat if we run many iterations. Is the accuracy sufficient for solving sequences from nonlinear solvers?

30 / 44

slide-67
SLIDE 67

Balanced incomplete factorization (BIF) experiments: III.

SPD experiments: II.

1 2 3 4 5 6 x 10

6

5 10 15 20 25 30 35 40 total time (in seconds) size of the preconditioner (in the number of nonzeros) RIF BIF 31 / 44

slide-68
SLIDE 68

Direct-inverse decomposition

Vector formulation of the shifted biconjugation can hide important details Bru, Mas, Marín, T. 2009 vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj,

32 / 44

slide-69
SLIDE 69

Direct-inverse decomposition

Vector formulation of the shifted biconjugation can hide important details Bru, Mas, Marín, T. 2009 vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj,

k p

32 / 44

slide-70
SLIDE 70

Direct-inverse decomposition

Vector formulation of the shifted biconjugation can hide important details Bru, Mas, Marín, T. 2009 vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj,

k p

vpi: just the entries of V with indices p + 1, . . . , i − 1 are involved

32 / 44

slide-71
SLIDE 71

Direct-inverse decomposition

Vector formulation of the shifted biconjugation can hide important details Bru, Mas, Marín, T. 2009 vi = (ai − sei)T −

i−1

  • j=1

zT

j (ai − sei)

dj vj,

k p

vpi: just the entries of V with indices p + 1, . . . , i − 1 are involved good, but not enough: the inverse factor still updated only by entries

  • f the inverse factor

32 / 44

slide-72
SLIDE 72

Direct-inverse decomposition: II.

Even more sophisticated computation possible Here we demonstrate the computation in the fully nonsymmetric case

33 / 44

slide-73
SLIDE 73

Direct-inverse decomposition: II.

Even more sophisticated computation possible Here we demonstrate the computation in the fully nonsymmetric case

  • p

p k k V Vt

33 / 44

slide-74
SLIDE 74

Direct-inverse decomposition: II.

Even more sophisticated computation possible Here we demonstrate the computation in the fully nonsymmetric case

  • p

p k k V Vt

v1:p−1 computed using fully filled areas

33 / 44

slide-75
SLIDE 75

Direct-inverse decomposition: II.

Even more sophisticated computation possible Here we demonstrate the computation in the fully nonsymmetric case

  • p

p k k V Vt

v1:p−1 computed using fully filled areas vp+1:n computed using dashed areas

33 / 44

slide-76
SLIDE 76

Direct-inverse decomposition: II.

Even more sophisticated computation possible Here we demonstrate the computation in the fully nonsymmetric case

  • p

p k k V Vt

v1:p−1 computed using fully filled areas vp+1:n computed using dashed areas direct and inverse factors influence each other

33 / 44

slide-77
SLIDE 77

Direct-inverse (NBIF) decomposition: experiments: II.

0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 2.4 x 10

5

200 400 600 800 1000 1200 number of iterations size of the preconditioner (in the number of nonzeros) ILU(tau) NBIF

Figure: Sizes of NBIF and ILU(τ) preconditioners versus iteration counts of the preconditioned BiCGStab method for the matrix CHEM_MASTER1.

34 / 44

slide-78
SLIDE 78

Scaling parameter

Choice of scaling parameter s / computational procedures should be coordinated Here we demonstrate the computation in the fully nonsymmetric case

−10 −8 −6 −4 −2 2 4 6 8 10 10

−10

10

−8

10

−6

10

−4

10

−2

10 10

2

k (s=10k) error norm ||A−LDU||2 Matrix Olm100 without Z with Z 35 / 44

slide-79
SLIDE 79

Outline

1

Limits of standard algebraic approaches

2 Standard biconjugation and matrix inverses 3 Direct-inverse decompositions 4 A flavor of applications different from preconditioning 5 Conclusions

36 / 44

slide-80
SLIDE 80

Condition number estimation

Condition number estimation in the 2-norm (Duintjer Tebbens et al., 2009) Two basic approaches: Incremental condition estimation using left singular vectors (ICE, Bischof, 1990) and Incremental norm estimation using right singular vectors (INE, Duff, Vömel, 2002)

37 / 44

slide-81
SLIDE 81

Condition number estimation

Condition number estimation in the 2-norm (Duintjer Tebbens et al., 2009) Two basic approaches: Incremental condition estimation using left singular vectors (ICE, Bischof, 1990) and Incremental norm estimation using right singular vectors (INE, Duff, Vömel, 2002) Availability of simultaneously computed inverse factor = ⇒: four possible ways to estimate the condition number

37 / 44

slide-82
SLIDE 82

Condition number estimation

Condition number estimation in the 2-norm (Duintjer Tebbens et al., 2009) Two basic approaches: Incremental condition estimation using left singular vectors (ICE, Bischof, 1990) and Incremental norm estimation using right singular vectors (INE, Duff, Vömel, 2002) Availability of simultaneously computed inverse factor = ⇒: four possible ways to estimate the condition number κ(R) ≈ σmaxL(R) σminL(R) (ICE) − → . . . − → κ(R) ≈ σmaxR(R)σminRI(R) yellow (green) curve blue curve

37 / 44

slide-83
SLIDE 83

Condition number estimation: II.

50 Random matrices A forming AAT

10 20 30 40 50 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 38 / 44

slide-84
SLIDE 84

Condition number estimation: III.

50 Random matrices A forming A + AT with an additional shift

10 20 30 40 50 0.78 0.8 0.82 0.84 0.86 0.88 0.9 0.92 0.94 0.96 0.98 39 / 44

slide-85
SLIDE 85

Condition number estimation: IV.

50 Random matrices A forming A + AT , different shift

10 20 30 40 50 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 40 / 44

slide-86
SLIDE 86

Condition number estimation: V.

6 Harwell-Boeing matrices, not via BIF

1 2 3 4 5 6 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 41 / 44

slide-87
SLIDE 87

Outline

1

Limits of standard algebraic approaches

2 Standard biconjugation and matrix inverses 3 Direct-inverse decompositions 4 A flavor of applications different from preconditioning 5 Conclusions

42 / 44

slide-88
SLIDE 88

Conclusions

Development of algebraic decompositions far from being finished.

43 / 44

slide-89
SLIDE 89

Conclusions

Development of algebraic decompositions far from being finished. In particular, a new direct-inverse decomposition BIF/NBIF for preconditioning was introduced.

43 / 44

slide-90
SLIDE 90

Conclusions

Development of algebraic decompositions far from being finished. In particular, a new direct-inverse decomposition BIF/NBIF for preconditioning was introduced. BIF/NBIF may be useful in other applications, e.g. in construction of condition estimators.

43 / 44

slide-91
SLIDE 91

Conclusions

Development of algebraic decompositions far from being finished. In particular, a new direct-inverse decomposition BIF/NBIF for preconditioning was introduced. BIF/NBIF may be useful in other applications, e.g. in construction of condition estimators. It is possible to construct a direct (GE-like) decomposition using

  • rthogonal projections.

43 / 44

slide-92
SLIDE 92

Conclusions

Development of algebraic decompositions far from being finished. In particular, a new direct-inverse decomposition BIF/NBIF for preconditioning was introduced. BIF/NBIF may be useful in other applications, e.g. in construction of condition estimators. It is possible to construct a direct (GE-like) decomposition using

  • rthogonal projections.

Do we really understand Gaussian elimination in the sense to expect all future improvements of GE-like decompositions from the inside?

43 / 44

slide-93
SLIDE 93

Conclusions

Development of algebraic decompositions far from being finished. In particular, a new direct-inverse decomposition BIF/NBIF for preconditioning was introduced. BIF/NBIF may be useful in other applications, e.g. in construction of condition estimators. It is possible to construct a direct (GE-like) decomposition using

  • rthogonal projections.

Do we really understand Gaussian elimination in the sense to expect all future improvements of GE-like decompositions from the inside? Of course, not.

43 / 44

slide-94
SLIDE 94

Conclusions

Development of algebraic decompositions far from being finished. In particular, a new direct-inverse decomposition BIF/NBIF for preconditioning was introduced. BIF/NBIF may be useful in other applications, e.g. in construction of condition estimators. It is possible to construct a direct (GE-like) decomposition using

  • rthogonal projections.

Do we really understand Gaussian elimination in the sense to expect all future improvements of GE-like decompositions from the inside? Of course, not.

The way from efficient rules of decomposition to fully GE-aware algorithms may be very long

43 / 44

slide-95
SLIDE 95

Last but not least Thank you for your attention!

44 / 44

slide-96
SLIDE 96

Last but not least Thank you for your attention!

44 / 44

slide-97
SLIDE 97

Last but not least Thank you for your attention!

44 / 44

slide-98
SLIDE 98

Last but not least Thank you for your attention!

44 / 44