Title: CS 240A Lecture 12 Parallel sparse matrix methods
1CS 240A Lecture 12Parallel sparse matrix methods
- Homework 3 due Mon 10 May
- Project progress report and web site due Mon 10
May - 1-2 pages, emailed to course staff
- Also email us a link to your project web page
- Sparse Gaussian elimination
- SuperLU direct methods
- Preconditioned conjugate gradients iterative
methods
2The Landscape of Axb Solvers
More Robust
Less Storage (if sparse)
3Complexity of linear solvers
Time to solve model problem (Poissons equation)
on regular mesh
4Complexity of direct methods
Time and space to solve any problem on any
well-shaped finite element mesh
5Column Cholesky Factorization
- for j 1 n
- for k 1 j-1
- cmod(j,k)
- for i j n
- A(i,j) A(i,j) A(i,k)A(j,k)
- end
- end
- cdiv(j)
- A(j,j) sqrt(A(j,j))
- for i j1 n
- A(i,j) A(i,j) / A(j,j)
- end
- end
- Column j of A becomes column j of L
6Sparse Column Cholesky Factorization
- for j 1 n
- for k lt j with A(j,k) nonzero
- sparse cmod(j,k)
- A(jn, j) A(jn, j) A(jn, k)A(j, k)
- end
- sparse cdiv(j)
- A(j,j) sqrt(A(j,j))
- A(j1n, j) A(j1n, j) / A(j,j)
- end
- Column j of A becomes column j of L
7Data structures
- Full
- 2-dimensional array of real or complex numbers
- (nrowsncols) memory
- Sparse
- compressed column storage
- about (1.5nzs .5ncols) memory
D
8Graphs and Sparse Matrices Cholesky
factorization
Fill new nonzeros in factor
Symmetric Gaussian elimination for j 1 to n
add edges between js higher-numbered
neighbors
G(A)chordal
G(A)
9Fill-reducing matrix permutations
- Minimum degree
- Eliminate row/col with fewest nzs, add fill,
repeat - Theory can be suboptimal even on 2D model
problem - Practice often wins for medium-sized problems
- Nested dissection
- Find a separator, number it last, proceed
recursively - Theory approx optimal separators gt approx
optimal fill and flop count - Practice often wins for very large problems
- Banded orderings (Reverse Cuthill-McKee, Sloan, .
. .) - Try to keep all nonzeros close to the diagonal
- Theory, practice often wins for long, thin
problems - Best modern general-purpose orderings are ND/MD
hybrids.
10Symmetric-pattern multifrontal factorization
11Symmetric-pattern multifrontal factorization
- For each node of T from leaves to root
- Sum own row/col of A with childrens Update
matrices into Frontal matrix - Eliminate current variable from Frontal matrix,
to get Update matrix - Pass Update matrix to parent
12Symmetric-pattern multifrontal factorization
- For each node of T from leaves to root
- Sum own row/col of A with childrens Update
matrices into Frontal matrix - Eliminate current variable from Frontal matrix,
to get Update matrix - Pass Update matrix to parent
13Symmetric-pattern multifrontal factorization
- For each node of T from leaves to root
- Sum own row/col of A with childrens Update
matrices into Frontal matrix - Eliminate current variable from Frontal matrix,
to get Update matrix - Pass Update matrix to parent
14Symmetric-pattern multifrontal factorization
T(A)
15Symmetric-pattern multifrontal factorization
- Really uses supernodes, not nodes
- All arithmetic happens on dense square matrices.
- Needs extra memory for a stack of pending update
matrices - Potential parallelism
- between independent tree branches
- parallel dense ops on frontal matrix
16MUMPS distributed-memory multifrontalAmestoy,
Duff, LExcellent, Koster, Tuma
- Symmetric-pattern multifrontal factorization
- Parallelism both from tree and by sharing dense
ops - Dynamic scheduling of dense op sharing
- Symmetric preordering
- For nonsymmetric matrices
- optional weighted matching for heavy diagonal
- expand nonzero pattern to be symmetric
- numerical pivoting only within supernodes if
possible (doesnt change pattern) - failed pivots are passed up the tree in the
update matrix
17GEPP Gaussian elimination w/ partial pivoting
P
x
- PA LU
- Sparse, nonsymmetric A
- Columns may be preordered for sparsity
- Rows permuted by partial pivoting (maybe)
- High-performance machines with memory hierarchy
18See Sherry Lis slides on SuperLU-MT
19Column Preordering for Sparsity
Q
P
- PAQT LU Q preorders columns for sparsity, P
is row pivoting - Column permutation of A ? Symmetric permutation
of ATA (or G?(A)) - Symmetric ordering Approximate minimum degree
Amestoy, Davis, Duff - But, forming ATA is expensive (sometimes bigger
than LU). - Solution ColAMD ordering ATA with data
structures based on A
20Column AMD Davis, G, Ng, Larimore
Matlab 6
row
col
A
I
row
AT
0
col
G(aug(A))
A
aug(A)
- Eliminate row nodes of aug(A) first
- Then eliminate col nodes by approximate min
degree - 4x speed and 1/3 better ordering than Matlab-5
min degree, 2x speed of AMD on ATA - Question Better orderings based on aug(A)?
21SuperLU-dist GE with static pivoting Li,
Demmel
- Target Distributed-memory multiprocessors
- Goal No pivoting during numeric factorization
- Permute A unsymmetrically to have large elements
on the diagonal (using weighted bipartite
matching) - Scale rows and columns to equilibrate
- Permute A symmetrically for sparsity
- Factor A LU with no pivoting, fixing up small
pivots - if aii lt e A then replace aii by
?e1/2 A - Solve for x using the triangular factors Ly
b, Ux y - Improve solution by iterative refinement
22Row permutation for heavy diagonal Duff,
Koster
1
5
2
3
4
1
2
3
4
5
A
- Represent A as a weighted, undirected bipartite
graph (one node for each row and one node for
each column) - Find matching (set of independent edges) with
maximum product of weights - Permute rows to place matching on diagonal
- Matching algorithm also gives a row and column
scaling to make all diag elts 1 and all
off-diag elts lt1
23Iterative refinement to improve solution
- Iterate
- r b Ax
- backerr maxi ( ri / (Ax b)i )
- if backerr lt e or backerr gt lasterr/2 then
stop iterating - solve LUdx r
- x x dx
- lasterr backerr
- repeat
- Usually 0 3 steps are enough
24SuperLU-dist Distributed static data structure
Process(or) mesh
Block cyclic matrix layout
25See Sherry Lis slides on SuperLU-dist
26Question Preordering for static pivoting
- Less well understood than symmetric factorization
- Symmetric bottom-up, top-down, hybrids
- Nonsymmetric top-down just starting to replace
bottom-up - Symmetric best ordering is NP-complete, but
approximation theory is based on graph
partitioning (separators) - Nonsymmetric no approximation theory is known
partitioning is not the whole story
27Remarks on (nonsymmetric) direct methods
- Combinatorial preliminaries are important
ordering, bipartite matching, symbolic
factorization, scheduling - not well understood in many ways
- also, mostly not done in parallel
- Multifrontal tends to be faster but use more
memory - Unsymmetric-pattern multifrontal
- Lots more complicated, not simple elimination
tree - Sequential and SMP versions in UMFpack and WSMP
(see web links) - Distributed-memory unsymmetric-pattern
multifrontal is a research topic - Not mentioned symmetric indefinite problems
- Direct-methods technology is also needed in
preconditioners for iterative methods
28Conjugate gradient iteration for Ax b
x0 0 approx solution r0 b
residual b - Ax d0 r0
search direction for k 1, 2, 3, . .
. xk xk-1
new approx solution rk
new residual dk
new search direction
29Conjugate gradient iteration for Ax b
x0 0 approx solution r0 b
residual b - Ax d0 r0
search direction for k 1, 2, 3, . .
. ak step
length xk xk-1 ak dk-1
new approx solution rk
new residual dk
new
search direction
30Conjugate gradient iteration for Ax b
x0 0 approx solution r0 b
residual b - Ax d0 r0
search direction for k 1, 2, 3, . .
. ak (rTk-1rk-1) / (dTk-1Adk-1) step
length xk xk-1 ak dk-1
new approx solution rk
new residual
dk
new search direction
31Conjugate gradient iteration for Ax b
x0 0 approx solution r0 b
residual b - Ax d0 r0
search direction for k 1, 2, 3, . .
. ak (rTk-1rk-1) / (dTk-1Adk-1) step
length xk xk-1 ak dk-1
new approx solution rk
new residual ßk
(rTk rk) / (rTk-1rk-1) dk rk ßk dk-1
new search direction
32Conjugate gradient iteration for Ax b
x0 0 approx solution r0 b
residual b - Ax d0 r0
search direction for k 1, 2, 3, . .
. ak (rTk-1rk-1) / (dTk-1Adk-1) step
length xk xk-1 ak dk-1
new approx solution rk rk-1 ak
Adk-1 new residual ßk
(rTk rk) / (rTk-1rk-1) dk rk ßk dk-1
new search direction
33Conjugate gradient iteration
x0 0, r0 b, d0 r0 for k 1, 2,
3, . . . ak (rTk-1rk-1) / (dTk-1Adk-1)
step length xk xk-1 ak dk-1
approx solution rk rk-1 ak
Adk-1 residual ßk
(rTk rk) / (rTk-1rk-1) improvement dk
rk ßk dk-1
search direction
- One matrix-vector multiplication per iteration
- Two vector dot products per iteration
- Four n-vectors of working storage
34Conjugate gradient Krylov subspaces
- Eigenvalues Av ?v ?1, ?2 ,
. . ., ?n - Cayley-Hamilton theorem
- (A ?1I)(A ?2I) (A ?nI) 0
- Therefore S ciAi 0 for some ci
- so A-1 S (ci/c0) Ai1
- Krylov subspace
- Therefore if Ax b, then x A-1 b and
- x ? span (b, Ab, A2b, . . ., An-1b) Kn (A, b)
0 ? i ? n
1 ? i ? n
35Conjugate gradient Orthogonal sequences
- Krylov subspace Ki (A, b) span (b, Ab, A2b, .
. ., Ai-1b) - Conjugate gradient algorithm for i 1, 2, 3,
. . . find xi ? Ki (A, b) such that ri
(b Axi) ? Ki (A, b) - Notice ri ? Ki1 (A, b), so ri ? rj for all
j lt i - Similarly, the directions are
A-orthogonal (xi xi-1 )TA (xj xj-1 ) 0 - The magic Short recurrences. . . A is symmetric
gt can get next residual and direction from
the previous one, without saving them all.
36Conjugate gradient Convergence
- In exact arithmetic, CG converges in n steps
(completely unrealistic!!) - Accuracy after k steps of CG is related to
- consider polynomials of degree k that are equal
to 1 at 0. - how small can such a polynomial be at all the
eigenvalues of A? - Thus, eigenvalues close together are good.
- Condition number ?(A) A2 A-12
?max(A) / ?min(A) - Residual is reduced by a constant factor by
O(?1/2(A)) iterations of CG.
37Conjugate gradient Parallel implementation
- Lay out matrix and vectors by rows
- Hard part is matrix-vector product
y Ax - Algorithm
- Each processor j
- Broadcast x(j)
- Compute y(j) A(j,)x
- May send more of x than needed
- Partition / reorder matrix to reduce communication
38Other Krylov subspace methods
- Nonsymmetric linear systems
- GMRES for i 1, 2, 3, . . . find xi ? Ki
(A, b) such that ri (Axi b) ? Ki (A,
b)But, no short recurrence gt save old vectors
gt lots more space - (Usually restarted every k iterations to use
less space.) - BiCGStab, QMR, etc.Two spaces Ki (A, b) and Ki
(AT, b) w/ mutually orthogonal basesShort
recurrences gt O(n) space, but less robust - Convergence and preconditioning more delicate
than CG - Active area of current research
- Eigenvalues Lanczos (symmetric), Arnoldi
(nonsymmetric)
39Preconditioners
- Suppose you had a matrix B such that
- condition number ?(B-1A) is small
- By z is easy to solve
- Then you could solve (B-1A)x B-1b instead of Ax
b - Each iteration of CG multiplies a vector by B-1A
- First multiply by A
- Then solve a system with B
40Preconditioned conjugate gradient iteration
x0 0, r0 b, d0 B-1 r0, y0
B-1 r0 for k 1, 2, 3, . . . ak
(yTk-1rk-1) / (dTk-1Adk-1) step length xk
xk-1 ak dk-1 approx
solution rk rk-1 ak Adk-1
residual yk B-1 rk
preconditioning
solve ßk (yTk rk) / (yTk-1rk-1)
improvement dk yk ßk dk-1
search direction
- One matrix-vector multiplication per iteration
- One solve with preconditioner per iteration
41Choosing a good preconditioner
- Suppose you had a matrix B such that
- condition number ?(B-1A) is small
- By z is easy to solve
- Then you could solve (B-1A)x B-1b instead of Ax
b - B A is great for (1), not for (2)
- B I is great for (2), not for (1)
- Domain-specific approximations sometimes work
- B diagonal of A sometimes works
- Better blend in some direct-methods ideas. . .
42Incomplete Cholesky factorization (IC, ILU)
- Compute factors of A by Gaussian elimination,
but ignore fill - Preconditioner B RTR ? A, not formed explicitly
- Compute B-1z by triangular solves (in time
nnz(A)) - Total storage is O(nnz(A)), static data structure
- Either symmetric (IC) or nonsymmetric (ILU)
43Incomplete Cholesky and ILU Variants
- Allow one or more levels of fill
- unpredictable storage requirements
- Allow fill whose magnitude exceeds a drop
tolerance - may get better approximate factors than levels of
fill - unpredictable storage requirements
- choice of tolerance is ad hoc
- Partial pivoting (for nonsymmetric A)
- Modified ILU (MIC) Add dropped fill to
diagonal of U or R - A and RTR have same row sums
- good in some PDE contexts
44Incomplete Cholesky and ILU Issues
- Choice of parameters
- good smooth transition from iterative to direct
methods - bad very ad hoc, problem-dependent
- tradeoff time per iteration (more fill gt more
time) vs of iterations (more
fill gt fewer iters) - Effectiveness
- condition number usually improves (only) by
constant factor (except MIC for some problems
from PDEs) - still, often good when tuned for a particular
class of problems - Parallelism
- Triangular solves are not very parallel
- Reordering for parallel triangular solve by graph
coloring
45Sparse approximate inverses
- Compute B-1 ? A explicitly
- Minimize B-1A I F (in parallel, by
columns) - Variants factored form of B-1, more fill, . .
- Good very parallel
- Bad effectiveness varies widely
46Multigrid (preview of next lecture)
- For a PDE on a fine mesh, precondition using a
solution on a coarser mesh - Use idea recursively on hierarchy of meshes
- Solves the model problem (Poissons eqn) in
linear time! - Often useful when hierarchy of meshes can be
built - Hard to parallelize coarse meshes well
- This is just the intuition lots of theory and
technology
47The Landscape of Axb Solvers
More Robust
Less Storage (if sparse)
48Complexity of linear solvers
Time to solve model problem (Poissons equation)
on regular mesh