Kalos3, Jaime Marian3 1Los Alamos National Laboratory - - PowerPoint PPT Presentation

kalos3 jaime marian3
SMART_READER_LITE
LIVE PREVIEW

Kalos3, Jaime Marian3 1Los Alamos National Laboratory - - PowerPoint PPT Presentation

Synchronous Parallel Kinetic Monte Carlo Application to Critical, billion-atom, 3D Ising Systems Enrique Martinez1, Paul R. Monasterio2, Mal Kalos3, Jaime Marian3 1Los Alamos National Laboratory 2Massachusetts Institute of Technology 3Lawrence


slide-1
SLIDE 1

Synchronous Parallel Kinetic Monte Carlo Application to Critical, billion-atom, 3D Ising Systems

Enrique Martinez1, Paul R. Monasterio2, Mal Kalos3, Jaime Marian3

1Los Alamos National Laboratory 2Massachusetts Institute of Technology 3Lawrence Livermore National Laboratory

slide-2
SLIDE 2

Lawrence Livermore National Laboratory

Motivation: Method

  • Kinetic Monte Carlo is widely used in many

scientific disciplines.

  • Several methods to (more or less) efficiently

parameterize kMC models.

  • Not so much work on speeding up kMC itself.
  • Parallelization

is commonplace in MD simulations and as a way to ‘slave’ computationally-demanding calculations

  • f

saddle points, attempt frequencies, etc, in kMC methods.

  • What about parallelization of kMC?

2

slide-3
SLIDE 3

Lawrence Livermore National Laboratory

Motivation: Application

  • Ising model can be used to map discrete-lattice

systems: lattice gas, binary alloys, magnetism, etc.

  • Ising system is interesting for studying second-
  • rder phase transformations.
  • Belongs to a universality class for systems with

long-range correlated disorder.

  • Very large systems must be considered to

capture the kinetic behavior, particularly in 3D.

  • No analytical solution for critical kinetics in 3D,
  • nly slow converging numerical solutions.
  • Many methods employed over the years.

3

slide-4
SLIDE 4

Lawrence Livermore National Laboratory

4

Master equation: Transition rates for Glauber dynamics:

Ising system kinetics

ΔEi follows from the Ising Hamiltonian

slide-5
SLIDE 5

Lawrence Livermore National Laboratory

Temperature behavior of Ising systems

JP Sethna (2009)

5

T >Tc T =Tc T < Tc

  • At the critical

temperature Tc, domains of aligned spins are created.

  • These domains are

defined by a correlation length ξ:

  • ν is the ‘scale’ critical

exponent.

  • Critical exponents not

converged for 3D. paramagnetic critical ferromagnetic

slide-6
SLIDE 6

Lawrence Livermore National Laboratory

The net magnetization is the order parameter of the Ising system

6

m(t) Time [λ– 1] The value of the critical exponents in 1D and 2D is analytically known and can be converged for 4 and higher dimensions. In 3D: no converged numerical solution.

Serial kMC not sufficient

slide-7
SLIDE 7

Lawrence Livermore National Laboratory

Parallel kMC algorithms to study large kinetic systems

7

  • Discrete event kinetics are inherently difficult to

parallelize.

  • Traditional parallelization approaches based on

asynchronous kinetics (Lubachevsky 1988, Jefferson

1985).

  • Causality errors arise with these approaches:

mutually affecting events occurring in different domains.

  • This requires ‘roll-back’ techniques to reconcile the

time evolution of different processors.

  • This leads to implementation complexity and regions
  • f low efficiency.
  • Rigorous and semi-rigorous algorithms have been

proposed (Amar and Shim 2003, Shim and Amar 2005,

slide-8
SLIDE 8

Lawrence Livermore National Laboratory

We use a novel synchronous parallel kMC algorithm to study very large systems

8

Assume a spatial domain containing N walkers: Each walker defined by a rate Perform spatial domain decomposition:

slide-9
SLIDE 9

Lawrence Livermore National Laboratory

We use a novel synchronous parallel kMC algorithm to study very large systems

9

Now, for parallel kMC, perform K (4) domain partitions and construct frequency lines:

1:

R1

2: 3:

R3

4:

R4 R2 qi

slide-10
SLIDE 10

Lawrence Livermore National Laboratory

We use a novel synchronous parallel kMC algorithm to study very large systems

10

Now, for parallel kMC, perform K (4) domain partitions and construct frequency lines:

1:

R1

2: 3:

R3

4:

R4 Rma x qi

slide-11
SLIDE 11

Lawrence Livermore National Laboratory

We use a novel synchronous parallel kMC algorithm to study very large systems

11

Now, for parallel kMC, perform K (4) domain partitions and construct frequency lines:

1:

R1

2: 3:

R3

4:

R4 Rma x r01 r03 r04 qi

slide-12
SLIDE 12

Lawrence Livermore National Laboratory

We use a novel synchronous parallel kMC algorithm to study very large systems

12

The r0k are the ‘dummy’ rates (no event) that ensure synchronicity: Now, for parallel kMC, perform K (4) domain partitions and construct frequency lines:

1:

R1

2: 3:

R3

4:

R4 Rma x r01 r03 r04 qi

slide-13
SLIDE 13

Lawrence Livermore National Laboratory

We use a novel synchronous parallel kMC algorithm to study very large systems

13

The r0k are the ‘dummy’ rates (no event) that ensure synchronicity: Now, for parallel kMC, perform K (4) domain partitions and construct frequency lines:

1:

R1

2: 3:

R3

4:

R4 Rma x r01 r03 r04 qi

slide-14
SLIDE 14

Lawrence Livermore National Laboratory

We use a novel synchronous parallel kMC algorithm to study very large systems

14

The r0k are the ‘dummy’ rates (no event) that ensure synchronicity: Now, for parallel kMC, perform K (4) domain partitions and construct frequency lines: For optimum scalability, perform domain decomposition subject to the following constraint:

1:

R1

2: 3:

R3

4:

R4 Rma x r01 r03 r04 qi

slide-15
SLIDE 15

Lawrence Livermore National Laboratory

1

Perform spatial decomposition into K domains.

2

Define partial aggregate rates in each :

3

Choose the maximum partial rate as:

4

Assign ‘null’ rates to each such that:

5

Sample event from each subdomain with probability

6

Execute event and advance time by

Parallel kMC algorithm

15

slide-16
SLIDE 16

Lawrence Livermore National Laboratory

Parallel boost

16

1 Non-interacting diffusers (homog) Non-interacting diffusers (inhomog) Interacting diffusers

slide-17
SLIDE 17

Lawrence Livermore National Laboratory

Solution of boundary conflicts via sublattice decomposition

17

Boundary conflicts appear when mutually-influencing events occur simultaneously on different domains A simple solution is to use a sublattice decomposition (chess method in 2D) Co-occurring events only on identically-colored subcells Amar et al. (2004, 2005)

slide-18
SLIDE 18

Lawrence Livermore National Laboratory

Sublattice decomposition introduces a bias

18

Standard deviation of serial runs

  • Reduced sampling space

introduces a systematic error (bias).

  • Bias can be controlled with

system size and numbers of processors.

  • One must also include

intrinsic statistical fluctuations of the parallel runs:

  • We find that σb is always less than the standard

deviation of the serial calculations

slide-19
SLIDE 19

Lawrence Livermore National Laboratory

Calculation of critical exponent z

19

1024×512×512 1024×1024×512 1024×1024×1024 Martinez, Monasterio, Marian, JCP (2010)

slide-20
SLIDE 20

Lawrence Livermore National Laboratory

Weak scaling of parallel kMC algorithm is good.

20

Parallel efficiency governed by local MPI calls:

K

slide-21
SLIDE 21

Lawrence Livermore National Laboratory

Conclusions

  • Parallel synchronous kMC algorithm suitable for

large systems.

Resolution of boundary conflicts Good scalability Controlled sampling errors

21

  • Critical behavior of Ising systems is well reproduced

and converged to the state of the art.

  • Current and future applications of the method include

solid solution precipitation, segregation, and in general situations where large systems are required.