monte carlo methods for magnetic systems
play

Monte Carlo methods for magnetic systems Zoltn Nda Babe-Bolyai - PowerPoint PPT Presentation

Monte Carlo methods for magnetic systems Zoltn Nda Babe-Bolyai University Dept of Theoretical and Computational Physics Cluj-Napoca, Romania Main objective of the lecture: To give an introduction for basic Monte Carlo methods in some


  1. Monte Carlo methods for magnetic systems Zoltán Néda Babeş-Bolyai University Dept of Theoretical and Computational Physics Cluj-Napoca, Romania Main objective of the lecture: To give an introduction for basic Monte Carlo methods in some simple models of magnetism . European School on Magnetism, Timişoara, Sept. 1-11, 2009

  2. Syllabus Syllabus •About Monte Carlo methods • Deterministic versus stochastic simulation methods • Elements of Stochastic Processes (Markov chains) • Monte Carlo integration • Theoretical approach to magnetic models • What we are interested in? applet made by R. Sumi • The Metropolis MC method for magnetic systems • Implementing the Metropolis MC method for the 2D Ising model • Finite-size effects • Efficient MC techniques

  3. What are Monte Carlo methods? - Molecular dynamics (deterministic Computer simulations, based on the integration of the equation of motion) simulation - Monte Carlo methods (Stochastic simulation methods: techniques, where the random number generation plays a crucial role) - In general we speak about Monte Carlo simulation methods whenever the use of the random numbers are crucial in the algorithm! MC: the art of using pseudo random numbers - Monte Carlo techniques are widely used in studying models of : statistical physics, soft condensed matter physics, material science, many-body problems, complex systems, fluid mechanics, biophysics, econo-physics, nonlinear phenomena, particle physics, heavy-ion physics, surface physics, neuroscience etc….

  4. Deterministic versus stochastic simulations the Galton table - used to exemplify the normal distribution Molecular dynamics approach: integrating in time the equation of motion of the particles. advantage  the realistic dynamics     x ( t ) x ( t 1 ) disadvantage  slow even on supercomputers, only short time-scales or small systems can be simulated Random number: 1 with p=1/2 and -1 with p=1/2 Monte Carlo approach: the result of many deterministic effects is handled as a stochastic (random) force. advantage  fast, easy to implement disadvantage  less realistic, many elements of the real phenomena are not in the model Molecular dynamics MC

  5. Some necessary elements of Stochastic Processes lements of Stochastic Processes Some necessary e Markov processes/ Markov chains Markov processes/ Markov chains Stochastic process : let x label the element of any state-space. A process that randomly visits in time these possible x states is a stochastic process: 2D random Example the 1D random walk: walk: P=1/2 P=1/2 -4 -3 -2 -1 0 1 2 3 Markov processes (chain) are characterized by a lack of memory (i.e. the statistical properties of the immediate future are uniquely determined from the present, regardless of the past) Example : random walk --> Markov process; self-avoiding walk is NOT a Markov process Let x i be the state of the stochastic system at step “i”, a stochastic variable The time- evolution of the system is described by a sequence of states: x 0 , x 1 , ….., x n , …. P ( x | x ,..... x ) The conditional probability that x n is realized if previously we had: x 0 , x 1 , ….., x n-1 :  n n 1 0  P ( x | x , x ,..., x ) P ( x | x ) Definition : For a Markov process we have:    n n 1 n 2 0 n n 1  P ( x ,..., x ) P ( x | x ). P ( x | x ).... P ( x , x ). a    0 n n n 1 n 1 n 2 1 0 0 one-step transition probabilities, elements of    P ( x , x ) P ( x x ) P m j m j m , j the stochastic matrix

  6.  w 0 ; m  Definition : A probability distribution over the possible  w 1 ; states (w k ) is called invariant or stationary for a given m m Markov chain if satisfy:   w w P s m ms  w The probability that x=k during an m k infinitely long process - A Markov chain is irreducible if and only if every state can be reached from every state! (the stochastic matrix is irreducible) -A Markov chain is aperiodic , if all states are aperiodic . A state x has a period T>1 if P ii (n) =0 unless n=zT (z: integer), and T is the smallest integer with this property. A state is aperiodic if no such T>1 exist. (Here we denoted by P ik (n) the probability to get from state i to state k through n steps) Definition : An irreducible and aperiodic Markov chain is ergodic The basic theorem for Markov processes : An ergodic Markov chain posses an invariant distribution w k over the possible states

  7. One dimensional Monte Carlo integration One dimensional Monte Carlo integration b   I f ( x ) dx Problem: given a function f(x), compute the integral: a The integral can be computed by choosing n points (x i ) randomly on the     ( x ) 1 /( b a ) Const [a,b] interval, and with a uniform distribution : b    ( x ) dx 1 normalization:     ( x ) probability density: P ( x , x dx ) ( x ) dx a  b a n Straightforward sampling      I f ( x ) ( b a ) f ( x ) i n  i 1 The strong law of large numbers guarantees us that for a sufficiently large sample one can come arbitrary close to the desired integral!     I f ( x ) ( x ) dx ; Let x 1 ,x 2 ,…,x n be random numbers selected according    ( x ) to a normalized probability density , then :   1 n      P lim f ( x ) I 1 (!) the above affirmation is also true if the   n i n    i 1 random numbers are correlated, or the interval is finite  x  How rapidly the sum converge? --> for very badly!!! ( ) Const  ( x ) Central limit theorem  the convergence improve if the shape of approximates f(x)  we are sampling in the neighborhood where f(x) is big

  8. Important sampling The important sampling MC method will calculate the I integral by sampling  on random points on the [a,b] interval according to a distribution ( x ) which approximates the shape of |f(x)| If one generates n points, x i ,according to an arbitrary  ( x ) b b f ( x ) 1 N f ( x )        I f ( x ) dx ( x ) dx i   ( x ) n ( x )  i 1 i a a  x  ( ) | f ( x ) | the convergence is infinitely fast if Before getting to excited….  one cannot simply choose ,  x  ( ) | f ( x ) |   since in this case one cannot normalize (normalization of is ( x ) ( x ) equivalent with the initial problem  one cannot generate thus random  x  numbers simply according to the desired distribution ( ) | f ( x ) |

  9. Theoretical approach to a magnetic ordering Usually canonic ensemble is used  T, N, h is fixed (T  temperature, h  external magnetic field, N  particle number) relevant energies heat bath [internal interactions + interaction with external magnetic field, h, is a stochastic effect +kinetic terms]  favor randomness a deterministic effect  could   T or 1 kT /( ) favor ordering Statistical thermodynamics Hamiltonian , H(x i ,H)=E i approach   E   Z exp  i  (x i  labels the microstates)  kT  i   F kT ln( Z )  H ( x )    Z exp   dx  kT   assuming that the density of F  the free energy; k  the Boltzmann constant state-space points is constant

  10. What are we interested in ? The primary goal of the MC type simulations in magnetic systems is to estimate some averages at various T, h and N values M average magnetization 2 M average square magnetization E average energy a sum with huge number of terms (number of terms 2 E average square energy increasing exponentially with 1 system size…ex: 2 N )…or very     X X exp( E ) high dimensional integrals i i Z in canonical ensemble i 1 X  2 2  M ; M ; E ; E     X X ( x ) exp [ H ( x )] dx Z  the problem is that these sums cannot be usually analytically calculated  MC methods !! exact enumeration is possible for small N (not of thermodynamic interest)  C we are also interested in measurable quantities like: ; V    E   heat capacity at 1          2 2 C E E   constant volume V  2 T kT N   V , N    M 1       2     2 ( M M ) susceptibility    h NkT    H 0

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend