Lectures Notes on Boltzmann's Equation

Total Page:16

File Type:pdf, Size:1020Kb

Lectures Notes on Boltzmann's Equation Lectures notes on Boltzmann's equation Simone Calogero 1 Introduction Kinetic theory describes the statistical evolution in phase-space1 of systems composed by a large number of particles (of order 1020). The main goal of kinetic theory, as far as the physical applica- tions are concerned, is to predict the evolution of those quantities associated to the system which depend only on the local average dynamics of the particles. These are called macroscopic quanti- ties. The most important examples of physical systems to which kinetic theory applies are dilute gases (where the molecules play the role of the particles) and in this case examples of macroscopic quantities are the temperature and the pressure of the gas. In fact, the temperature of a gas is a measure of the mean kinetic energy of the molecules in small regions of the gas, while the pressure is a macroscopic manifestation of the exchange of momentum between the particles and the walls of the gas container (of with any body inserted into the gas). The relation between kinetic and macroscopic quantities will be discussed in Section 7. The statistical description of the dynamics is given in terms of the one particle distribution function, denoted by f, which is a function of time, position and velocity (or momentum2), that is 3 3 f = f(t; x; v); t ≥ 0 ; x 2 R ; v 2 R : By definition, f(t; x; v) is the probability density to find a particle at time t, in the position x, with velocity v. Thus the integral Z Z f(t; x; v) dv dx V Ω gives the probability to find a particle in the region V at time t with velocities v 2 Ω. Since the number of particles is so large, the above quantity can also be interpreted as the relative number of particles in the region V at time t with velocities v 2 Ω. Moreover, denoting by M the total mass of the system, the above integral multiplied by M can be interpreted as the total mass of the system in the region V . We shall use freely all these interpretations of the function f, which lead naturally to require that 1 3 3 0 ≤ f(t; ·; ·) 2 L (R × R ) and that kf(t)k1 = const = 1. The Boltzmann equation is a (integro-)partial differential equation for the one particle distribution function f. Let us \derive" it in the most simple case, that is for a system of particles moving with constant velocities. Suppose (for notational simplicity) that all particles are identical (otherwise we should introduce a particles distribution for each species). If x(t) denotes the position vector at time t and v the (constant) velocity vector of a particle, the position at time t + will be given by x(t + ) = x(t) + v : 1That is the manifold of all possible positions and velocities 2In non-relativistic mechanics, the momenum and velocity of a particle differ only by a multiplicative constant| the mass of the particle. 1 Now if instead of knowing the exact position of the particle at time t we only know the probability f(t; x(t); v) to find the particle there, it is natural to assume that f(t + , x(t + ); v) = f(t; x(t); v) ; since we know with certainty that the particle has moved with constant velocity v in the -interval of time. Provided f is a sufficiently smooth function, the above identity is equivalent to the free transport equation @tf + v · rxf = 0 : (1) The solution of (1) with initial data f(0; x; v) = fin(x; v) is given by f(t; x; v) = fin(x − vt; v). Note that fin ≥ 0 implies f(t; x; v) ≥ 0 and that kf(t)k1 = kf0k1. This is consistent with the interpretation of f as a probability density (i.e., of f(t; x; v)dvdx as a probability measure). Next we make the assumption that in the (infinitesimal) interval of time [t; t + ] the particle undergoes a collision with the other particles of the system. Here the word \collision" is used to refer to a general interaction which takes place in such a short time interval and small region of space that we can safely say that it occurs at time t in the position x. If the result of the collision is to increase the number of particles with velocity v, equation (1) modifies to @tf + v · rxf = G: Here G = G(t; x; v) ≥ 0 is called the gain term and gives the probability density that a new particle with velocity v is gained after the collision. Likewise, if a particle with velocity v is lost after the collision with probability L, we have @tf + v · rxf = −L: The function L = L(t; x; v) ≥ 0 is called loss term and the minus sign indicates that it makes f decrease (obviously, if a particle looses velocity v, the probability to have a particle with this velocity after the collision will be smaller). In general we have @tf + v · rxf = G − L: (2) Hence the Boltzmann equation is a balance equation: It says us how the particle distribution f changes as a consequence of the collision. In order to give an explicit form to G and L we have to give more information on how collisions occur. We shall consider only the case of the Boltzmann equation for binary elastic collisions. Exercise 1. Consider a system of particles moving under the influence of an external force field F = F (t; x). What should the equation for the one particle distribution function be in this case? Prove that the resulting equation (Vlasov equation) is consistent with the interpretation of the solution as a probability density (i.e., the solution is non-negative and its L1 norm is constant). 2 The Boltzmann equation for binary elastic collisions Binary elastic collisions means that we take into account only collisions between pair of particles (binary), i.e. the simultaneous collisions between three or more particles are assumed to occur with negligible probability. Moreover the collisions are assumed to be elastic, meaning that, besides the total momentum of the particles, also the total kinetic energy is conserved during the collision3. 3If the kinetic energy is not conserved, the collision is called inelastic. 2 0 0 Let v; v∗ denote the velocities before the collision of two particles and v ; v∗ their velocities after the collision (see Fig. 1). We also assume the particles to have the same mass (which is also conserved in the collision). Then the conservation laws of momentum and kinetic energy take the form 0 0 v + v∗ = v + v∗ (cons. of momentum) ; (3) 2 2 0 2 0 2 jvj + jv∗j = jv j + jv∗j (cons. of kinetic energy) : (4) Suppose the pre-collisional velocities (v; v∗) are given and we want to derive the post-collisional 0 0 velocities (v ; v∗) from (3) and (4). Since we have 6 unknowns but only 4 equations, the above system is underdetermined. However one can prove that the manifold of the solution is a 2-sphere. Precisely we have the following 0 0 Lemma 1. A quadruple (v; v∗; v ; v∗) solves (3)-(4) if and only if 0 v = v − [(v − v∗) · !]!; (5) 0 v∗ = v∗ + [(v − v∗) · !]!; (6) for some ! 2 S2. 0 0 Proof. The claim is true for the trivial solution (v = v, v∗ = v∗) (since it is obtained for ! 0 0 orthogonal to v − v∗). Otherwise, set v∗ = v∗ + a !, v = v + b !; from the conservation of 0 0 momentum, equation (3), we have a = −b. Substituting v∗ = v∗ + a ! and v = v − a ! in (4) and solving for a we obtained the desired result with 0 0 v∗ − v∗ v − v ! = 0 = − 0 : jv∗ − v∗j jv − vj 3 The triple (v; v∗;!) is called a collision configuration. The direction of ! is the scattering direction of the collision. Some important geometrical properties of binary elastic collisions are collected in the following lemma, whose proof is left as exercise. Lemma 2. The following holds: 0 0 0 0 (i) jv − v∗j = jv − v∗j; j(v − v∗) · !j = j(v − v∗) · !j; 0 0 0 0 0 0 (ii) v = v − [(v − v∗) · !]!; v∗ = v∗ + [(v − v∗) · !]!; 0 0 (iii) The Jacobian of the transformation (v; v∗) ! (v ; v∗) is equal to 1. Exercise 2. Prove Lemma 2. The meaning of (ii) is that a binary elastic collision is a reversible process (see Fig. 2). Recall from the Introduction that the gain term measures the probability that a new particle with velocity v results from the collision of two particles. By (ii) of Lemma 2, this is the case if the 0 0 particles collide with velocities v = v − [(v − v∗) · !]! and v∗ = v∗ + [(v − v∗) · !]! (see Fig. 2 0 0 again). The probability to find a particle with velocity v (resp. v∗) in the point x at time t is 0 0 given by f(t; x; v ) (resp. f(t; x; v∗)). The probability to have both particles at the same time in the same position (so that they may collide) is given by the product 0 0 f(t; x; v )f(t; x; v∗) provided we assume that the occurrence of two particles in the same point at the same time with given velocities are two independent events. This is called the molecular chaos assumption. Now if the two particles had always the same probability (say one) to collide independently from their velocities and scattering angle, then the product above would already give the probability density to obtain a new particle with velocity v from the collision of two particles with velocities 0 0 v and v∗.
Recommended publications
  • A Simple Method to Estimate Entropy and Free Energy of Atmospheric Gases from Their Action
    Article A Simple Method to Estimate Entropy and Free Energy of Atmospheric Gases from Their Action Ivan Kennedy 1,2,*, Harold Geering 2, Michael Rose 3 and Angus Crossan 2 1 Sydney Institute of Agriculture, University of Sydney, NSW 2006, Australia 2 QuickTest Technologies, PO Box 6285 North Ryde, NSW 2113, Australia; [email protected] (H.G.); [email protected] (A.C.) 3 NSW Department of Primary Industries, Wollongbar NSW 2447, Australia; [email protected] * Correspondence: [email protected]; Tel.: + 61-4-0794-9622 Received: 23 March 2019; Accepted: 26 April 2019; Published: 1 May 2019 Abstract: A convenient practical model for accurately estimating the total entropy (ΣSi) of atmospheric gases based on physical action is proposed. This realistic approach is fully consistent with statistical mechanics, but reinterprets its partition functions as measures of translational, rotational, and vibrational action or quantum states, to estimate the entropy. With all kinds of molecular action expressed as logarithmic functions, the total heat required for warming a chemical system from 0 K (ΣSiT) to a given temperature and pressure can be computed, yielding results identical with published experimental third law values of entropy. All thermodynamic properties of gases including entropy, enthalpy, Gibbs energy, and Helmholtz energy are directly estimated using simple algorithms based on simple molecular and physical properties, without resource to tables of standard values; both free energies are measures of quantum field states and of minimal statistical degeneracy, decreasing with temperature and declining density. We propose that this more realistic approach has heuristic value for thermodynamic computation of atmospheric profiles, based on steady state heat flows equilibrating with gravity.
    [Show full text]
  • LECTURES on MATHEMATICAL STATISTICAL MECHANICS S. Adams
    ISSN 0070-7414 Sgr´ıbhinn´ıInstiti´uid Ard-L´einnBhaile´ Atha´ Cliath Sraith. A. Uimh 30 Communications of the Dublin Institute for Advanced Studies Series A (Theoretical Physics), No. 30 LECTURES ON MATHEMATICAL STATISTICAL MECHANICS By S. Adams DUBLIN Institi´uid Ard-L´einnBhaile´ Atha´ Cliath Dublin Institute for Advanced Studies 2006 Contents 1 Introduction 1 2 Ergodic theory 2 2.1 Microscopic dynamics and time averages . .2 2.2 Boltzmann's heuristics and ergodic hypothesis . .8 2.3 Formal Response: Birkhoff and von Neumann ergodic theories9 2.4 Microcanonical measure . 13 3 Entropy 16 3.1 Probabilistic view on Boltzmann's entropy . 16 3.2 Shannon's entropy . 17 4 The Gibbs ensembles 20 4.1 The canonical Gibbs ensemble . 20 4.2 The Gibbs paradox . 26 4.3 The grandcanonical ensemble . 27 4.4 The "orthodicity problem" . 31 5 The Thermodynamic limit 33 5.1 Definition . 33 5.2 Thermodynamic function: Free energy . 37 5.3 Equivalence of ensembles . 42 6 Gibbs measures 44 6.1 Definition . 44 6.2 The one-dimensional Ising model . 47 6.3 Symmetry and symmetry breaking . 51 6.4 The Ising ferromagnet in two dimensions . 52 6.5 Extreme Gibbs measures . 57 6.6 Uniqueness . 58 6.7 Ergodicity . 60 7 A variational characterisation of Gibbs measures 62 8 Large deviations theory 68 8.1 Motivation . 68 8.2 Definition . 70 8.3 Some results for Gibbs measures . 72 i 9 Models 73 9.1 Lattice Gases . 74 9.2 Magnetic Models . 75 9.3 Curie-Weiss model . 77 9.4 Continuous Ising model .
    [Show full text]
  • Entropy and H Theorem the Mathematical Legacy of Ludwig Boltzmann
    ENTROPY AND H THEOREM THE MATHEMATICAL LEGACY OF LUDWIG BOLTZMANN Cambridge/Newton Institute, 15 November 2010 C´edric Villani University of Lyon & Institut Henri Poincar´e(Paris) Cutting-edge physics at the end of nineteenth century Long-time behavior of a (dilute) classical gas Take many (say 1020) small hard balls, bouncing against each other, in a box Let the gas evolve according to Newton’s equations Prediction by Maxwell and Boltzmann The distribution function is asymptotically Gaussian v 2 f(t, x, v) a exp | | as t ≃ − 2T → ∞ Based on four major conceptual advances 1865-1875 Major modelling advance: Boltzmann equation • Major mathematical advance: the statistical entropy • Major physical advance: macroscopic irreversibility • Major PDE advance: qualitative functional study • Let us review these advances = journey around centennial scientific problems ⇒ The Boltzmann equation Models rarefied gases (Maxwell 1865, Boltzmann 1872) f(t, x, v) : density of particles in (x, v) space at time t f(t, x, v) dxdv = fraction of mass in dxdv The Boltzmann equation (without boundaries) Unknown = time-dependent distribution f(t, x, v): ∂f 3 ∂f + v = Q(f,f) = ∂t i ∂x Xi=1 i ′ ′ B(v v∗,σ) f(t, x, v )f(t, x, v∗) f(t, x, v)f(t, x, v∗) dv∗ dσ R3 2 − − Z v∗ ZS h i The Boltzmann equation (without boundaries) Unknown = time-dependent distribution f(t, x, v): ∂f 3 ∂f + v = Q(f,f) = ∂t i ∂x Xi=1 i ′ ′ B(v v∗,σ) f(t, x, v )f(t, x, v∗) f(t, x, v)f(t, x, v∗) dv∗ dσ R3 2 − − Z v∗ ZS h i The Boltzmann equation (without boundaries) Unknown = time-dependent distribution
    [Show full text]
  • A Molecular Modeler's Guide to Statistical Mechanics
    A Molecular Modeler’s Guide to Statistical Mechanics Course notes for BIOE575 Daniel A. Beard Department of Bioengineering University of Washington Box 3552255 [email protected] (206) 685 9891 April 11, 2001 Contents 1 Basic Principles and the Microcanonical Ensemble 2 1.1 Classical Laws of Motion . 2 1.2 Ensembles and Thermodynamics . 3 1.2.1 An Ensembles of Particles . 3 1.2.2 Microscopic Thermodynamics . 4 1.2.3 Formalism for Classical Systems . 7 1.3 Example Problem: Classical Ideal Gas . 8 1.4 Example Problem: Quantum Ideal Gas . 10 2 Canonical Ensemble and Equipartition 15 2.1 The Canonical Distribution . 15 2.1.1 A Derivation . 15 2.1.2 Another Derivation . 16 2.1.3 One More Derivation . 17 2.2 More Thermodynamics . 19 2.3 Formalism for Classical Systems . 20 2.4 Equipartition . 20 2.5 Example Problem: Harmonic Oscillators and Blackbody Radiation . 21 2.5.1 Classical Oscillator . 22 2.5.2 Quantum Oscillator . 22 2.5.3 Blackbody Radiation . 23 2.6 Example Application: Poisson-Boltzmann Theory . 24 2.7 Brief Introduction to the Grand Canonical Ensemble . 25 3 Brownian Motion, Fokker-Planck Equations, and the Fluctuation-Dissipation Theo- rem 27 3.1 One-Dimensional Langevin Equation and Fluctuation- Dissipation Theorem . 27 3.2 Fokker-Planck Equation . 29 3.3 Brownian Motion of Several Particles . 30 3.4 Fluctuation-Dissipation and Brownian Dynamics . 32 1 Chapter 1 Basic Principles and the Microcanonical Ensemble The first part of this course will consist of an introduction to the basic principles of statistical mechanics (or statistical physics) which is the set of theoretical techniques used to understand microscopic systems and how microscopic behavior is reflected on the macroscopic scale.
    [Show full text]
  • Entropy: from the Boltzmann Equation to the Maxwell Boltzmann Distribution
    Entropy: From the Boltzmann equation to the Maxwell Boltzmann distribution A formula to relate entropy to probability Often it is a lot more useful to think about entropy in terms of the probability with which different states are occupied. Lets see if we can describe entropy as a function of the probability distribution between different states. N! WN particles = n1!n2!....nt! stirling (N e)N N N W = = N particles n1 n2 nt n1 n2 nt (n1 e) (n2 e) ....(nt e) n1 n2 ...nt with N pi = ni 1 W = N particles n1 n2 nt p1 p2 ...pt takeln t lnWN particles = "#ni ln pi i=1 divide N particles t lnW1particle = "# pi ln pi i=1 times k t k lnW1particle = "k# pi ln pi = S1particle i=1 and t t S = "N k p ln p = "R p ln p NA A # i # i i i=1 i=1 ! Think about how this equation behaves for a moment. If any one of the states has a probability of occurring equal to 1, then the ln pi of that state is 0 and the probability of all the other states has to be 0 (the sum over all probabilities has to be 1). So the entropy of such a system is 0. This is exactly what we would have expected. In our coin flipping case there was only one way to be all heads and the W of that configuration was 1. Also, and this is not so obvious, the maximal entropy will result if all states are equally populated (if you want to see a mathematical proof of this check out Ken Dill’s book on page 85).
    [Show full text]
  • Boltzmann Equation II: Binary Collisions
    Physics 127b: Statistical Mechanics Boltzmann Equation II: Binary Collisions Binary collisions in a classical gas Scattering out Scattering in v'1 v'2 v 1 v 2 R v 2 v 1 v'2 v'1 Center of V' Mass Frame V θ θ b sc sc R V V' Figure 1: Binary collisions in a gas: top—lab frame; bottom—centre of mass frame Binary collisions in a gas are very similar, except that the scattering is off another molecule. An individual scattering process is, of course, simplest to describe in the center of mass frame in terms of the relative E velocity V =Ev1 −Ev2. However the center of mass frame is different for different collisions, so we must keep track of the results in the lab frame, and this makes the calculation rather intricate. I will indicate the main ideas here, and refer you to Reif or Landau and Lifshitz for precise discussions. Lets first set things up in the lab frame. Again we consider the pair of scattering in and scattering out processes that are space-time inverses, and so have identical cross sections. We can write abstractly for the scattering out from velocity vE1 due to collisions with molecules with all velocities vE2, which will clearly be proportional to the number f (vE1) of molecules at vE1 (which we write as f1—sorry, not the same notation as in the previous sections where f1 denoted the deviation of f from the equilibrium distribution!) and the 3 3 number f2d v2 ≡ f(vE2)d v2 in each velocity volume element, and then we must integrate over all possible vE0 vE0 outgoing velocities 1 and 2 ZZZ df (vE ) 1 =− w(vE0 , vE0 ;Ev , vE )f f d3v d3v0 d3v0 .
    [Show full text]
  • Boltzmann Equation
    Boltzmann Equation ● Velocity distribution functions of particles ● Derivation of Boltzmann Equation Ludwig Eduard Boltzmann (February 20, 1844 - September 5, 1906), an Austrian physicist famous for the invention of statistical mechanics. Born in Vienna, Austria-Hungary, he committed suicide in 1906 by hanging himself while on holiday in Duino near Trieste in Italy. Distribution Function (probability density function) Random variable y is distributed with the probability density function f(y) if for any interval [a b] the probability of a<y<b is equal to b P=∫ f ydy a f(y) is always non-negative ∫ f ydy=1 Velocity space Axes u,v,w in velocity space v dv have the same directions as dv axes x,y,z in physical du dw u space. Each molecule can be v represented in velocity space by the point defined by its velocity vector v with components (u,v,w) w Velocity distribution function Consider a sample of gas that is homogeneous in physical space and contains N identical molecules. Velocity distribution function is defined by d N =Nf vd ud v d w (1) where dN is the number of molecules in the sample with velocity components (ui,vi,wi) such that u<ui<u+du, v<vi<v+dv, w<wi<w+dw dv = dudvdw is a volume element in the velocity space. Consequently, dN is the number of molecules in velocity space element dv. Functional statement if often omitted, so f(v) is designated as f Phase space distribution function Macroscopic properties of the flow are functions of position and time, so the distribution function depends on position and time as well as velocity.
    [Show full text]
  • Nonlinear Electrostatics. the Poisson-Boltzmann Equation
    Nonlinear Electrostatics. The Poisson-Boltzmann Equation C. G. Gray* and P. J. Stiles# *Department of Physics, University of Guelph, Guelph, ON N1G2W1, Canada ([email protected]) #Department of Molecular Sciences, Macquarie University, NSW 2109, Australia ([email protected]) The description of a conducting medium in thermal equilibrium, such as an electrolyte solution or a plasma, involves nonlinear electrostatics, a subject rarely discussed in the standard electricity and magnetism textbooks. We consider in detail the case of the electrostatic double layer formed by an electrolyte solution near a uniformly charged wall, and we use mean-field or Poisson-Boltzmann (PB) theory to calculate the mean electrostatic potential and the mean ion concentrations, as functions of distance from the wall. PB theory is developed from the Gibbs variational principle for thermal equilibrium of minimizing the system free energy. We clarify the key issue of which free energy (Helmholtz, Gibbs, grand, …) should be used in the Gibbs principle; this turns out to depend not only on the specified conditions in the bulk electrolyte solution (e.g., fixed volume or fixed pressure), but also on the specified surface conditions, such as fixed surface charge or fixed surface potential. Despite its nonlinearity the PB equation for the mean electrostatic potential can be solved analytically for planar or wall geometry, and we present analytic solutions for both a full electrolyte, and for an ionic solution which contains only counterions, i.e. ions of sign opposite to that of the wall charge. This latter case has some novel features. We also use the free energy to discuss the inter-wall forces which arise when the two parallel charged walls are sufficiently close to permit their double layers to overlap.
    [Show full text]
  • Lecture 7: Boltzmann Distribution & Thermodynamics of Mixing
    Prof. Tibbitt Lecture 7 Networks & Gels Lecture 7: Boltzmann distribution & Thermodynamics of mixing Prof. Mark W. Tibbitt { ETH Z¨urich { 14 M¨arz2019 1 Suggested reading • Molecular Driving Forces { Dill and Bromberg: Chapters 10, 15 2 Boltzmann distribution Before continuing with the thermodynamics of polymer solutions, we will briefly discuss the Boltzmann distri- bution that we have used occassionally already in the course and will continue to assume a working knowledge of for future derivations. We introduced statistical mechanics earlier in the course, showing how a probabilistic view of systems can predict macroscale observable behavior from the structures of the atoms and molecules that compose them. At the core, we model probability distributions and relate them to thermodynamic equilibria. We discussed macrostates, microstates, ensembles, ergodicity, and the microcanonical ensemble and used these to predict the driving forces of systems as they move toward equilibrium or maximum entropy. In the beginning section of the course, we discussed ideal behavior meaning that we assumed there was no energetic contribution from interactions between atoms and molecules in our systems. In general, we are interested in modeling the behavior of real (non-ideal) systems that accounts for energetic interactions between species of the system. Here, we build on the introduction of statistical mechanics, as presented in Lecture 2, to consider how we can develop mathematical tools that account for these energetic interactions in real systems. The central result is the Boltzmann distribution, which provides probability distributions based on the underlying energy levels of the system. In this approach, we now want to consider simple models (often lattice models) that allow us to count microstates but also assign an energy Ei for each microstate.
    [Show full text]
  • Statistical Thermodynamics (Mechanics) Pressure of Ideal Gas
    1/16 [simolant -I0 -N100] 2/16 Statistical thermodynamics (mechanics) co05 Pressure of ideal gas from kinetic theory I co05 Molecule = mass point L Macroscopic quantities are a consequence ¨¨* N molecules of mass m in a cube of edge length L HY HH of averaged behavior of many particles HH Velocity of molecule = ~ , , 6 H = ( , ,y ,z) HH HH ¨¨* After reflection from the wall: , , y ¨ → − ¨ Next time it hits the wall after τ = 2L/, Force = change in momentum per unit time - Momentum P~ = m~ Change of momentum = ΔP = 2m, Averaged force by impacts of one molecule: 2 ΔP 2m, m, F, = = = τ 2L/, L → Pressure = force of all N molecules divided by the area N N 2 F, 1 m, p =1 = = 2 = 3 P L P L Kinetic energy of one molecule: 1 2 1 2 1 2 2 2 m ~ m = m(, + ,y + ,z) 2 | | ≡ 2 2 3/16 4/16 Pressure of ideal gas from kinetic theory II co05 Boltzmann constant co05 Kinetic energy = internal energy (monoatomic gas) 1 N 3 N 2 2 pV = nRT = NkBT Ekin = m = m, 2 1 2 1 X= X= N = nNA ⇒ N 2 1 m, 2 Ekin p = = 3 = P L 3 V R k 1.380649 10 23 JK 1 Or B = = − − NA × 2 ! pV = Ekin = nRT 3 Note: Summary: since May 20, 2019 it is defined: Temperature is a measure of the kinetic energy ( 0th Law) k 1.380649 10 23 JK 1, B = − − ∼ × 23 1 Pressure = averaged impacts of molecules NA = 6.02214076 10 mol , × − We needed the classical mechanics therefore, exactly: 1 1 R = 8.31446261815324 J mol K Ludwig Eduard Boltzmann (1844–1906) Once more: − − credit: scienceworld.wolfram.com/biography/Boltzmann.html N R 3n 3N 3 n = , kB = U Ekin = RT = kBT, CV,m = R NA NA ⇒ ≡ 2 2
    [Show full text]
  • Deterministic Numerical Schemes for the Boltzmann Equation
    Deterministic Numerical Schemes for the Boltzmann Equation by Akil Narayan and Andreas Klöckner <{anaray,kloeckner}@dam.brown.edu> Abstract This article describes methods for the deterministic simulation of the collisional Boltzmann equation. It presumes that the transport and collision parts of the equation are to be simulated separately in the time domain. Time stepping schemes to achieve the splitting as well as numerical methods for each part of the operator are reviewed, with an emphasis on clearly exposing the challenges posed by the equation as well as their resolution by various schemes. 1 Introduction The Boltzmann equation is an equation of statistical mechanics describing the evolution of a rarefied gas. In a fluid in continuum mechanics, all particles in a spatial volume element are approximated as • having the same velocity. In a rarefied gas in statistical mechanics, there is enough space that particles in one spatial volume • element may have different velocities. The equation itself is a nonlinear integro-differential equation which decribes the evolution of the density of particles in a monatomic rarefied gas. Let the density function be f( x, v, t) . The quantity f( x, v, t)dx dv represents the number of particles in the phase-space volume element dx dv at time t. Both x and 3 3 + v are three-dimensional independent variables, so the density function is a map f: R R R R 6 × × → 0 R 0, . The theory of kinetics and statistical mechanics gives rise to the laws which the density func- ∪ ∞ tion {f must} obey in the absence of external forces.
    [Show full text]
  • Boltzmann Equation 1
    Contents 5 The Boltzmann Equation 1 5.1 References . 1 5.2 Equilibrium, Nonequilibrium and Local Equilibrium . 2 5.3 Boltzmann Transport Theory . 4 5.3.1 Derivation of the Boltzmann equation . 4 5.3.2 Collisionless Boltzmann equation . 5 5.3.3 Collisional invariants . 7 5.3.4 Scattering processes . 7 5.3.5 Detailed balance . 9 5.3.6 Kinematics and cross section . 10 5.3.7 H-theorem . 11 5.4 Weakly Inhomogeneous Gas . 13 5.5 Relaxation Time Approximation . 15 5.5.1 Approximation of collision integral . 15 5.5.2 Computation of the scattering time . 15 5.5.3 Thermal conductivity . 16 5.5.4 Viscosity . 18 5.5.5 Oscillating external force . 20 5.5.6 Quick and Dirty Treatment of Transport . 21 5.5.7 Thermal diffusivity, kinematic viscosity, and Prandtl number . 22 5.6 Diffusion and the Lorentz model . 23 i ii CONTENTS 5.6.1 Failure of the relaxation time approximation . 23 5.6.2 Modified Boltzmann equation and its solution . 24 5.7 Linearized Boltzmann Equation . 26 5.7.1 Linearizing the collision integral . 26 5.7.2 Linear algebraic properties of L^ ............................. 27 5.7.3 Steady state solution to the linearized Boltzmann equation . 28 5.7.4 Variational approach . 29 5.8 The Equations of Hydrodynamics . 32 5.9 Nonequilibrium Quantum Transport . 33 5.9.1 Boltzmann equation for quantum systems . 33 5.9.2 The Heat Equation . 37 5.9.3 Calculation of Transport Coefficients . 38 5.9.4 Onsager Relations . 39 5.10 Appendix : Boltzmann Equation and Collisional Invariants .
    [Show full text]