2 - INFN

Transcription

2 - INFN
Solving the collisionless Boltzmann equation
using N-body simulations
• Gravitational dynamics
• N-body simulations
The distribution of galaxies in the
local Universe. Distances
(redshifts) are measured using
Hubbles velocity-distance relation.
The galaxies are not distributed
randomly but are clustered along
filaments that connect together
the groups and clusters.
This clustering pattern is repeated
on larger scales and can be
quantified using correlation
functions, topology, fractal
dimensions, power spectrum
analysis….
cfa survey (1980)
1000 galaxies
1 00
Mp
c
2df survey (2001)
100,000 galaxies
0
0
10
M
pc
Globular clusters
Omega-Cen
A cluster of 10^4 – 10^6 stars roughly 1-10
parsecs across.
Galaxies are surrounded by a swarm of
globular clusters – the brighter the galaxy
then the more globular clusters.
Spectroscopy of the stars reveals that they
are moving at a few km/s.
d2 r
GM
acceleration =
=− 2
2
dt
r
Integrating this equation with the boundary condition r=R at t=0:
Tfreefall
1
1 R3
−
=
∝ (Gρ ) 2
2 GM
= 1/ 2 (10 x3.1e16)3 /(6.67e − 11x1.0e6 x 2.0e30)
= 0.0003Gyrs
ρstar = 250 stars/pc3
ρ proton = 104 cm −3
Tcollision = 1/ (π nσ v ) = 1015 yrs >> age of the universe
Globular clusters are fascinating objects for many reasons:
•Short dynamical time – undergo core collapse, perhaps sites where black holes form.
•Formation before the main galaxy – they tell us something about the initial conditions.
•They can be used as “test particles” to weigh the galactic systems.
Many body gravitating systems
Systems with N>2 gravitating objects (stars, protons etc) can’t be
solved analytically.
Two ways to proceed:
Statistical treatment of dynamics:
-determine the gravitational potential
-estimate/observe shape of orbits
-solve for the past and future evolution
Numerical simulations:
-direct summation of gravitational forces
-integrate equations of motion
-update positions, velocities
-repeat process on timestep dt
Equation of motion:
Many body systems
d 2x
m 2 = f ( x),
dt
Phase space:
dx
dv
= v, m = f ( x )
dt
dt
The flow of particles through phase
space is quite intricate. Within a small
volume dxdydz the phase space
3
3
density f ( x, y, z , vx , v y , vz , t )d x, d v
is conserved.
The evolution of the entire system can
be followed using the collisionless
Boltzmann equation and Poissons eq.
It’s a 6 dimensional equation that is
difficult to solve. Gravitating systems
do not have an equation of state that
relates pressure and density.
The origin of structure
in the Universe
The origin of cosmic
structures is are complex
non-linear problems best
solved by computational
methods
1
T
10
n
dy
1
10
^
0
10
^
0
N
10
Galaxy formation, the stability of
the solar system, the core
collapse of a globular cluster into
a black hole all require about
10^18 floating point operations.
The distribution function
Within a smooth potential the number of stars with positions in
volume d3x centered on x and velocities d3v centred on v is:
f ( x, v, t )d 3 xd 3v
is the distribution function (phase space density)
Given f ( x, v, t ) we can predict f ( x, v, to ) at any time in the future or past
Assume that f is a smooth continuous function such that particles move
smoothly through phase space. i.e. N >> 106
The coordinates in phase space:
w ≡ ( x, v) ≡ ( x, y, z , vx , v y , vz ) ≡ ( w1 , w2 , w3 , w4 , w5 , w6 )
The velocity of the flow
w& ≡ ( x& , v&) = (v, −∇Φ )
Therefore f ( w, t ) must satisfy a continuity equation analogous
to ρ (x,t) of an ordinary fluid. i.e. stars are conserved in the flow
and drift slowly through phase space.
In the most general definition, a Boltzmann Equation (BE) is a differential
equation for the evolution of a probability density during time furnished
with an initial condition. This means that there are BEs for problems with
discrete and continuous states allowing discrete and continuous time
evolution.
The Boltzmann equation appears in many areas of science:
Kinetic treatment of gas
Material science
Cell Biology
Fluid dynamics
Electron transport in nuclear reactors
Radiative transport in planetary atmospheres
Astrophysics
The fluid analogy
Rate of change of mass with time:
dm
∂ρ 3
= ∫ d x = − ∫ ρ v ⋅d 2 S
dt V ∂t
S
volume V
S
(v, ρ )
Use the divergence theorem:
⎛ ∂ρ
⎞ 3
∫V ∇ ⋅ fd x = ∫S fd S → V∫ ⎜⎝ ∂t + ∇ ⋅ ( ρ v) ⎟⎠d x = 0
This must hold for any volume V therefore:
∂ρ
(Continuity equation)
+ ∇ ⋅ ( ρ v) = 0
∂t
Now consider the analogous stellar phase space distribution function
3
2
6
∂ ( fw& α )
∂f
+∑
=0
∂t α =1 ∂wα
6
∂ ( fw& α )
∂f
+∑
=0
∂t α =1 ∂wα
Simplify this equation:
volume V
ρ
3
⎛ ∂vi ∂v&i ⎞
∂w& α
(
v
,
)
=
+
⎜
⎟
S
∑
∑
∂vi ⎠
α =1 ∂wα
i =1 ⎝ ∂xi
∂vi
= 0 since x and v are independent coordinates of phase space
but
∂xi
6
and since ∇Φ does not depend on velocity we have:
⎛ ∂Φ ⎞
⎜
⎟=0
⎝ ∂xi ⎠
Now the continuity equation becomes the CBE:
∂w& α 3
∂
=∑ −
∑
∂vi
α =1 ∂wα
i =1
6
6
3
⎡ ∂f ∂Φ ∂f ⎤
∂f
∂f
∂f
+ ∑ w& α
= 0 or
+ ∑ ⎢vi
−
⎥=0
∂t α =1 ∂wα
∂t i =1 ⎣ ∂xi ∂xi ∂vi ⎦
or
3
⎡ ∂f ∂Φ ∂f ⎤
∂f
+ ∑ ⎢vi
−
⎥=0
∂t i =1 ⎣ ∂xi ∂xi ∂vi ⎦
in vector notion:
∂f
∂f
+ v ⋅∇f − ∇Φ ⋅ = 0
∂t
∂v
This is the fundamental equation of stellar dynamics.
What does it mean?
Transform to Lagrangian coordinates that travel along with a star.
x = xo = const.
d
∂
= + w& ⋅∇ 6 →
dt ∂t
6
df ∂f
∂f
=
+ ∑ w& α
=0
dt ∂t α =1 ∂wα
The fluid analogy
volume V
S
(v, ρ )
As in the fluid case we can integrate the CBE to see that the
first term is the rate of change of stars in V and the second term is
the rate of change of outflow/inflow of stars from/into V .
df
is the rate of change of phase space density seen by an observer
dt travelling along with a star
df
= 0 → that the flow of stellar points through phase space is
dt
incompressible. The space density near the star is constant.
Example of incompressible flow: idealised marathon race where all the
runners run at a constant speed. At the start of the race they travel at a
wide variety of speeds. At the finish the density is low but the runners
going past have nearly the same speed.
∂f
In the absence of collisions then
= 0 i.e. phase space is conserved
∂t
along a given trajectory through 6-d space. This also implies that any maximum
of the distribution always remains the same. If the force is conservative, as
the gravitational force, then the 6 dimensional flow is incompressible, which
means that the phase space volume of a contour f ′ is always conserved.
The N-body technique
The collisionless Boltzmann equation is the fundamental
equation of a collisionless system
∂f
∂f
∂f
+ v ⋅ − ∇Φ ⋅ = 0
∂t
∂r
∂v
Where f = ( x, y, z, vx , v y , vz ) is the phase space density.
This can be combined with Poissons equation to completely
describe a gravitating system
∇ 2 Φ = 4π G ρ
Where ρ = ∫ fd 3v
Solving these equations using finite
difference techniques is impractical
therefore solve the equations of motions of
N particles directly. The N particles are a
Monte-Carlo realisation of the true initial
conditions.
dr
dv
= v,
= −∇Φ
dt
dt
Gm j
Φ = −∑
2
i≠ j
rj − ri + ε 2
(
)
1/ 2
To represent the mass distribution function f(r,v,t) one uses a set of N
bodies each possessing mass, position and velocity – i.e. one replaces the
continuous distribution function with a set of delta functions.
N
f (r , v) → ∑ miδ 3 (r − ri )δ 3 (v − vi )
i =1
For this substitution to work the expected mass of the bodies within any
phase space volume V must be equal to the integral of the distribution
function over that volume:
3
3
d
rd
vf (r , v) =
∫
V
∑
( ri ,vi )∈V
mi
Where the angle brackets indicate an average of statistically equivalent
realizations and the sum includes all bodies with phase space coordinates
within the volume V
The potential is calculated using the particle distribution so that
Poisson’s eq. looks like:
N
∇ Φ = 4π G ∑ miδ 3 (r − ri )
2
i =1
It is a standard trick to use orders of magnitude fewer particles than a
real system may contain.
i.e. air turbulence has a length scale of about 10cm. A cube of this size
would contain about 10^22 particles. The largest volume of air that could
be simulated exactly is 0.01mm^3.
Weather prediction may wish to resolve this 10cm length scale on the scale
of the earth which would require a grid of order 10^18 or 10^20 particles.
To make these problems manageable then use a macroscopic fluid
treatment where small scale turbulence is parameterised in viscosity and
diffusion terms.
For some cosmology problems we can use one particle per star – i.e. a
globular cluster with 10^6 stars can be followed exactly.
However other problems, such as the 10^70 dark matter particles in a
galactic halo can’t be treated exactly and we use “super massive” particles
to represent the phase flow – as long as particle-particle collisional effects
do not become important.
Constructing Initial Conditions
Before we discuss the numerical techniques of how one evolves N particles, first
we need to set up some initial conditions – i.e. a Monte Carlo representation of a
mass distribution using a set of points.
This is non-trivial. For example, assume we want to evolve a Globular Cluster of
stars to see if the central region will collapse to a black hole. We need to initialise
positions and velocities of particles such that our model globular cluster is in
equilibrium.
In other words, we have to assign velocities to particles as a function of radius
such that the flow of particles through any small volume is conserved. This ensures
that the model is in equilibrium and satisfies the collisionless Boltzmann equation.
The CBE can only be solved exactly for a few density-potential pairs that are
usually spherical systems with isotropic velocity dispersion profiles.
King model, Plummer sphere, Hernquist halo, Evans disks…
(The infinite isothermal sphere is also simple but needs to be an infinite model to
be stable.)
Usually we are trying to model a system with an observed density profile
therefore we wish to solve for the velocity distribution of particles.
Assigning positions to particles to create a random realisation of a 3d
density is reasonably easy. (Use accept/reject method or faster to draw
radii from the cumulative mass distribution.)
Assigning velocities is the tricky part. Use Poissons equation to relate the
observed density profile to the integral over the phase space density.
∇ 2 Φ = 4π G ρ = 4π G ∫ fd 3v
Jean’s theorem states that for a spherically symmetric system the
distribution function f is a function of energy alone. Thus:
∇ 2Φ =
1 d ⎛ 2 dΦ ⎞
⎜r
⎟ = 4π G ∫
2
r dr ⎝ dr ⎠
⎛1
⎞
f ⎜ v 2 + Φ ⎟d 3v
⎝2
⎠
Assume some form for the distribution function and solve Poisson’s
equation for the density profile – this will be a solution to the CBE.
Alternatively, can solve this equation directly for the energy dist.
Approximate the distribution function for all components using moments of
the CBE. For a spherical halo the 2nd moment is:
d ( ρ h vr2 ) ρ h ⎡ 2
dΦ
+
2vr − (vθ2 + vΦ2 ) ⎤ = − ρ h
⎦
dr
r ⎣
dr
If the halo is isotropic (random velocity vectors) then we can integrate:
∞
1
dΦ
v =
dr
ρ h (r )
∫
dr
ρ h (r ) r
2
r
This gives the velocity dispersion as a function of radius. We can now
draw velocities randomly from a function whose 2nd moment is v 2
r
i.e. a Gaussian:
⎛ 1 ⎞
F (v, r ) = 4π ⎜
2 ⎟
⎝ 2πσ ⎠
3/ 2
σ 2 exp(−σ 2 / 2vr2 )
This will have incorrect higher order moments!
More complex models - galaxies
ρ halo (r ) =
ρ bulge
M
2π
3
2
2
)
α exp(−r / rcore
rcore (r 2 + β 2 )
M
1
,
=
3
2
2π ac [t (1 + t ) ]
ρ disk ( R, z ) =
x2 + y 2 z 2
t =
+ 2
2
a
c
2
M
2
exp[
−
R
/
h
]
sech
( z / zo )
4π h 2 zo2
Approximate all the components using moments of the CBE, so that the
potential is the sum of the disk, bulge and halo components….
∞
1
dΦ
v =
dr
ρ h (r )
∫
dr
ρ h (r ) r
2
r
Nice example is to study the effects of the stellar bar
on the central dark matter distribution
Simulating N-Body systems
Simulating N-Body systems
N-Body codes calculate the gravitational
force between particles and update the
positions and velocities on a sequence of
discrete timesteps.
Simulating N-Body systems
Here we are simply
following the
Newtonian gravitational
force between N stars
as they orbit within
their collective
potential.
The hard part is doing
this quickly and
accurately for many
particles.
In 3d each two body interaction requires 10 flops.
Total number of operations for a small simulation:
T force = 10.N timesteps N particles ( N particles −1 )τ
where τ is the wallclock time per flop
N timesteps ≈ 1000
N particles ≈ 106
τ CrayT 3 E ≈ 10−11 sec
⇒ T force ≈ 14, 000hours=2 processor years
Galaxy formation, the stability of the solar system, the core collapse of a globular
cluster into a black hole all require about 10^18 floating point operations.
Hardware and software gains
8
Ghigna
7
6
Log(N)
Carlberg, Dubinski,
Warren etal
5
4
White
3
White
2
Holmberg
Moore’s law
Aarseth, van
Albada, Peebles
1
0
1940
1950
1960
1970
1980
1990
2000
2010
Year
Consider the increase in the resolution of collisionless simulations of clusters
The world’s highest resolution N-body simulation?
Andreas Adelmann, PSI (Paul Scherrer Institute)
1 particle per proton = 10^11 particle simulation
Ways to speed up the force calculation
Direct summation
Start with initial positions and velocities
Compute gravitational potential from particle distribution
Compute accelerations from potential gradients
Update positions and velocities for one step.
Goto step 1
Robust, accurate but expensive – scales as n^2
Can use special purpose hardware that does 1/r^2 on hardware PCI board
Field Expansion (PM)
Poissons equation can be expressed in terms of Fourier components.
Compute density at a point (interpolate onto a grid)
Compute potential on the grid by FFT
Compute accelerations by finite differencing the potential
Update particle positions and velocities
Goto step 1
Scales as 0(n)
Can use P3M where P-P forces are calculated within a cell
Ways to speed up the force calculation
Hierarchical methods (TREES)
The higher order multipoles of an objects gravitational field decays rapidly
with respect to the dominant monopole term.
The long range gravitational potential of a region can be approximated as
1/r therefore replace the sum over N-1 bodies by a sum over (0)LogN
regions.
Create a TREE data structure by recursively dividing the computational
volume.
OCT-Trees:
KD-Tree:
Split space into successively smaller octants by volume.
Each node has 8 children or branches.
Divide by one dimension at a time. (Fail for gravity since
one ends up with long cells that can’t be approximated with
even hexadecipole expansion.
Trees – building and walking
OCT-Trees:
Construct Tree from top down in time proportion to Nln(N)
Building
Root node is the entire volume of the simulation
Split volume in 8
Loop over all particles and sum mass in each quadrant
If octant has > 1 particles, create new node
If octant has =1 particle, create pointer
If octant has =0 particles, create null pointer
Repeat until no active quadrants with >1 particle
For a given particle, walk down the tree to determine the interaction list
Open cells according to theta criteria
Sum up the accelerations and multipoles
Walking
Each node of the Tree has:
Pointers to 8 children in 3D (node,
particle or null)
Pointer to the parent node
Position of the centre of mass and
physical centre
Mass
Higher order multipoles of the node
The active tree on a single particle
Hierarchical methods (TREES)
The force on a given particle is
calculated by moving up the tree,
opening cells according to a simple
opening angle criteria.
r
θ
d
If r/d<0.5 then open the cell and sum
the contributions to the force from
the particles within the subcells,
otherwise do not open the cell.
This is the entire
two dimensional tree
for a system
modelled with 1000
particles.
Next we will look at
the “active” tree for
a single particle.
Grids
Assign particles to the grid to determine the
density at grid points via interpolation
Compute gravitational potential at grid points by
solving Poisson’s equation on the grid (FFT)
Compute gravitational accelerations at grid points
by finite differencing the potential
Calculate accelerations at grid points using the
same interpolation method as in step 1.
Update positions and velocities using leap-frog
integrator.
Invented in Los Alamos in 1950’s – now used for:
Plazmas, Galactic dynamic, fluids, molecular
dynamics, weather…….
Interpolation techniques:
Nearest grid point (NGP)
Simple
Quick
Density discontinuous across the grid
Cloud in cells (CIC)
Most popular
Each particle is a cube of size the grid spacing
Most particles intersect 2^d grid cells
First derivative of the density is discontinuous,
but density is smooth across grid cells
Triangular clouds (TC)
Rarely used
Expensive
Each particle is a cube of (tri) linearly rising density
Density and derivative of density are continuous
Spherical Profiles (SP)
Each particle has some linearly rising density profile
Useful for P3M since automatically have a smoothing radius.
Determining the potential
N
Gα = ∑ Gαβ M β
β =1
M β is the mass in the β cell
Gαβ is the potential at the center of the α cell generated by unit
mass at the β cell
−1
Gαβ ∝
2
ε 2 + xα − xβ
Gαβ is the "Greens" function for the grid that needs to be calculated only once at t = 0
Then Φ ( xα ) can be estimated at each step using FFT's
In Fourier space Φ (k ) = G (k ) ρ (k )
(This is a discrete Fourier transform convolution)
Φ (α ,β ,γ )= ∑ G i,j,k ρ i , j ,k e 2π i(iα +jβ +kγ )/M
i,j,k
Perform FFT to get ρ (k)
Multiply each array element by G ( K ) to get Φ (k )
Reverse FFT to transform result Φ (k ) → Φ ( x ) in real space
FFT's can be executed in time ∝ N cell Ln(N cell ) steps
FFT's are periodic which is good for cosmological volumes
Determining the potential
Gα =
N
Gαβ M β
∑
β
=1
M β is the mass in the β cell
Gαβ is the potential at the center of the α cell
generated by unit mass at the β cell
−1
Gαβ ∝
2
ε 2 + xα − x β
Gαβ is the "Greens" function for the grid that needs to be calculated only once at t = 0
Then Φ ( xα ) can be estimated at each step using FFT's
In Fourier space Φ ( k ) = G ( k ) ρ ( k )
(This is a discrete Fourier transform convolution)
Φ ( α , β , γ )= ∑ G i,j,k ρ i , j , k e 2 π i(iα +jβ +k γ )/M
i,j,k
Perform FFT to get ρ (k)
M ultiply each array element by G ( K ) to get Φ ( k )
Reverse FFT to transform result Φ ( k ) → Φ ( x ) in real space
FFT's can be executed in time ∝ N cell Ln(N cell ) steps
FFT's are per iodic which is good for cosmological volumes
The Greens function
This plot shows the force errors
that occur when single particles
are placed randomly in the
computational box at a random
distance from the origin.
Error on scales <3 cells due to geometry and discretisation.
Error on box scale due to periodic boundaries
N
Gα = ∑ Gαβ M β
β =1
M β is the mass in the β cell
Gαβ is the potential at the center of the α cell
generated by unit mass at the β cell
Gαβ ∝
−1
ε 2 + xα − xβ
2
Gαβ is the "Greens" function for the grid that needs to be calculated only once at t = 0
Then Φ ( xα ) can be estimated at each step using FFT's
In Fourier space Φ ( k ) = G ( k ) ρ ( k )
(This is a discrete Fourier transform convolution)
Φ (α ,β ,γ )= ∑ G i,j,k ρ i , j , k e 2π i(iα +jβ +kγ )/M
i,j,k
Perform FFT to get ρ (k)
Multiply each array element by G ( K ) to get Φ (k )
Reverse FFT to transform result Φ ( k ) → Φ ( x ) in real space
FFT's can be executed in time ∝ N cell Ln(N cell ) steps
FFT's are periodic which is good for cosmological volumes
Determining the acceleration
Usually calculated by finite differencing the grid of potential energies:
acceleration ∝ ∇Φ i.e.
x
a i,j,k
= −(Φ i +1, j ,k − Φ i −1, j ,k ) / 2, similar for a y and a z
This is a simple 2 point scheme.
Finally interpolate the accelerations back onto the locations of particles.
Important to use the same interpolation scheme used to place ρ on the grid.
Last step is to update ( x, v) using leap frog.
Resolution poor within 3 grid cells
Memory limits calculation < 1000 grid cells
Can use P3M code that evalulates P-P forces directly in cells that contain
many particles
AP3M codes use grids within grids
Simulating N-Body systems
Several numerical
parameters are
important during
these simulations.
•Force softening
•Discreteness noise
•Force
approximations
•Time-stepping
•Integration errors
•Machine round-off
errors
Gravitational force softening
dri
= vi
dt
and
m j (rj − ri )
dvi
=∑
2
2 3/ 2
dt
j ≠i ( r − r
+ε )
j
i
The particles represent mass points moving in phase space.
The force becomes infinite at zero separation which does
not occur in real systems.
Forces are softened to prevent scattering (collisional
processes) during close encounters. This is particularly
important when we are trying to follow collisionless systems.
(In some situations, such as to follow the core collapse and
black hole formation, star-star encounters are important and
need to be followed accurately.)
Timestepping
The orbits are calculated using discrete timesteps. One must
choose these shorter than the local dynamical time. A large
speedup can be obtained using variable timesteps – each
particle moves on a locally determined timestep.
However this destroys the nice symplectic nature of the
leapfrog integrator.
Variable timesteps can be based on various local quantities:
∂t (v, ρ , −∇Φ, ε )
Consider solar system: Mercury orbits every 88 days but Pluto
orbits every 250 years. For Keplerian systems can save
factors of 1000 by scaling timesteps with velocity.
Multistepping increases the complexity of the parallel domain
decomposition – end up with few active particles.
Timestepping and integrators
Recall that an 2nd order differential equation can be reduced to a set of 1st
order ODE’s
Equation of motion:
d 2x
m 2 = f ( x),
dt
Phase space:
dx
dv
= v, m = f ( x )
dt
dt
The Euler method
Split orbit into a series of timesteps ∂t
xi +1 = xi + ∂t. f ( xi , ti )
Taylor expansion of f ( x) → error ∝ O∂t 2
i+1
i
dt
Timestepping and integrators
Runga-Kutta
∂x1 = ∂t. f ( xi , ti )
∂x2 = ∂t. f ( xi + ∂xi / 2, ti + ∂ti / 2)
∂x3 = ∂t. f ( xi + ∂x2 / 2, ti + ∂t2 / 2)
∂x4 = ∂t. f ( xi + ∂x3 , ti + ∂t3 )
xi +1 = xi + ∂x1 / 6 + ∂x2 / 3 + ∂x3 / 3 + ∂x4 / 6 + Ο∂t 5
Leapfrog
Offset positions and velocities by ∂t / 2
xi +1 = xi + vi +1/ 2 ∂t
vi +3/ 2 = vi +1/ 2 + f ( xi +1 )∂t
Error of the order O∂t 2
But it is time reversable → can replace ∂t=-∂t
Time reversibility is important since it guarantees conservation of energy,
angular momentum – any conserved quantity.
Consider again the problem of a simple elliptical solution of the
gravitational two-body problem. Imagine that our integrator makes an
energy error of +E as it integrates the system forward through 1 orbital
period. Now imagine reversing the integration. You might guess that the
energy error in the reverse integration would be -E , but this is not the
case. In any system where the equations of motion are unchanged by time
reversal (and, specifically, in any case where the function f depends only on
the coordinates), the time-reversed orbit is itself a solution of the original
ODE (with v simply replaced by -v), so the energy error is still +E. But if
our integration scheme is time-reversible, we know that the final energy
error is zero (because we return to our starting point). The only possible
way that this can occur is if E=0 (i.e. energy is exactly conserved!).
As illustrated below, even the small error (exaggerated here by choosing a
timestep 5 times greater than in the other two integrators) made by
Runge-Kutta-4 is systematic, leading to a long-term drift in the orbital
parameters, the energy error in the Leapfrog scheme has no such longterm trend. There is a periodic error over the course of an orbit, at the
same level as the error in the Mid-point scheme, but the errors incurred
over the outgoing portion of the orbit exactly cancel those produced on
the incoming segment, so no net error results. The Leapfrog method is only
second-order accurate, but it is very stable.
Force approximations
The direct and most exact approach calculates forces between all N
particles. The calculation scales as N^2. The largest N^2 calculations
approach 10^5 particles.
Special purpose hardware (pci “grape boards”) can be used to do 1/r^2
calculations in hardware.
Treecodes, mesh codes or particle-mesh codes all attempt to speed up
the force calculations by approximating long range forces.
This introduces additional errors – but they can be quantified and
controlled. These codes scale as order Nl.n(N) or even as order N.
Discreteness effects
In many cases we can’t use one particle per real star. For example, a dark
matter halo surrounding a galaxy may contain 10^70 particles! We can fit
at most 10^10 particles in the memory of a parallel supercomputer.
Discreteness leads to collisional processes and sets a limit to the phase
space density that can be resolved.
One can esimate the “relaxation time” of a system of N particles
analytically or by following the actual energy changes of individual
particles.
What we really need to simulate the Universe, is the processing
power of a single human brain.
•Each cell is an elaborate chemical computer with power management,
read only and random access memory.
•Each cell has a backup of the genetic code of the entire body.
•Cell reproduction alone requires terabytes/second of DNA copying.
•The 10^10 neurons in the celebral cortex can each exhange
information via several mechanisms leading to a processing power of
about 10^14 bits/second.
•The power consumption of a human is only about 30 Watts.
Let’s build a supercomputer!
The zBox can process 10^12
operations per second and
transfer information to the
server at 10^9 bits/second.
This is about 1% of a human
brain. But it also consumes and
dissipates 1000 times as much
power….
zBox: (Stadel & Moore)
288 AMD MP2200+ processors, 144 Gigs ram
Compact, easy to cool and maintain
Very fast Dolphin/SCI interconnects - 4 Gbit/s, microsecond latency
A teraflop supercomputer for $500,000
Roughly one cubic meter, one ton and requires 40 kilowatts of power
The beginning…
The horizon project
A 15 Teraflop
2000 processor
Petabyte disk store
3d SCI network
Unizh/PSI/CSCS
The largest supercomputer in Europe
Astrophysics – computation biology – computational chemistry
Need to make the codes scale in parallel to thousands of processors
The power spectrum of density fluctuations in three different dark matter models
CMB
Large scales
(galaxy clusters)
Small scales
(dwarf galaxies)
How does dark matter change P(k)?
Most models for the generation of fluctuations predict a scale free
spectrum of perturbations. i.e. P(k) proportional to k.
This is imprinted on the mass distribution but the linear growth of
fluctuations alters the primordial power spectrum through various
physical “filtering” processes.
The final processed power spectrum is curved due to two main
phenomenon: horizon crossing and free streaming.
(Also see wiggles on large scales due to the coupled baryon-photon
plasma between matter-radiation equality and recombination.)
Free streaming
When the particles “freeze out” from equilibrium with the radiation then they
may be moving relativistically. If the particles were within some density
perturbation as it came through the horizon then they would move out of the
fluctuation in some random direction.
This effectively damps out fluctuations below some mass scale.
The comoving distance which a free-streaming particle can travel by time t:
t
v(t ')
dt '
R(t )
0
rFS = ∫
Split this into terms that come from relativistic era and to t equality
t EQ
t NR
v(t ')
v(t ')
dt ' + rFS = ∫
rFS = ∫
dt '
R (t )
R(t )
t NR
0
The first term is the horizon scale over the time the particle is relativistic and
the second term is the period over which the streaming velocities are damped
due to the expansion of the Universe (vpeculiar ∝ R −1 )
m
⎞
rFS ≈ 28Mpc ⎛⎜ ν
⎟
⎝ 30eV ⎠
−1
−2
m FS
m
⎞ M
≈ 10 ⎛⎜ ν
⎟
⎝ 30eV ⎠
15
Creating cosmological initial conditions
Need to model the correct power spectrum of the linear density field
Start with an initial 3-d distribution of particles that have zero power on
all scales: GRID or a GLASS
Perturb particles from there initial conditions so that they have a
clustered distribution that represents the P(k).
Use non-linear perturbation theory in comoving (Lagrangian) coordinates
Zeldovich approximation:
x = R(t )r + b(t ) p (r )
x = proper coordinate r = comoving coordinate
The first term is the uniform expansion of the background model
The second term represents the perturbation of a particles position about
its Lagrangian or comoving coordinate r
b(t ) is the growth rate of the displacements of particles from their initial
displacements p (r )
Goals:
Model the development of large scale structure
Predict the structure of dark matter halos surrounding
galaxies and galaxy clusters. (Sizes, masses, shapes,
density profiles, substructure)
Make predictions for direct and indirect detection
experiments that are sensitive to the phase space
structure of dark matter halos on parsec scales.
Constrain the nature of particle dark matter
candidates.
N=700 (White 1976)
This overmerging problem
was part of the motivation
for White & Rees (1978)
and the beginning of semianalytic galaxy formation
models.
The formation of large scale structure in the universe
The colours show
the local density
of dark matter
plotted using
comoving
coordinates that
expand with the
Universe.
This volume is a
substantial
fraction of the
observable
universe.
We are able to
reproduce the
same clustering
pattern as
observed in the
galaxy
distribution.
100Mpc
A theoretical prediction for the mass distribution in the Universe on very large scales
Large scale distribution of matter does not test
the real nature of the dark matter particle.
(Apart from hot dark matter, the DM particle
may have a diverse range of properties that
would not affect the development of large scale
structure, but may change the structure of nonlinear objects. i.e. WDM, a particle that
interacts weakly with photons, a particle that
interacts with some strong interaction with
itself, etc….)
Want to look at individual objects at a resolution
probed by the stars – i.e parsec-kpc scales
within a 100Mpc volume.
Lets look at the
formation of a single
cluster of galaxies.
This movie shows
the density of dark
matter in physical
coordinates so that
you can see the
expansion of the
universe.
This is worlds
highest resolution
calculation of
structure formation
that took several
months running in
parallel on hundreds
of processors.
Structure
formation in
“comoving
coordinates”
that expand
with the
universe.
The power spectrum of density fluctuations in three different dark matter models
CMB
Large scales
(galaxy clusters)
Small scales
(dwarf galaxies)
Next class will confront these and other dark matter candidates with observations
z=3
z=0
HDM
WDM
CDM