Thermal lensing in Virgo and Polynomial Search: an all-sky

Transcription

Thermal lensing in Virgo and Polynomial Search: an all-sky
Thermal lensing in Virgo and Polynomial
Search: an all-sky search for gravitational
waves from spinning neutron stars in
binary systems
Sipho van der Putten
Front and back cover designed by Chiara Farinelli and Hegoi Garitaonandia. Back cover
based on ‘Fiets 1’ (1988) by Mark Raven all rights reserved.
This work is part of the research program of the ‘Stichting voor Fundamenteel Onderzoek
der Materie (FOM)’, which is financially supported by the ‘Nederlandse organisatie voor
Wetenschappelijke Onderzoek (NWO)’.
VRIJE UNIVERSITEIT
Thermal lensing in Virgo and
Polynomial Search: an all-sky search
for gravitational waves from spinning
neutron stars in binary systems
ACADEMISCH PROEFSCHRIFT
ter verkrijging van de graad Doctor aan
de Vrije Universiteit Amsterdam,
op gezag van de rector magnificus
prof.dr. L.M. Bouter,
in het openbaar te verdedigen
ten overstaan van de promotiecommissie
van de faculteit der Exacte Wetenschappen
op dinsdag 6 december 2011 om 9.45 uur
in het auditorium van de universiteit,
De Boelelaan 1105
door
Sipho van der Putten
geboren te Amsterdam
promotoren:
prof.dr. J.F.J. van den Brand
prof.dr. F.L. Linde
Contents
Introduction
1
1 Theory
1.1 General relativity and gravitational waves . . . . . . . . .
1.2 Sources of gravitational waves . . . . . . . . . . . . . . . .
1.2.1 Compact binary coalescence . . . . . . . . . . . . .
1.2.2 Bursts . . . . . . . . . . . . . . . . . . . . . . . . .
1.2.3 Stochastic background . . . . . . . . . . . . . . . .
1.2.4 Continuous waves . . . . . . . . . . . . . . . . . . .
1.3 Neutron stars and pulsars . . . . . . . . . . . . . . . . . .
1.3.1 Continuous gravitational waves from neutron stars .
2 Gravitational wave detection
2.1 Laser-interferometric detectors
2.2 Virgo . . . . . . . . . . . . . .
2.2.1 Optical design . . . . .
2.2.2 Suspensions . . . . . .
2.2.3 Noise sources . . . . .
2.3 Calibration of h(t) in Virgo .
2.4 Detector beam-pattern . . . .
2.5 Status of Virgo . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
3 Thermal Lensing
3.1 The finite element model . . . . . . . . . .
3.2 Simulation results . . . . . . . . . . . . . .
3.2.1 Steady state thermal analysis . . .
3.2.2 Eigenmodes in thermal equilibrium
3.2.3 Transient temperature analysis . .
3.3 Optical path length . . . . . . . . . . . . .
3.4 Thermal lensing effects in Virgo . . . . . .
3.5 Conclusion and outlook . . . . . . . . . . .
i
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
3
3
5
6
7
8
8
8
9
.
.
.
.
.
.
.
.
15
15
20
20
22
24
25
27
28
.
.
.
.
.
.
.
.
31
33
36
36
39
40
43
44
48
Contents
4 Continuous waves analysis methods
51
4.1 The gravitational wave signal model . . . . . . . . . . . . . . . . . . . . . 52
4.2 Coherent searches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
4.3 Semi-coherent searches . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
5 Polynomial Search
5.1 Doppler shifts of binary systems . . . . . . . . . . . .
5.2 Binary system test cases . . . . . . . . . . . . . . . .
5.3 Constructing the filters and the correlations . . . . .
5.4 Coherence length . . . . . . . . . . . . . . . . . . . .
5.5 Step size in parameter space . . . . . . . . . . . . . .
5.6 Constructing the hit maps . . . . . . . . . . . . . . .
5.6.1 False-alarm probability . . . . . . . . . . . . .
5.6.2 False-dismissal probability . . . . . . . . . . .
5.7 Combination of hits . . . . . . . . . . . . . . . . . . .
5.7.1 Number count statistics . . . . . . . . . . . .
5.7.2 Renormalized number count . . . . . . . . . .
5.7.3 The sensitivity . . . . . . . . . . . . . . . . .
5.7.4 Consistent filters . . . . . . . . . . . . . . . .
5.7.5 The sensitivity with consistent filters . . . . .
5.8 Results of the simulations . . . . . . . . . . . . . . .
5.9 Discussion and outlook . . . . . . . . . . . . . . . . .
5.9.1 Follow-up and parameter estimation strategies
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
61
61
63
67
71
74
81
82
85
87
88
91
92
96
101
103
113
115
6 Software and computing
117
6.1 The AnaPulsar framework . . . . . . . . . . . . . . . . . . . . . . . . . . 117
6.2 Computing costs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
7 Summary
127
A Fourier transform definitions
131
B The correlation statistic
135
C Stepsize plots
139
D AnaPulsar inheritance diagrams
145
References
160
Samenvatting
161
Acknowledgements
165
ii
Introduction
The publication of Einstein’s first paper on general relativity in 1916 [1] paved the
way for the prediction of the existence of gravitational waves. It was first thought that
gravitational waves were nothing more than mathematical artifacts originating from the
linearization of the Einstein field equations. In 1957 Bondi [2] (and independently Weber
and Wheeler [3]) published papers in which gravitational waves were shown to be actual
physical phenomena which could be measured. The prediction from Einstein’s theory of
general relativity is that any mass distribution with a time-dependent quadrupole (or
higher multipole) moment will emit gravitational waves.
Gravitational waves are perturbations in spacetime which travel at the speed of light.
These perturbations locally disturb the spacetime metric such that measured distances
between freely falling test-masses will vary as a function of time. General relativity states
that spacetime is an elastic but stiff medium meaning that in order to produce measurable differences from flat space in the metric, the associated quadrupole moment must
be large. Furthermore, due to this stiffness of spacetime, gravitational wave signals will
not interact strongly with matter and thus the Universe will be largely transparent to
gravitational waves. This means that gravitational waves will yield a unique insight into
various astrophysical processes. Since the amplitude of a gravitational wave is related to
the magnitude of the variations in the quadrupole moment, compact objects like binary
neutron stars and black holes are the most promising sources.
In 1975 Hulse and Taylor discovered a neutron star binary system PSR 1913+16 [4].
Over the course of decades they measured the time evolution of the orbit of this system
which revealed that the system decayed (e.g. the system loses energy). Furthermore,
this decay has been shown to be consistent with the loss of energy due to the emission
of gravitational waves. This result experimentally proved for the first time since the
formulation of general relativity that gravitational radiation exists and carries energy
from a system. For this discovery Hulse and Taylor received the Nobel prize in physics
in 1993.
While Hulse and Taylor did prove the existence of gravitational radiation, their determination of the energy loss of the system constitutes an indirect measurement of
gravitational waves. Direct detection of the amplitude, phase and polarization of gravitational waves will yield much more information on the physical processes of the source.
In order to directly detect gravitational waves, the relative distance between two freely
1
Introduction
falling test masses must be measured as accurately as possible. This can be done by
using laser interferometric gravitational wave detectors like Virgo.
The Virgo detector is a Michelson interferometer with 3 km long Fabry-Perot resonant cavities as arms and is located in Italy near the city of Pisa. The detector is capable
of measuring the relative distance of two test masses up to about 6 × 10−23 at approximately 200 Hz in 1 second of observation time. With Virgo (and other interferometric
gravitational detectors like LIGO, LCGT and GEO600) it will be possible to directly
measure gravitational wave signals.
These signals, when detected, could be used amongst other things to test general relativity in the strong field regime. They can also be used to gain (additional) insight into
various astrophysical processes like supernovae or the physics of compact objects like
neutron stars and black holes. Gravitational waves can even be used to measure various
cosmological parameters and to study the primordial gravitational wave background.
Since the Virgo detector uses resonant cavities as the detection medium, the laser
power that builds up will heat up the optics. One of these thermal effects is called
‘Thermal Lensing’ and a simulation study of the thermal lensing effect in the Virgo
input mirrors is presented in this thesis. Furthermore, a new analysis method for continuous waves from rapidly rotating neutron stars in binary systems, called ‘Polynomial
Search’, is presented together with the developed analysis framework.
Outline
Chapter 1 gives a short introduction into general relativity and shows the linearization
of the Einstein field equations giving rise to gravitational waves. Furthermore, the different sources of gravitational waves are discussed with emphasis on the mechanisms of
emission of continuous waves from rotating neutron stars or pulsars.
The principles of interferometric gravitational wave detection are discussed in chapter 2. Also, in this chapter the main components of the Virgo detector are presented
as well as the current and future status of Virgo together with the status of the LIGO
detectors.
Chapter 3 presents the simulation of the heating effects of the Virgo input mirrors
of the main arm cavities. This is done by constructing a 3D model of the Virgo input
mirrors together with their recoil masses. The heating effects are computed by solving
the heat equation in this model by finite element analysis. These effects are then related
to the various time constants for heating up the mirror as well as to the change in optical
path length and the change in eigenmode frequencies as a result of this heating.
Chapter 4 gives an overview of the current developed analysis algorithms designed
for searching for continuous waves in interferometric data. Chapter 5 presents the Polynomial Search method and the tests performed on simulated data together with the
sensitivity estimate of the search in comparison to other existing methods.
Finally, chapter 6 gives an overview of the developed framework in which Polynomial
Search is implemented and the computing requirements for performing such a search on
the data obtained with LIGO and Virgo.
2
Chapter
1
Theory
1.1
General relativity and gravitational waves
Gravity is one of the four fundamental forces of Nature. In 1916 Albert Einstein published a theory which describes the gravitational interactions as curvature of the spacetime continuum. This theory is known as General Relativity (GR) [1] and relates the
energy-momentum tensor, Tµν , to the metric tensor, gµν , in a series of coupled non-linear
differential equations called the Einstein field equations.
The theory of GR predicts the existence of gravitational waves (GWs) which are
strains in spacetime propagating at the speed of light. Since these GWs are small perturbations on the background curvature, they are difficult to observe directly. Indirect
observations have been performed by following the evolution of the orbit of binary pulsar system PSR 1913+16 [4]. R.A. Hulse and J.H. Taylor received the Nobel prize in
physics in 1993 for this discovery. Recently [5], it has been shown that the measured
rate of change of the orbital period agrees with that expected from the emission of gravitational radiation, according to general relativity, to within about 0.2 percent.
Equation (1.1) shows the Einstein field equations
8πG
1
Tµν = Rµν − gµν R,
(1.1)
4
c
2
where G is the gravitational constant, c the speed of light, Tµν the energy-momentum
tensor, Rµν the Ricci tensor and R the Ricci scalar. The Ricci tensor is related to the
Riemann tensor by Rµν ≡ Rλµλν and the Ricci scalar is defined as R ≡ g µν Rµν . The
Riemann tensor can be written in terms of the connection Γρνσ , as
Rρσµν = ∂µ Γρνσ − ∂ν Γρµσ + Γρµλ Γλνσ − Γρνλ Γλµσ .
The connection can be written as a series of derivatives of the metric as
1
Γσµν = g σρ (∂µ gνρ + ∂ν gρµ − ∂ρ gµν ).
2
(1.2)
(1.3)
Note that the metric tensor, gµν , is related to the proper distance ds2 as
ds2 = gµν dxµ dxν .
(1.4)
3
Chapter 1. Theory
GR relates the energy-momentum tensor to the metric through Eq. (1.1), (1.2) and
(1.3). Equation (1.1) can be solved analytically in certain circumstances, for instance in
the weak-field limit. For weak gravitational fields the metric can be approximated by a
time-dependent perturbation of the Minkowski metric,
gµν ≈ ηµν + hµν ,
where ηµν is the Minkowski metric written as

−1 0
0 1
ηµν = 
0 0
0 0
|hµν | 1,
0
0
1
0

0
0
.
0
1
(1.5)
(1.6)
Combining Eqs. (1.1), (1.2) and (1.3) with the weak field approximation (1.5) together
with the requirement |hµν | 1 and thus neglecting terms with O(h2 ) and higher, the
Einstein equations can be linearized. When also defining the trace-reversed perturbation
h̄µν by
1
h̄µν = hµν − ηµν hλλ
2
and with the Lorentz gauge condition
∂µ h̄µλ = 0,
the Einstein equations become
h̄µν =
−16πG
Tµν .
c4
(1.7)
Note that is the D’Alambertian which, in Cartesian coordinates and flat space is
denoted by
1 ∂2
∂2
∂2
∂2
= − 2 2 + 2 + 2 + 2 .
c ∂t
∂x
∂y
∂z
Finally, taking Tµν = 0 (far away from the source), Eq. (1.7) becomes
h̄µν = 0,
(1.8)
which, when realizing that the perturbation is small and thus the space can be considered
flat, can be recognized as the wave equation and h̄µν represents the GW. Substituting
the plane wave solution
σ
h̄µν = Cµν eikσ x ,
(1.9)
where Cµν a 4 × 4 amplitude matrix and kσ is the wave vector, yields the condition
kσ k σ = 0 meaning that GWs travel at the speed of light. Choosing the transversetraceless gauge implies h̄TµνT = hTµνT . Furthermore, when also orienting the propagation
4
1.2. Sources of gravitational waves
of the wave along the z-axis, the GW strain can be written as


0 0
0
0
0 h+ h× 0 iω(t−z)

,
h(t, z) = 
0 h× −h+ 0 e
0 0
0
0
(1.10)
where the h+ and h× are the ‘plus’ and ‘cross’ polarizations of the GW. The effect of
the two different polarizations on a ring of test masses is shown in Fig 1.1.
Figure. 1.1: The effect of a GW with only the plus polarization (top) and the cross
polarization (bottom). The direction of the GW is perpendicular to the
paper. The phase of the wave is shown over one full period of time T.
Contrary to electromagnetic radiation, which can be produced by accelerating charge
dipoles, GWs can only be produced by a varying mass-quadrupole moment. This is due
to the fact that conservation of mass/energy prevents a time-dependent monopole mass
moment and conservation of momentum forbids a time-dependent dipole mass moment
from emitting GWs. Thus the first multipole moment that allows for the emission of
GWs is the quadrupole moment.
1.2
Sources of gravitational waves
There are roughly four categories of promising astrophysical sources of gravitational
waves. Binary systems of compact objects (either neutron stars or black holes), which
due to energy loss coalesce, fall in the ‘Compact Binaries Coalescence’ (CBC) category.
These CBCs have a transient but well modeled waveform which can be divided into three
5
Chapter 1. Theory
parts called the Inspiral, Merger and Ringdown phases. Unmodeled transient waveforms
fall in the ‘Burst’ category. Residual GW background from the Big Bang and/or a
large number of weak astrophysical sources fall in the ‘Stochastic Background’ category.
Finally long-lived continuous GWs originating from spinning neutron stars are called
‘Continuous Waves’ (CW).
1.2.1
Compact binary coalescence
Compact binary coalescence is the most well-known and promising potential source for
GW detection. This is due to the fact that the strains involved are relatively high. The
optimal horizon distance for neutron star neutron star (NS-NS) inspiral, neutron star
black hole (NS-BH) and BH-BH have been computed in Ref. [6] to be 33/445 Mpc,
70/927 Mpc and 161/2187 Mpc for Initial and Advanced detectors, respectively. As an
example, the predicted rate, computed by extrapolating from observed binary pulsars,
for a NS-NS inspiral event is 100 Myr−1 per Milky Way Equivalent Galaxy (MWEG).
However, this rate can plausibly range from 1 to 1000 Myr−1 MWEG−1 . These numbers
have been converted to detection rates ranging from 2×10−4 to 0.2 yr−1 for Initial LIGO
and Virgo and from 0.4 to 400 yr−1 for the Advanced detectors.
A CBC consists of three distinct phases of GW emission. First, due to loss of angular momentum the compact objects spiral towards each other in the inspiral phase.
Next, the two objects merge in the merger phase. Finally, the distorted black hole relaxes to a stationary Kerr state through emission of GWs in the ringdown phase. All
of these phases of the waveform are shown schematically in figure 1.2. The GW signal
Figure. 1.2: Cartoon indicating the three phases of an inspiral together with a
schematic of the inspiral waveform (courtesy of [7]).
produced by a CBC will increase in frequency and amplitude during the inspiral phase.
This means that the signal will enter the detector band from about 10 Hz only in the
6
1.2. Sources of gravitational waves
last few minutes or seconds (depending on the masses involved) before the merger.
The data analysis algorithms to find CBC signals in the data are based on ‘matched
filtering’. Matched filtering is done by computing analytical or phenomenological waveforms as signal templates and comparing them to the data stream. The output of each
filter-data comparison is assigned a signal-to-noise ratio (SNR). Events whose SNR
crosses a threshold are subjected to a coincidence analysis between the available detectors. The consistency of the resulting coincidence is verified by using a χ2 -test defined
in Ref. [8].
1.2.2
Bursts
A second category of potential GW sources consists of unmodeled transient phenomena,
or ‘bursts’. A wide variety of astrophysical sources is expected to produce such signals at
some level. Sophisticated models have been made in some cases, while in other cases the
amplitude and phase of the GW signals is highly uncertain. The only thing the potential
sources have in common is the fact that the duration of the signals is less than several
seconds.
Possible sources targeted by the burst search include, but are not limited to, supernovae core (SN) collapses [9] and ringdowns of black-hole mergers [10] whose inspiral
signal does not lie within the detectable region for ground-based interferometers. Furthermore, astrophysical data on bursts like Gamma-Ray Bursts (GRBs) are used as
triggers for a targeted burst search which can be run with increased sensitivity.
The rate of SN core collapse is predicted to be 0.5 yr−1 for a horizon at 5 Mpc [9] but
the actual range of burst searches from SN depends on the amplitude and phase evolution of the GWs. This strongly depends on the preferred model and thus no accurate
prediction of the event rate of these types of bursts can be made. The detection rates
of intermediate mass (20M ≤ M ≤ 106 M ) binary black hole (IMBBH) inspirals are
about 10 events per year for the Advanced detectors [10]. The rate of detection for GRBs
depends on observation distance and the field of view of the satellite used to generate
triggers as well as the amount of energy converted to GW emission. GRBs come in two
broad classes: short (commonly associated with NS-NS or NS-BH mergers) and long
(commonly associated with core collapse supernovae). Furthermore, a local population
of under-luminous short GRBs [11] seems to exist which has an observed rate density
approximately 103 times that of the high-luminosity population. When assuming that
the energy emitted in GWs equals 0.01M c2 , a field of view of 9.5 sr of the satellite
(Fermi), and Advanced Virgo and LIGO sensitivity, the expected rates are 5.38 × 10−3
yr−1 , 5.38 yr−1 and 0.08 yr−1 for the long, local and short type of GRB [12], respectively.
Since the burst search is meant to find unmodeled burst-like signals, glitches in the
detector which mimic burst signals will produce background. The search techniques for
burst searches always include correlation between detectors to reduce this background.
There are numerous techniques developed to perform these burst searches like templatebased searches [13], cross-correlation [14] and coherent methods [15].
7
Chapter 1. Theory
1.2.3
Stochastic background
A stochastic background of GWs is expected to arise as a superposition of a large number
of unresolved sources, from different directions in the sky, and with different polarizations. It is usually described in terms of the GW spectrum ΩGW (f ) ∝ f α [16] where α
is different for different emission mechanisms.
Many possible sources of stochastic GW background have been proposed. Some of
the proposed theoretical models are cosmological in nature, for example pre-big-bang
models [17]. Others are astrophysical in nature, such as rotating neutron stars, supernovae and low-mass X-ray binaries.
The analysis method used is the cross-correlation method. This method uses the output of multiple interferometers and uses the cross-correlation to filter out any common
signals from the background.
1.2.4
Continuous waves
Continuous GWs can be emitted by rapidly rotating neutron stars. Some neutron stars
are observed as pulsars and have rotational frequencies ranging from a few millihertz
up to a kilohertz. Furthermore, in order to emit GWs the neutron star must have a
quadrupole moment. This implies that there must be a non-axisymmetric asymmetry
present. This asymmetry can either be due to a geometrical asymmetry in the crust, due
to fluidic asymmetry or induced by accretion of matter onto the neutron star. These GW
emission mechanisms are discussed in some more detail in section 1.3.1. Furthermore,
the various existing data analysis methods will be summarized in chapter 4.
1.3
Neutron stars and pulsars
A neutron star is a type of stellar remnant that can result from the gravitational collapse
of a massive star during a supernova. As the core of a massive star is compressed during
a supernova and collapses into a neutron star, it retains most of its angular momentum.
The neutron star has only a fraction of its parent’s radius (around 10 km) and is formed
with high rotation speed. It gradually slows down. Neutron stars are known to have
rotation periods between about 1.4 ms to 30 seconds. Observationally, some neutron
stars are known to have very strong magnetic fields, which as a function of their age
and composition can range from 108 Gauss (milli-second pulsars) to 1015 Gauss (young
pulsars/magnetars).
The magnetic field together with the rotation of the neutron star results in the acceleration of protons and electrons on the star surface and the creation of a beam of
radiation emanating from the poles of the magnetic field. If the neutron star is positioned
in such a way that this beam sweeps the Earth, it is observed as a pulsating source of
electromagnetic radiation. Neutron stars with such properties are commonly referred to
as pulsars.
Pulsars are only observed when their radiation beam is swept into the direction of
Earth during a revolution of the neutron star. Pulsar emission is usually too weak to
8
1.3. Neutron stars and pulsars
be detected extragalactically. Due to these effects, only a fraction of the pulsars in our
Galaxy is known and more of these pulsars are close by than far away. Taking these
observational biases into account over 105 active pulsars are expected to exist in our
Galaxy [18]. However, since active pulsars last only a few tens of millions of years and
not all neutron stars pulse, the number of neutron stars in the Galaxy is expected to be
orders of magnitude larger.
Pulsars have been known to spin up or undergo sudden increases of their rotation
speed, called ‘glitches’, which range in fractional size from 10−11 to 10−4 of the spin
frequency. Most models for pulsar glitches build on the vortex unpinning model [19].
Superfluid vortices pin to defects in the crystalline crust and are prevented from migrating outward as the crust spins down electromagnetically. At some stage, many vortices
unpin catastrophically, transferring angular momentum to the crust. It is however unknown what causes this collective unpinning, this process is also thought to generate
GWs [20].
Pulsars emit a lot of energy, i.e. the Crab pulsar emits about 5 × 1038 ergs/s [21]
in wide-band electromagnetic emission. Since most of this energy is drawn from the
rotational energy of the pulsar, the rotational frequency should slowly decrease. This
decrease in rotation is characterized by the so-called spindown age τs ,
τs =
f
,
|f˙|
(1.11)
where f is the rotation frequency and f˙ is the first order time derivative of the rotation
frequency. Incidentally, when assuming1 that f˙ = −Kf n , where K is a constant and the
‘braking index’ n > 1, the spindown age is related to the age of the pulsar by
t=
τs
.
n−1
(1.12)
Experimentally, the breaking index typically has values n ' 1 − 3 [22] depending on the
specific pulsar (i.e. Crab: n ' 2.5). From the ATNF pulsar catalog [21], the left panel
in figure 1.3 shows f˙ versus f . It can be seen from this panel that all known pulsars
have f˙ < 10−9 Hz/s. Furthermore, the right panel in figure 1.3 shows the frequency
distribution of the pulsars with rotational frequency above 20 Hz. The cut off of 20 Hz
has been chosen since it roughly corresponds to the beginning of the sensitive region
of ground-based interferometers like LIGO and Virgo. The figure also shows that most
known millisecond pulsars are located in binary systems.
1.3.1
Continuous gravitational waves from neutron stars
In the frequency band relevant to ground-based interferometers there are three predicted
mechanisms for producing continuous GWs. All of these mechanisms involve neutron
stars.
1
This assumption is reasonable since it can be computed that for electromagnetic dipole radiation
˙
fEM ∝ f 3 and for GW emission f˙GW ∝ f 5 .
9
- df (Hz/s)
dt
10-8
10-9
10-10
10-11
10-12
10-13
10-14
10-15
10-16
10-17
10-18
# pulsars
Chapter 1. Theory
Isolated pulsars
Pulsars in binary systems
16
All pulsars
14
Pulsars in binary systems
12
10
8
6
4
2
10-1
1
10
102
103
f (Hz)
0
0
100 200 300 400 500 600 700
f (Hz)
Figure. 1.3: The left panel shows the first order spin-down versus the frequency of all
known pulsars. The right panel shows the frequency distribution of known
pulsars with rotational frequency over 20 Hz for isolated pulsars (total
37) and pulsars in binary systems (total 75) The data are taken from the
ATNF pulsar catalog (January 2011).
ˆ Non-axisymmetric distortions of the solid part of the star [23].
ˆ Unstable r-modes in the fluid part of the star [24].
ˆ Free precession of the whole star [25].
Furthermore, neutron stars in binary systems can emit GWs due to accretion [26].
Non-axisymmetric distortions
Non-axisymmetric distortions cannot exist in a perfect fluid star, but in realistic neutron
stars such distortions can be supported either by elastic stresses or by magnetic fields.
These non-axisymmetric distortions are akin to mountains on the Earth’s surface and
are expressed in terms of the ellipticity
=
Ixx − Iyy
,
Izz
(1.13)
where Ixx , Iyy and Izz are the moments of inertia along the x, y and z axes, respectively.
A non-axisymmetric neutron star rotating along the z-axis, at distance d with frequency frot emits GWs with amplitude ([27])
h0 =
4π 2 G If 2
,
c4 d
(1.14)
where G is Newton’s constant, c the speed of light, I the moment of inertia along the axis
of rotation and f is the GW frequency (in the case of an optimally orientated source,
f = 2frot ).
The maximum ellipticity that can be sustained by a neutron star depends on the
10
1.3. Neutron stars and pulsars
model of the neutron star used. This makes the estimation of the maximum ellipticity
highly uncertain and predictions range from 10−4 to 10−6 [23]. For a typical neutron
star the moment of inertia is in the order of I = 1038 kg m2 . Equation (1.14) can then
be parameterized as
−27
h0 = 1.068 × 10
1038
I
kg m2
10 kpc
r
f
100 Hz
2 .
10−6
(1.15)
Note that the GW amplitude scales with f 2 . It has been normalized to 100 Hz, while
the distance has been normalized to 10 kpc corresponding to a NS in our Galaxy. When
considering pulsars in a binary system frot can be as high as 700 Hz (see Fig. 1.3).
Furthermore, the nearest known pulsar in a binary system (J1045-4509) is located at
0.3 kpc. Taking these values, the GW amplitude can be as high as h0 ≈ 10−24 .
Unstable r-modes
An alternative way of generating GWs from neutron stars is through the so-called r-mode
oscillations. An r-mode is a particular torsional mode which is due to material motion
tangentially to the neutron star’s surface. Essentially, they can be seen as horizontal
currents associated with small density variations. The r-modes exist only in rotating
neutron stars and are caused by the Coriolis force generated by the rotation. Gravitational radiation couples to these modes primarily through the current multipoles, rather
than the usual mass multipoles.
It has been shown by Chandrasekhar [28] (and later generalized by Friedman and
Schutz [29]) that for these r-modes the gravitational radiation reaction amplifies the
oscillation of the rotating star as opposed to damping it. This effect can be understood
qualitatively by considering that the r-mode period is larger than the rotational period
of the neutron star (this is an intrinsic property of the r-mode oscillation). The sign of
the angular momentum carried away by the gravitational waves is the same as that of
the angular momentum of the mode pattern as measured in the inertial frame. Here, the
gravitational wave will carry positive angular momentum away from the system and to
infinity. As a result the star will spin down.
In the co-rotating frame (i.e. rotating with the neutron star) the neutron star is
at rest and the r-mode will rotate in the opposite direction. In this frame, the angular
momentum will be negative and gravitational waves will still be emitted. These gravitational waves will carry angular momentum away from the system, making the total
angular momentum even more negative for the observer in the co-rotating frame. This
implies that what started as perturbations with small negative angular momentum in
the co-rotating frame, have been transformed (in the same frame) into large amplitude
oscillations with a progressively larger negative angular momentum that will emit increasingly large amounts of gravitational waves. Therefore, these modes are unstable
and this instability is known at the Chandrasekhar-Friedman-Schutz (CFS) instability.
Ultimately, the internal viscosity of the star will stop the instability from growing.
The r-modes have been proposed as a source of GWs from newborn and/or accreting
11
Chapter 1. Theory
neutron stars [24]. Unlike the GWs originating from a non-axisymmetric mass distribution, the GWs from r-mode instabilities will be emitted at
4
f = frot .
3
(1.16)
The production mechanism of GWs from r-modes depends on the equation of state and
on the amplitude of the oscillations. The CFS instability of the r-modes in newborn
neutron stars is not a good candidate for the detection of continuous GWs since the
emission is short lived or has low amplitude. On the other hand, accreting neutron stars
are considered a better prospect for a detection of r-mode gravitational radiation since
the emission might be long-lived due to the transfer of angular momentum [30].
Free precession
When a rotating neutron star has a symmetry axis which does not coincide with its
rotation axis, the star will precess. Since this precession has a quadrupole moment, it will
cause the neutron star to emit GWs. This GW emission would occur at f = frot + fprec
where fprec frot is the precession frequency.
The GW amplitude has been parameterized in Refs. [23, 25] as
−30
h0 ≈ 4 × 10
θw
0.1 rad
10 kpc
r
frot
100 Hz
2
,
(1.17)
where θw is the angle between the axis of symmetry and the axis of deformation. It is
clear from comparing Eq. (1.17) with Eq. (1.15) that GWs from free precessing neutron
stars will even be weaker than those from non-axisymmetric neutron stars.
Accretion
When a neutron star is located in a binary system, the situation may arise that matter
is transfered from the companion star to the neutron star. These companions can range
from supergiants to white dwarves. Accretion is the process via which old pulsars are
‘recycled’ and spun-up to become milli-second pulsars. Due to the high gravitational
pull of the neutron star, accretion leads to the formation of an accretion disk around
the neutron star. Since the matter from the companion will be accelerated in this disk,
it will emit X-rays.
In the context of GW generation, accretion is a natural way of generating and maintaining crustal deformations. Accretion is not an isotropic process thus there will be
many ‘hot spots’ on the neutron star surface at which the accreted material accumulates. These hot spots can lead to the formation of crustal asymmetries, which would
lead to the emission of GWs which will carry away angular momentum. This can lead
to a situation in which the accretion spin-up is balanced by the spin-down, leading to
the continuous emission of GWs as suggested in Ref. [26].
Rapidly accreting stars are of particular interest for this sort of emission mechanism, for example in low-mass X-ray binaries (LMXBs). Under the assumption that the
12
1.3. Neutron stars and pulsars
GW emission is in equilibrium with the accretion torque, the GW amplitude is directly
related to the X-ray luminosity. The GW amplitude is parameterized in Ref. [23] as
−27
h0 ≈ 5 × 10
300 Hz
frot
1/2 10−8
Fx
erg cm−2 s−1
1/2
,
(1.18)
where Fx is the X-ray flux. The spin frequencies, frot , of LMXBs can go up to about 620
Hz.
Furthermore, according to Ref. [31], the magnetic field of a neutron star can help to
localize the accreting matter to a small area of the surface. The magnetic field lines can be
compressed into a narrow belt at the magnetic equator, which then confines the accreted
material to the poles of the neutron star. Since the magnetic axis is not generally aligned
to the angular momentum axis, this accumulation of material takes place asymmetrically,
thus creating a mass quadrupole and consequently GWs are emitted. The dissipation
of these mountains is slow because the matter is highly conductive and thus crosses the
field lines slowly. This leads to a scenario in which the pile up is matched by steady GW
emission. Reference [32] identifies the most promising targets, and lists specific actions
that would lead to significant improvements in detection probability.
13
Chapter 1. Theory
14
Chapter
2
Gravitational wave detection
As discussed in section 1.1, gravitational waves are perturbations of the spacetime curvature. These perturbations induce a strain in spacetime resulting in a time-dependent
variation of the physical distance between test masses. This strain h is in essence a
relative distance variation defined as
∆L
h≡2
,
(2.1)
L
where L and ∆L are the distance and change in distance between two test masses respectively. The detection of gravitational waves is based on measuring distances between test
masses. Note that the gravitational wave amplitude is typically a small number (10−27 )
for GWs from neutron stars (see section 1.2.4).
2.1
Laser-interferometric detectors
A laser interferometer can be used to measure the distance between test masses to great
accuracy. The simplest example of a laser interferometer is a Michelson interferometer,
shown schematically in Fig. 2.1. The figure shows that in a Michelson interferometer laser
light is split into two separate beams by using a 50/50% beam splitter. The mirrors act
as test masses and reflect the light back. The beams form an interference pattern of
which a single fringe is transmitted towards the output photodiode.
The essential idea is to maintain the interferometer in a locked position, with the
difference of its arm length fixed by measuring the light at the output and feeding
the signal back to the mirrors. The arm lengths are kept fixed in such a way that the
light interferes destructively. The correction, or error-signal, would contain the potential
gravitational wave signal since it measures the deviation from the working point or dark
fringe.
When taking the expression for the proper distance (Eq. (1.4)) and substituting the
metric tensor by the perturbed metric as shown in Eq. (1.5) together with the assumption
that the gravitational wave is +-polarized and propagating in the z-direction, the proper
distance of a photon traveling along the x-axis is given by
ds2 = −c2 dt2 + (1 + h+ (t))dx2 = 0,
(2.2)
15
Chapter 2. Gravitational wave detection
Figure. 2.1: A schematic representation of a Michelson interferometer (left) and the
effect of a passing gravitational wave (right). In the right panel, the wave
is assumed to be of plus polarization and is propagating perpendicular to
the interferometer, into the page.
where c is the speed of light. Since h is considered small, the approximation |h+ (t)| 1
can be made and will yield
1
(1 − h+ (t))c dt = dx.
(2.3)
2
Integrating Eq. (2.3) over the time it takes the light to complete a single round rip, will
yield the phase change along the x and y direction given by
Z
2Lx,y Ωl Ωl t
φx,y = Ωl ∆tx,y =
±
h+ (t0 )dt0 ,
(2.4)
c
2 t− 2Lcx,y
where Ωl = 2πfl is the angular frequency of the laser light. The phase difference δφGW =
φy − φx caused by a gravitational wave can be computed from Eq. (2.4) to be
Z t
δφGW = Ωl
h+ (t0 )dt0 .
(2.5)
t−
2Lx,y
c
Assuming that the arm lengths are equal (L = Lx = Ly ) and changes in the incident gravitational wave are much slower than the storage time in the arms, the phase
difference becomes
4πL
δφGW =
h+ ,
(2.6)
λl
where λl = 2πc/Ωl . Equation (2.6) shows that the phase difference is proportional to
the length of the arms of the Michelson interferometer. Note that Eq. (2.6) is only valid
when the wavelength of the gravitational wave is large compared to the arm length.
16
2.1. Laser-interferometric detectors
The sensitivity of a Michelson interferometer can be computed by substituting the
Fourier transform of the gravitational wave,
Z ∞
h+ (t) =
h+ (ω)e−iωt dω,
−∞
into Eq. (2.5) which yields
Z
δφGW =
∞
−∞
HMI (ω)h+ (ω)eiωt dω,
(2.7)
where ω is the angular frequency of the GW and the sensitivity of the Michelson interferometer as a function of ω, HMI (ω), is written as
2Ωl
Lω −i Lω
HMI (ω) =
sin
e c .
(2.8)
ω
c
From Eq. (2.8) it can be seen that the gravitational wave response of the Michelson
interferometer is maximized when L satisfies
Lω
π
= ,
c
2
for a given gravitational wave angular frequency ωGW . This would imply that the optimal arm length for a 1 kHz gravitational wave signal is about 75 km. Since this is
an unrealistic arm length, other ways of improving the optical design to increase the
effective arm length have been developed.
One way of accomplishing this, is the so-called delay-line Michelson interferometer.
Figure 2.2 shows a delay-line Michelson interferometer in a schematic way. In this de-
Figure. 2.2: Schematic of a delay-line Michelson interferometer. In this case the beam
is reflected 4 times (NDL = 4).
17
Chapter 2. Gravitational wave detection
sign another mirror is placed in between the beam splitter and the end mirror of each
arm. These mirrors are designed in such a way that the laser beam is reflected multiple
times before it is transmitted onto the beam splitter. The sensitivity of this delay-line
interferometer is given by
2Ωl
NDL Lω −i NDL Lω
2c
HDLMI (ω) =
sin
e
,
(2.9)
ω
2c
where NDL is the number of bounces between the two mirrors (a simple Michelson interferometer will have NDL = 2). As Eq. (2.9) shows, the sensitivity of the interferometer
increases with the number of bounces. However, it is non-trivial to produce mirrors
which allow for multiple bounces.
A more practical way of increasing the arm length is by using Fabry-Perot resonant
cavities as arms for a Michelson interferometer. A Fabry-Perot cavity is an arrangement
of mirrors that forms a standing wave cavity resonator for laser light. The sharpness of
the resonance is determined by the reflection coefficients of both the input (rI ) and end
mirrors (rE ) of the cavity. The finesse, F, is defined as the ratio of the cavity resonance
width (FWHM) to the free spectral range ∆f = c/2L, the interval between resonances.
For a general cavity, the finesse is given by
√
π rI rE
F=
.
(2.10)
1 − rI rE
In the limit that the time scale for the metric change is long compared to the storage
time of the light in the cavity, the light stored in a Fabry-Perot cavity of finesse F will
have
2F
Nφ =
(2.11)
π
times the phase shift in response to a GW of a one-bounce interferometer arm. Figure
2.3 shows the phase of the reflected beam and the fraction of the transmitted power of
a Fabry-Perot cavity. From Fig. 2.3 it can be seen that the phase of the reflected beam
is a steep function of ∆φ for large values of the finesse. Thus a small deviation from
∆φ = 0 will induce a large phase shift in the reflected light. Furthermore, due to the
sharpness of the Fabry-Perot resonance, small deviations from ∆φ = 0 will cause a large
decrease in transmitted power (or an increase in reflected power).
Two Fabry-Perot cavities can be combined in a Michelson interferometer as shown
schematically in Fig. 2.4 This setup will produce an interference pattern which is directly related to the reflected phase and power of the light from both of the Fabry-Perot
cavities.
The frequency response of a Fabry-Perot-Michelson interferometer HFPMI (ω) is written as
sin Lω
2acav Ωl
−i Lω
c
c ,
(2.12)
HFPMI (ω) =
Lω e
−2i
ω 1 − rI rE e
c
where
acav =
18
t2I rE
.
1 − rI rE
(2.13)
3
Finesse 10
Finesse 50
Finesse 100
Finesse 500
2
1
Ptrans
Phase (rad)
2.1. Laser-interferometric detectors
1
Finesse 10
Finesse 50
Finesse 100
Finesse 500
0.8
0.6
0
0.4
-1
0.2
-2
-3
-0.1
-0.05
0
0.05
0.1
∆φ (rad)
0
-0.4
-0.2
0
0.2
0.4
∆φ (rad)
Figure. 2.3: The left panel shows the phase of the reflected beam as a function of the
accumulated phase difference ∆φ during the storage time of the light. The
right panel shows the fraction of transmitted power as a function of ∆φ.
Both panels show curves for different values of the finesse.
Figure. 2.4: Schematic of a Michelson interferometer with Fabry-Perot cavities.
Figure 2.5 shows the frequency response of a delay-line Michelson interferometer, a
Fabry-Perot Michelson interferometer and a simple Michelson interferometer. As can be
seen in Fig. 2.5, the response of the Fabry-Perot interferometer shows a more smooth
response compared to that of the delay-line interferometer. Also it can be seen that
the sensitivity of the Fabry-Perot interferometer is lower than that of the delay-line interferometer in the high frequency range fGW > 100. However, the smoothness of the
frequency response as well as the fact that delay-line interferometers are more susceptible to scattered light noise makes the Fabry-Perot interferometer the most adopted
configuration for gravitational wave detectors.
Currently, there are several ground-based interferometers operational in the world:
LIGO [33], GEO600 [34] and Virgo [35]. LIGO consists of two separate interferometers
located in Livingston (Louisiana) and Hanford (Washington) in the USA. GEO600 is
19
|H(ω)|2
Chapter 2. Gravitational wave detection
HDLMI
HFPMI (finesse 50)
HFPMI (finesse 500)
HMI
104
103
102
10
1
10-1
10-2
10-3
10-4
1
10
102
103
104
105
fGW (Hz)
Figure. 2.5: The magnitude of the frequency response for the Michelson interferometer (MI), delay-line Michelson Interferometer (DLMI) and Fabry-Perot
interferometer (FPMI) as a function of the gravitational wave frequency
fGW . All frequency responses are shown relative to the Michelson interferometer response. For all curves, the base line arm length is 3 km, the laser
wavelength has been chosen to be 1 µm the finesse of the cavities is set to
50 and 500 and the number of bounces for the delay-line interferometer is
computed with Eq. (2.11).
an interferometer with 600 m long arms located near Hannover in Germany. The Virgo
project consists of a single interferometer which is located in Italy near the city of Pisa.
2.2
Virgo
Virgo [35] is a Michelson interferometer with Fabry-Perot cavities as arms each 3 km
long, located in Italy near the city of Pisa. Its objective is to directly detect gravitational
waves in coincidence with other interferometers like LIGO [33]. The most important
components of Virgo are the optics and the suspension system.
2.2.1
Optical design
The optical layout of the Virgo interferometer is shown schematically in Fig. 2.6. The
interferometer consists of a laser system, an input mode cleaner, a power recycling
mirror, two Fabry-Perot cavities and a detection system.
20
2.2. Virgo
Figure. 2.6: A diagram of the Initial Virgo design optical layout.
The laser system
The laser is a Nd:Yag system emitting at λ = 1.064 µm. This wavelength has been
chosen because lasers emitting in the infrared region are more powerful, reducing the
shot noise. Furthermore, optical components for this wavelength are readily available.
The output power in the TEM00 mode is about 20 W. The laser system is placed on an
optical table within a clean room outside of the vacuum.
The input mode cleaner
Any positional noise (e.g. coupling of seismic ‘beam jitter’ noise to the input bench) of the
laser system induces additional higher order modes to the fundamental TEM00 mode.
The input mode cleaner is a triangular cavity, 144 m long with a finesse of about 1000
which is tuned to resonate on the fundamental TEM00 mode and thus filters out these
higher order modes. It consists of two plane mirrors placed on an optical bench known as
the ‘injection bench’ and a third concave mirror suspended by a superattenuator located
in the mode-cleaner tower. The end-mirror is connected through a dedicated vacuum
pipe to the injection bench.
21
Chapter 2. Gravitational wave detection
Power recycling
Shot noise is one of the fundamental noises which limit the sensitivity of an interferometer. Shot noise is a photon counting error, which is proportional to the square-root of the
laser power incident in the interferometer. This noise is dominant at high frequencies
(>500 Hz). Since the gravitational wave signal is proportional to the incident power,
the shot-noise level of an interferometer is improved by increasing the laser power. In
order to increase the laser power in the Fabry-Perot cavities, a technique called power
recycling has been used. This involves the placement of an extra power recycling mirror
before the beam splitter. This mirror forms a resonant cavity together with the interferometer. The recycling factor, defined as the ratio between the power stored in the cavity
and the input power, is designed to be 50. With this design value a power of about 1
kW is expected on the beam splitter.
Fabry-Perot cavities
Each arm is a Fabry-Perot cavity consisting of an input mirror and an end mirror. The
cavities are 3 km long and in Initial Virgo designed with a finesse of 50. The input
mirrors and end mirrors are 350 mm in diameter. The input mirrors are flat and have
an approximate reflectivity of 98% while the output mirrors are 100% reflective (within
a few p.p.m.). Furthermore, in order to compensate for the beam divergence over the 3
km length, both end mirrors are curved with an approximate radius of curvature of 3.5
km. Both input and output mirrors (as well as the beam splitter, power recycling mirror
and the mode cleaner end mirrors) are suspended by a series of coupled pendula called
‘superattenuators’.
Detection system
The detection system is composed of the suspended detection bench placed in vacuum
and the external detection bench which is not suspended and is located outside the
vacuum. The suspended detection bench mainly holds the output telescope and the
output mode cleaner (OMC), a monolithic optical cavity 2.5 cm long of finesse F = 50,
which filters higher order modes of the beam. The main beam leaves the suspended
bench to reach the external bench, where a set of InGaAs photodiodes are located in
order to detect the signal. The output of these diodes are used for the locking of the
interferometer as well as for the signal detection.
2.2.2
Suspensions
The detection of gravitational waves is limited at low frequencies, where the seismic
noise of the ground is transmitted through the suspensions √
to the mirrors. This seismic
noise would induce position disturbances of about 10−9 m/ Hz at 10 Hz. Specific seismic isolation systems, consisting of coupled pendula called ‘superattenuators’ [36], have
been developed for Virgo. The purpose of these superattenuators is to suspend the test
masses (i.e. the mirrors) in such a way that they are virtually freely falling (e.g. the
22
2.2. Virgo
amount of external force on the test masses is minimized).
Figure
2.7
shows
a
schematic
of
a
superattenuator.
It can be seen from this figure
that a superattenuator consists
of five filter platforms. The filters are connected by suspension wires and a support marionetta. The mirror and recoil
mass are suspended from the
last filter (the so-called ‘filter 7’).
Each filter consists of a series of
blade springs which isolate the
test mass in the vertical direction. The entire chain of filters
is suspended from the ground by
a so-called ‘inverted pendulum’
which introduces yet another step
in the seismic isolation.
The superattenuator is capable of inhibiting the transmission
of seismic vibrations to the mirror above a few Hz in all 6 degrees of freedom. The superattenuators are designed to reduce the
residual motions of the mirrors.
In order to reach √
a target sensi−18
tivity of 10
m/ Hz at 10 Hz,
a local control system for the test
masses has been installed in the
superattenuators. Six full superattenuators (tall towers) are suspending the mirrors of the FabryPerot cavities, the beam splitter
and the power recycling mirror.
Because of the lesser seismic isolation requirement, three shorter
version (short towers) are utilized
to suspend the suspended injec- Figure. 2.7: Left: A schematic view of the Virgo superattenuator taken from Ref. [36]
tion bench, the far mirror of the
mode cleaner and the bench of
the detection system. See Ref. [36] for a more detailed description and performance
of the different seismic isolation systems in use in Virgo.
23
Chapter 2. Gravitational wave detection
2.2.3
Noise sources
Sn (Hz-1/2)
Several sources of noise can introduce spurious signals capable of masking the weak
effects due to an incident gravitational wave. The detector sensitivity is limited by the
intensities of these noise sources. The total expected noise is the sum of the power of
all these noise sources and results in the sensitivity curve √
of an interferometer. This
sensitivity curve is given as the amplitude spectral density, Sn , (see appendix A and
Eq. (A.15)), with Sn the sum of the powers of all these noise sources.
Figure 2.8 shows the design sensitivity of the Virgo detector broken down into the
three most prominent noise sources. Reference [37] gives a more detailed description of
all known noise sources. The main noise sources shown in Fig. 2.8 are seismic noise,
10-16
Total noise
Seismic noise
Thermal noise
Shot noise
10-17
10-18
10-19
10-20
10-21
10-22
10-23
10-24
1
10
102
103
104
f (Hz)
Figure. 2.8: Breakdown of the design sensitivity curve of the Virgo detector in the
most relevant noise components.
thermal noise and shot noise.
Seismic noise
The seismic noise curve is computed by taking the transfer function of the superattenuator and assuming a seismic noise spectral amplitude x̃seis = 10−7 /f −2 for 1 Hz < f < 103
Hz. Figure 2.8 shows that seismic noise is expected to be dominant up to about 3 Hz.
Seismic noise is the limiting factor for the sensitivity in the low frequency regime. Any
improvements to the superattenuators (e.g. lowering the resonant frequencies) will not
yield a significant improvement in the sensitivity. This is due to the so-called Newtonian
noise or gravity gradient noise [38]. This noise is an irreducible background noise caused
by stochastic gravitational field modulations due to seismic noise.
24
2.3. Calibration of h(t) in Virgo
Thermal noise
Thermal noise is due to energy loss processes of the suspension wires and test masses.
The thermal noise curve shown in Fig. 2.8 is the sum of the noise spectra of all these
processes. As can be seen, thermal noise dominates from 3 to about 500 Hz. Up to about
30 Hz the noise is due to thermal fluctuations of the pendula formed by the mirror and
its suspension wires. Above 30 Hz the tail contribution is due to the high frequency
internal modes of the mirrors. The spike at approximately 7 Hz is due to the thermal
noise associated to the main vertical mode of the marionetta-mirror system. The various
spikes in the spectrum are due to contributions to the thermal noise of the violin modes
of the suspension wires. Note that the thermal modes of the cavity mirrors (around 5.5
kHz) play a major role in chapter 3.
Shot noise
Since the arrival times of the photons at the photodiodes are random and can be considered Poisson distributed, the detected power will fluctuate. This noise decreases with
the square root of the number Nγ of photons and thus with the power contained in the
interferometer. An accurate measurement of a gravitational wave requires an accurate
measurement of the phase of the output laser light. For high frequency gravitational
waves, a more accurate measurement of the phase is needed compared to lower frequency gravitational waves. Since the responses of the MI, DLMI and FPMI (Eqs. (2.8),
(2.9) and (2.12)) interferometers are inversely proportional to the gravitational wave
frequency, the accuracy of the phase measurement degrades accordingly. In this region,
the shot noise will dominate due to the intrinsic photon-counting error, known as shot
noise.
2.3
Calibration of h(t) in Virgo
In order to produce a measurement of the strain amplitude of a gravitational wave with
the Virgo detector, a calibration of the output signal is necessary. This calibration is
needed to measure the sensitivity of the interferometer and to reconstruct the strain
amplitude, h(t), from the data stream. The calibration method is described in Ref.
[39] and the implementation in Virgo is described in Ref. [40]. In order to obtain the
sensitivity curve and the reconstructed strain amplitude, the change in arm length must
be related to the variation of the output power of the interferometer. This is done by
first calibrating the length measurement and then measuring the interferometer response
function.
To achieve optimal sensitivity the positions of the mirrors are actively controlled
in such a way that the Fabry-Perot cavities remain on resonance. Furthermore, this
control is designed to keep the output locked on the dark fringe. The mirrors can only
be controlled up to frequencies of a few hundred Hertz. This implies that the active
control of the mirrors will also attenuate the effect of a passing gravitational wave with
frequencies below these few hundreds Hertz. Above a few hundreds Hertz, the mirrors
25
Chapter 2. Gravitational wave detection
behave as free falling masses in the longitudinal direction.
The mirror active control is calibrated by actuating the mirrors and monitoring
the effect in the detector output. The actuation will induce a time dependent length
change of the interferometer arms. This length change is related to the measured phase
change and is given by Eq. (2.5). When purposefully misaligning the input mirrors the
interferometer becomes a simple Michelson interferometer. The measured phase change
is then related to the length change by
δφGW =
4π
∆L(t),
λl
(2.14)
where λl is the laser wavelength and ∆L(t) is the induced differential arm length change
of the interferometer. Since the power on the output photodiode is a function of the
phase difference, Eq. (2.14) can be used to calibrate the length measurement ∆L.
In order to measure the sensitivity of Virgo, the interferometer response function
must be measured. This response function is in essence the transfer function from the
differential arm length variation, ∆L(t), to the output power on the photo diode. This
response function has dimension m/W and is measured by injecting white noise into
the end mirror actuators. Due to the calibration of the actuators described above, the
relation between actuation and length change is known. However, since the longitudinal
controls do not act above a few hundred Hertz, the response function can only be
measured up to these few hundred Hertz. In order to measure the response function
for higher frequencies, a model is used to extrapolate the effect to higher frequencies.
This model is based on an ideal Fabry-Perot interferometer with a finesse of 50. The
Virgo sensitivity curve is measured by taking the spectrum of the output signal (in
W) and multiplying it with the response function. This yields the sensitivity in meters
which will become the sensitivity in strain amplitude when dividing by the Virgo arm
length of 3 km. Note that the sensitivity is measured a few minutes after the response
function to avoid variations of the optical gain from the noise injection. During the first
Virgo science run (VSR1) the errors on the sensitivity below 1 kHz were 4% due to the
mirror actuation calibration. The errors above 1 kHz were higher (5 to 10%) due to the
inaccuracy of the response function model at these frequencies (see Ref. [40]).
For the data analysis the output data stream of the interferometer must be converted
to a gravitational wave strain amplitude stream, h(t). Equation (2.1) states that the
strain is the relative distance variation between two test masses, in this case the endmirrors. Therefore, in order to reconstruct the strain amplitude from the detector output,
the effective differential arm length variation (∆L(t)) is required. As stated before, ∆L(t)
can be computed. However, first the effect of the active controls on the mirrors must be
corrected for. This is done by subtracting the contribution of the injected control signals
from all mirrors from the output signal. This control free signal is then multiplied by
the response function to obtain ∆L(t). Finally this is converted to h(t) by dividing by
the 3 km Virgo arm length. The systematic errors on the h(t) data stream is found to
be in the order of 4% in amplitude and 50 mrad in phase during VSR1 (see Ref. [40]).
26
2.4. Detector beam-pattern
2.4
Detector beam-pattern
Since the detector is not equally sensitive in all directions and gravitational waves have
two polarizations (‘plus’ and ‘cross’), any gravitational wave arriving at the detector can
be written as
h(t) = F+ (t)h+ (t) + F× (t)h× (t),
(2.15)
where h(t) is the detector output, h+ is the plus polarization component of the gravitational wave, h× is the cross polarization component, and F+ and F× are the so-called
beam-pattern functions. These beam-pattern functions represent the detector’s acceptance for any point in the sky. Because of the motion of Earth the beam-pattern functions are periodic functions of time with a period of one sidereal day. The detector
beam-pattern functions are computed in Ref. [41] and can be written as
F+ (t) = sin ζ [a(t, α, δ) cos 2ψ + b(t, α, δ) sin 2ψ] and
F× (t) = sin ζ [b(t, α, δ) cos 2ψ − a(t, α, δ) sin 2ψ] ,
(2.16)
(2.17)
where ζ is the angle between the arms in the interferometer (90◦ in Virgo and LIGO)
and ψ is the polarization angle of the gravitational wave source. The functions a(t, α, δ)
and b(t, α, δ) both depend on the time and location of the source (right ascension α and
declination δ) as well as on the orbital parameters of Earth and the detector coordinates.
Figure 2.9 shows the location parameters of the source as well as the polarization
angle. The coordinate system is such that the gravitational wave propagates in the z
y
Z0
x
ψ
Source
z
δ
Y0
α
X0
Equatorial
referential
Figure. 2.9: The coordinate system (x, y, z) of the source relative to the equatorial coordinate system (X0 , Y0 , Z0 ). Here, α and δ are the equatorial coordinates
of the source and ψ is the polarization angle.
direction. The x and y axes are oriented along the principal axes of deformation of the
+ polarization. The polarization angle, ψ, is the angle between the x axis and the (α,Z0 )
plane and the plane perpendicular to the source.
The acceptance of an interferometer as a function of sky position, time and polarization is often shown as the so-called antenna pattern function. The polarization-averaged
27
Chapter 2. Gravitational wave detection
antenna pattern, < F >ψ , is defined as
1
< F >ψ ≡
2π
Z
2π
0
q
F+2 + F×2 dψ.
(2.18)
δ (rad)
Note that Eq. (2.18) will yield a number between 0 and 1 as a function of sky location.
When combining Eqs. (2.16) and (2.17) into Eq. (2.18) together with the detector coordinates of Virgo, the antenna pattern of Virgo can be computed. Figure 2.10 shows
this antenna pattern as a function of right ascension (α) and declination (δ). Note that
Fig. 2.10 also depends on the time due to Earth’s motion in the solar system. It can be
3
0.9
2
0.8
0.7
1
0.6
0
0.5
0.4
-1
0.3
0.2
-2
0.1
-3
-3
-2
-1
0
1
2
3
α (rad)
Figure. 2.10: The polarization-averaged antenna pattern for the Virgo detector.
seen clearly that there are parts of the sky where the detector is blind and this must be
taken into account when performing any kind of data analysis.
In order to give an estimate of the loss of signal-to-noise ratio for a signal at a random
position in the sky, Eq. (2.18) can be averaged over sky-location as well as over time.
This computation for Virgo yields an average value of 0.42 with a variation of 0.16 over
one sidereal day as computed in Ref. [42].
2.5
Status of Virgo
The Initial Virgo and LIGO detectors have concluded their first science runs. For Virgo
this was called Virgo Science Run 1 (VSR1) which ran from May 18, 2007 to October
1, 2007. After this run the Virgo detector was shut down for an upgrade, designated
Virgo+, and returned to operation in 2009. The second science run (VSR2) ran from July
7, 2009 to January 8, 2010. There was a short science run (VSR3) from August 11, 2010
to October 19, 2010. After the end of VSR3 various small commissioning runs as well
28
2.5. Status of Virgo
Figure. 2.11: The sensitivity curves for Virgo and LIGO during VSR2/S6.
as a minor science run together with GEO600 have been planned. Finally, in September
2011 Virgo+ is planned to be taken off line for a major upgrade called Advanced Virgo.
Advanced Virgo will have a sensitivity increase of an order of magnitude across the
entire frequency interval. In 2014 Advanced Virgo is planned to have first lock meaning
that after a certain commissioning period the science data taking of Advanced Virgo
will start in 2015.
In 2007 both Virgo and LIGO collaborations signed a memorandum of understanding
which effectively merged both collaborations. The goal of this merger was to perform
coincident analyses, increasing the confidence of a possible detection of gravitational
waves. Therefore, each Virgo science run has a LIGO counterpart. For VSR1 it was
LIGO science run 5 (S5) and for VSR2 it was S6.
Figure 2.11 shows the sensitivity of LIGO and Virgo during VSR2/S6 as a function
of frequency. As can be seen in this figure, the Virgo detector is superior to the LIGO
detectors for frequencies below 50 Hz due to the superattenuators. For the intermediate
frequencies, 50 < f < 500 Hz, LIGO has a better sensitivity than Virgo while at high
frequencies the Virgo and LIGO detectors are of comparable sensitivity.
29
Chapter 2. Gravitational wave detection
30
Chapter
3
Thermal Lensing
Stable operation of the Virgo interferometer is limited by wavefront deformations caused
by thermal effects. Even with the use of highest possible quality optics available, the
absorption of laser power within transmissive and reflective optical components together
with the large optical power stored in the various parts of the interferometer will result
in thermal instabilities of the system. The thermal transients complicate the capability
to lock the interferometer.
When injecting the full power of the laser into the interferometer, the stored power
causes the mirrors to heat up in a non-uniform way. Such local heating induces a variation in the optical path length inside the mirror which results in the distortion of the
wavefront. This thermal effect, called ‘thermal lensing’, sets a limit on the laser power
which can be injected into the interferometer. Thermal lensing is an important limiting
effect which will become even more important since Advanced Virgo requires an increase
of both the injected power and the finesse of the Fabry-Perot arm cavities.
The amount of allowable power stored in the interferometer is limited by the optical absorption in the mirrors and their coatings. The temperature dependent index
of refraction and the thermal expansion coefficient of the optical materials (e.g. fused
silica) ensures that non uniform temperature gradients induced by the absorption of the
laser power will result in non uniform optical path length distortions. These path length
distortions will induce a non uniform phase distortion of the laser light. This effect may
limit both the controllability and sensitivity of Virgo.
In general, the total phase distortion, φtrans , for a collimated probe beam transmitted
through a heated optical component and propagating along the z-axis is
Z
2π h(r)
n(x, y, z, t)dz
φtrans (x, y, t) =
λ 0
Z h0
Z h0
2π
≈
n0 h0 +
∆nT (x, y, z, t)dz +
∆nE (x, y, z, t)dz
λ
0
0
2π
+ (n0 uz (x, y, h0 , t) − (n0 − 1)uz (x, y, 0, t))
λ
= φ0 + φT (x, y, t) + φE (x, y, t) + ψ(x, y, t),
(3.1)
31
Chapter 3. Thermal Lensing
where h0 is the nominal distance through the optical component [43], uz (x, y, z, t) is
the net local axial displacement of the material under thermoelastic deformation at
the coordinates (x, y, z) at time t, ∆nT is the local refractive index change induced by
temperature change (the thermooptic effect), ∆nE is the local refractive index change
induced by local thermal strain (the elastooptic effect).
For temperature increases small compared to the ambient temperature, the thermooptic effect can be written as
∆nT =
dn
∆T,
dT
(3.2)
dn
where dT
is the refractive index derivative with respect to temperature (thermooptic
coefficient) and ∆T is the temperature deviation from the ambient temperature. When
assuming cylindrical symmetry and polarization along the x-axis, the elastooptic effect
can be approximated by
∆nE = −pxx α∆T,
(3.3)
where pxx is the component of the elastooptical tensor in the direction of the polarization
and α is the thermal expansion coefficient. Combining Eqs. (3.1), (3.2) and (3.3) yields
∆S ≡
λ
(φtrans − φ0 ) = ∆ST + ∆u + ∆SE
2π
Z h0
dn
∆T dz,
=
+ α − αpxx
dT
0
(3.4)
where ∆S is the total change in optical path length, ∆ST is the change in optical path
length induced by the thermooptic effect, ∆u is the change in optical path length induced
by the thermal expansion of the component and ∆SE is the elastooptic effect.
Finite Element Analysis (FEA) yields a general, precise, and rapid means of solving
the partial differential equations which govern heat transfer and thermal expansion for
individual optical components.
The results of a FEA to compute the temperature field T (x, y, z, t) in the Virgo input
mirrors are presented in this chapter. The temperature field has been calculated for two
cases.
ˆ The steady state case: The temperature of the mirror is stationary as absorbed
power heating of the mirror is in equilibrium with the heat transfer by radiation.
ˆ The transient case: The temperature of the mirror is not stationary.
The computation of the temperature field allows for the calculation of the induced
change in optical path length as well as the temperature dependence of the mirror resonance modes. This Finite Element Model (FEM) has been developed with the COMSOL
multiphysics package [44].
32
3.1. The finite element model
3.1
The finite element model
Thermal lensing of the mirror is analyzed by solving the heat equation,
ρC
∂T
+ ∇ · (−k∇T ) = Q,
∂t
(3.5)
for a given power density due to absorption of laser power1 where T is the temperature, ρ the density, C the heat capacity, k the thermal conductivity and Q is the heat
source. Note that heat transfer by radiation is not shown in this equation but is taken
into account as boundary conditions in the full simulation. The power emitted through
radiation by an object of surface A and with emissivity is given by Stefan-Boltzmann’s
law
Prad = Aσ T 4 − T04 ,
(3.6)
where σ = 5.67×10−8 Wm−2 K−4 is the Stefan-Boltzmann constant and T0 is a reference
temperature which is taken to be room temperature (300 K).
A dynamical thermal analysis is performed which includes the heat radiation exchange with its environment, including the recoil mass of the mirror. The resulting
temperature distribution evolution is used to predict the change in optical and mechanical properties of the mirror at different time scales. These include its thermal expansion,
optical path length and the frequencies of the vibrational modes.
Figure 3.1 shows the meshed 3D-models of the geometry used for the analysis with
and without recoil mass. A mapped mesh has been used to mesh the mirror geometry
with cubic elements. The element density has been chosen higher in the center since
that is where the thermal gradient will be the highest. Finally, the recoil mass has been
meshed in a coarse way since it is only used as a heat-reflective shield. The geometry
Y
Y
Z
X
Z
X
Figure. 3.1: 3D models used in the FEA. The left panel shows the mesh model of the
input mirror and the right panel shows the mesh model for the input mirror and recoil mass.
of the input mirror is modeled as a solid cylinder without any internal structure like
1
Reference [45] gives an analytical solution to a similar problem.
33
Chapter 3. Thermal Lensing
coating layers2 . This is justified by considering that the combined radial heatflow is
much smaller than the combined axial heatflow since the summed thickness of the coating layers is much smaller than the radius of the mirror (Table 3.1). Furthermore, the
emissivity of the mirror is treated as a measured intrinsic quantity which means that
the internal structure of the mirror becomes irrelevant in terms of radiation.
In the actual setup, the mirror is suspended from a large mass by two thin steel wires.
The heat transfer through these wires has been neglected as well as the effect they might
have on the eigenfrequencies of the mirror. The power dissipated in the substrate is given
by
dPs
Ps = Plaser
ts R ≈ 1.94 mW,
(3.7)
dz
s
where Plaser represents the laser power, dP
the fractional loss (in ppm/cm) in the subdz
strate, ts is the thickness of the mirror and R the power recycling factor of the power
recycling mirror. The beam splitter contributes a factor of 12 , which is canceled by the
fact that the beam passes twice through the substrate. The power dissipated in the
coating is given by
F
Pc = Plaser δPc R ≈ 5.73 mW,
(3.8)
π
where δPc is the fractional power loss (in ppm/bounce) in the coating and F the finesse
of the Fabry-Perot cavity.
From the calculated amount of dissipated power, the heat flow for the two losses can
be calculated. The power density in the substrate is modeled in accordance with the
Gaussian shape of the beam with waist w. The power density in the coating is similarly
modeled. The total absorbed power relates to the respective heat flows as
ZZZ
Ps =
2 +y 2
w2
−2 x
Qs e
ZZ
dV,
and
Qc e−2
Pc =
V
x2 +y 2
w2
dS.
S
Here, V and S denote a volume and surface element, respectively and Qs and Qc are the
induced power densities for the substrate and the coating, respectively. A calculation
with as input the parameters from Table 3.1 yields
Qs =
W
2Ps
≈
26.6
ts πw2
m3
and
Qc =
2Pc
W
≈
7.50
,
πw2
m2
(3.9)
Note that these estimated power densities are based purely on the nominal design values
of Virgo for the given finesse and expected losses in the substrate and coating. It has
been shown that the actual losses are significantly higher than expected. This may be
due to damage and/or pollution of the mirrors and warrants further investigation.
2
The reflective coating of the input mirror consists of alternating layers made of SiO2 and Ta2 O5
each approximately 0.2 µm thick.
34
Value
0.175
0.096
8
0.022
36
50
300
2203
73.2
0.164
739
1.36
0.89
0.7
0.54 × 10−6
9.80 × 10−6
1.5 × 10−4
7.1 × 10−5
33
0.75
1.25
2700
160
900
0.2
Symbol
r
ts
Plaser
w
R
F
Tbg
ρs
E0
ν0
cs
ks
dPs /dz
α
dn/dT
(1/E0 )dE/dT
(1/ν0 )dν/dT
kc
tc
δPc
ρr
kr
cr
r
W/(m K)
µm
ppm/bounce
kg/m3
W/(m K)
J/(kg K)
K−1
ppm/cm
J/(kg K)
W/(m K)
K
kg/m3
GPa
W
m
[46]
m
The variation of the Poisson ratio has been calculated taking the derivative of the formula for ν given in Ref. [51], which yields
1 dG
1 dG
(ν + 1) E1 dE
dT − G dT , where G represents the shear modulus. The value for G dT is also taken from Ref. [51].
a
dν
dT
=
[52]
[48]
[49]
[50]
[50, 51]a
[48]
[46]
[48]
[47]
Source
Unit
Table. 3.1: The parameters used in the FEA. The value for the emissivity of oxidized aluminum was used for the recoil mass.
Recoil mass properties (Al)
Coating properties (Ta2 O5 & SiO2 )
Substrate properties (SiO2 )
ITF properties
Geometry
Property
Mirror radius
Mirror thickness
Recoil mass parameters
Laser power
Beam width
Power recycling factor
Fabry-Perot finesse (end cavity)
Background temperature
Density
Young modulus
Poisson ratio
Heat capacity
Thermal conductivity
Emissivity
Power loss
Thermal expansion
Thermooptic coefficient
Young modulus per Kelvin
Poisson ratio per Kelvin
Thermal conductivity
Summed thickness
Relative power loss
Mass density
Thermal conductivity
Heat capacity
Emissivity
3.1. The finite element model
35
Chapter 3. Thermal Lensing
3.2
Simulation results
3.2.1
Steady state thermal analysis
Results of the 3D FEA for the equilibrium state for both the power densities Qc and Qs
separately and summed are presented in Fig. 3.2. The figure shows that the contributions
x
x
z
z
x
z
Figure. 3.2: Temperature increase half-slices at y=0. The top left panel shows the temperature increase due to losses in the coating, the top right panel shows
the temperature increase due to losses in the substrate and the bottom
panel shows the total temperature increase induced by the sum of the two.
to the total temperature distribution for the coating and for the substrate differ by
almost an order of magnitude. The maximum temperature increase is obtained when
both Qc and Qs are turned on: Tmax − T0 = ∆Tmax ≈ 75 mK for the mirror with recoil
mass.
36
3.2. Simulation results
The average mirror temperature increase < ∆T > is defined as
RRR
∆T dV
V
< ∆T >= RRR
.
dV
V
(3.10)
Table 3.2 presents the average temperatures for the different cases with or without recoil
mass. It is shown that the recoil mass has a significant effect on the global temperature
Without recoil mass Including recoil mass
< ∆T >s (mK)
< ∆T >c (mK)
< ∆T >c+s (mK)
1.47
4.08
5.54
2.26
6.28
8.52
Table. 3.2: Average mirror temperatures < ∆T >c , < ∆T >s and < ∆T >c+s calculated under Qc , Qs and Qc + Qs loads.
increase. Assuming a linear response to the small power densities we can calculate the
respective thermal resistances using < ∆T >c+s from Table 3.2 and the value for Tmax
mentioned earlier. It follows that
d < ∆T >
≡ 1.11 K/W
dQ
and
dT max
= 9.78 K/W.
dQ
In order to quantitatively show the temperature distributions along key areas on
and in the mirror, four sample curves have been chosen where the (one dimensional)
equilibrium temperature field will be drawn. Figure 3.3 schematically shows these key
positions as bold lines. Figure 3.4 shows the equilibrium temperature distributions for
Figure. 3.3: The four mirror lines where the temperature distribution is calculated for
the case of thermal equilibrium: front radial, back radial, lateral and axial.
the mirror with and without recoil mass. It can be seen from the radial profiles that the
temperature of the mirror locally increases a few mK due to the thermal shielding of
37
Chapter 3. Thermal Lensing
Figure. 3.4: One dimensional temperature profiles when the mirror is in thermal equilibrium. The top left and right panels show the radial front and radial
back profiles while the bottom left and right panels show the lateral and
axial profiles as shown schematically in Fig. 3.3. The red curves with the
box-markers are the profiles computed with the recoil mass included in the
FEA and the blue curves with the circle-markers are computed without
the recoil mass.
the recoil mass. For the mirror’s outer edge (the lateral profile shown in Fig. 3.4) this
effect is more pronounced as its temperature increase is roughly doubled by the recoil
mass shielding.
38
3.2. Simulation results
3.2.2
Eigenmodes in thermal equilibrium
The eigenmodes of the Virgo input mirrors have been calculated. Figure 3.5 outlines the
two principal eigenmodes, the ‘Butterfly’ and the ‘Drum’ mode. Note that these results
are in agreement with those of previous calculations [46]. To first order the eigenfre-
X
X
Z
Z
Figure. 3.5: Principal resonance modes for a uniform temperature distribution. The
left panel shows the butterfly resonance mode predicted at 3904.0 Hz and
the right panel shows the drum resonance mode predicted at 5578.1 Hz.
√
quency is proportional to the square root of the stiffness: f ∝ E. When defining
∆f ≡ f (T = T0 + < ∆T >) − f (T = T0 ) where f (T ) is the eigenfrequency for temperature T , < ∆T > is the average mirror temperature increase as defined in Eq. (3.10) and
T0 is room temperature, it can be derived that
∆f ≈ f (T0 ) < ∆T >
1 1 dE
.
2 E0 dT
(3.11)
In Eq. (3.11) E0 is the Young modulus at room temperature. With the parameters from
Table 3.1, Eq. (3.11) yields
∆f = f (T0 ) < ∆T > × 7.5 × 10−5 K−1 ,
(3.12)
where f (T0 ) is the eigenfrequency without thermal load. From Eq. (3.12) it is possible
∆f
, for both modes
to predict the frequency dependency on the average temperate, <∆T
>
as
∆f
∆f
= 0.42 Hz/K and
= 0.29 Hz/K.
< ∆T > drum
< ∆T > butterfly
The eigenfrequencies can also be studied in a FEA with the temperature load obtained
from the equilibrium analysis. The resulting frequency shifts together with the change
in frequency per Kelvin are listed in Table 3.3. From the results shown in Table 3.3
it can be seen that the difference in frequency due to the temperature increase of the
eigenmodes is in the order of a few millihertz for initial Virgo operation. Also there is
∆f
for the drum mode with or without recoil
a noticeable difference in the value for <∆T
>
mass. Considering the fact that the only difference between the two situations is that
the temperature distributions are different, this suggests that the drum mode is more
39
Chapter 3. Thermal Lensing
Mode
Drum
Butterfly
Drum + recoil mass
Butterfly + recoil mass
∆f (mHz) ∆f / < ∆T > (Hz/K)
3.38
1.81
4.57
2.64
0.61
0.34
0.53
0.30
Table. 3.3: Temperature dependence of the resonance frequency of the various eigenmodes.
sensitive to a non-uniform temperature distribution.
Comparing the calculated numbers originating from Eq. (3.12) with the numbers in
Table 3.3 shows that the values obtained without recoil mass differ from the theoretical
values. The difference seems to decrease when considering the addition of the recoil
mass.
Previous work on the temperature dependence of the eigenfrequencies of the two
principal eigenmodes has been done and is presented in Refs. [48, 53]. In that work the
recoil mass was omitted and uniform heating of the mirrors was assumed. The present
work improves on the previous results by including the recoil mass and computing the
temperature dependency of the eigenfrequencies of both principal modes for non-uniform
temperature distributions. The values found here are in qualitative agreement with the
previous work done.
3.2.3
Transient temperature analysis
The temperature as a function of time for different points of the mirror has been calculated using the FEM. The transient temperature behavior is sampled at three different
points on the mirror, shown in Fig. 3.6. Figure 3.7 shows this transient behavior for each
Figure. 3.6: The three mirror positions where the temperature as a function of time is
shown in Fig. 3.7.
of the mirror positions listed. From this figure it can be seen that thermal equilibrium
40
3.2. Simulation results
Figure. 3.7: Evolution of the temperature in time on various points on the mirror. The
top left, top right and bottom panels show the temperature evolution on
the circumference, the beam spot and the rear of the mirror, respectively.
The data points are obtained from the transient FEA while the curves are
fits of Eq. (3.13). The blue curves represent the mirror without recoil mass
and the red curves the mirror with recoil mass.
is reached after a certain time. This equalization process can be modeled by
f (t) = A(1 − e−t/τ ),
(3.13)
where A is the equilibrium value and τ the time constant. Table 3.4 shows the resulting
time constants when fitting Eq. (3.13) to Fig. 3.7. Due to the complex structure of the
mirror and recoil mass system various time constants are expected. These time constants
pose a challenge for the control systems used to operate the Virgo detector. The time
constant of the temperature in the beam spot is determined by the beam waist and the
mirror’s thermal conductivity. The front side of the mirror will dissipate the heat much
faster through conduction to the substrate than by thermal radiation. This implies that
the time constant in the beam spot is much smaller than on the other places places on the
mirror. The rest of the mirror’s heat dissipation is dominated by radiation and therefor
the other time constants will be larger (in the order of hours). The relative difference
between the time constants with and without recoil mass on the circumference of the
mirror is smaller than at the back of the mirror. This is because the recoil mass does not
41
Chapter 3. Thermal Lensing
location
Circumference
Beam spot
Rear
τno recoil (h) τrecoil (h)
5
0.08
3
6
0.08
4
Table. 3.4: Time constants for the temperature stabilization on various points on the
mirror. These time constants have been computed by fitting Eq. (3.13) to
the temperature evolution which have been obtained from the FEA for the
various locations on the mirror.
cover the entire back of the mirror (as can be seen in the geometry in Fig. 3.2) and thus
has a smaller effect on the thermalization. Also, the equilibrium temperature is lower
than on the front side, which is to be expected for the same reason.
The time evolution of the average temperature with and without recoil mass is shown
in Fig. 3.8. From Figs. 3.7 and 3.8 it can be concluded that the recoil mass has a non
Figure. 3.8: The average temperature of the mirror as function of time. The blue curve
represents the mirror without recoil mass and red curve the mirror with
recoil mass. The time constants are computed to be τrecoil = 6 h and
τno recoil = 4 h.
negligible effect on the long term behavior of the mirror as well as on the equilibrium
temperature.
42
3.3. Optical path length
3.3
Optical path length
The local heating of the mirror will cause a variation in optical path length for the light
interacting with the mirror. This may cause problems for the interferometer locking
system at relatively hight cavity powers. The change in optical path length (∆S) can be
written in terms of the thermooptic effect and the thermoelastic effect as described in
Eq. (3.1). For this analysis the thermooptic effect is much stronger than the elastooptic
effect and the latter can be neglected. The change in optical path length at distance r
from the axis can be calculated by numerically integrating the temperature distribution,
∆T , along z by using Eq. (3.14).
∆S(r) =
dn
+α
dT
Z
ts
∆T (z, r)dz.
(3.14)
0
A measure for the severity of the thermal lensing effect in the input mirror is the effective
radius of curvature [54]. A first order approximation for the effective radius of curvature3
for the heated mirror and assuming that the mirror is initially flat, is given by
R=
w2
,
2δS
(3.15)
where w is the waist of the beam and δS is the relative change in optical path length
between r = 0 and r = w given by
δS ≡ ∆S(0) − ∆S(w).
Figure 3.9 shows the change in optical path length versus the radial coordinate for
a beam traveling through the mirror parallel to the optical axis. By using Eq. (3.15)
in combination with the data shown in Fig. 3.9, the effective radius of curvature is
calculated to be 21 km4 .
Figure 3.10 shows the transient behavior of the change in relative optical path length
(δS) together with that of the radius of curvature. The value of the time constant
roughly corresponds to the time constant calculated on the beam spot of the mirror,
τbeamspot = 0.08 h (see Table 3.4). Figure 3.10 also shows that the influence of the recoil
mass is negligible on the surface of the mirror in the region between r = 0 and r = w.
This means when using Eq. (3.15) to calculate the effective radius of curvature, the recoil
mass can be neglected. This can be understood by comparing the steady-state change
in optical path length as a function of r for the case with and without the recoil mass
as shown in Fig. 3.9. In this figure the two curves seem to differ by a constant offset,
especially for r < w. This implies that the radius of curvature will be the same for both
cases which is to be expected since the addition of the recoil mass mainly increases the
temperature in a uniform way.
3
This approximation assumes that the change in optical path length in the region 0 < r < w can be
described by a circle with radius R.
4
For comparison: the end-mirrors have a radius of curvature of 3.5 km.
43
Chapter 3. Thermal Lensing
Figure. 3.9: Change in optical path length, ∆S as function of the radial coordinate
in thermal equilibrium for the mirror with and without the recoil mass
included in the simulation.
Figure. 3.10: Left: the time evolution of δS with τrecoil = τno
time evolution of effective radius of curvature.
3.4
recoil
= 0.1 h. Right: the
Thermal lensing effects in Virgo
In section 3.2.2 the frequency of the two principal resonant modes (drum and butterfly)
of the Virgo input mirrors have been computed. Furthermore, the change in frequency
as a function of average mirror temperature has been computed. With this result it is
possible to infer the average mirror temperature by monitoring the frequency change of
these modes.
Figure 3.11 shows the power spectra of the output data of the Virgo detector. The
drum modes of both input mirrors, shown in the left panel, can be seen as the double
44
3.4. Thermal lensing effects in Virgo
10-32
10-33
10-34
10-35
10-36
10-37
10-38
10-39
10-40
10-41
10-42
10-43
10-44
Sn (Hz-1)
Sn (Hz-1)
10-32
10-33
10-34
10-35
10-36
10-37
10-38
10-39
10-40
10-41
10-42
10-43
10-44
3860
3880
3900
3920
3940
f (Hz)
5540
5560
5580
5600
f (Hz)
Figure. 3.11: The PSDs taken from Virgo at GPS time 947284200 (VSR2) created
from the h-reconstructed data, channel ‘V1:h 16384Hz’. The left panel
shows the frequency interval corresponding to the butterfly modes and
the right panel shows the frequency interval corresponding to the drum
modes.
peaks centered around 5543.2 Hz and 5545.5 Hz, respectively. The double peaks centered around 5584.5 Hz and 5586.3 Hz, respectively represent the drum modes of the
end mirrors. The difference between the two sets of peaks are due to the fact that the
end mirrors are slightly thinner. In the right panel of Fig. 3.11 the same double peak
structures corresponding to the butterfly modes can be seen. In this case the input mirrors modes are centered around 3884.2 Hz and 3885.2 Hz while the end mirror modes are
centered around 3917.7 Hz and 3918.8 Hz. The simulated values for these modes, shown
in section 3.2.2 deviate at maximum 0.6 % and 0.5 % from the actual values for the
drum and butterfly modes, respectively. Note that the difference in absolute frequency
between the end and input mirrors of the modes are small and thus the temperature of
all mirrors can be inferred by using the results of this simulation.
From the temperature dependence of the modes together with the output of the interferometer, the average temperature of the mirror can be inferred. A method to measure
the frequency of the modes and convert them to average temperatures is presented in
Ref. [48]. In order to verify the simulated values of the temperature dependence of the
mirror modes shown in section 3.2.2, the long term temperature evolution of the West
End (WE) building is compared to the temperature of the mirror as computed from the
resonant modes. The result is shown in Fig. 3.12. It can be seen that the simulated values of the frequency evolution of the drum and the butterfly modes are consistent with
the global temperature. Furthermore, plots like Fig. 3.12 can be used to identify which
resonance corresponds to which mirror as the frequency of the modes can be linked to
temperature sensors in the various buildings.
For comparison it is noted that when measuring the power contained in the FabryPerot cavities and using this as input for the FEM, it is possible to simulate the heating
and cooling cycles of the mirrors. This has been done with a more simple FEM discussed
in Ref. [55] (Punturo’s model) which did not include a recoil mass. The result of this
45
∆ T (K)
Chapter 3. Thermal Lensing
0.4
0.35
0.3
WE temperature
Drum mode
Butterfly mode
0.25
0.2
0.15
0.1
0.05
0
-0.05
0
20
40
60
80
100
120
time (h)
Figure. 3.12: The deviation of the temperature from room temperature (300 K) of the
West End (WE) mirror computed from the frequency of the drum mode
(‘+’-sign) and the butterfly mode (the dots). The solid curve shows the
global temperature evolution in the WE building as measured by temperature sensor ‘EM TE SUWE10’. In this plot, the 0 hour point corresponds to GPS time 840240000. The data from the Drum and Butterfly
modes have been taken from Ref. [55] and have been rescaled such that
the more accurate results of the temperature dependence of the eigenfrequencies computed for non-uniform heating and with a recoil mass have
been used.
analysis compared to the temperature of the mirror which is computed from the drum
mode for the West Input (WI) mirror is shown in Fig. 3.13. From this figure it can be
seen that the simulation agrees reasonably well with the data. However, the measured
mirror temperature if roughly a factor of 10 larger than the simulated temperature.
A similar study shows that the temperatures of the North Input (NI) mirror are approximately a factor 4 higher than the nominal value. These results indicate a larger
absorption of laser light than previously assumed (Table 3.1). This discrepancy should
be investigated with a more accurate FEM which includes the recoil mass. Furthermore,
in order to estimate the actual losses in the mirror, more FEM studies are needed which
should include the recoil mass.
To actively control the length of the cavities, the input beam is phase modulated
and the reflected light is examined to obtain an error signal. This modulation is done at
a certain frequency ωm and will result in taking power from the carrier at frequency Ωl
and splitting this into two sidebands at frequencies Ωl + ωm and Ωl − ωm . This modulation frequency is chosen such that the sidebands are antiresonant in the optical cavity
when the carrier is resonant. This implies that if the cavity is slightly shorter than on
resonance, the reflected carrier suffers a small phase shift and will combine with the
46
3.4. Thermal lensing effects in Virgo
Figure. 3.13: The average temperature deviation from room temperature as computed
with Punturo’s model (solid curve) and the average mirror temperature
computed from the frequency of the drum mode. This FEM uses the
measured power stored in the Fabry-Perot cavity as input and the mirror temperature is computed with the old uniform value of 0.61 Hz/K, as
derived in Ref. [53].
fully reflected sidebands. The combined signal will oscillate at ωm out of phase with the
original modulation signal. If the cavity is slightly longer it will oscillate in phase with
the modulation signal.
It has been shown in Ref. [43] and [56] that the sidebands are especially sensitive to
thermal lensing effects. The heating of the mirrors will cause the sidebands to be lost
which will unlock the interferometer. These losses are especially strong in the recently
proposed Marginally Stable Recycling Cavities (MSRC) optical scheme for Advanced
Virgo. The loss of these sidebands seems to be governed by a series of time constants.
These could be linked to the simulated time constants shown in section 3.2.3. However,
since this loss of sideband power has to do with the actual phase of the carrier and sidebands a more advanced study including the phase of both beams should be performed.
A preliminary study is performed in Ref. [57].
47
Chapter 3. Thermal Lensing
In order to compensate these thermal lensing effects in Virgo, a ‘Thermal Compensation System’ (TCS) has been developed (see Ref. [58]). This TCS consists of a CO2
laser which is tuned to emit a ring-shaped beam which is incident on the mirror in
order to locally heat the mirror such that the wavefront distortions will be negated. A
first TCS for the input mirrors has been installed in Virgo in April 2008. Note that
more stringent constraints must be placed on Virgo’s thermal compensation system in
order to guarantee successful operation at high power with the MSRC optical scheme of
Advanced Virgo.
3.5
Conclusion and outlook
A finite element model has been described that models the Virgo input mirrors as a
cylinder and its recoil mass with the dimensions as realistic as possible. With the assumed power dissipated in the substrate and coating, Ps = 1.94 mW and Pc = 5.73 mW,
the corresponding power densities on the substrate and the coating can be calculated:
Qs = 25.6 W/m3 and Qc = 7.50 W/m2 , respectively. With these numbers, the temperature evolution in the mirror can be obtained taking into account the heat transfer via
radiation.
In thermal equilibrium the volume-average temperature increase is 8.52 mK and 5.54
mK for a model with or without recoil mass. From this difference it can be inferred that
the recoil mass acts as a heat-reflective shield and must be included in realistic calculations.
The eigenmodes of the mirror are a function of the temperature distribution, since
both the Young modulus and the Poisson ratio depend on temperature. Eigenfrequencies
with and without a thermal load have been calculated for the non-uniform temperature
distributions. The change in frequency per unit temperature has been calculated for the
drum and butterfly mode. The results are 0.53 Hz/K for the drum mode and 0.30 Hz/K
for the butterfly mode.
It is important to realize that the FEA result significantly differs from the case of
uniform heating of the mirror: 0.42 Hz/K for the drum mode and 0.29 Hz/K for the
butterfly mode. The value for the butterfly mode only deviates 3% from the uniform
heating value while the drum mode deviates 26%. The butterfly mode is less sensitive
to a Gaussian-like non-uniform temperature distribution compared to the drum mode
due to the spatial shape of the mode. These results allow measurement of the average
mirror temperature by measuring the frequency change of the drum and butterfly modes
together with the computed temperature dependence of the frequency of these modes.
The simulation has shown that the addition of the recoil mass has a non-negligible effect
on this temperature dependence.
The transient temperature analysis reveals the different time scales of the system,
varying from several minutes for the temperature in the beam spot to many hours for
the global temperature. The presence of the recoil mass increases the long term time
constant from 4 to 6 hours.
With the temperature information gained in the equilibrium and transient analyses,
48
3.5. Conclusion and outlook
the change in optical path length and the effective radius of curvature can be calculated.
The change in optical path length in equilibrium along the optical axis varies between
0 and 24.5 nm. The effective radius of curvature in equilibrium is calculated to be 21
km. The transient analysis shows that the time constant involved is 0.1 h. This result
has been obtained by using an approximation for the effective radius of curvature which
assumes circular deformation of the optical path length around the optical axis.
All simulations have been done assuming that = 0.2 for the emissivity of the recoil
mass (a typical number for the emissivity of oxidized aluminum). This number should
be verified since it impacts the simulation.
It has been shown that according to the measurement of the average temperature of
the mirrors that the optical losses are significantly higher than the expected values (up
to an order of magnitude). In order to measure the actual optical losses in the mirror
it may be possible to use the FEA together with the measured power in the cavity as a
function of time. The resulting temperature can then be compared to the measured average temperature of the mirror. However, such an analysis would require further study.
The sidebands have been shown to be extremely sensitive to the thermal lensing effect
due to the observation of various time constants in the measured optical power in the
sidebands. Further study is needed to determine the exact nature of this behavior and
the origins of these time constants.
49
Chapter 3. Thermal Lensing
50
Chapter
4
Continuous waves analysis methods
In section 1.3.1 it has been discussed that non-axial symmetric neutron stars are the
most promising continuous sources of GWs in terms of expected amplitude. From Eq.
(1.15) it is apparent that the typical strain may be in the order 10−27 to 10−24 . Figure
2.11 shows that the power spectral density of the LIGO and Virgo interferometers is
3 × 10−23 Hz−1/2 and 6 × 10−23 Hz−1/2 respectively in the most sensitive region. Thus
the GW amplitude is weak compared to the noise level.
Continuous GWs can be integrated over longer periods of time. This is optimally
done via matched filtering as explained in appendix B. The signal-to-noise (SNR) ratio
for the optimal filter h̃(f ) is
2
Z ∞
S
|h̃(f )|2
2
SNR =
=4
df,
(4.1)
N
Sn (f )
0
where Sn is the single sided power spectral density of the noise. Matched filtering is often
referred to as ‘coherent integration’. For a search based √
on this coherent integration
technique, run over a coherent time T , the SNR scales as T .
The scaling defined above can be used to find signals with amplitudes which are
weak compared to the noise level of the detector. For example, an integration
time of 1
√
year allows for the detectability of GW amplitudes hmin = (2 × 10−4 ) × Sn . However,
this sensitivity will only be achieved if an optimal filter, also called a ‘template’, can
be constructed and compared to the data. In order to construct this template, the
physical parameters of the system must be known to sufficient precision such that it will
remain coherent with the signal over the entire observation time. When the parameters
are not known, or not known well enough, a series of templates can be constructed,
compared to the data and a significance can be assigned to each of these templates.
For a small number of templates and with certain parameters constrained well enough,
this template-based search is the preferred analysis method. These types of searches are
referred to as ‘coherent searches’ and will be covered in more detail in section 4.2.
The template-based search is not ideal when performing a search for GWs from
neutron stars of which the parameters are (partially) unknown. The number of templates
scales as a power of the coherence time and renders such a search computationally
unfeasible for coherence times above a few days. In order to perform a blind all-sky
51
Chapter 4. Continuous waves analysis methods
search, shorter coherent sections
√ of the data must
√ be combined. In general, the sensitivity
of such methods will scale as 4 T instead of T in the case of coherent searches. These
methods are known as ‘semi-coherent searches’ and will be covered in section 4.3.
4.1
The gravitational wave signal model
If the GW signal from a neutron star were truly monochromatic, then a simple delta
function in the frequency domain would be the optimal filter. The data analysis would
then be trivial as a sufficiently long Fourier transform would be able to pick the signal
out of the background. However, as discussed in section 1.3, pulsars suffer from glitches
and spin-down. Furthermore, the signal originating from the neutron star will be Doppler
shifted by Earth’s motion (daily and yearly). Additionally, when the neutron star is in a
binary system, the orbit of the neutron star will introduce Doppler shifts. Moreover, the
detector does not have a uniform sensitivity over the entire sky, as discussed in section
2.4. A signal model can be constructed which includes all these previously mentioned
effects. This signal model can be used to construct templates for a coherent search.
When introducing χ as the angle between the deformation and rotation axis of a
neutron star and ι as the angle between the rotation axis and the propagation direction,
the GW signal from Eq. (2.15) can be written in terms of the GW amplitude h0 (see
Eq. (1.15)) as shown in Ref. [41] as
h(t) = F+ (t)(h1+ (t) + h2+ (t)) + F× (t)(h1× (t) + h2× (t)),
(4.2)
where h1 and h2 are used to decompose the signal into Φ and 2Φ components and are
given by
1
h0 sin 2χ sin 2ι cos Φ(t),
8
1
h2+ (t) =
h0 sin2 χ(1 + cos2 ι) cos 2Φ(t),
2
h1+ (t) =
(4.3)
(4.4)
and
1
h0 sin 2χ sin ι sin Φ(t),
4
h2× (t) = h0 sin2 χ cos ι sin 2Φ(t).
h1× (t) =
(4.5)
(4.6)
In Eq. (4.2), t is the time in the detector frame, Φ(t) is the phase of the gravitational
wave and F+ (t) and F× (t) are the beam-pattern functions of the interferometer and are
discussed in section 2.4. Note that the beam-pattern functions depend on the polarization
angle ψ as well as on the sky-coordinates of the source, α and δ. Furthermore, the
functions h1+ , h2+ , h1× and h2× depend on the phase Φ(t) of the GW.
Within good approximation, isolated neutron stars (e.g. neutron stars without a
companion) will have a constant velocity with respect to the solar system barycenter
(SSB). For these systems it is sufficient to consider only the Doppler shifts due to Earth’s
daily and yearly motion through the solar system. Figure 4.1 shows a diagram the
52
4.1. The gravitational wave signal model
Figure. 4.1: Diagram of the relevant position vectors of an isolated neutron star relative to the detector. The origin is located at the SSB, ~k is the GW vector,
~rd is the position of the detector on Earth, ~vd is the velocity vector of the
detector relative to the SSB and ~n0 is the position vector of the neutron
star N S.
relevant position vectors for an isolated neutron star system. When assuming a neutron
star emitting GWs at constant frequency f0 (e.g. no spin-down), the instantaneous
Doppler shifted frequency f is
f = f0
~vd · n̂0
1+
c
,
(4.7)
where ~vd is the velocity vector of the observer with respect to the source and c is the speed
of light. If the product ~vd ·n̂0 is constant (e.g. the neutron star has a constant velocity and
Earth would stand still), the analysis would be trivial again since the detected frequency
would simply gain a constant offset. However, the velocity of the detector with respect
to the source changes in time due to Earth’s rotation and its revolution around the Sun.
The change in detected frequency over integration time T can be computed to be
(∆f )Doppler = f0
(∆vd )T
,
c
(4.8)
where (∆v)T is the change in velocity during the integration time. In order to retain the
strength of the signal in a single frequency bin with size ∆f = 1/T one can calculate the
maximally allowed coherence time. This is done by requiring that the frequency shift is
at most half a frequency bin. Equating the Doppler shift from Eq. (4.8) to ∆f /2 and
solving for T together with taking the detector at latitude 40◦ and R⊕ = 6.38 × 106 m
53
Chapter 4. Continuous waves analysis methods
will yield
1/2
1 kHz
≈ 30 min
,
f0
1/2
1 kHz
≈ 120 min
,
f0
rot
Tmax
orb
Tmax
(4.9)
(4.10)
rot
orb
for the daily (Tmax
) and yearly (Tmax
) motion. Equation (4.9) shows that the maximum
integration time when performing a simple Fourier transform is about 30 minutes for a
1 kHz GW. Furthermore, when choosing the location of the source in such a way that
the Doppler shift is maximal, the maximum frequency variation for both the yearly and
daily motion can be computed to be
f0
rot
−3
(∆f )max ≈ 2.4 × 10 Hz
,
(4.11)
1 kHz
f0
orb
(∆f )max ≈ 0.2 Hz
.
(4.12)
1 kHz
The signal model should contain the phase modulation introduced by these Doppler
shifts. This it done by computing the phase with the relation
Z
Φ(t) = 2π f (t)dt.
(4.13)
In the case of binary systems the orbit of the neutron star around its companion
introduces additional Doppler shifts, depending on the orbital parameters of the neutron
star. The maximum coherence time such that the signal from a neutron star in a binary
system will remain in a single frequency bin is computed in Ref. [59] to be
2/3 −1/6 −1/2
(1 + q)1/3
P
mNS
f
kepler
Tmax = 131.6 sec
, (4.14)
√
q
1 day
1.4 M
1 kHz
where q = m2 /mNS represents the ratio of the masses of the two objects, P the period of
the binary, mNS the mass of the neutron star and f the GW frequency. For an example
binary system with parameters m2 = mNS = 1.4 M , P = 2 h and f0 = 1 kHz we find
kepler
= 20 s, much less than for isolated neutron stars. This coherence time allows for
Tmax
only a marginal gain of the signal-to-noise ratio compared to the PSD of the detector.
Section 5.1 gives a detailed description of the Doppler shifts induced by a neutron star
in a binary system.
As discussed in section 1.3, pulsars tend to lose angular momentum and spin-down.
These spin-down values are given in terms of the derivative of the rotation frequency, f˙.
This means that the frequency of the signal changes by an amount of ∆f = T f˙ during
observation time T . The maximum coherence time can then be parameterized as
s
10−9 Hz/s
spindown
≈ 8.78 hour
Tmax
.
(4.15)
−f˙
54
4.1. The gravitational wave signal model
An optimal filter has to take into account all Doppler shifts as well as the spin-down
in case a coherence time larger than 9 hours is required. This spin-down can be taken
into account by adding so-called spin-down parameters to the phase model. These spindown parameters are simply the direct derivatives of the rotational frequency and can
be included up to certain order.
The phase of the GW takes it simplest form when the time coordinate used is tNS ,
the proper time in the neutron star rest frame. The phase in this frame, ΦNS , is written
as
s
(n)
X
fNS
tn+1
(4.16)
ΦNS (tNS ) = Φ0 + 2π
NS ,
(n
+
1)!
n=0
(0)
where Φ0 is the initial phase, fNS is the instantaneous frequency of the GW in the
(n)
NS rest system, fNS are the spin-down parameters in the NS rest frame and s are the
number of spin-down parameters included in the phase model. In order to include all
previously discussed Doppler shifts in the phase model, the phase shown in Eq. (4.16)
must be rewritten in terms of the detected phase on Earth.
In the coordinate system of which the origin is located in the solar system barycenter
(SSB), the position of the detector is given by ~rd , the GW wave vector is given by ~k and
the neutron star is denoted by NS. This is shown in Fig. 4.2. The detected phase, Φd (t),
Figure. 4.2: Diagram of the relevant position vectors of a neutron star in a binary system relative to the detector. The origin is located at the SSB, ~k is the GW
vector, ~rd is the position of the detector on Earth, ~n0 is the direction towards the center of mass µ of the Keplerian orbit and ~rK is the position
vector of the neutron star N S relative to µ.
55
Chapter 4. Continuous waves analysis methods
can be computed by rewriting the proper time tNS in terms of the time in the detector
frame t. The phase observed at the detector at local time t was emitted by the NS at
coordinate time t0 such that
|~k(t0 )|
t = t0 +
,
(4.17)
c
where ~k is defined in Fig. 4.2 and c is the speed of light. Finally, the relation between
the time t0 and tNS is given in Ref. [41] by
p
k0
0
tNS = 1 − βNS t +
,
(4.18)
c
where βNS = vNS /c and k0 = |~k(t = −~k(0)/c)|. Note that Eq. (4.18) takes special
relativity effects into account. Also note that in order to rewrite the proper time as
a function of the detector time, Eq. (4.17) must be inverted. This inversion is highly
non-trivial since ~k in Eq. (4.17) contains all the orbital parameters (Earth’s daily and
yearly motion together with the Keplerian parameters) thus the functional dependency
of Eq. (4.17) can be extremely complicated. The detected phase in the interferometer,
Φd , will be of the form
Φd (t) = ΦNS (tNS (t)).
(4.19)
Even though Eq. (4.19) has no straightforward analytical form, it does contain all the
Doppler shifts discussed before.
In Ref. [41] an analytical approximation of the detected phase for the case of isolated
pulsars has been derived from Eq. (4.19). Using the coordinate system shown in Fig.
4.1, this phase is
Φd (t) = Φ0 + 2π
s
X
n=0
(n)
f0
s
n
X
tn+1
2π
(n) t
+
n̂0 · ~rd (t)
f0
,
(n + 1)!
c
n!
n=0
(4.20)
(n)
where f0 denotes the n-th order derivative of the frequency (spin-down parameter1 ),
c the speed of light. Note that since Eq. (4.20) is derived for the isolated neutron star
case, n̂0 = n̂0 (α, δ) is a unit vector pointing towards the neutron star originating from
the SSB. Equation (4.20) has been derived under the assumption that relativistic effects
can be neglected as well as the assumption that f0 < 1 kHz and that the spin-down
age τs > 40 yr. The phase model for isolated NS (substituting Eq. (4.20) into Eq. (4.2))
contains all the necessary parameters to describe the GW induced by a neutron star with
a fixed quadrupole moment. Note that glitches have not been included in the model.
The phase model shown in equation (2.15) depends on 8 + s parameters: h0 , Φ0 ,
f0 , α, δ, ι, χ, ψ and f (1) . . . f (s) . This implies that the parameter space for performing
a template-based coherent search is 8 + s dimensional. When adding (non-relativistic)
binary systems to this signal model, five more parameters need to be taken into account.
These additional parameters as well as a modified phase model will be discussed in
section 5.1.
1
(n)
The spin-down parameters f0 are measured in the SSB and are not equal to the instantaneous
(n)
spin-down parameters fNS from Eq. (4.16). The relation between the two is given in Ref. [41].
56
4.2. Coherent searches
4.2
Coherent searches
As explained before, the matched filtering procedure will be optimal. Such a search is
called a ‘fully coherent search’ and requires the precise signal shape to√be known in order
to create templates. The SNR of such a template will then scale as T where T is the
observation time.
In the case of targeting known isolated pulsars, radio observations can be used to constrain parameters like α, δ, f0 and the spin-down values. For these pulsars it is possible
to confine a GW search to a few and in some cases a single signal template. This allows
for the coherent integration of all available data with this matched filtering algorithm to
achieve the best possible sensitivity. For sources where a single template is sufficient, the
integration is done in the time domain and the parameters of the source are calculated
with Bayesian inference. When the parameters are not known with sufficient precision
through astrophysical observations, the search can still be performed in a coherent way.
However, the number of templates required will scale as T 5 [60]. This implies that such
coherent searches are limited by allowing only a small mismatch between the observed
parameters and the template parameters. This ‘wide parameter’ search is performed
in the frequency domain for computational reasons. The resulting detection statistic is
called the F-statistic (see Ref. [41]).
Both the narrow and wide parameter coherent search has been applied to the Crab
pulsar [61]. For the single-template search the joint (i.e. multi-detector) 95% upper limit
on the GW amplitude is measured to be h95%
= 2.7 × 10−25 . From this upper limit one
0
can derive that less than 4% of the energy loss of the Crab pulsar, observed in the
spin-down, is due to the emission of GWs. For the wide-band search the upper limit will
be worse due to the number of templates applied and in this analysis less than 73% of
the observed energy loss is due to GWs. These results have been obtained by analyzing
LIGO’s fifth science run (S5).
The template or F-statistic search has also been applied in the case of unknown
neutron stars. Such a search is called an all-sky search and assumes no prior knowledge
about the sky position of the neutron stars which will potentially emit GWs. This search
has been performed with the Einstein@Home distributed infrastructure [62]. At the time
of the analysis, the Einstein@Home project contained about 105 computers. The search
was performed with a coherence time of 30 hours and the templates were applied with
50 Hz < f0 < 1500 Hz and −f /τs < f˙ < f /τs where τs > 1000 years for signals below
400 Hz and τs > 8000 years for signals above 400 Hz. In the most sensitive region of the
LIGO detector, between 125 Hz and 225 Hz, the 90% upper limit has been measured to
be h90%
= 3 × 10−24 for all continuous GW sources. These results have been obtained
0
by analyzing S5.
The F-statistic search has also been applied to the known Low Mass X-Ray Binary
(LMXB) Scorpius X-1 (Sco-X1) [23]. Sco-X1 is a neutron star orbiting a low-mass companion with a period of 18.9 hours. The sky-location of Sco-X1 is well known due to
X-ray observations but the binary orbital parameters have large uncertainties. Also, the
rotational frequency of the neutron star is unknown as well as the associated frequency
derivatives. This search is limited to 6 hours of coherent integration due to the compu-
57
Chapter 4. Continuous waves analysis methods
tational constraints and has been performed on the frequency bands 464 − 484 Hz and
= 1.7 × 10−22 and
604 − 624 Hz 2 . The upper limits computed from this search are h95%
0
h95%
= 1.3 × 10−21 for both bands respectively. Note that this result has been produced
0
by analyzing the second LIGO science run (S2) which had limited sensitivity.
When attempting to perform an all-sky search for GWs from neutron stars in binary
systems, one must be able to create templates in a 13 + s dimensional parameter space.
It has been estimated [59] that the number of templates required for performing an
all-sky search is about 1026 . This number applies to a search for neutrons stars in binary
systems where sky-coordinates α and δ are fixed. Furthermore, there are constraints on
the maximum eccentricity emax = 0.8, semi-major axis ap = 1011 cm, orbital angular
velocity ωmax = 10−4 rad/s and maximum GW frequency fmax = 1 kHz. The assumed
5
coherence time is 107 seconds. In this case the number of templates scales with a5p , fmax
and linearly with T . When attempting to perform an actual all-sky search, the number
of templates will scale as a large power of T . This clearly shows that a coherent all-sky
search for GWs from neutron stars in binary systems is computationally unfeasible.
4.3
Semi-coherent searches
Since all-sky searches for GWs are computationally challenging or in some cases unfeasible, several computationally less demanding search analysis methods have been developed. Such methods cannot obtain the same sensitivity as the coherent searches. These
analysis methods generally utilize short stretches of coherent integration and combine
the results per coherent stretch in a incoherent way. Two different semi-coherent analysis
methods have been developed, called the ‘Hough search’ [63] and the ‘PowerFlux’ search
[64].
These methods are based on summing measures of strain power from many coherent stretches. Each method also corrects explicitly for sky-position dependent Dopplermodulations of the apparent source frequency due to Earth’s motion as well as for the
spin-down as shown schematically in Fig. 4.3. This requires a search in a four-dimensional
parameter space over the parameters f0 , f (1) , α and δ. The PowerFlux method also includes the polarization angle ψ in the search.
The PowerFlux and Hough methods mainly differ in the way the measures of strain
power are defined. PowerFlux measures strain power as the measured power per frequency bin weighted with the beam-pattern functions for circular and linearly polarized
gravitational waves. The detection statistic here is the estimated amplitude for linear
or circular polarized gravitational waves. The Hough method measures strain power by
selecting frequency bins where the power exceeds a predetermined threshold. The Hough
detection statistic is the number count of the selected frequency bins. The Hough and
PowerFlux methods of combining separate FFTs are based on the same principle and
thus the sensitivity will scale in similar ways as a function of observation time.
The Hough search is based on dividing the total observation time, Tobs , into N seg2
The interval has been chosen since Sco-X1 exhibits quasi-periodic oscillations and as discussed in
Ref. [23], these QPOs might be linked to the rotational frequency of the neutron star.
58
4.3. Semi-coherent searches
Figure. 4.3: An illustration of the discrete frequency bins of the short Fourier transforms (SFTs) of the data versus the discrete start time of the SFT. The
dark pixels represent the signal in the data. Its frequency changes with
time due to the Doppler shifts and spin-down. By sliding the different
SFTs in frequency, the signal bins can be lined up and summed after appropriate weighting to create a detection statistic.
ments of length T which are Fourier transformed. By applying a threshold on the power
spectrum certain bins will be selected for each segment. These selected bins produce a
distribution in the time-frequency plane. This distribution will contain all the Doppler
shifts and spin-downs discussed in section 4.1. The Doppler modulation of the signal
depends on the sky position of the source. Since this dependence is known a priori (Eq.
(4.8)), the sky position (α, δ) consistent with a certain measured f (t) can be computed3 .
Furthermore, the spin-down parameters introduce another shift in the time-frequency
plane. Also for this effect the functional dependency is known: a linear decrease of the
frequency for a single spin-down parameter, a quadratic decrease for two spin-down parameters, etc. The Hough search uses the Hough transform4 to sum the frequency bins
consistent with a certain sky position and spin-down in order to generate a detection
statistic.
The sensitivity of a Hough search with a false alarm rate of 1% and a false dismissal
rate of 10% is computed [63] to be
r
r
8.54 Sn
Sn
1/4
= 8.54N
.
(4.21)
h0 = 1/4
N
T
Tobs
The second equality is valid under the assumption that Tobs = N T which is not the case
if there are gaps in the data. Note that the power of the signal must be contained in a
3
Strictly speaking the set of sky positions consistent with a certain f (t) form a circle in the celestial
sphere centered around the vector ~v .
4
The Hough transform is a robust parameter estimator of multidimensional patterns in images
initially developed by Paul Hough [65] to analyze bubble chamber pictures at CERN.
59
Chapter 4. Continuous waves analysis methods
single frequency bin. Section 4.1 shows that the maximum coherence time is about 30
minutes for a 1 kHz signal. It is also possible to perform the Hough search on so-called
demodulated data which are data which have been resampled such that the Doppler
shifts have been corrected for. In that case there is no such restriction on T . However
such an approach would require much more CPU power as it involves generating signal
templates for each point in the sky.
The best Hough upper limit result for the fourth LIGO science run (S4) is published
in Ref. [64] to be h95%
= 4.25 × 10−24 . This result has been obtained with T = 30
0
minutes, Tobs = 29.5 days and N = 2966 in the band 140.00 to 140.25 Hz using the data
from H1, H2 and L1.
As stated before, a coherent all-sky search for GWs from neutron stars in binary
systems is computationally unfeasible. A semi-coherent approach based on the principle
of the signal not leaving a single frequency bin like the Hough search will yield a short
coherence time. Furthermore, the parameter space will increase to 13 + s parameters.
There are currently two analysis methods which target unknown neutron stars in binary
systems: the ‘TwoSpect’ method and the ‘Polyomial Search’ method.
TwoSpect, described in Ref. [66], uses short Fourier transforms such that the power
of the signal remains in a single frequency bin. From these spectra, a time-frequency
plot is computed which will contain the Doppler shift induced frequency variation of the
signal. The method then exploits the periodicity of the Keplerian orbit by performing
another Fourier transform on this time-frequency diagram. The signal will be concentrated around the central GW frequency and its Doppler shift induced harmonics. When
first demodulating the Doppler shifts induced by Earth’s motion, these harmonics are
purely determined by the binary orbit of the neutron star. These harmonics are then
summed with an incoherent harmonic summing technique which identifies certain regions of interest in the parameter space. Subsequently, templates are applied to these
regions of interest and a detection statistic is generated by weighted summing of the
second Fourier power. The parameter space used for these templates are the GW frequency f , the binary period P and the modulation depth (e.g. the frequency difference
of the signal between the apastron and periastron) ∆f .
The TwoSpect search is based on the same principles as the other incoherent searches:
weighted summing of a statistic where the weights are defined by the templates applied.
The sensitivity will scale as
r
Sn
,
(4.22)
h0 ∝ N 1/4
Tobs
where the proportionality constant is defined by how many templates are applied, what
the false alarm and false dismissal probability is, etc. Note that TwoSpect is less sensitive
to eccentric orbits and is limited to relatively slow binaries since the search is dependent
on containing the signal in a single frequency bin. This thesis presents ‘Polynomial
Search’ developed at Nikhef. Chapter 5 gives the details of Polynomial Search together
with the sensitivity estimates and tests on simulated gravitational wave signals in white
noise.
60
Chapter
5
Polynomial Search
In section 1.3.1 it has been discussed that non-axial symmetric neutron stars emit continuous gravitational waves. The amplitude of these waves are orders of magnitude lower
than the noise level of the LIGO and Virgo detectors. However, when the continuous
signal√is integrated over an observation time T , the signal-to-noise ratio will increase
with T which will allow us to obtaining a lower detection limit.
From the ATNF pulsar catalog shows that the majority of the known pulsars with
rotational frequencies above 10 Hz are in binary systems as can be seen in Fig. 1.3. Furthermore, the gravitational wave amplitude for a neutron star with a given quadrupole
mass moment scales with f 2 as shown in Eq. (1.15). Thus an all-sky analysis method
specifically targeted towards the detection of gravitational waves from neutron stars in
binary systems is desirable. However as stated in section 4.1, the Keplerian orbit of the
neutron star introduces a Doppler shifts to the signal in addition to the Doppler shift
from the detector motion, causing a coherent analysis method to be unfeasible.
In order to target these unknown binary systems, the ‘Polynomial Search’ method has
been developed. Polynomial Search employs a limited number of templates (or filters)
and aims to increase the possible coherence time by more than an order of magnitude
compared to existing semi-coherent methods. This is accomplished by taking the templates to be Taylor expansions of the phase of the gravitational wave which will match
the signal over a certain part of the orbit. Furthermore, Polynomial Search is designed
to include highly eccentric orbits as well as completely circular orbits.
5.1
Doppler shifts of binary systems
The phase of the gravitational wave is Doppler shifted. Equation (4.20) shows the phase
as a function of all the Doppler shifts for isolated neutron stars. The magnitude of these
shifts (the Earth’s daily and yearly motion as well as the spin-down) have been discussed
in section 4.1. The fourth contribution to the Doppler shift is given by the Keplerian
orbit of the neutron star and the magnitude of the modulation will depend on the orbital
parameters.
Polynomial Search is based on the fact that the detected phase of the gravitational
61
Chapter 5. Polynomial Search
wave signal can be written as a Taylor expansion up to a certain order N during an
observation time T ,
N
X
φn n
Φ(t) ≈ 2π
t ,
(5.1)
n!
n=0
where φn are the Taylor components. The required number of orders to be taken into
account N , strongly depends on the coherence time T as well as on the binary parameters
which are to be covered by the search. To make statements about the limits and errors on
the Taylor parameters in the expansion shown in Eq. (5.1), a description of the relative
motion of the neutron star with respect to the center of mass of the binary system and
the relative motion of the detector with respect to the Solar System Barycenter (SSB)
must be made.
When assuming non-relativistic motion, the orbit of a neutron star can be computed
with Kepler’s laws. Figure 5.1 schematically shows the orbit of a pulsar in an eccentric
orbit. In this coordinate system, the position of the neutron star in polar coordinates r
and θ can be computed with the following equations
t−τ
2π
= E − e sin(E)
(5.2)
P
r
θ
E
1+e
tan
=
tan
(5.3)
2
1−e
2
(1 − e2 )a
r(t) =
.
(5.4)
1 + e cos(θ)
The polar coordinate θ is computed by by solving Kepler’s equation (Eq. (5.2)) for the
angle E (the eccentric anomaly) and substituting it into Eqs. (5.3). Note that in Eq.
(5.2) τ is the time of periastron passage (i.e. the time the NS passes point P in Fig.
5.1). The r coordinate is computed by solving Eq. (5.3) for θ and substituting it into
Eq. (5.4). Finally, the length of the semi-major axis, a can be computed from Kepler’s
third law:
r
2
3 G(M + mns )Pb
a=
,
(5.5)
4π 2
where G is Newton’s constant, M the mass of the companion object, mns the mass of the
neutron star and Pb is the period of the orbit. It can be seen that the (non-relativistic)
description of the orbit of a neutron star depends on four orbital parameters (P , a, e
and τ ). Combining these orbital parameters with an additional orbital inclination angle
i, makes the parameter space for binary orbits five dimensional.
In order to set upper limits on the possible observable phase changes in the signal
due to the Doppler shifts, the extreme values of the orbit and their time derivatives
must be computed. These extreme values occur in the periastron and can be computed
by combining Eqs. (5.2), (5.3) and (5.4) together with the fact that the Keplerian-orbit
induced phase modulation, ΦK , is given by
ΦK (t) = 2πf0 t +
62
2πf0
n̂ · ~rK (t).
c
(5.6)
5.2. Binary system test cases
Auxillary circle
y
NS orbit
NS
r
E
θ
µ
O
P
x
ae
a
Figure. 5.1: Schematic representation of the motion of a neutron star in an eccentric
orbit in the orbital xy plane. O denotes the center of the elliptical orbit
with the semi-major axis length a and eccentricity e as well as the center
of the circle with radius a. P is the periastron (the point of closest approach). NS is the neutron star which has polar coordinates coordinates
(r(t), θ(t)) as a function time. The origin of this polar and corresponding
Cartesian coordinate system is located at the center of mass, µ.
Note that the n̂ and ~rK are defined in Fig 4.2 and that |~rK | = r(t) as computed by
solving the Kepler equations.
5.2
Binary system test cases
In order to define a region of parameter space where Polynomial Search can be applied,
the maximum values of the Doppler-shifts will be computed for two extreme cases.
Figure 5.2 shows the orbital periods and eccentricities of known pulsars in binary systems
with spin frequencies above 20 Hz. To illustrate the computation of the limits of the
63
Chapter 5. Polynomial Search
P (days)
histo
M > 0.5 Msol
M < 0.5 Msol
103
102
10
1
10-1
10-7
10-6
10-5
10-4
10-3
10-2
10-1
1
Eccentricity
Figure. 5.2: Orbital period versus eccentricity for the known pulsars with frequencies
above 20 Hz in binary systems (total 78) from the ATNF pulsar catalog
[21]. The triangles indicate neutron stars with a companion with a median
mass M < 0.5M . The squares indicate neutron stars with a companion
with M > 0.5M .
parameters used in Polynomial Search, two extreme systems are considered with fgw = 1
kHz for both cases.
1. A 1.4 M neutron star orbiting an object with the same mass in a circular orbit
with P = 2 h.
2. A 1.4 M neutron star orbiting an object of 14 M in an eccentric orbit with
e = 0.6 and P = 6 h.
These two cases roughly correspond to the binaries located in lower right corner (high
eccentricity, small orbital period and high companion mass) and lower left corner (low
eccentricity and small orbital periods) of Fig. 5.2.
The parameters of the Keplerian orbit are displayed in Table 5.1. Figure 5.3 shows
parameter binary 1 binary 2
h0
f0
P
ap
e
10−27
1000 Hz
2h
1.65 ls
0
10−27
1000 Hz
6h
9.68 ls
0.6
Table. 5.1: The parameters used for the different test binaries.
64
5.2. Binary system test cases
y (ls)
x (ls)
the position of the neutron star in these binary orbits as a function of time. The x and
y coordinates of the neutron star are defined in Fig. 5.1. The origin of this system is
set on the center-of-mass of the binary system (see Fig. 5.1). Figure 5.4 shows the first
4
2
0
-2
-4
-6
-8
-10
-12
-14
0
6
4
2
0
-2
-4
-6
2
4
6
8
10
12
t (h)
0
2
4
6
8
10
12
t (h)
Figure. 5.3: The left panel shows the x coordinate of the neutron star in the circular
orbit (solid curve) and in the eccentric orbit (dashed curve). The right
panel shows the situation for the y coordinate. Note that both positions
are computed in light seconds (ls).
four derivatives of these orbits. The phase modulation of the gravitational wave signal
for these orbits is dominated by the Doppler shift of the binary motion. By using Eq.
(5.6) the maximum change in the n-th order derivative of the phase can be written as

2πf0


2πf0 t +
n̂ · ~rK |max for n = 0,


c


2πf0
d~rK dn ΦK for n = 1,
= 2πf0 + c n̂ · dt (5.7)
dtn max 
max

n


2πf0
d ~rK 

n̂ ·
for n > 1.
c
dtn max
n
In Eq. (5.7) d dtΦnK max is the maximum value for the nth derivative of the detected phase
n
on Earth and ddt~rnK max is the nth derivative of the position of the neutron star in the
direction of maximum variation and at the time of maximum variation (i.e. in the
periastron). Equation (5.7) allows for the interpolation of limits on Polynomial Search
parameters.
The maximum values for the first, second and third-order derivative of the phase
can be computed with the information from Fig. 5.4 together with Eq. (5.7). With
these values, the limits on Polynomial Search parameters, φn |max up to fourth order are
calculated and are shown in Table 5.2. From Fig 5.4 it can be seen that for the eccentric
binary the absolute values of the derivatives are smaller than for the circular system for
about 80 percent of the time. The maxima occur close to the periastron, resulting in
higher values for the extremal Taylor components.
The maximum values of the consecutive derivatives decrease as powers of the orbital
65
1000
×103
d2x(t)
(m/s2)
dt2
dx(t)
(m/s)
dt
Chapter 5. Polynomial Search
500
400
200
0
-200
-400
-600
-800
-1000
-1200
-1400
0
0
-500
2
4
6
8
10
12
t (h)
d4x(t)
(m/s4)
dt4
d3x(t)
(m/s3)
dt3
-1000
0
1.5
1
0.5
2
4
6
8
10
12
t (h)
2
4
6
8
10
12
t (h)
0.005
0.004
0.003
0.002
0
0.001
-0.5
0
-1
-0.001
2
4
6
8
10
12
t (h)
0
×103
d2y(t)
(m/s2)
dt2
dy(t)
(m/s)
dt
-1.5
0
1500
1000
500
0
2
4
6
8
10
12
t (h)
0.5
d4y(t)
(m/s4)
dt4
d3y(t)
(m/s3)
dt3
-500
0
0
-0.5
2
4
6
8
10
12
t (h)
0.004
0.003
0.002
0.001
0
-1
-0.001
-1.5
-0.002
-0.003
-2
0
800
600
400
200
0
-200
-400
-600
-800
0
2
4
6
8
10
12
t (h)
-0.004
0
2
4
6
8
10
12
t (h)
Figure. 5.4: The first four derivatives of the x and y coordinates as a function of time
for the neutron star in the circular orbit (solid curve) and the eccentric
orbit (dashed curve).
frequency ωorbit |max . The latter is evaluated at a certain point in the orbit, torbit , where
66
5.3. Constructing the filters and the correlations
Parameter
Circular orbit
φ1 |max (Hz)
1000 ± 1.45
φ2 |max (Hz/s) ±1.26 × 10−3
φ3 |max (Hz/s2 ) ±1.10 × 10−6
φ4 |max (Hz/s3 ) ±9.62 × 10−10
Eccentric orbit
1000 ± 5.27
±2.77 × 10−3
±5.07 × 10−6
±1.74 × 10−8
Table. 5.2: The extremal values of the polynomial search parameters for the circular
and eccentric orbit (see text) for gravitational wave frequencies up to 1
kHz.
the derivatives are maximal. This maximal orbital velocity is given by
~r(torbit ) × ~v (torbit ) ,
ωorbit |max = |~ω |max = r(torbit )2
(5.8)
where ~r and ~v are the velocity vector and position vector of the neutron star, respectively.
In the circular case ωorbit = 2π
= ωorbit |max . When the coherence time Tcoh is sufficiently
P
1
smaller than ωorbit | , the higher-order contributions can be ignored1 .
max
5.3
Constructing the filters and the correlations
Instead of attempting to create templates in the 13 + s dimensional parameter space,
Polynomial Search employs an empirical phase model given by the Taylor expansion of
the phase and creates templates in a parameter space given by the components φn . As
stated in section 5.1, the Doppler shift induced phase modulation is proportional to the
powers of the maximum orbital frequency
n d φ
∝ (ωorbit |max )n ,
(5.9)
dtn max
where ωorbit |max represents the angular frequency at a certain point, torbit , in the orbit
where the velocity is maximal. For circular orbits the maximum contributions, occurring
in either the peri- or apastron, of the nth and the (n + 2)th derivative coincide with the
roots of the (n + 1)th derivative2 . Therefore, the maximum of the first and third-order
derivatives occur at the same value of torbit , where the third-order contribution is reduced
with a factor (ω)3 . This implies that the next contributing order will be the fifth order
which will be reduced by (ω)5 . This, together with the fact that in Ref. [67] is has been
computed that the number of templates to consider scales3 as T n(n+1)/2 , leads to the
P
.
In the case of the circular orbit this statement simply says Tcoh 2π
For example: in the periastron, the orbital velocity (the first derivative) is maximal, but the derivative of the orbital velocity is zero.
3
Strictly speaking this scaling is valid for a full bandwidth search for all allowed values for the filter
parameters which obey the Shannon-Nyquist theorem.
1
2
67
Chapter 5. Polynomial Search
choice to truncate the expansion at the third-order in the current implementation of
Polynomial Search. The phase, up to third-order, is given by
α 2 β 3
~
Φ(t; ξ) = 2π φ0 + f0 t + t + t ,
(5.10)
2
6
where the vector ξ~ = (φ0 , f0 , α, β) lives in the four dimensional template (or filter)
~ is given by
parameter space. Note that the frequency f (t, ξ)
~
~ = 1 dΦ(t; ξ) = f0 + αt + β t2 .
f (t; ξ)
2π dt
2
(5.11)
The filters (based on Eq. (4.2))
~ = sin Φ(t; ξ)
~
F (t; ξ)
(5.12)
are created in this parameter space and are to be applied to the data in stretches of
time T . These filters are valid representations of the signal shown in Eq. (4.2) as long
as T is short enough such that the beam-pattern functions are constant. The maximum
values for α and β can be calculated from the Kepler orbits assuming that the dominant contribution in the frequency dependence on time arises from the Doppler shifts
as discussed in section 5.1. The values were calculated for the two specific binaries and
are shown in Table 5.2.
Polynomial Search is performed by stepping in parameter space and creating a collection of filters. The filters are then applied to the data by shifting the data over the
~ could be computed in the
filter and computing the correlation. The correlation C(τ ; ξ)
time domain as shown in Eq. (A.3) but this requires the integral to be evaluated for each
value of the lag. This computation will be much more computationally efficient when
done in the frequency domain. When combining the computation of the correlation in
the frequency domain with the expression of the signal-to-noise for matched filters, an
expression for the correlation is derived in appendix B. The result of this derivation is
Z ∞
~
D̃(f )F̂ ∗ (f ; ξ)
~
q
C(τ ; ξ) =
e−2πif τ df,
(5.13)
1
−∞
S
2 n
where τ is the lag, D̃(f ) is the Fourier-transformed data and F̂ ∗ denotes the complex
conjugate of the Fourier-transformed filter normalized to its total power. Furthermore,
the power spectral density of the noise has been taken to be independent of frequency
(white). In the case of discretely sampled data with sample frequency fs,data during time
T
Nf
1 X D̃(fk )F̂ ∗ (fk ) −2πifk τj
q
C(τj ) =
e
,
(5.14)
T k=0
1
S
2
n
where τj is the j th time sample, k denotes the frequency bin number, Nf = fs,data T /2 is
the number of frequency bins and fk is the k th frequency bin.
68
5.3. Constructing the filters and the correlations
Equation (5.13) shows that the correlation between a filter and the data can be
computed for all lags by a single inverse Fourier transform. Note that by using the
correlation theorem (Eqs. (A.3) and (A.4)) integration time T , Eq. (5.13) is equivalent
to the expression
r Z T
2
~ =
~
C(τ ; ξ)
D(t)F̂ (t − τ ; ξ)dt
(5.15)
Sn 0
as explained in appendix B. It can be seen from Eq. (5.15) that the correlation statistic
can be considered Gaussian distributed with unit variance and can also be interpreted
as a signal to noise ratio. Furthermore, the filter is a function of t − τ , implying that the
filter is effectively shifted in time with lag τ . With the substitution t0 = t − τ , the phase
of each shifted filter can be written as

!
!


α
β
β
2
3
2
Φ(t0 ) = 2π 
 φ0 − f0 τ + 2 τ − 6 τ + f0 − ατ + 2 τ t

|
{z
} |
{z
}
φ0 (τ )
f0 (τ )

1
+
2
!
α − βτ
|
{z

1 3
t + βt 
,
6 

2
}
(5.16)
α(τ )
From Eq. (5.16) it follows that the lag parameter creates an interdependency of the filter parameters. This property can be exploited to avoid generating filters for each point
in parameter space and will drastically reduce the amount of CPU time and memory
needed.
In generating and applying the filters it is important to minimize the number of computational steps. This is accomplished in two ways. Firstly, the dependence of the lag
parameter τ is calculated by time shifting (employing a phase shift in frequency space;
see appendix B). Secondly, each filter is frequency shifted to obtain the dependence of
f0 . The computational steps will be elaborated upon next.
In order to use the lag parameter as a time shift, it must be possible to shift the filter
within the block of data between 0 and T . Thus the filters are zero-padded for t > T /2
and only the lags 0 < τ < T /2 are considered, implying that Eq. (5.16) is valid for
τ < t0 < τ + T /2 and that values for τ > T /2 should be ignored. For each value of the
lag a new shifted filter will be generated. The dotted curve in figure 5.5 schematically
shows the frequency of a typical filter as a function of time by using Eq. (5.11). The
dashed curve in Fig. 5.5 represents the frequency of a potential gravitational wave signal
as a function of time. As can be seen, the filter does not match the frequency evolution
of the signal. However, the filter matches the signal when it is shifted in time by τ = 300
s (solid curve) as the horizontal arrow indicates. Furthermore, while the parameters of
the signal (shown in Fig. 5.5) differ from those of the filter, the time shift causes the
filter to match due to the functional dependencies of the filter parameters on the lag
69
f (Hz)
Chapter 5. Polynomial Search
data: f0 = 4.42, α = -0.0029, β = 1e-05
8
time-shifted filter: f0 = 4.0, α = 0.0001, β = 1e-05, τ = 300.0
filter: f0 = 4.0, α = 0.0001, β = 1e-05, τ = 0
7.5
frequency-shifted filter: f0 = 5.0, α = 0.0001, β = 1e-05, τ = 0
7
6.5
6
5.5
5
frequency shift
time shift
4.5
4
0
200
400
600
800
1000
1200
t (s)
Figure. 5.5: An example of the frequency and lag evolution of a typical filter. The dotted curve represents the unshifted filter. The dashed curve is an example
of what a gravitational wave signal may look like during observation time
T = 1200 s. The horizontal arrow indicates the time-shift (τ = 300 s)
needed for the filter to match the signal, resulting in the solid curve overlapping with the signal. The vertical arrow indicates a potential frequencyshift (a shift in f0 ) such that the same filter can be used to match a signal
with higher frequency. The legend shows the actual filter parameters used
in this figure.
shown in Eq. (5.16).
Since the gravitational wave signal shown in Fig. 5.5 can match arbitrary filter parameters, Polynomial Search must be able to search over the parameter space (e.g. the
stepping in parameter space is discussed in more detail in section 5.5). This is done
by creating a series of filters with different parameters each representing a point in parameter space. Naively speaking, each point in parameter space requires the numerical
re-computation of the filter function F (t). While this is true for each new value for α and
β, the f0 parameter can be varied by simply shifting the power spectrum of the filter
in the frequency direction. This is shown in Fig. 5.5 by the vertical arrow, depicting
the same filter as before but shifted in the frequency direction by 1 Hz. In practice, the
filters are sampled at a sample frequency fs,filter fs,data . In this way, the filters can be
applied to different frequency intervals of the data by comparing a different part of the
power spectrum of the data to the filter. This process is shown schematically in Fig. 5.6.
In order to make sure the frequency of the filter does not fall in the 0 Hz or fN bins, a
buffer zone has been chosen.
~ for
When discretizing the filter, nlags = T fs,filter /2 values of the correlation C(τ, ξ)
the lags between 0 and T /2 are obtained, where each value of τ represents a new filter
with parameters φ(τ ), f0 (τ ) and α(τ ) as shown in Eq. (5.16). In other words: in this
70
5.4. Coherence length
Figure. 5.6: A schematic representation of the power spectrum of a typical filter (left)
and the power spectrum of a block of data (right). In the left panel, the
hatched region represents the ‘buffer’ zone, fN,filter = fs,filter /2 the Nyquist
frequency and ∆ffilter is the active frequency range of the filter. In the
right panel the gray area represents the active frequency range of the filter
and fN,data = fs,data /2 the Nyquist frequency of the data.
~ yields a collection of nlags correlations which are obtained by
way a single filter F (t; ξ)
a single computation of the inverse Fourier transform.
In order to reduce the number of correlations that must be processed, only the maximum value of C from all the lags from Eq. (5.13) and the corresponding values for
τ, f0 , α and β are stored. Note that the value for φ0 has been chosen to be 0. The
freedom to choose an arbitrary value for φ0 arises from the fact that all values of φ0
are sufficiently covered when all lags τ are considered. Since C can also be interpreted
as a signal-to-noise ratio, a threshold in terms of the SNR can be set. When applying
such a threshold to the collection of maxima the resulting numbers are called hits. Hits
are displayed by plotting the frequency of the filter as a function of time f (t; ξ~hit ) by
using Eq. (5.11), where ξ~hit represents the parameters of the filters for which the correlation exceeds the threshold. Such plots are called hit maps and examples of hit maps
of simulated gravitational wave data are shown in Figs. 5.32 and 5.34.
5.4
Coherence length
In order to determine the optimal coherence length of the filters used in Polynomial
Search, a criterion for matching a filter to a known signal must defined. As the objective
of this study is to determine the optimal coherence length of the filters it is advantageous
to consider the correlation in the time domain. The correlation between a normalized4
~ over a coherent integration
signal (without noise) ŝ(t) and a normalized filter F̂ (t; ξ)
4
The definition of a normalized signal is given in appendix A.
71
Chapter 5. Polynomial Search
time of Tcoh is given by
Z
cŝF̂ =
Tcoh
~
ŝ(t)F̂ (t; ξ)dt.
(5.17)
0
Note that since the lag is not a filter parameter (e.g. the lag just allows for an efficient
computation of the correlation) it is set to 0 in Eq. (5.17). The functions ŝ and F̂ are
periodic functions of time. This implies that −1 < cŝF̂ < 1; the correlation will be +1
if F̂ matches ŝ during Tcoh and −1 if F̂ matches ŝ but is π out of phase during Tcoh .
Furthermore, if ξ~ is such that the filter does not match the signal, the correlation will lie
in between these extremal
values. As discussed in appendix B, the SNR of an optimal
√
filter will scale as Tcoh . Therefore, the quality factor is defined as
Z Tcoh
p
~
~
Q(Tcoh ; ξ) ≡ Tcoh
ŝ(t)F̂ (t; ξ)dt.
(5.18)
0
~ is the normalized
In Eq. (5.18) ŝ(t) is the normalized signal without noise and F̂ (t; ξ)
√
filter. Note that if a filter correlates fully during the integration time, then Q = Tcoh .
This corresponds to the SNR scaling for a fully coherent search.
As a demonstration, the optimal length of the coherence time Topt of the filters was
calculated for the binary systems mentioned in section 5.1, where Topt is defined as the
time length of the filter for which the quality factor is maximal. The orientation of
the binary system with respect to the detector is in all cases chosen for the worst-case
scenario: the contribution of the neglected orders in the phase derivatives is maximized.
For example, for a first order filter (f0 6= 0 and α = β = 0) the second order derivative
of the phase (n = 2 in Eq. (5.7)) is chosen to be maximal. This is done by maximizing
2
n̂ · ddt~r2K which would result in taking n̂ in the direction of the acceleration. Furthermore,
the filter parameters ξ~ are computed to match the derivatives of the phase at the time
in the orbit torbit , where these derivatives are maximal. The integration over the filter
and data multiplication shown in Eq. (5.18) is then taken from torbit − Tcoh
to torbit + Tcoh
.
2
2
Note that this represents the worst case scenario in terms of the phase modulation; the
actual phase modulation will always be less.
For the aforementioned circular orbit, Fig. 5.7 shows the quality factor versus the
coherence time for filters up to fourth order (γ/24t4 ). From this figure the values for
Topt have been inferred and are displayed in Table 5.3. These numbers give the lower
limit for the optimal coherence time for a Polynomial Search for neutron stars in binary
systems with circular orbits with periods larger than 2 hours for different orders of the
filter.
For a real search, the filters can be applied with a longer coherence length and
still remain coherent. From Table 5.3 it can be seen that the gain in quality factor
when adding a third-order term to the phase model is about a factor of 1.6. Adding an
additional fourth-order term will increase the quality factor with a factor 1.3, while the
10
number of extra filters that must be computed will scale as Tcoh
[67].
For the second, eccentric binary, Q depends on torbit since the values of the derivatives
depend on the position in the orbit. Figure 5.8 shows the three-dimensional plots of the
time in the orbit versus the coherence time with the Q-factor as the greyscale for the
72
Q
5.4. Coherence length
30
4th order
3rd order
2nd order
1st order
25
20
15
10
5
0
200
400
600
800 1000 1200 1400 1600 1800 2000
Tcoh (s)
Figure. 5.7: The quality factor versus the coherence time of the filter for different orders of the filter. The quality factor has been computed with the worstcase orientation of the system.
Maximum order considered Topt (s) Qmax
1
2
3
4
32.7
195
511
955
5.07
12.9
21.2
29.4
Table. 5.3: The optimal coherence length of the filters for neutron stars in binary systems with a circular orbit (first binary).
eccentric orbit. It can be seen from this figure that the value of Q depends on where
in the orbit the filter is applied. This is to be expected since Fig. 5.4 shows that the
derivatives of the phase are strongly dependent on the orbital phase.
Figure 5.9 shows that the optimal coherence length is approximately 250 seconds in
the periastron for a third-order filter. With this coherence time, a third-order polynomial
filter will be coherent at any point in the orbit for 250 seconds. In other words, when
taking Tcoh = 250 seconds all gravitational waves with Taylor components up to 3rd order
given in Table 5.2 are guaranteed to be covered. However, when relaxing this constraint
the coherence time can be taken longer (1900 seconds for a 3rd order filter). Another
point is that Figs. 5.8 and 5.9 have been created assuming the worst case orientation
for each point in the orbit and each applied order of parameter space which means that
in reality the actual Topt should be higher.
The coherence length chosen in the analysis is Tcoh = 600 seconds. Therefore, the
length of the data stretches is T = 2Tcoh = 1200 seconds, since the filter is zero padded
73
Chapter 5. Polynomial Search
50
103
60
Tcoh (s)
Tcoh (s)
60
50
103
40
40
30
2
10
20
30
102
20
10
10000
15000
20000
torbit (s)
0
60
50
103
10
0
5000
10000
15000
20000
torbit (s)
50
103
40
40
30
30
102
10
0
0
60
Tcoh (s)
5000
Tcoh (s)
10
0
10
102
5000
10000
15000
20000
torbit (s)
20
20
10
10
0
10
0
5000
10000
15000
20000
torbit (s)
0
Figure. 5.8: The quality factor Q, as a function of the time in the orbit torbit and the
coherence length Tcoh computed for the eccentric binary. The Q-factor for
a first, second, third and fourth order filter are shown in the top left, top
right, bottom left and bottom right panel, respectively. The quality factor
has been computed for the worst-case orientation of the system.
for half the time.
5.5
Step size in parameter space
Polynomial Search uses filters constructed from a four-dimensional parameter space.
The aim of the search is to reconstruct the phase evolution of a gravitational wave by
approximating it with the phase evolution of the filters that give a hit. These filters
are applied to the data by computing the correlation statistic defined in Eq. (5.13).
This correlation statistic can be interpreted as an inner product between two vectors
in parameter space. In order to determine a step size in the parameter space, the inner
product between two neighboring filters, represented by ξ~i and ξ~j , is computed. This
inner product gives rise to a metric for the parameter space. The separation between
74
Topt (s)
5.5. Step size in parameter space
104
1st order
2nd order
3rd order
4th order
103
102
10
0
5000
10000
15000
20000
torbit (s)
Figure. 5.9: The optimal coherence length Topt versus the time in the orbit for maximum Q at that time in the orbit.
two filters in parameter space is defined as
∆ξ~ij = ξ~i − ξ~j = (∆f0 , ∆α, ∆β),
(5.19)
where ∆f0 , ∆α and ∆β represent the difference in their respective parameter values.
Since Polynomial Search is based on the assumption that a potential gravitational wave
signal will behave like a polynomial during the coherent time |∆ξ~ij | can be interpreted
as the mismatch between the potential signal and the nearest filter.
In order to perform Polynomial Search, a filter bank must be created which implies
that the parameter space must be discretized. To find the optimal step size one can
attempt to compute the metric corresponding to the inner product. This metric has
been computed in the literature [67] for a binary search and restricted to low eccentric
orbits with a different way of parameterizing the phase. Their phase expansion was
written as
(m)
∞
j
X
uj (m)
(m)
t − tmid ,
(5.20)
φ (t) =
j!
j=0
(m)
where m is the data segment index, the uj parameters are the instantaneous phase
(m)
derivatives and tmid is the midpoint of the mth segment. It can be seen from Eq. (5.20)
(m)
that this Taylor expansion is centered around a certain point tmid . The difference with
Polynomial Search is that the τ parameter to match the filter with the data at the
optimal time is obtained by taking the maximum correlation of each filter-data compar(m)
ison. This essentially implies that tmid would be an additional parameter in Polynomial
Search. Furthermore, the step size in this parameter is implicitly chosen by the sample
frequency and the coherence time of the filter. These reasons make the metric computed
75
Chapter 5. Polynomial Search
in Ref. [67] unsuited for usage in Polynomial Search. The lag makes analytic computation of the parameter space metric non-trivial. Consequently, a different approach has
been adopted.
In order to demonstrate the correlation of the filters as a function of the step sizes,
slices from the parameter space metric are computed by considering the correlations of
a filter bank with a data set with known filter parameters. This filter bank has been
generated with filters with small step sizes in α, β and f0 .
Table 5.4 shows the different data sets to which the filter banks were applied and
Table 5.5 shows the different filter banks which have been applied to the data sets.
Data set f0 (Hz) α (Hz/s) β (Hz/s2 )
1
2
3
4
600
600
600
600
4 × 10−6
4 × 10−5
1 × 10−3
1 × 10−3
1 × 10−7
2 × 10−6
2 × 10−6
1 × 10−7
Table. 5.4: Data sets used to determine the step sizes in parameter space by applying a
filter bank to them. The data are sampled at 4000 Hz for a stretch of 1200
seconds.
Filter bank frange , δf0 (Hz) αrange , δα (Hz/s)
1
2
3
4
0.5, 10−3
2, 10−3
1, 10−3
1, 10−3
10−4 ,
10−3 ,
10−4 ,
10−3 ,
10−6
10−5
10−6
10−5
βrange , δβ (Hz/s2 )
4 × 10−7 , 8 × 10−9
10−6 , 5 × 10−8
10−6 , 5 × 10−8
3 × 10−7 , 5 × 10−8
Table. 5.5: The parameters of the filter banks applied to the different data sets shown
in Table 5.4. The range and step size are given for each parameter per filter
bank which was applied to a particular data set. The central value of each
parameter is chosen to be the true value shown in Table 5.4. The data are
sampled at 4000 Hz for a stretch of 1200 seconds.
In order to visualize the parameter space, the filters from the filters banks from Table
5.5 are applied to the data sets from Table 5.4. The results are drawn as slices of parameter space, for example the correlation statistic as a function of the value for f0 versus α
for different slices of β as is shown for set 1 in Fig. 5.10. The remaining permutations of
the parameters are shown in Figs. 5.11 and 5.12. In these plots the gray scale shows the
correlation normalized such that a correlation of 1 means that the filter and the data
are equal. The plots for the remaining data sets can be found in appendix C.
When considering Figs. 5.10, 5.11 and 5.12 two observations can be made. Firstly,
around the true values for data set 1 (f0 = 600 Hz, α = 4 × 10−6 Hz/s and β = 1 × 10−7
Hz/s2 ) a band of values appears where the correlation is close to one. This is due to
76
5.5. Step size in parameter space
-3
×10 β (Hz/s^2) = -3e-07
0.1
0.5
0
599.9
600
×10
0.1
-3
600.1
β (Hz/s^2) = -6e-08
599.9
600
600.1
×10-3 β (Hz/s^2) = 1.8e-07
0.1
0.5
0
599.9
600
600.1
599.9
600
×10
0.1
600.1
β (Hz/s^2) = 2e-08
599.9
600
×10
0.1
600.1
β (Hz/s^2) = 2.6e-07
600
1
0.5
0
599.9
600
×10
0.1
600.1
β (Hz/s^2) = 1e-07
1
0.5
0
599.9
600
600.1
×10 β (Hz/s^2) = 3.4e-07
0.1
-3
1
0.5
0
-3
×10 β (Hz/s^2) = -1.4e-07
0.1
-3
1
0.5
0
599.9
1
0.5
-3
1
β (Hz/s^2) = -2.2e-07
0
-3
1
0.5
0
×10
0.1
-3
1
600.1
0.5
0
599.9
1
600
600.1
Figure. 5.10: The step size plots for data set 1 (f0 = 600 Hz, α = 4 × 10−6 Hz/s
and β = 1 × 10−7 Hz/s2 ). The parameters of the filter bank used in
this figure are shown in Table 5.5. The plots show α (vertical axis) versus
f0 (horizontal axis) for different values of β. The color codes shows the
correlation, normalized to unity.
the fact that the interdependency shown in Eq. (5.16) together with the lag as a free
parameter can be used to compensate for a certain amount of mismatch. This effect can
be clearly seen in Fig. 5.11 when comparing the panel where α = 4 × 10−6 (the middle
one) to the following 4 panels: the mismatch in α can be (partially) compensated by
taking a lower value for β and a higher value of f0 . This compensation only works if τ is
left to be a free parameter and the maximum value of the correlation is taken for each
computation of the correlation of a filter with the data.
Another observation from Figs. 5.10, 5.11 and 5.12 is that the parameter space slices
are not symmetrical around the true injected value. This is due to the fact that the
injected signal has a positive α and β and thus filters with negative parameters will not
fit as well as filters with positive parameters regardless of the value of the lag as no
amount of shifting in time will increase the correlation.
The requirements on the step sizes are such that for only signal in the data, at least
77
Chapter 5. Polynomial Search
×10
-6
0.5
α (Hz/s) = -7.6e-05
×10
-6
1
0.5
α (Hz/s) = -5.6e-05
0.5
0
600
×10
-6
0.5
600.1
α (Hz/s) = -1.6e-05
599.9
600
0.5
×10
600.1
α (Hz/s) = 4e-06
600
×10
-6
0.5
600.1
α (Hz/s) = 4.4e-05
599.9
600
0.5
×10
599.9
600.1
α (Hz/s) = 6.4e-05
0.5
0
600.1
0.5
×10
600.1
α (Hz/s) = 2.4e-05
1
0.5
599.9
600
×10
-6
1
0.5
599.9
600.1
α (Hz/s) = 8.4e-05
0.5
0
600
600
0
-6
1
599.9
0.5
0
599.9
1
0.5
-6
1
0.5
0
α (Hz/s) = -3.6e-05
0
-6
1
0.5
0.5
0
599.9
×10
-6
1
1
0.5
0
600
600.1
599.9
600
600.1
Figure. 5.11: The step size plots for data set 1 (f0 = 600 Hz, α = 4 × 10−6 Hz/s
and β = 1 × 10−7 Hz/s2 ). The parameters of the filter bank used in
this figure are shown in Table 5.5. The plots show β (vertical axis) versus
f0 (horizontal axis) for different values of α. The color codes shows the
correlation, normalized to unity.
one filter should yield a correlation larger than a certain cutoff value. This optimal step
should be independent of the shape of the potential signal. When comparing the step
size plots made with data from data set 1 (Figs. 5.10, 5.11 and 5.12) to the step size
plots of the other data sets (appendix C), it can be seen that the shape of the signal
also changes the regions in parameter space where the correlation is non-zero. Therefore,
there is no single step size where the correlation is optimal and independent of the signal.
Choosing a fixed sample frequency for the filter effectively sets a step size in the lag. The
correlation is obtained for all lags in one operation (i.e. the inverse Fourier transform
in Eq. (5.13)) and due to taking the maximum of C, the filter parameters f0 , α and β
are computed for a certain value of τmax . Thus the step sizes have been computed in a
more empirical way.
By recording the maximum correlation, Polynomial Search automatically finds the
value of the lag where the filter matches best with a potential signal. When |∆ξ~ij | is
78
5.5. Step size in parameter space
×10
-6
0.5
f 0 (Hz) = 599.96
×10
-6
1
0.5
f 0 (Hz) = 599.97
0.5
0
0.5
×10
f 0 (Hz) = 599.99
×10
0.1
×10
-6
0.5
f 0 (Hz) = 600.0
0.5
0
0
×10
f 0 (Hz) = 600.02
0.5
0
0.5
×10
f 0 (Hz) = 600.03
×10
0.1
0
×10
-6
1
0.5
0
×10
0.1
×10
0.1
-3
f 0 (Hz) = 600.04
0.5
0.5
1
0.5
0
-3
0
1
0.5
-3
-6
0
f 0 (Hz) = 600.01
0
×10
0.1
1
×10
-6
1
×10
0.1
-3
0
0.5
-3
0.5
×10
0.1
0
-6
0.5
-3
0
1
1
0
-3
-6
f 0 (Hz) = 599.98
0.5
0
0
×10
-6
0.5
1
×10
0.1
-3
0
×10
0.1
-3
0
Figure. 5.12: The step size plots for data set 1 (f0 = 600 Hz, α = 4 × 10−6 Hz/s and
β = 1 × 10−7 Hz/s2 ). The parameters of the filter bank used in this figure
are shown in Table 5.5. The plots show α versus β for different values of
f0 and the color codes show the correlation, normalized to unity.
small, the assumption is made that at this lag the filter and the adjacent filter (or the
data) are matched in phase, frequency and (to a lesser extend) in α. Using these assumptions, the step in β is defined as the step that leads to a phase difference of π/2
radians at the end of the filters. This yields a step size of
1
π
= 2π δβ
2
6
3
T
96
→ δβ = 3 ,
4
T
(5.21)
where δβ is the step size in β and T is the duration of one FFT of the data in seconds.
The step size shown in Eq. (5.21) means that the phase of the filter can maximally
deviate π/4 from the signal at the end of the filter. Note that T = 2Tcoh where Tcoh is
the coherence time of the filter as described in section 5.4.
The step in α can be derived from Eq. (5.16) by noting that α(τ ) = α − βτ , meaning
that the lag will step through all α values between α and α − βτ . The step in α thus
79
Chapter 5. Polynomial Search


 |β|T for |β| > δβ
2
δα =
(5.22)
8


for
|β|
<
δβ
T2
where δα is the step size in α and the absolute value of β is taken since all step sizes are
defined to be positive. When |β| < δβ, i.e. when the β term does not contribute to the
phase difference anymore, the step in α becomes the minimal step. This minimal step is
defined, like the step in β, to lead to a phase difference of π/2 at the end of the filters.
When following Eq. (5.16) it would seem that f0 (τ ) gives the step size in f0 to be
2
T
α 2 + β2 T2 . However, this step size would allow for the possibility that a signal will
be missed which can be verified by considering Figs. 5.10, 5.11 and 5.12 and the ones
in appendix C. The figures show that it is not possible to have the maximum step in β
(δβ), the maximum step in α (δα) and the maximum step in f0 at the same time.
The way the step in f0 is computed is to assume that at a certain value for the lag,
τi , the filter will match φ0 and f0 in the middle. The phase difference at the end of the
filter will be
"
2
3 #
∆α T
∆β T
∆Φend = 2π
+
,
(5.23)
2
4
6
4
becomes
where ∆α and ∆β are the differences in α and β, respectively. By construction
∆β
∆Φend − 2π
6
3
T
π
<
4
2
(5.24)
due to the definition of the step in β. Note that α will traverse all values between α and
α + βτi , thus ∆α = βτi . In order to prevent the step in α becoming too large such that
a potential signal is missed, τi is limited to such a value that the phase difference at the
end of the filter does not exceed π/2. Substituting Eq. (5.23) in (5.24) and solving for
τi yields
8
τi = 2 ,
(5.25)
T |β|
which represents the maximum value the lag can take such that the phase difference is
less than π/2. The step in f0 is then defined as the step which allows for at most a phase
difference of π/2. Assuming symmetry around the point where f0 and φ0 are matched
(e.g. in the middle of the filter at t = T /4), the step in f0 can then be computed with
Eq. (5.11) to be
δf0 = f (T /4 + τi ) − f (T /4 − τi ).
(5.26)
Equation (5.25) yields two cases which lead to two different step sizes in f0 . The first one
is τi < T /4 meaning that the value of β is too large to take the maximum step in f0 . In
this case the step in f0 is given by the substitution of Eq. (5.25) in Eq. (5.26). When τi >
T /4 it means that the maximum step size can be taken so the lag-dependence of f0 (τ )
from Eq. (5.16) will give the step in f0 . Finally, when either of the previously discussed
cases yields a step in f0 which is smaller than the intrinsic resolution of the FFT, the
80
5.6. Constructing the hit maps
step in f0 automatically becomes this minimum step. These arguments combined gives
the step in f0 as

T
T


for τi <
2|α|τi + |β|τi


2 4

 2
T
δf0 = |α| T + |β| T
(5.27)
for τi >


2
2
2
4



1
1
for δf0 < .
T
T
When considering Eq. (5.25) together with Eq. (5.27) it can be seen that the maximum
step in frequency, δf0 |max , occurs when τi = T /4. When taking T = 1200 s, the β value
where the maximum step in frequency occurs can be computed by inverting Eq. (5.25)
and yields β = 1.85 × 10−8 Hz/s2 . When taking αmax = 4 × 10−3 Hz/s (see Table 5.6)
then δf0 |max = 2.4 Hz. The minimum step in frequency δf0 |min = 1/T becomes 8.4×10−4
Hz.
5.6
Constructing the hit maps
In order to discriminate between noise and signal, a threshold will be set on the collection
of maximum correlations. Each filter of which the correlation exceeds this threshold is
called a ‘hit’. The threshold depends on the amount of trials done and the desired number
of noise hits. The false-alarm probability is used to quantify this threshold.
Assuming that a gravitational wave signal is present in the data, the threshold set
will give rise to a probability that the signal will be missed. This probability is known as
the false-dismissal probability and can be used, together with the false-alarm probability,
to estimate the sensitivity of Polynomial Search per stretch of data.
A filter bank (Table 5.6) is constructed by using the limits on the polynomial search
parameters given in Table 5.2 together with the circular and eccentric binary (parameters
given in Table 5.1). A Polynomial Search with these parameters covers the parameter
parameter
Value
f0
α
β
Nfil
T
1000 ± 8 Hz
±4 × 10−3 Hz/s
±2 × 10−5 Hz/s2
4830694
1200 s
32 Hz
fs, filter
Table. 5.6: The filter bank used to perform Polynomial Search. Here, f0 ,
the parameters of Polynomial Search and the values represent
value and the range of each parameter, Nfil are the number of
ated with the supplied ranges in the parameters, T = 2Tcoh is
the data and fs,filter is the sample frequency of the filter.
α and β are
the central
filters crethe length of
81
Chapter 5. Polynomial Search
space such that the signals with frequency derivatives smaller than the values in the
table will be detected. This filter bank is used to verify the false-alarm probability and
compute the false-dismissal probability of Polynomial Search applied to a single stretch
of data. The search bandwidth, ∆fPS is defined as the frequency interval of the data
to which the filters are applied. In the case of the parameters shown in Table 5.6, the
search bandwidth is
∆fPS = (f0 )max − (f0 )min = 16 Hz.
(5.28)
The hit maps themselves are constructed by taking the filters which passed the
threshold and superimposing them. These hit maps will describe the frequency evolution
of the gravitational wave signal.
5.6.1
False-alarm probability
The probability that, given a threshold and in absence of a signal, at least one filter
will give a hit is called the false-alarm probability. Note that as described in section 5.3,
the maximum correlation for all the lags is taken. The correlation C for each lag is a
random number drawn from a probability density function (pdf) f (C). Reference [68]
shows that the pdf, ρ(C), of the order statistic for the maximum of N random numbers
drawn from f (C) can be written as
ρ(c) = N F (C)N −1 f (C),
(5.29)
where F (C) is the cumulative distribution function of f (C).
In the case of Polynomial Search the number of lags N = nlags is given by
nlags =
fs,filter T
,
2
(5.30)
where the factor 2 arises because of the fact that half the filter is zero-padded. This
number of lags is in essence the number of sample points in the filter as each sample
point represents a filter which is compared to data. Furthermore, the pdf f (C) is assumed
to be Gaussian,
2
(C−µ)
− √
1
2σ 2
f (C) = √
e
,
(5.31)
2πσ
where σ 2 is the variance of the noise and µ is a measure for the amplitude of a potential
signal hidden in the data.
With the Gaussian assumption together with Eqs. (5.29) and (5.30), the pdf for
finding a maximum correlation of C in nlags lags becomes
"
#nlags −1
2
C−µ
(C−µ)
− √
nlags 1 + erf( √2σ2 )
2σ 2
ρ(C|µ, σ, nlags ) = √
×e
.
(5.32)
2
2πσ
In the case of pure noise (no signal), µ = 0 and σ = 1. Figure 5.13 shows ρ(C) from Eq.
(5.32) for Polynomial Search applied to white Gaussian noise with parameters shown in
82
ρ
5.6. Constructing the hit maps
1.4
1.2
1
0.8
0.6
0.4
0.2
0
0
1
2
3
4
5
6
C
Figure. 5.13: The probability density function, ρ, for finding a correlation C in a series
of nlags = 19, 200 lags in the case of pure noise (µ = 0 and σ = 1).
Table 5.6 (nlags = 19, 200). From this figure it can be clearly seen that the pdf is nonGaussian. This type of pdf is known as a ‘Gumbel’ distribution [69] and has a positive
mean with nonzero skewness.
The probability that in absence of a signal none of the maximum correlations out of
nlags lags passes a threshold Cthr can be computed from Eq. (5.32) to be
!nlags
Z Cthr
)
1 + erf( C√thr
2
pnone =
ρ(C, µ = 0, σ = 1) dC =
.
(5.33)
2
−∞
The step sizes defined in section 5.5 will yield a total number of filters, Nfil . Assuming
that the filters are independent it is advantageous to consider the number of filters per
Hz of searched bandwidth Nfil0 = Nfil /∆fPS where ∆fPS is defined in Eq. (5.28). The
false-alarm probability per Hz searched bandwidth pfa then becomes
0
Nfil
pfa (Cthr ) = 1 − (pnone )
=1−
where
Ntrials =
1 + erf( C√thr
)
2
Nfil nlags
.
∆fPS
2
!Ntrials
,
(5.34)
(5.35)
In order to verify Eq. (5.34), Polynomial Search is applied to Gaussian white noise.
The parameters employed in the polynomial search are given in Table 5.6. The correlation of the filters with the data are computed and the false-alarm probability has then
been computed for each value of the threshold. In order to do this the filters are counted
83
Chapter 5. Polynomial Search
false alarm probability
by binning the frequency of the middle of the filter, in this case, in 1 Hz bins. The
false-alarm probability is then computed by registering as a function of the threshold 1
if at least one filter out of Nfil filters gave a hit or 0 if no hit occurred. These 0s and 1s
have then been averaged over 1 Hz slices of the search bandwidth (e.g. ∆fPS = 16 Hz
for this particular filter bank) for each FFT and the results are then averaged over 24
FFTs. The false-alarm and the error were given by the mean and the standard deviation,
respectively.
In practice the filters are correlated, meaning that the false-alarm probability calculated with Eq. (5.34) is overestimated. This is due to the fact that the internal correlations between the filters reduce the amount of independent trials. In order to calculate
the false-alarm probability with the correlations between the filters taken into account,
one needs to compute all the correlations between the filters individually. This computation would be impractical when dealing with a large number of filters due to the
combinatorics. Therefore, these correlations have been taken into account by using an
effective number of trials, Neff , fitting Eq. (5.34) to the data points. Figure 5.14 shows
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
5.5
Uncorrelated: Ntrials = 5.797 × 109
Simulation: Neff = 2.998 × 109
6
6.5
7
7.5
Cthr
Figure. 5.14: The false-alarm probability for the uncorrelated case where Ntrials =
5.797 × 109 (dashed curve) which is computed with Eqs. (5.34) and
(5.14). The data points represent the false-alarm probability calculated
by applying the filters to Gaussian noise and the solid curve is a fit to
these points yielding Neff = 2.998 × 109 .
the false-alarm probability as a function of Cthr for two cases. The ‘uncorrelated’ case
is the false-alarm probability assuming that the filter bank is completely uncorrelated
and that Ntrials can be computed with Eq. (5.35). The second case, termed ‘simulation’
is the result of applying Polynomial Search to the simulated data.
From Fig. 5.14 it can be seen that the expected, theoretical, false-alarm probability
(Eq. (5.34)) overestimates the actual false-alarm probability. The difference between the
84
5.6. Constructing the hit maps
theoretical curve and the simulated curve in terms of the effective number of trials is a
measure of the degree of internal correlations of the filters within the filter bank. The fit
value of the number of trials can be interpreted as the effective number of independent
trials, Neff = 2.998 × 109 .
In order to be able to separate noise hits from signal hits, a threshold on the falsealarm probability must be set. For any desired false-alarm probability the threshold is
given by inverting Eq. (5.34) which yields
i
h
√
1
−1
Neff
Cthr = 2 erf
−1 .
(5.36)
2(1 − pfa )
When taking a false-alarm probability (per Hz) of 1%, 10% or 50%, the corresponding
values for Cthr for the parameters in Table 5.6 are 6.81, 6.51 and 6.24, respectively for
Gaussian noise.
5.6.2
False-dismissal probability
The false-dismissal probability is defined as the probability that no correlation between
filter and data passes a given threshold in the presence of a signal. The false-dismissal
probability can be calculated when the correlations of all filters with the signal with
amplitude Asig are known. Also the assumption is made that out of the nlags (Eq. (5.30))
lags, the maximum correlation will give the correlation with the signal while the others
can be neglected.
The probability that none of the Nfil filters will cross the threshold Cthr , while a
signal is present in the data can be computed. The false-dismissal probability becomes
Neff −Nfil

Z Cthr
1 + erf C√thr
2
fil

pfd (SNR, Cthr ) = ΠN
ρ(C, µi , σ = 1)dC × 
, (5.37)
i=1
2
−∞
where the first factor represents the Nfil filters which yield a mean correlation of µi .
The second factor represents all the other lags which were not considered since only
the maximum correlation out of nlags lags was taken. Also, according to the assumption
stated above, these correlations were assumed to be too small to cross the threshold.
The fact that the filters in the filter bank are correlated is also taken into account by
using the effective number of trials, Neff , instead of Ntrials .
After computing the integral in Eq. (5.37), the false-dismissal probability becomes
 
Neff −Nfil

1 + erf Cthr −µ√i2(SN R)
1 + erf C√thr
2
fil
×

pfd (SNR, Cthr ) = ΠN
. (5.38)
i=1
2
2
From Eq. (5.38) it can be seen that in the limit of SNR → 0, the false-dismissal probability reduces to 1 − pfa . This is to be expected as when SNR → 0 only the noise
contributes. This implies that the probability to miss a (weak) signal is equivalent to
not having a false alarm (1 − pfa ).
85
Chapter 5. Polynomial Search
The pdf from which the correlations are drawn is shown in Eq. (5.32) and is a function of the expectation value µ of the correlation statistic. In the presence of a signal
s(t), the expectation value for the i-th filter µi is given by Eqs. (B.16) and (B.18)
h
i
(5.39)
µi = SNR × ci = SNR × F −1 ŝ(f )F̂i∗ (f ) ,
r
where
Asig √
T
√ N
=
(5.40)
Sn
Anoise 2
as derived in appendix B. Note that 0 < i < Nfil .
In order to compute the false-dismissal probability with Eq. (5.38), the actual signal
shape is needed. Even though this can be accomplished by means of simulations, the
false-dismissal probability obtained this way will be only relevant for the specific signal
shape and filter bank used in the simulation. A simplistic estimate can be given when
assuming that the filter bank is orthonormal meaning that all filters are independent
from each other. This will imply that only a single of these orthonormal, filters will give
a correlation of c = 1 and all the other filters will have negligible correlations with the
signal. This assumption will decouple the choice of filter bank and signal shape from the
false-dismissal probability. Equation (5.38) will then reduce to
SNR ≡ Asig

pfd (SN R, Cthr ) = 
1 + erf
Cthr√
−SN R
2
2
 

1 + erf
2
C√thr
2
Neff −1

.
(5.41)
Since Eq. (5.41) is based on the assumption that the filters are uncorrelated, it will be
defined as the ‘uncorrelated’ case, while Eq. (5.38) is based on the actual filter bank.
Figure 5.15 shows both the uncorrelated case and the actual case for the filter bank
given in Table 5.6 for both the circular and eccentric binary. The correlations resulting
from the application of the filter bank to the signal directly give ci described in Eq. (5.39).
In order to compute the data points shown in the figure, the false-dismissal probability
given by Eq. (5.38) has been computed for each FFT and subsequently averaged over a
single period of the signal (11 FFTs in the circular case and 35 FFTs in the eccentric
case). The error bars represent the standard deviation to the mean value over this single
period.
When comparing the two panels in Fig. 5.15 it can be seen that the false-dismissal
probabilities for both signals are similar. However, in the eccentric case the variation
of the false-dismissal probability is somewhat larger than in the circular case. This can
be explained by the fact that the coherence time has been set such that the periastron
will not be covered optimally with this particular filter bank as explained in section 5.4.
Despite this minor difference, it is deduced that the filter bank covers both the eccentric
and the circular binary sufficiently so that the false-dismissal probabilities for both cases
will be identical.
From Fig. 5.15 it can be seen that the uncorrelated false-dismissal probability does
not match the actual false-dismissal probability. The filter bank was chosen to be dense
86
1
1% FAP
10% FAP
50% FAP
Uncorrelated
0.8
0.6
1
1% FAP
10% FAP
50% FAP
Uncorrelated
0.8
0.6
0.4
0.4
0.2
0
0
false dismissal probability
false dismissal probability
5.7. Combination of hits
0.2
1
2
3
4
5
6
7
8
9 10
SNR
0
0
1
2
3
4
5
6
7
8
9 10
SNR
Figure. 5.15: The false-dismissal probability for the circular (left panel) and eccentric
binary (right panel) averaged over a single period of the binary. The data
points represent the average and the error bars the standard deviation
of the false-dismissal probability. The dashed curve represents the falsedismissal probability assuming perfect correlation for a single filter. The
data points represent the actual situation where the simulated signals
and the filter bank are used. The false-alarm probability has been set to
1%, 10% and 50%, respectively.
enough so that several filters (at least 20) will produce a correlation larger than 0.7.
This means that the false-dismissal probability, calculated according to Eq. (5.38), is
reduced with respect to the false-dismissal probability for an orthonormal filter bank.
5.7
Combination of hits
In sections 5.6.1 and 5.6.2 the statistics of the hits for a single FFT of data have been
explained. However, the gain in sensitivity for detecting gravitational waves from periodic sources lies in the integration of the signal over long stretches of time. Therefore,
the hits from each SFT must be combined. Polynomial Search employs incoherent combination by constructing a number count statistic from the number of filters which cross
a certain threshold. The statistical treatment of this statistic is analogous to that of the
Hough search described in Ref. [63].
The number count is constructed per frequency band. Then, there are two ways of
constructing the number count, either counting the number of hits for each FFT and
adding them up, or counting the number of FFTs which have one or more hits. The former will be less robust since non-stationarities and instrumental lines will have a larger
effect on the number count. The latter will suffer from a lack of statistics when dealing
with fewer FFTs to process.
87
Chapter 5. Polynomial Search
5.7.1
Number count statistics
The probability that khits out of Nfil filters cross the threshold per Hz of search bandwidth
is defined as P (khits ). The probability that kffts out of Nfft FFTs have at least one
hit in them per Hz is defined as P (kfft ). Both probabilities are given by the binomial
distribution as
Nfil !
pkhits (1 − phit )Nfil −khits and
khits !(Nfil − khits )! hit
Nfft !
P (kffts ) =
pkffts (1 − pfft )Nfft −kffts .
kffts!(Nfft −kffts )! fft
P (khits ) =
(5.42)
(5.43)
Here, phit is the probability that a single filter will cross the threshold on the correlation
and pfft is the probability that at least one hit will be found in a single FFT. When no
signal is present the false-alarm probability as defined in section 5.6.1 (e.g the probability
that at least one filter or FFT gives a hit while no signal is present) is related to phit and
pfft . In the case of counting hits, Eq. (5.42) gives the probability of finding khits which
can be related to the false-alarm probability by considering that
pfa = 1 − P (khits = 0) = 1 − (1 − phit )Nfil .
→ phit = 1 − (1 − pfa )
(5.44)
1
Nfil
Inverting Eq. (5.44) will give phit as a function of pfa . When counting FFTs with hits in
them the relation between the pfa and pfft is simply
pfa = pfft .
(5.45)
The relation given in Eq. (5.44) implies that when the false-alarm probability for a single
FFT is computed, the probability for finding an arbitrary number of hits in an arbitrary
number of FFTs is fully determined by Eq. (5.42). The same holds true for Eq. (5.45)
when counting FFTs with hits in them. However, Eq. (5.44) implies that the probability
for each filter to cross the threshold is independent. It has been shown in section 5.6.1
that this situation can be approximated by using the effective number of independent
filters. On the other hand when deriving Eq. (5.45) no such assumption has been made.
The probability of finding khits hits in a single FFT is given by substituting Eq. (5.44)
into Eq. (5.42). When applying Polynomial Search to Nffts independent FFTs and search
bandwidth ∆fPS , the expectation value µn and the variance σn2 of the number of noise
hits are
µn = Nfil phit Nfft ∆fPS ,
σn2 = Nfil phit (1 − phit )Nfft ∆fPS .
(5.46)
(5.47)
It can be seen in Eq. (5.46) that the expectation value consists of the term Nfil phit which
is the mean of the binomial distribution (5.42). The term Nfft ∆fPS arises from the fact
that the FFTs and frequency bands are independent. The same argument holds for the
variance (Eq. (5.47)). Note that phit depends on the false alarm probability as shown in
88
5.7. Combination of hits
Eq. (5.44).
When counting FFTs with hits in them, Eq. (5.45) together with Eq. (5.43) gives
the mean and variance for this counting strategy as
µn = Nfft pfft ∆fPS ,
σn2 = Nfft pfft (1 − pfft )∆fPS .
(5.48)
(5.49)
Here, the false alarm probability is identical to pfft .
Because Polynomial Search employs a scheme with FFTs having a 50% overlap, the
total observation time as a function of the number of FFTs can be written as
Tobs =
Nfft + 1
T.
2
(5.50)
For the filter bank with parameters shown in Table 5.6, the expected number of noise
hits is computed for the different total observation times shown Table 5.7. The result is
Tobs
Nfft
1200 seconds
1
1 days
143
30 days
4,319
1 year
52,595
Table. 5.7: The total observation times for which the number count statistics are computed together with the associated number of overlapping FFTs according
to Eq. (5.50).
shown in Fig. 5.16. This figure shows the linear scaling of the expected number of noise
counts as a function of the number of FFTs or the observation time. Furthermore, it
shows that the number of expected noise hits falls exponentially when the threshold is
raised. When performing Polynomial Search and counting the hits, these number counts
can be analyzed for statistical excess in a particular frequency band, depending on which
counting strategy is used.
In the case that a signal is also present in the data, the probability of finding k hits
is still the same as given in Eq. (5.42) and (5.43). However, phit and pfft are now related
to the false-dismissal probability since this is the probability that no filters (or FFTs)
cross the threshold. In this case pfd is related to phit and pfft as
pfd = P (khits = 0) = (1 − phit )Nfil
pfd = 1 − pfft ,
(5.51)
(5.52)
for counting hits or FFTs with hits in them, respectively. The observed number count
is related to the false-dismissal probability by inverting Eq. (5.51) or Eq. (5.52) and
substituting the result in Eq. (5.48). This results in the mean number of hits when a
signal is present µs :
1
µs = Nfil Nfft (1 − pfd ) Nfil ∆fPS
(5.53)
89
106
1200 seconds of data
1 day of data
30 days of data
1 year of data
105
104
µn
µn
Chapter 5. Polynomial Search
106
30 days of data
104
103
103
2
102
10
10
10
1
1
-1
-1
1 year of data
10
10
-2
10
1200 seconds of data
1 day of data
105
6
6.2
6.4
6.6
6.8
7
Cthr
10-26
6.2
6.4
6.6
6.8
7
Cthr
Figure. 5.16: The expected number of noise hits (µn ) for 1200 seconds (1 FFT), one
day, one month and one year observation time as a function of Cthr . The
left panel shows the expected number of noise hits in the case of counting the number of filters and the right panel shows the number of noise
hits when counting the number of FFTs which contain at least one noise
hit.
for counting filters and
µs = Nfil Nfft (1 − pfd )∆fPS
(5.54)
107
µs
µs
for counting FFTs. Note that pfd depends on SNR through the relation given in Eq.
(5.38). The mean number of hits in the presence of signal as a function of SNR and
observation time is shown in Fig 5.17.
1200 seconds of data
1 day of data
30 days of data
1 year of data
6
10
105
107
105
104
104
103
103
102
102
10
10
1
1
-1
-1
10
0
1
2
3
4
5
1200 seconds of data
1 day of data
30 days of data
1 year of data
106
6
7
8
9 10
SNR
10
0
1
2
3
4
5
6
7
8
9 10
SNR
Figure. 5.17: Expected number of hits as a function of the SNR of a potential signal
for a false-alarm threshold of 10%. The left panel shows the result when
counting individual filters and the right panel shows the result when
counting FFTs with hits in them. The results are computed for 1200 seconds, 1 day, 30 days and 1 year of data, respectively.
The threshold for both panels has been set to a false-alarm probability of 10% (Cthr =
6.51). The expected number of hits is related to (1−pfd )1/Nfil and (1−pfd ) for the left and
90
5.7. Combination of hits
right panel, respectively. With these two remarks, it can be confirmed by considering
Fig. 5.15 that signals with SNR above about 6 will have a non-zero value for µs for a
single FFT (1200 seconds of data). Furthermore, the scaling as a function of observation
time is again linear as in Fig. 5.16. In the low SNR limit, the number of expected hits
will reduce to the false-alarm probability. Finally, the number of expected hits in the
high SNR limit will reduce to the total number of filters or the total number of FFTs
for the left and right panel, respectively.
5.7.2
Renormalized number count
In order to acquire possible gravitational wave candidates, a threshold on the number
count must be set. This threshold will be of the form
nthr = µn + λσn ,
(5.55)
where nthr denotes the number count threshold, µn the expectation value of the noise
hits, λ the significance level and σn the standard deviation of the noise hits. When
taking η to be the observed number count in a particular frequency interval, it is more
convenient to define the renormalized number count as
Ψ(SNR) ≡
η − µn
.
σn
(5.56)
In this way, the significance level λ is a direct threshold on the renormalized number
count. Given the the fact that the noise is Gaussian and white, the renormalized number count is normally distributed. Substituting Eqs. (5.53) and (5.54) for the observed
number of hits (η) in (5.56) will give the expectation value of the renormalized number
count as a function of SNR. This renormalized number count is shown in Fig. 5.18 for
different observation times. Note that if NFFT and/or Nfil are large enough, Eq. (5.56)
is drawn from a normal distribution with an expectation value is related to the SNR of
a potential signal (Fig. 5.18) and variance 1.
The number count false-alarm probability, p?fa , is defined as the probability that at
least one renormalized number filter or FFT crosses the significance level λ defined in
Eq. (5.55) in the case that SNR= 0. As the renormalized number count for this case is
a standard normal distributed (mean 0 and variance 1), the number count false-alarm
probability can be written as a function of the threshold as
Z ∞
1 − 1 x2
1
λ
?
2
√ e
pfa =
.
(5.57)
dx = 1 −
1 + erf √
2
2π
2
λ
Which means that a significance level of λ = 3 corresponds to p?fa = 0.13% and λ = 6
corresponds to p?fa = 9.9 × 10−8 %
When setting this threshold λ, and with the false-alarm and false-dismissal probabilities given in the previous section, the sensitivity of Polynomial Search can be computed
for the filter bank from Table 5.6.
91
Chapter 5. Polynomial Search
105
Ψ
Ψ
105
1200 seconds of data
1 day of data
30 days of data
1 year of data
104
104
103
103
102
102
10
10
1
0
1
2
3
4
5
1200 seconds of data
1 day of data
30 days of data
1 year of data
6
7
8
9 10
SNR
1
0
1
2
3
4
5
6
7
8
9 10
SNR
Figure. 5.18: The expectation value of the renormalized number count as a function
of the SNR of a potential signal for a false-alarm threshold of 10%. The
left panel shows the result when counting individual filters and the right
panel shows the result when counting FFTs with hits in them. The results are computed for 1200 seconds, 1 day, 30 days and 1 year of data,
respectively.
5.7.3
The sensitivity
The sensitivity of Polynomial Search is computed by setting the threshold λ on the
renormalized number count and by solving Eq. (5.56) for SNR. Since the false-dismissal
probability is computed with simulations, the sensitivity must also be computed with
these simulations. When making the assumption that the SNR of any search for continuous waves will go as a certain power of the observation time, the sensitivity of the
search can be parameterized as
−p2
SNR = p1 Nfft
,
(5.58)
where p1 and p2 are parameters to be determined from the simulated binaries. Note that
SNR here means the minimum SNR which can be detected given a value for λ and Cthr .
Figure 5.19 shows the sensitivity of Polynomial Search as a function of the number
of FFTs for different false-alarm probabilities at λ = 6 and λ = 3. It can be seen from
the difference between the top and bottom panels in Fig. 5.19 that the sensitivity for
counting FFTs with hits in them is slightly worse at fixed number of FFTs. This follows from the fact that the simulation shows that a higher SNR is needed for the same
significance, number for filters and false alarm probability. However, the scaling as a
function of FFT number is slightly better. Furthermore, as stated before, the top panel
shows that due to a lack of statistics a certain number of FFTs are required in order to
surpass the significance level. This can be seen by comparing the curves made for λ = 6
to the ones made for λ = 3. The former will require more FFTs for a signal to trigger
enough hits (FFTs in this case) to become significant. Also for the λ = 6 curves it can
also be seen that for higher false-alarm probability more FFTs are needed to produce a
significant number count. On the other hand it might be more advantageous to count
FFTs with hits instead of individual hits when the noise is non stationary. This count-
92
SNR
5.7. Combination of hits
7
FAP = 1%, λ = 6
FAP = 10%, λ = 6
FAP = 50%, λ = 6
FAP = 1%, λ = 3
FAP = 10%, λ = 3
FAP = 50%, λ = 3
6.5
6
5.5
5
4.5
4
SNR
3.5
1
10
102
103
6.5
104
105
Nfft
FAP = 1%, λ = 6
FAP = 10%, λ = 6
FAP = 50%, λ = 6
FAP = 1%, λ = 3
FAP = 10%, λ = 3
FAP = 50%, λ = 3
6
5.5
5
4.5
4
3.5
3
1
10
102
103
104
105
Nfft
Figure. 5.19: The SNR as a function of the number of FFTs for different values of the
false-alarm probability and for significance level of λ = 6 and λ = 3. The
solid curves are fits of the sensitivity parameterization. The top panel
shows the SNR curve when counting FFTs with hits in them and the
bottom panel shows the SNR curve for counting individual filters.
ing strategy will be more robust since a single noisy FFT will count only once in the
number count statistic while when counting filters a noisy FFT will produce many more
hits. This effect is due to the excess power with respect to the estimated PSD will cause
the C-statistic to be artificially increased and thus many filters will cross the threshold.
Also, when counting individual hits, the internal correlations of the filter bank are not
properly described which will result in a non-Gaussian distributed renormalized number
93
Chapter 5. Polynomial Search
count. This is due to the fact that Eq. (5.51) is derived under the assumption that the
internal correlations in the filter bank can be neglected.
Table 5.8 shows the values for the parameters p1 and p2 for this filter bank for different values of the false-alarm probability and of the significance level. From this table it
FAP (%) λ
1
10
50
1
10
50
6
6
6
3
3
3
{p1 , p2 (×10−2 )}filcount
5.35,
5.58,
5.81,
5.04,
5.23,
5.40,
4.15
4.59
5.10
4.15
4.56
5.06
{p1 , p2 (×10−2 )}fftcount
6.20,
6.70,
7.39,
5.73,
6.21,
6.74,
4.58
5.21
5.97
4.31
5.10
5.77
Table. 5.8: The parameters of the sensitivity model as a function of the false-alarm
probability and the significance
are given for the case
level. The parameters
−2
of counting individual filters ( p1 , p2 ×10
) and for counting
filcount
FFTs with hits in them ( p1 , p2 ×10−2 fftcount ).
can be seen that when lowering the significance level from λ = 6 to λ = 3, the minimal
SNR of a detectable signal goes down by about 7% at fixed number of FFTs for both
the counting strategies. Furthermore, lowering the threshold on the correlation (e.g. by
increasing the false-alarm probability) will increase the scaling of the sensitivity as a
function of the number of FFTs. However, it will also reduce sensitivity at fixed number of FFTs with respect to setting a higher correlation threshold. The cross-over is at
about 2, 000 FFTs for counting individual filters while the crossover for counting FFTs
is > 105 due to the lower statistics.
Another observation from Fig. 5.19 is that there is some gain in increasing the observation time by adding the hits over a set of FFTs. The sensitivity of other semi-coherent
−1/4
searches as mentioned in section 4.3, scales as Nfft . The scaling of the sensitivity of
Polynomial Search as a function of the number of FFTs is given by the p2 parameter in
Eq. (5.58). As can be seen in Table 5.8 Polynomial Search scales with a lower power of
the number of FFTs with respect to the traditional semi-coherent searches. This apparent disadvantage of Polynomial Search can be explained by the following two effects.
Firstly, the threshold on the false-alarm probability per filter given in Eq. (5.44)
(phit ≈ 1.9 × 10−7 for pfa = 10%) is much lower than in the Hough search (typically
phit = 0.1). This implies that the threshold on the correlation is much higher than for
a Hough search. This effect can be mitigated by lowering the threshold and indeed the
slope of the curves in Fig. 5.19 will increase. Since Polynomial Search employs a bank
of filters, the probability that a hit will be a false-alarm will depend on the number of
filters applied. This imposes a lower limit on the threshold which will be higher than in
the Hough case. Also, maximizing the correlation introduces extremal value statistics.
In that case setting the threshold to for example Cthr = 3, will cause virtually all the
filters to be a hit (see also Fig. 5.13), making such thresholds unusable. This implies
that the Hough-like scaling of the signal-to-noise ratio will not be achieved by simply
94
5.7. Combination of hits
lowering the threshold.
Secondly, when considering a Hough search, the frequency bins which have been
selected by setting the threshold on the power. These bins must be combined in such
a way that the signal will add incoherently. In the case of isolated neutron stars this
combination is done by binning in the sky positions (α, δ) and spindown parameters
(f 1 , f 2 , . . . , f s ). If the same incoherent combination for Polynomial Search is to be
achieved, then the filters have to be combined in a way which is consistent with a
binary system with the Keplerian orbital parameters (P, ap , e, ω0 and τ ). This would
imply that the hits have to be binned not only in the sky position and spindown parameters, but also in the 5 additional
Keplerian parameters which would be computationally
√
4
unfeasible. Note that the T sensitivity gain in the incoherent summation of the Hough
search is obtained for each of the (α, δ, f 1 , . . . , f s ) bins. The number of these bins can
be seen as a number of trials and will increase as a function of T . Furthermore, if one
would want to apply a Hough search to binary systems, the coherent integration time
would be limited to about 20 seconds due to computational arguments as explained in
section 4.1.
When combining the definition of SNR given in Eq. (5.40) and Eq. (5.58), an expression for the minimal detectable gravitational wave signal can be constructed. This
expression is of the form
r
Sn
−p2
h=
p1 Nfft
.
(5.59)
T
Note that the h parameter is equivalent to the Asig parameter in Eq. (5.40). This implies
that the actual gravitational wave amplitude will have to be higher since the beampattern function has been absorbed in this parameter.
When assuming a false-alarm probability of 10%, a significance level of 6, together
with the parameterization of the SNR, the sensitivity curve shown in Fig. 5.19 and
counting FFTs with hits in them, the minimal detectable gravitational wave amplitude
is
p
−5.21×10−2
h = 0.19 × Nfft
× Sn .
(5.60)
This minimal detectable gravitational wave amplitude can computed with the Advanced
Virgo design sensitivity by inserting the design PSD in Eq. (5.60). Figure 5.20 shows the
sensitivity of Polynomial Search when assuming Advanced Virgo sensitivity for various
observation times.
The sensitivities shown in Fig. 5.20 are valid for the case that the amplitude modulation can be ignored during a single FFT. This assumption is reasonable in case the
length of the FFT is small compared to the Earth’s daily motion. Furthermore, Poly
<
nomial Search with these
parameters will cover all gravitational waves with df
dt
2 4 × 10−3 Hz/s and ddt2f < 2 × 10−5 Hz/s2 as long as the higher-order terms obey
3 d f 3
dt3 < 3.7 × 10−10 Hz/s . This includes residual Doppler effects like the Earth’s motion
(yearly and daily) as well as other frequency modulation induced by glitches and spindown.
As stated before, the advantage of adding multiple FFTs to increase the signal-to-
95
h
Chapter 5. Polynomial Search
10-21
1200 seconds of data
1 day of data
30 days of data
1 year of data
10-22
10-23
10-24
10-25
10
102
103
f (Hz)
Figure. 5.20: The sensitivity of Polynomial Search as a function of the frequency for
the Advanced Virgo design sensitivity. The sensitivity estimates have
been computed with Cthr = 6.51 (F AP = 10 %), λ = 6 (a 6σ significance
level), counting FFTs with hits in them and ignoring the effects of the
beam-pattern function of the detector.
noise ratio is rather limited. This means that in order to get a statistical significant
sample, more FFTs have to be included in the analysis. Furthermore, in order not to
be limited by the antenna pattern of the interferometer, it is advantageous to have an
observation time of at least a few days. Since this beam pattern function is periodical
over one day, the effect it has on the renormalized number count will be averaged out
when enough FFTs are taken.
5.7.4
Consistent filters
The gain of adding more FFTs to the analysis can be increased when the threshold on
the correlation can be lowered. This lowering can be achieved by exploiting the fact that
a potential signal is continuous and thus the filters which correspond to this signal should
be continuous over the FFT boundaries. When imposing constraints on the difference
of the filter derivatives between hits in consecutive FFTs, inconsistent hit pairs can be
discarded. By keeping the consistent hit pairs, the threshold on the correlation can be
lowered and this improves the sensitivity of Polynomial Search.
A consistent hit pair is defined as a hit from the first FFT and a hit from the second
FFT of which the frequency, first and second order derivatives match within a certain
tolerance. These derivatives are compared by extrapolating to a time tm . The quantity
tm is defined as
tend,1 + tbegin,2
tm =
,
(5.61)
2
96
5.7. Combination of hits
where tend,1 and tbegin,2 are the end time of hit 1 and the start time of hit 2, respectively.
The tolerance T is defined in an empirical manner as the allowed frequency mismatch
between two filters at this intermediate time. With this definition and in combination
with the step sizes in the parameters defined in section 5.5, the allowed mismatch of the
higher-order derivatives will scale as T −n . Consistent filters obey the relation
δf (n) (tm ) < ξn
T
,
Tn
(5.62)
false alarm probability
where the tolerance T is some empirical value, δf (n) (tm ) represents the difference of the
n−th order derivatives of the filters evaluated at time tm and ξn is a constant for each
derivative. Ideally, ξn depends on the actual step size of the filter bank for each parameter. However, since the step size depends on the parameters of the filters themselves
and ξn might also depend on tm , optimal selection of ξn is non-trivial. Therefore, ξn has
been set to unity for each of the derivatives making Eq. (5.62) very restrictive.
Testing for consistency between the hits in two adjacent FFTs is a combinatorial
operation (e.g. each hit in the first FFT is matched to all of the hits in the second
FFT). To keep the combinatorics involved in checking the consistency of filter under
control an initial threshold on the correlation of Cthr = 4.5 has been set. Figure 5.21
shows the false-alarm probability when applying the consistency criterion to each pair of
FFTs with the filter bank described in Table 5.6. The data points have been computed
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
4.5
Tolerance 0.5
Tolerance 0.1
Tolerance 0.01
Tolerance 0.001
No tolerance
5
5.5
6
6.5
7
7.5
Cthr
Figure. 5.21: The revised false-alarm probability by using the consistency criterion
(see Eq. (5.62)) for different values of T . The dashed curve represents the
fitted value of the false-alarm probability when the consistency criterion
is not applied, see also Fig. 5.14.
by using the same procedure as described in section 5.6.1. However, in this case the
false-alarm probability is the probability that at least one valid pair of filters crosses the
97
Chapter 5. Polynomial Search
threshold. A pair of filters is considered valid when both filters have a correlation larger
than the threshold and adhere to the consistent filter criterion. This different interpretation means that Eq. (5.34) does not apply. Modeling the false-alarm probability in
this case would require a way to model the probability that two random filters match in
parameters and that both filter pass the threshold. Since this is highly non-trivial as the
filter bank has internal correlations and the filters are not spaced at equidistant points
in parameter space, the values for the threshold at a given false-alarm probability will
be computed by interpolation.
It can be seen from Table 5.9 that by applying the consistency criterion, the thresholds can be reduced while maintaining the same false-alarm probability. The lowering
T
1%
Cthr
10%
Cthr
50%
Cthr
−
0.5
0.1
0.01
0.001
6.81
6.14
5.60
5.06
4.72
6.51
5.87
5.44
4.88
−
6.24
5.66
5.26
4.69
−
1%
10%
50%
Table. 5.9: Thresholds for various consistent filter criteria. Cthr
, Cthr
and Cthr
are
the thresholds for false-alarm probability 1%, 10% and 50%, respectively.
of the threshold will result in a more favorable scaling of the sensitivity as a function of
the number of FFTs considered. However, the tolerance cannot be arbitrarily decreased
for a the reasons listed below
ˆ The lower limit on the tolerance is defined by the frequency (and higher order)
resolution given by the FFT time. In reality, the chosen step size in section 5.5
determines the lower limit for the value of T for each order of the derivatives.
ˆ The filters are an approximation of the actual signal. Therefore, the filters will have
a mismatch in their derivatives at the end and beginning which will increase when
noise is present. When the tolerance is set too stringent, this mismatch will cause
the false-dismissal probability to increase which in turn will negatively impact the
sensitivity.
ˆ The number of hits strongly increases with the lowering of the threshold as can
be seen in Fig. 5.13. Since applying the consistency criterion is a combinatorial
problem, the amount of computing power scales with the number of hits squared.
This implies that the lowering of the threshold will be also limited by the amount
of computing power available.
ˆ A single filter tests for correlations for all the lags between 0 < τ < T /2. Since the
maximum correlation for all these different τ values is taken, the correlations will
be Gumbel distributed [69] and have a positive mean and nonzero skewness (see
Fig. 5.13). Setting a low threshold (i.e. Cthr = 3) will then yield practically all the
filters, negating the noise-discriminating properties of computing the correlation.
98
5.7. Combination of hits
The false-dismissal probability can be computed by running Polynomial Search, including the consistency criterion, on the simulated binaries. The resulting correlations
between the filters and the signals are inserted in Eq. (5.38) which yields false-dismissal
probability. Figures 5.22 and 5.23 show the false-dismissal curves for different values of
T . Note that due to computational restrictions, the threshold for T = 0.001 can only be
computed for the case of a 1% false-alarm probability, as shown in Fig. 5.21. Therefore,
only this value for the false-alarm probability can be used to compute the false-dismissal
probability shown in the lower left panels of Figs. 5.22 and 5.23. The aforementioned
1
false dismissal probability
false dimissal probability
Tolerance = 0.01
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
0
0
1
Tolerance = 0.1
1
2
3
4
5
6
1
7
8
9 10
SNR
0
0
1
2
3
4
5
6
7
8
9 10
SNR
false dismissal probability
Tolerance = 0.001
0.8
1% FAP
10% FAP
50% FAP
0.6
FDP without tolerance
0.4
0.2
0
0
1
2
3
4
5
6
7
8
9 10
SNR
Figure. 5.22: The false-dismissal probability for the circular binary evaluated for different values of T . The upper left plot corresponds to T = 0.1, the upper
right plot corresponds to T = 0.01 and the lower left plot corresponds to
T = 0.001. The error bars and shaded regions represent the variation of
the false-dismissal probability over a single period of the binary system.
effect of increasing the false-dismissal probability can be seen in the difference between
the curves of the different panels in Figs. 5.22 and 5.23. This effect is most clearly seen
in the third panel in both figures as the variation of the 1% curve is larger than the
no-tolerance curve. This is due to the fact that there is a larger chance to miss the
signal in the FFTs where the signal has large derivatives and thus the mismatch of the
filters between the FFTs will be larger. In the eccentric case this effect is even more pro-
99
Chapter 5. Polynomial Search
1
Tolerance = 0.01
False dismissal probability
false dismissal probability
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0
0
1
Tolerance = 0.1
0.2
1
2
3
4
5
6
1
7
8
9 10
SNR
0
0
1
2
3
4
5
6
7
8
9 10
SNR
false dismissal probability
Tolerance = 0.001
0.8
1% FAP
10% FAP
50% FAP
0.6
FDP without tolerance
0.4
0.2
0
0
1
2
3
4
5
6
7
8
9 10
SNR
Figure. 5.23: The false-dismissal probability for the eccentric binary evaluated for different values of T . The upper left plot corresponds to T = 0.1, the upper
right plot corresponds to T = 0.01 and the lower left plot corresponds to
T = 0.001. The error bars and shaded regions represent the variation of
the false-dismissal probability over a single period of the binary system.
nounced since the variation was intrinsically higher due to the fact that the periastron is
not matched well by the chosen filter bank. Also, the average false-dismissal probability
at high SNR is larger than when not using the consistent filter criterion. This is due to
the fact that these filters will be rejected by the strict tolerance. At lower SNR there is
still some gain when employing the consistent filter criterion with T = 0.001, because
the threshold can be set at a lower value and thus the search will be more sensitive to
low SNR signals.
Both Fig. 5.22 and 5.23 show that when setting T = 0.001 the false-dismissal probability will increase compared to the larger values of T implying that a tolerance of 0.001
is too strict. However, setting T = 0.01 does not have this effect and therefore T has
been set to 0.01 for the remainder of the analysis. For this value of T a gain in sensitivity
of about 30 − 40 % is observed for false-dismissal probabilities between about 0.05 − 0.95
with respect to the sensitivity without using the consistent filter argument. Note that
taking T = 0.01 is an ad-hoc choice for the tolerance. Conceivably, one could refine this
100
5.7. Combination of hits
value of the tolerance, however the relative gain of such a fine tuning is expected to be
smaller than 5%.
5.7.5
The sensitivity with consistent filters
The expectation value of the renormalized number count for Polynomial Search with
a false-alarm probability of 10% which, for Fig. 5.21 and Table 5.9, corresponds to a
threshold of 5.40 is shown in Fig. 5.24. When comparing the results to those presented
105
Ψ
Ψ
105
1200 seconds of data
1 day of data
30 days of data
1 year of data
104
104
103
103
102
102
10
10
1
0
1
2
3
4
5
1200 seconds of data
1 day of data
30 days of data
1 year of data
6
7
8
9 10
SNR
1
0
1
2
3
4
5
6
7
8
9 10
SNR
Figure. 5.24: The expectation value of the renormalized number count as a function of
the SNR of a potential signal for a false-alarm threshold of 10% and for
T = 0.01. The left panel shows the result when counting individual filters
and the right panel shows the result when counting FFTs with hits in
them. The different curves are computed for 1200 seconds, 1 day, 30 days
and 1 year of data, respectively. All curves have been computed assuming
that the middle frequency of the filters have been counted in 1 Hz bands.
in Fig. 5.18 it can be seen that Polynomial Search will be more sensitive to lower SNR
signals when employing the consistent filters criterion. This sensitivity is shown in Fig.
5.25 and in Table 5.10.
From the difference between the values in Tables. 5.8 and 5.10 and Figs. 5.19 and
5.25 it can be seen that there is an improvement in sensitivity when testing filters
for consistency. Lowering the threshold on the correlation could possibly improve the
sensitivity as well. However, as stated before, the computational burden would become
too high. Another improvement could be to extend the consistency argument over more
than two FFTs (i.e. demanding that the filter derivatives are consistent over 2 FFT
borders instead of 1). This would reduce the false-alarm probability. This would also
increase the computational power required and would be more susceptible to potential
holes in the data stream. When comparing Figs. 5.19 to 5.25 it can also be seen, especially
in the top panels, that the SNR is not well fitted to the model shown in Eq. (5.58).
This can be explained by the fact that the model applies to uncorrelated filters. When
applying the consistency criterion, the interdependencies of the filter bank become more
pronounced (e.g. the probability of finding a random consistent filter pair with low
101
SNR
Chapter 5. Polynomial Search
6
FAP = 1%, λ = 6
FAP = 10%, λ = 6
FAP = 50%, λ = 6
FAP = 1%, λ = 3
FAP = 10%, λ = 3
FAP = 50%, λ = 3
5
4
3
2
1
SNR
0
1
10
102
6
103
104
105
Nfft
FAP = 1%, λ = 6
FAP = 10%, λ = 6
FAP = 50%, λ = 6
FAP = 1%, λ = 3
FAP = 10%, λ = 3
FAP = 50%, λ = 3
5
4
3
2
1
0
1
10
102
103
104
105
Nfft
Figure. 5.25: The SNR as a function of the number of FFTs for different values of the
false-alarm probability and for a significance level of λ = 6 and λ = 3.
The solid curves are fits of sensitivity parameterization. The top panel
shows the SNR curve when counting FFTs with hits in them and the
bottom panel shows the SNR curve for counting filters. The figure has
been produced for T = 0.01.
values of α and β is higher since more filters have a low α and β value). This becomes
less of an issue when considering FFTs with hits in them since an FFT can only count
once and thus setting the threshold from λ = 3 to λ = 6 will cut away most of the
non-Gaussianity.
The sensitivity of Polynomial Search for the consistent filter criterion with T = 0.01,
102
5.8. Results of the simulations
FAP (%) λ {p1 , p2 (×10−2 )}filcount
1
10
50
1
10
50
6
6
6
3
3
3
3.25,
3.57,
3.90,
3.00,
3.26,
3.15,
9.98
9.90
10.4
17.7
11.3
11.7
{p1 , p2 (×10−2 )}fftcount
4.03,
4.81,
5.94,
3.60,
4.24,
5.14,
8.71
9.69
11.0
9.09
9.95
12.2
Table. 5.10: The parameters of the sensitivity model as a function of the false-alarm
probability and the significance level. These values are computed for
T = 0.01. The parameters
are given for the case of counting individ−2
ual filters ( p1 , p2 ×10
) and for counting FFTs with hits
filcount
( p1 , p2 ×10−2 fftcount ).
λ = 6, Cthr = 4.88 (FAP = 10%), counting FFTs and with Table 5.10 together with Eq.
(5.59) can be written as
p
−9.69×10−2
h = 0.14 × Nfft
× Sn .
(5.63)
Fig. 5.26 shows the sensitivity of the Polynomal Search for the Advanced Virgo sensitivity as input for different total observation times.
5.8
Results of the simulations
In section 5.1 two binary systems have been defined in order to test Polynomial Search.
A Virgo tool called SIESTA [70] has been used to generate the waveforms. SIESTA is
a data analysis tool which is used among other things to generate waveforms originating from a variety of sources. For the two binaries described in Table 5.1, the SIESTA
module MEBinaryPulsar was used.
Both binaries have been simulated at a sample frequency of 4000 Hz. For these simulations, the detector beam pattern function is taken to be unity and the source is taken to
be aligned with the detector (e.g. ψ = 0 χ = π/2). Figure 5.27 shows the periodograms
of both binaries in the absence of noise. Note that since the FFTs are overlapping and
thus plotting all the FFTs will complicate the periodograms, Fig. 5.27 only shows the
odd-numbered FFTs. From Fig. 5.27 it can be seen that the power of the binary will be
spread out over a significant number of frequency bins. This is due to the Doppler shifts
and is most pronounced in the periastron.
As stated before, the consistent filter argument requires an a priori threshold on the
correlation such that the number of filters to be checked for consistency is manageable.
In order to set this threshold, Fig. 5.28 shows the distribution of the values for the
correlation without noise (e.g. the correlation between filter and signal is given by ci in
Eq. (5.39)) for both binaries averaged over one period of the signal. From this figure
it can be seen that both the circular and the eccentric binary have the same shape for
103
h
Chapter 5. Polynomial Search
10-21
1200 seconds of data
1 day of data
30 days of data
1 year of data
10-22
10-23
10-24
10-25
10
102
103
f (Hz)
Figure. 5.26: The sensitivity of Polynomial Search as a function of frequency compared
to the Advanced Virgo design sensitivity. The sensitivity estimates have
been computed with the consistent filter criterion for T = 0.01, Cthr =
4.88 (F AP = 10 %), λ = 6 (a 6σ significance level), counting FFTs with
hits in them and ignoring the effects of the beam-pattern function of the
detector.
correlations 0 < c < 0.15. The eccentric binary appears to have less filters with correlation 0.15 < c < 0.7 with respect to the circular case which is due to the fact that the
higher order terms contribute more for the eccentric binary. Furthermore, it can be seen
that there is a sharp rise in the number of filters with correlation c > 0.01 with respect
to filters with c < 0.01 and then an exponential decay up to c ≈ 0.15. The filters with
c < 0.01 are the filters that have a large mismatch in f0 with respect to the original
signal. The fact that these filters give a low correlation can be explained by considering
the expression for the correlation in the time-domain Eq. (A.3). The large mismatch
in f0 will imply that either g(t) or s(t) will be a fast oscillating term with respect to
the other causing the integral to reduce to a small number. For example computing the
correlation between a filter with f0 of 1004 Hz to a signal with central frequency of 1000
Hz will yield a small number for the correlation (c < 0.01). When a filter has parameters
which cause the filter to intersect with the signal in the f-t plane, but the values of the
other parameters have the wrong sign, the correlation will still be small, however not as
small as in the previous case. These filters end up in the 0.01 < c < 0.15 range. Finally,
there is a sharp rise in the filter count with c ≈ 0.15. This is due to the fact that these
filters will have parameters with the correct sign and values which correspond to the
actual signal.
Figure 5.29 shows the hit maps for both binaries for c > 0.1 (grey) and c > 0.8
(black). These hit maps are made by drawing the frequency as a function of time for
each filter passing the threshold. When comparing this figure to Fig. 5.27 it can be
104
5.8. Results of the simulations
1003
10-1
1002
1006
1005
1004
1003
1002
1001
1000
999
998
997
996
10-2
1001
10-3
1000
10-4
999
10-5
998
10-6
997
996
0
5000
10000
15000
20000
time (s)
1
frequency (Hz)
1
frequency (Hz)
1004
10-7
10-1
10-2
10-3
10-4
10-5
10-6
0
5000
10000
15000
20000
time (s)
10-7
Counts
Figure. 5.27: The periodograms for both test binaries. The left panel shows the circular binary and the right panel shows the eccentric binary. The FFT time
is taken to be T = 1200 seconds and the gray scale shows the power normalized to the total power of the signal per FFT. The total observation
time shown in these plots is 6 hours. Only the odd-numbered FFTs are
plotted.
106
Circular binary
105
Eccentric binary
104
103
102
10
1
10-1
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
Correlation
Figure. 5.28: The correlation distributions, without noise, for the circular and the eccentric binary. Both distributions have been averaged over a single period
(11 FFTs for the circular and 35 FFTs for the eccentric one).
seen that the frequency resolution of Polynomial Search is far superior to simply dealing
with the power spectra in the form of periodograms. When raising the threshold, the
frequency evolution becomes clearer. In the eccentric case (right panel of Fig. 5.29) the
large frequency derivatives in the periastron of the signal lead to no hits with c > 0.8.
As discussed in section 5.4, the coherence time is chosen such that the filter bank does
105
Chapter 5. Polynomial Search
Figure. 5.29: Hit maps (without noise) for both binaries where the gray area indicates
hits for filters with c > 0.1 and the black area corresponds to hits with
c > 0.8.
not contain filters capturing these most extreme derivatives.
While it is desirable to keep the correlation threshold high in order to have a good
frequency resolution, the sensitivity to low SNR signals will decrease. As stated in section 5.7.4 by using the consistent filter criterion, the threshold can be lowered. When
lowering the threshold, the number of filters passing this threshold will increase roughly
exponentially as can be seen in Fig. 5.28. One of the drawbacks of employing the consistent filter criterion is that it is a combinatorial operation which is a CPU intensive
task. Figure 5.30 shows hit maps of before and after applying the consistency criterion
on the filters for both binaries for a total time of 6 hours. The initial threshold on the
correlations has been set to 0.15 which leads to an average number of hits per FFT to
be tested for consistency of about 5 × 103 . It can be seen from these hit maps that the
Figure. 5.30: The hit maps of the circular and eccentric test binary where the gray
area represents all the hits which passed the initial correlation threshold
of 0.15 and the black area represents the hits which remain after applying the consistent filter criterion with T = 0.01.
106
5.8. Results of the simulations
consistent filter criterion increases the frequency resolution as was predicted in section
5.7.4. Note that the actual gain in sensitivity lies in the possibility to lower the threshold when computing the consistent filter pairs, making the number count statistic more
sensitive to potential gravitational wave signals.
In order to test the sensitivity of the search, Gaussian white noise with different amplitudes has been added to simulated binaries and Polynomial Search is applied to these
data. Figure 5.31 shows the periodograms for both binaries for a noise amplitude in the
time domain of An = 50 × h0 and An = 250 × h0 , where h0 represents the amplitude of
the gravitational wave which is, according to Table 5.1, set to 10−27 .√When expressing
the
of the
√
√ density (ASD) Sn they become
√ noise levels in terms
√ amplitude spectral
Sn = 1.11 × h0 1/ Hz and Sn = 5.59 × h0 1/ Hz, respectively. From Fig. 5.31 it
1003
1002
105
1001
1000
999
104
998
997
996
0
5000
10000
15000
20000
time (s)
frequency (Hz)
1004
103
106
1003
1002
105
1001
1000
999
104
998
997
996
0
5000
10000
15000
20000
time (s)
103
1006
1005
1004
1003
1002
1001
1000
999
998
997
996
106
frequency (Hz)
106
1006
1005
1004
1003
1002
1001
1000
999
998
997
996
105
104
0
5000
10000
15000
20000
time (s)
103
106
frequency (Hz)
frequency (Hz)
1004
105
104
0
5000
10000
15000
20000
time (s)
103
Figure. 5.31: Periodograms normalized to the power spectral density of the noise.
The left and right top panels show the periodograms of the circular
and eccentric
binary, respectively
√ with a noise amplitude spectral den√
sity of Sn = 1.11 × h0 1/ Hz. The bottom left and right panels
show
but with an amplitude spectral density
√
√ the same periodograms,
of Sn = 5.59 × h0 1/ Hz.
can be seen that the spread of the power induced by the Keplerian Doppler shifts causes
the signal to be buried in the noise. It also shows that searches for gravitational waves
from pulsars in binary systems which rely on applying a threshold on the (normalized)
107
Chapter 5. Polynomial Search
power will be quite limited in sensitivity due to these Doppler shifts.
Figure 5.32 shows the hit maps
√ for both binaries with noise
√ and hit distributions
amplitude An = 250 × h0 or ASD Sn = 5.59 × h0 1/ Hz. The SNR is computed to be
6.20. The false-alarm probability has been set to 10 % corresponding to Cthr = 4.88 and
the tolerance for the consistent filter criterion has been set to T = 0.01. From Fig. 5.31
it can be seen that the noise amplitude is too high to resolve the signal from the power
spectrum. The hit maps shown in Fig. 5.32 however, show that Polynomial Search can
resolve the gravitational wave signal from the background with relative ease.
The hit distributions, shown in Fig. 5.32, are calculated by counting the number
of FFTs in which at least one filter will give a hit in a certain frequency band. The
hits themselves are counted by computing the frequency in the middle of the filter and
binning the result in 1 Hz bins. Polynomial Search has been applied to the data between 960 and 1040 Hz in blocks of 16 Hz. The hit maps are shown only for the interval
992 Hz < f < 1008 Hz but the FFT distribution is shown for the entire interval. From
these hit distributions it can be seen clearly that Polynomial Search can distinguish the
signal from the background as expected when considering the SNR.
In order to verify that the expected number of noise counts as described in section
5.7 is obtained, it is insightful to compare the mean and standard deviation of the noise
counts from Eqs. (5.48) and (5.49) to the observed noise counts in Fig. 5.32. Figure 5.33
shows the count distributions taken from the FFT distributions shown in Fig. 5.32. From
this figure the mean and variance of the noise distributions can be calculated by averaging the Gaussian-like distribution from approximately 0 to 20. The noise distributions
are independent of the signal and both plots are statistically independent realizations
of the noise. On average the number of noise hits taken from Fig. 5.33 is 8.2 ± 2.3
10%
hits/Hz. When taking Cthr
= 4.88 the number of noise hits is expected to be 7.1 ± 2.7
hits/Hz, implying an underestimate of the noise counts of about 12%. This discrepancy
is due to the fact that the threshold corresponding to the 10% false-alarm probability
has been obtained by linear interpolation of Fig. 5.21. Since these curves have a steep
slope, a small variation in threshold will have a large effect on the value of the falsealarm probability. Taking the slope of these curves into account, a 12% underestimate
of noise counts leads to a 0.3% overestimation of the threshold. In order to reach the
same 6σ significance, taking into account this underestimate, the threshold should be
put at λ ≈ 8.3. However, when comparing this value to the expectation value of the
renormalized number count shown in Fig. 5.24 the effect on the overall sensitivity is
minimal.
for both binaries with noise
Figure 5.34 shows the hit maps√and the hit distributions
√
amplitude An = 400 × h0 or ASD Sn = 8.94 × h0 1/ Hz. The SNR is computed to be
3.87. The signal cannot be recognized easily by eye in the hit maps for this particular
noise amplitude. However, when counting FFTs with hits in them, a clear excess is observed in the hit distribution.
From the noise counts and the estimated variance of the noise, the renormalized
number count can be calculated for both binaries as a function of frequency. The renormalized number count Ψ, as defined in section 5.7, will be the final search statistic of
Polynomial Search. Figure 5.35 shows ψ by using both the estimated noise count com-
108
5.8. Results of the simulations
#FFTs with hits
frequency (Hz)
1008
1006
1004
1002
1000
70
60
50
40
998
30
996
20
994
10
992
0
10000
20000
30000
0
960
40000
time (s)
#FFTs with hits
frequency (Hz)
1008
1006
1004
1002
1000
994
10
40000
time (s)
1000
1020
1040
frequency
40
20
30000
980
50
996
20000
1020
1040
frequency (Hz)
60
30
10000
1000
70
998
992
0
980
0
960
Figure. 5.32: The
√ distributions for both binaries with noise ASD of
√ hit maps and hit
Sn = 5.59 × h0 1/ Hz taken over a period of 12 h (72 FFTs). The top
and bottom left panels show the hit maps of the circular and eccentric
test binary, respectively. The hit maps have been made with T = 0.01
and Cthr = 4.88. The top and bottom right panels show the distributions of the FFTs in which one or more filters had a hit in a particular
frequency band for both binaries.
puted with Eq. (5.48) and the mean of the distributions shown in Fig. 5.33 together
with Eq. (5.56). It can be seen that the renormalized number count distribution for the
eccentric binary case is narrower compared to that of the circular binary. This seems
counter intuitive since the Doppler shifts for the eccentric binary are larger. However,
these large Doppler shifts only occur in the periastron, which was not covered completely
by the filter bank due to the choice of coherence time as described in section 5.4. Thus
this section of the orbit will not contribute much in Polynomial Search. On the other
hand since most of the time the eccentric binary has relatively low Doppler shifts, more
FFTs will be counted and Polynomial Search is actually more sensitive to the eccentric
case.
By setting the significance level λ = 6, a signal with SNR = 3.87 can be recovered,
but a signal with SNR = 3.44 will be missed depending on the accuracy of the estimate of the noise counts. These numbers can be compared to the predicted sensitivity
109
counts
counts
Chapter 5. Polynomial Search
10
8
10
8
6
6
4
4
2
2
0
0
10
20
30
40
50 60 70
#FFTs with hits
0
0
10
20
30
40
50 60 70
#FFTs with hits
Figure. 5.33: The renormalized number count distributions for the circular (left panel)
and eccentric case (right panel).
of the search for λ = 6 and FAP = 10 % shown in Fig. 5.25. Since the number of noise
counts has been underestimated, causing the false-alarm probability to be slightly off,
an accurate direct comparison of the SNR values with the theoretical sensitivity curve
is non-trivial. Taking the 11% underestimate into account, the computed SNR values
can be considered in agreement with those presented in Fig. 5.25.
110
5.8. Results of the simulations
frequency (Hz)
#FFTs with hits
1008
1006
1004
1002
1000
45
40
35
30
25
20
998
15
996
10
994
5
992
0
10000
20000
30000
0
960
40000
time (s)
frequency (Hz)
#FFTs with hits
1008
1006
1004
1002
1000
1000
1020
1040
frequency (Hz)
980
1000
1020
1040
frequency (Hz)
45
40
35
30
25
20
998
15
996
10
994
5
992
0
980
10000
20000
30000
40000
time (s)
0
960
Figure. 5.34: The
√ distributions for both binaries with noise ASD of
√ hit maps and hit
Sn = 8.94 × h0 1/ Hz taken over a period of 12 h (72 FFTs). The top
and bottom left panels show the hit maps of the circular and eccentric
test binary, respectively. The hit maps have been made for T = 0.01 and
Cthr = 4.88. The top and bottom right panels show the distributions of
the FFTs in which one or more filters had a hit in a particular frequency
band for both binaries.
111
Ψ
Ψ
Chapter 5. Polynomial Search
25
25
20
20
15
15
10
10
5
5
0
0
960
980
1000
960
1020
1040
frequency (Hz)
20
15
15
10
10
5
5
0
0
1000
1020
1040
frequency (Hz)
Ψ
980
960
980
1000
1020
1040
frequency (Hz)
Ψ
25
25
20
20
15
15
10
10
5
5
0
0
960
1020
1040
frequency(Hz)
25
20
960
1000
Ψ
Ψ
25
980
980
1000
1020
1040
frequency (Hz)
960
980
1000
1020
1040
frequency (Hz)
Figure. 5.35: The renormalized number count as a function of frequency for the circular (left panel) and eccentric binary (right panel). The first row of
panels shows the renormalized number count computed
with 5.59 × h
√0
√
(SNR = 6.20), the second row is computed with √
Sn = 8.94 × h0 1/ √Hz
(SNR = 3.87) and the final row is computed with Sn = 10.1×h0 1/ Hz
(SNR = 3.44). The hatched region represents the renormalized number
count computed from the observed number count from Fig. 5.33 and the
white filled region represents the renormalized number count when using the computed number count from Eq. (5.48). All number counts have
been computed running Polynomial Search with T = 0.01, Cthr = 4.88
and NFFTs = 72.
112
5.9. Discussion and outlook
5.9
Discussion and outlook
A data analysis method for an all-sky search for gravitational waves from neutron stars
in binary systems, called the ‘Polynomial Search’, has been presented. Polynomial Search
is designed for detecting gravitational wave candidates in data from interferometry experiments like Virgo and LIGO. This method employs a bank of polynomial phase templates, or filters, which approximate the phase of continuous gravitational wave signals
by a Taylor expansion. These filters, of order n, remain fully coherent for short stretches
of duration T if the higher order derivatives of the phase obey
dn+1 φ
<
dtn+1
which is equivalent to
dn f
1
<
dtn
2π
2π
Tcoh
n+1
2π
Tcoh
,
(5.64)
n+1
.
(5.65)
Note that as the filters are zero padded, the integration time T = 2Tcoh .
The data are split into short, overlapping stretches and the filter bank is applied to
each of these stretches by computing the correlation. A threshold is applied to the correlations, and the filters which do not match the signal will be discarded. The remaining
‘hits’ are collected and counted for each stretch of data. A significance is assigned to the
resulting number count statistic and another threshold on this significance is applied
to select the gravitational wave candidates. The details of Polynomial Search have been
presented and sensitivity estimates in terms of the minimal detectable gravitational wave
amplitudes are given.
A filter bank has been constructed that allows for the analysis of all binaries with
orbital parameters in between two extremal cases. These two extreme binaries are
1. A double neutron star (1.4 M ) system with a 2 hour circular orbit.
2. A 1.4 M neutron star orbiting a 14 M star in a 6 hour eccentric orbit with
eccentricity 0.6.
The integration time of the data stretches (T = 1200 s) has been chosen such that all
gravitational waves with frequency derivatives
2 df −3
< 4 × 10 Hz/s and d f < 2 × 10−5 Hz/s2
(5.66)
dt2 dt will be covered with optimal correlation as long as the third order terms obey
3 d f 3
−9
dt3 < 1.9 × 10 Hz/s .
(5.67)
Note that the range in df /dt and d2 f /dt2 can be compared to the allowed spin down
parameters in other all-sky searches for isolated neutron stars. For instance the Einstein
113
Chapter 5. Polynomial Search
Home search F-statistic search applied to S5 data [62] covered a single spin down parameter f˙ with ranges 3 × 10−12 f Hz/s < f˙ < 3 × 10−13 f Hz/s for f < 400 Hz and
−8 × 10−12 f Hz/s < f˙ < 8 × 10−13 f Hz/s for f > 400 Hz. The Hough and PowerFlux
[64] semi-coherent all-sky search have been done with −2.2 × 10−9 Hz/s < f˙ < 0 and
−1 × 10−8 Hz/s < f˙ < 0, respectively. This shows that Polynomial search can cover a
much larger parameter space in terms of spin down parameters than previous searches.
Even though Polynomial Search is designed to target binary systems, these limits
also include residual Doppler effects like the Earth’s motion (yearly and daily) as well
as other frequency modulations induced by glitches and spindown. For this filter bank,
when counting FFTs with hits and with a false alarm probability of 10% together with
a significance threshold of λ = 6 (e.g. a 6σ threshold on the significance), the estimated
sensitivity is
p
−5.21×10−2
h = 0.19 × Nfft
× Sn .
(5.68)
Furthermore, an addition to Polynomial Search called the ‘consistent filter criterion’ has
been shown to improve the sensitivity. This additional algorithm exploits the continuity
of the signal across the data stretch border to select consistent filter pairs. Two filters
form a consistent filter pair if the difference between their filter parameters is smaller
than a certain tolerance. Setting this tolerance to T = 0.01, together with the previous
values for the false alarm probability and significance, the sensitivity equals
p
−9.69×10−2
h = 0.14 × Nfft
× Sn .
(5.69)
The application of this consistent filter criterion yields an improvement of about a factor
of 2 (for a 1 year total observation time). The sensitivity estimates have been verified
by analyzing Gaussian noise with software injections.
The aforementioned sensitivities have all been computed assuming Gaussian and
stationary noise. However, Polynomial Search is to be applied to data from experiments
like LIGO and Virgo for which the noise in general cannot be considered Gaussian
distributed and stationary over large frequency bands. A reliable method to characterize
the power spectral density, Sn , of the noise has to be applied when analyzing real data5 .
In order to verify the sensitivity estimates given in this thesis for more realistic
data behavior it is necessary to redo the analysis in section 5.8 for the case of nonGaussian noise. The most accurate way of doing this is to inject simulated gravitational
wave signals into the data stream of the detectors. The issue with this method is that
simply using the two discussed cases of binaries (circular and eccentric) will not be
enough. In order to fully verify the sensitivity, a large series of binaries with different
values for the 12 + s Keplerian, orbital and gravitational wave parameters must be
constructed. Polynomial Search is then to be run on the combined data plus signal
stream. The computing power required may be problematic as the verification will take
more CPU time than the search itself. Also, due to the high dimensionality of the
parameter space the result can only be an average sensitivity of the search. A possible
way of implementing such a verification could be to inject a limited set of binaries with
5
The Ligo Analysis Library (LAL) [71] provides various methods to estimate the PSD and will
possibly be used in future developments of Polynomial Search.
114
5.9. Discussion and outlook
randomized parameters.
Another point of discussion is the fact that all tests have assumed that the detector
antenna pattern function can be considered constant for the duration of a single FFT. In
reality this is not the case and the sensitivity of Polynomial Search will be a function of
sky position of the source. Since Polynomial Search is designed to be an omni-directional
search, there is no a priori way to fold the detector beam pattern function into the
sensitivity. However, when the coherent time of the search is limited to T 6 h, it
is possible to take the beam pattern function into account by weighing the filter bank
with the sky and polarization-averaged beam pattern function. Since this sky-average
still depends on the detector position as a function of time, each time the filter bank
is applied to a stretch of data, it has to be weighed with a different sky-average. When
averaging the sky and polarization-average over a single day the mean value of the beam
pattern function is computed to be 0.42 and the variation is 0.16 [42]. This implies that
the reduction of the sensitivity of Polynomial Search will roughly be around this average
value. However, for a given polarization and sky position, the sensitivity can be easily
recomputed.
5.9.1
Follow-up and parameter estimation strategies
Polynomial Search, after setting a threshold on the significance, will give candidates in
the form of frequency intervals where the significance traverses this threshold. These
candidates can already be distinguished from the background with good sensitivity (depending on the significance threshold).
The sensitivity of Polynomial Search can be improved by adding a follow-up step,
where the coherence time of Polynomial Search is increased for the candidates. This
will require the filter bank to be recomputed and this new filter bank will then only be
applied to the frequency range of the √
candidates. Since increasing the overall sensitivity
of Polynomial Search will scale with T , an increase in coherence time will be a larger
gain in sensitivity than adding more FFTs to the original search. On the other hand
the limit on the coherence time will be dictated by the orbital period of the binary
(the Taylor expansion of the phase must remain valid during time T ). In addition, the
Earth’s rotational period will impose a limit on T of no more than about 6 h due to
the fact that the induced Doppler-shift of the gravitational wave signal is sinusoidal and
that the beam pattern function will vary too much. At maximum such a follow-up step
will gain about a factor of 4 in sensitivity.
When a gravitational wave candidate is found, the next step will be to derive an estimate of the physical parameters of this candidate. One approach could be to perform
a Bayesian parameter estimation method together with a Markov Chain Monte Carlo
technique. A similar strategy is used in the targeted isolated neutron star search [72, 73].
However, physical templates must be generated in order to compute the likelihood for a
particular parameter set which is a highly CPU intensive task. Since a template-based
all-sky search for gravitational waves from neutron stars in binary systems was shown
to be unfeasible, it would require further study to see if such templates could be used
for parameter estimation of the candidates. An alternative approach would be to fit the
115
Chapter 5. Polynomial Search
phase model of the gravitational wave directly to the parameters of the hits of the candidates. This method could possibly be less CPU intensive, however such an approach
has never been attempted before in the LIGO-Virgo data analysis community.
Both of these methods and possible follow-up steps are outside the scope of this
thesis.
116
Chapter
6
Software and computing
6.1
The AnaPulsar framework
Polynomial Search has been implemented in a larger framework called AnaPulsar. This
is meant to be a multi-purpose environment in which different types of analyses can be
implemented. In order to keep the framework versatile and easily extendable by future
developers, an object-oriented (OO) approach has been chosen. AnaPulsar is entirely
written in C++ and includes classes for defining analysis pipelines, creating and handling
data containers, exception handling, IO and controlling the framework by parsing option
files. AnaPulsar uses the following external libraries:
ˆ BOOST v1.38.0 [74]: A general purpose collection of portable C++ source libraries.
ˆ FrLib v8.11 [75]: The official data format for various gravitational wave experiments, including Virgo and LIGO.
ˆ fftw3 v3.2.2 [76]: The ‘Fastest Fourier Transform in the West’ (FFTW) is a free
collection of fast C routines for computing the Discrete Fourier Transform.
ˆ ROOT v5.25.02 [77]: A set of OO frameworks with all the functionality needed to
handle and analyze large amounts of data in an efficient way.
The AnaPulsar framework has been developed with certain initial design choices:
ˆ A gravitational wave analysis consists of a series of operations on data elements.
ˆ Each data element has an associated data container.
ˆ Each analysis on a data container has an associated metadata container.
ˆ Each analysis has an input and output stream from which the data are read and
to which the data are written.
117
Chapter 6. Software and computing
ˆ The framework should be able to run the different types of implemented analyses by supplying an option file in which the desired analyses together with their
parameters are specified.
ˆ Each class in this framework should inherit from the base class:
– Each derived object has an associated state which determines whether the
object is valid.
– Any derived object must have a method which prints information about its
class.
Figure 6.1 shows the inheritance graph of AnaPulsar. As can be seen, the framework
has been compartmentalized into pieces with similar behavioral properties. These commonalities have been captured into a series of (major) abstract base classes called Options, AnalysisBase, DataObj, Generator, GenericException, GenericOperation
and IOobj. Furthermore, some minor utility classes exist which do not have any obvious commonalities with other classes. These classes are called OptionFileReader,
LogStream, AnaPulsarTime and AnaPulsarTimeDuration. Appendix D gives a more
detailed view of the inheritance relationships between the classes and their data members and methods. The following text will briefly highlight the more important classes
of the framework. A more detailed description can be found in the Doxygen-generated
code manual from the source code [78].
The AnaPulsar framework is controlled by supplying a text file containing the desired parameters and (partial) analyses the user desires. These so-called option files have
a specific syntax which can be found in the code documentation. An option file is processed by the OptionFileReader. All Options-derived classes contain the information
parsed from the option file, to initialize the relevant analyses. An analysis requires one
or more Option-derived classes to be initialized (i.e. an SFTCreation-analysis requires
an FFTOptions class in order to be initialized).
All AnalysisBase-derived classes implement the various analyses in AnaPulsar.
These analysis classes are initialized by using the information from the option file. This
information has been converted by the OptionFileReader to a series of Option-derived
classes which are collected in one of the Analysis-option classes (i.e. SFTCreationOptions for creating FFTs. As can be seen in Fig. 6.1 there are currently three AnalysisBase-derived classes implemented. Table 6.1 gives a short description of each of these
classes.
All DataObj-derived classes contain various types of data. These types can be metadata, THitTuple classes and time and frequency domain data containers. These time
and frequency domain containers derive from the DataContainer class since the FrequencyDomainData and TimeDomainData both contain a pointer to a data block. Table
6.2 gives a short description of all the DataObj-derived classes.
All Generator-derived classes are used to generate DataObj-derived objects. Table
6.3 shows all the Generator-derived objects together with a short description.
All GenericOperation-derived classes are (complex) operations which are to be performed on the data. GenericOperation-derived classes are meant to operate on one sin-
118
6.1. The AnaPulsar framework
Figure. 6.1: The inheritance graph of AnaPulsar. The boxes represent the various
classes, the solid curves represent the inheritance relations and the dashed
curves represent template instantiations.
119
Chapter 6. Software and computing
This analysis takes the input data stream, divides it up in
chunks of a specified length and performs a fast fourier transform by using fftw3.
AvgPSDCreation:
This analysis takes the FFTs which have been created by
SFTCreation and estimates the PSD of the data by averaging
the FFTs per supplied frequency range.
PolynomialSearch: This analysis performs Polynomial Search on an SFT-by-SFT
basis by taking filters from the FilterGenerator and applying
them one-by-one to the SFT.
SFTCreation:
Table. 6.1: A short description of the AnalysisBase-derived classes currently implemented in AnaPulsar.
AvgPSDContainer: Contains the computed PSD from the data split into different
frequency ranges as supplied by the user.
THitTuple:
Contains the hits from Polynomial Search in a ROOT TNtuplederived object.
DataContainer:
Is the base class of the classes FrequencyDomainData, TimeDomainData, FrequencyDomainPolynomialFilter and TimeDomainPolynomialFilter which contain the associated time and
frequency domain data.
SFT_MetaData:
Contains the metadata associated with a single SFT resulting
from the SFTCreation-analysis.
AvgPSDMetaData:
Contains the metadata associated with the result of the AvgPSDCreation-analysis.
PSMetaData:
Contains the metadata associated with the THitTuple resulting
from the PolynomialSearch-analysis applied to a single SFT.
Table. 6.2: A short description of DataObj-derived objects.
gle DataObj-derived object at a time. The different operation-derived classes can be split
into two categories: the metadata operations and the operations which are performed
on the data. The metadata operation classes (AvgMetaDataOperation, PSMetaDataOperation and SFT_MetaDataOperation) are used to generate the metadata for each of
the AnalysisBase-derived classes discussed in Table 6.1. Table 6.4 shows the operation classes besides the metadata operation classes together with a short description.
Additional documentation of discussed classes as well as the remaining classes is available with the AnaPulsar source distribution.
The polynomial search algorithm up to the registration of the hits is implemented
in AnaPulsar. Further analysis on these hits is implemented in a series of PyROOT [77]
scripts. The counting of the hits as well as applying the consistent filter criterion are
considered part of the post processing and are not implemented directly in AnaPulsar.
Figure 6.2 shows the implementation of the first part of the polynomial search algorithm schematically. In this figure, the arrows with the solid line indicate the flow of the
120
6.1. The AnaPulsar framework
FakeDataGenerator: This class generates white, Gaussian noise with a certain mean
and variance supplied by the user in the time domain. Additionally it is possible to add a signal of the form
h(t) = As sin(2π(f0 t +
FilterGenerator:
α 2 β 3
t + t )) + φ0 ),
2
3
where As , f0 , α, β and φ0 are parameters specified in the
option file.
This class generates the filters used by Polynomial Search. It
is possible to either generate filters by using a rectangular, regular grid in parameter space or by using the step sizes defined
in section 5.5.
Table. 6.3: A short description of the Generator-derived objects.
This class computes the mean and variance of the
FFTs in the desired frequency range. This is done in
the AvgPSDCreation class by passing all the FFTs
through this operation twice (first to compute the
mean and the second time to compute the variance).
FFTFrequencyDomOperation: This class computes the FFT of frequency domain
data, wrapped in a FrequencyDomainData container
resulting in a TimeDomainData container.
FFTTimeDomOperation:
This class computes the FFT of a time series wrapped
in a TimeDomainData container resulting in a FrequencyDomainData container.
FilterApplicator:
This class uses the FilterGenerator to generate
polynomial filters and applies them to the FFT.
FilterSearchOperation:
This class is used by the PolynomialSearch analysis
class to perform Polynomial Search on all the data.
AvgPSDOperation:
Table. 6.4: A short description of the GenericOperation-derived objects.
data, the square boxes represent the different data containers where D(t) denotes the
time domain data, D(f ) the frequency domain data, ‘Sn (f )’ the estimated PSD, ‘Filter’
the polynomial filter, ‘Cmax ’ the maximum correlation for each filter-data-correlationcomputation and ‘Collection of Cthr ’ are the maximum correlations for all the filters and
for all the FFTs. The elliptical balloons and arrows with the dashed lines indicate the
different operations and the data stream they affect. SF T is the SFT creation operation
done by using SFTCreation, AV G the estimate of the PSD done by AvgPSDCreation,
P S Polynomial Search done by PolynomialSearch and F ILGEN the filter generator
done by FilterGenerator.
The polynomial search algorithm shown in Fig. 6.2 starts with the time domain data
stream from the detector. With the desired FFT length, supplied by the user, the FFTs
121
Chapter 6. Software and computing
Figure. 6.2: The polynomial search algorithm shown schematically. First, the data are
Fourier transformed and the frequency domain data, D(f ), are averaged
using the AV G operation resulting in the estimate of the PSD, Sn (f ).
Next, the individual chunks of frequency domain data are combined with
the estimated PSD and the filters from the filter generator in the PS operation. This is done by computing the correlation as in Eq. (5.13). Finally,
the maximum correlation is chosen per filter-data combination (Cmax )
and collected for each filter in the output of the algorithm (Collection of
Cmax ).
are generated from this data stream. These FFTs are then averaged in order to estimate
the PSD. The estimated PSD is then used to normalize the frequency domain data.
Then the filters are generated one-by-one and applied to the normalized data and the
maximum correlations are recorded.
Figure 6.3 shows the post processing part of Polynomial Search including the consistent filter criterion. Again, the data stream is indicated by the solid arrows, the square
boxes represent the different data containers and the oval shapes the different operations
on the data. The post processing analysis runs over all hits from all FFTs and starts
with the ‘Collection of Cmax ’ which is the output of the previous analysis. The hits are
generated by applying a threshold Cthr to the data. Next, the consistent filter pairs are
computed by using the tolerance supplied by the user, the consistent filter pairs are
counted and the number count statistic is generated. The number of noise hits are computed either from these number counts or from first principles as discussed in section
5.8 and from these values the significance is computed for each frequency bin. Finally a
threshold is applied to the significance. This yields a list of candidate gravitational wave
signals.
122
6.2. Computing costs
Figure. 6.3: The post processing algorithms to create candidates from the hits.
6.2
Computing costs
Polynomial Search on a single FFT can be split into five steps which are shown in Table
6.5 together with the scaling of the number of operations required for each step. Note
1)
2)
3)
4)
5)
Computation of the FFT of the data:
Computation of the filters:
Computation of the filter-data product:
FFT back to time-domain:
Computing the maximum correlation:
∝ ND log ND
∝ Nfil Nf log Nf
∝ Nfil? Nf
∝ Nfil? Nf log Nf
∝ Nf Nfil?
Table. 6.5: The scaling of the number of operations per separate step of the polynomial
search algorithm.
that the computation of the PSD has been omitted since the computing cost for this is
negligible compared to Polynomial Search itself. In Table 6.5 ND represents the number
of sample points in a single FFT of the data, Nfil the number of explicitly computed
filters (e.g. the number of α and β steps), Nf the number of sample points in the filter
and Nfil? the number filters including the f0 steps. Since all these operations are applied
to NFFTs FFTs, everything scales linearly with NFFT . The computation of the FFT of
a data block has to be done only once per entire filter bank. This implies that the
most computationally intensive part of the algorithm depicted in Fig. 6.2 lies in the P S
operation (steps 2 to 5).
123
Chapter 6. Software and computing
Increasing the total observation time (equivalent to increasing NFFT ) will increase
the number of operations in a linear way. Increasing the coherence time however, will
increase NF in a linear way and Nfil? in a non-linear way due to the way the step sizes
in parameter space are chosen. This implies that wanting to improve the sensitivity by
increasing the coherence time will run into computational limitations much sooner than
simply adding more FFTs.
The application of Polynomial Search with parameters from Table 5.6 took on average
≈ 1.2 h per FFT on an Intel(R) Xeon(R) CPU L5520 2.27GHz. As shown in the table
this filter bank covers a frequency range of 16 Hz, the filters are sampled at 32 Hz, the
data are sampled at 4000 Hz and the FFT length is 1200 s. A month of data can be
searched over a band from 10 Hz to 1 kHz in ≈ 36 CPU year on this particular CPU.
Of the algorithm depicted in Fig. 6.3 the most computational-intensive part is the
‘Consistent filter criterion’ step. When applying the consistent filter criterion the number
of operations per pair of FFTs will scale as
X
NFFT /2
Nopt ∝
Nhits,2i Nhits,2i+1 ,
(6.1)
i=0
where Nhits represents the number of hits per FFT given a certain threshold. The number
of operations grows rapidly for lower thresholds due to the fact that the false-alarm
probability as a function of the threshold approaches a step function (see Fig. 5.14).
When comparing the computational costs shown in Table 6.5 to Eq. (6.1) it can be seen
that the number of operations for applying the consistent filter criterion will be less than
that for the actual polynomial search. This has to do with the fact that in order to have
a good separation between signal and noise correlations it is required that Nfil? Nhit .
10%
When demanding consistent filter pairs with a threshold of Cthr
= 4.88, on average
32, 484 filters will pass the threshold per Hz per FFT. Since at fixed threshold the
number of hits registered does not vary more than 1%, the assumption is made that the
computing power scales linearly with the number of FFTs and 16 Hz frequency bands.
When running the consistent filter code on these hits on the same Intel(R) Xeon(R)
CPU L5520 2.27GHz CPU, the average CPU time is about 6.8 minutes per FFT pair.
Analyzing the entire frequency band (about 1 kHz) for one month of data requires about
10 days of CPU time. This is negligible compared to the CPU time required to perform
the initial filter search.
In order to perform Polynomial Search in a timely fashion, the framework is run on
the Grid [79]. Running analyses on the Grid requires tools for data and job-management
in the form of Python scripts. These have been included with the AnaPulsar sourcedistribution. When running the analysis on the Grid, the total number of CPUs needed
to run Polynomial Search including the consistent filter criterion in real-time is computed
waves from neutron stars in
to be about1 433. Hence an all-sky search for gravitational
2 df d
f
binary systems (with dt < 4 × 10−3 Hz/s and dt2 < 2 × 10−5 ) Hz/s2 can be performed
in a reasonable time with moderate computational resources.
1
Note that this an approximate number since it depends on various circumstances such as the precise
CPUs used, data transfer speeds, etc.
124
6.2. Computing costs
The computing requirements can be compared to the most sensitive all-sky search
for isolated pulsars, the Einstein@Home search [62, 80]. This search is a template-based
all-sky search using the F-statistic. When run on the S5 data set the search ran over 840
hours of LIGO data and for coherent stretches of 30 hours the search took about 1 × 104
CPU years. Even though this search’ sensitivity is roughly an order of magnitude better
than Polynomial Search, it is meant for isolated pulsars. Thus it covers a single spin
down parameter f˙ (comparable to the α-parameter in Polynomial Search) of −f /τ <
f˙ < 0.1f /τ where τ = 103 yr for f < 400 Hz and τ = 4 × 103 yr for f > 400 Hz.
125
Chapter 6. Software and computing
126
Chapter
7
Summary
General relativity predicts the emission of gravitational waves by any mass distribution
which has a time-dependent quadrupole (or higher order multipole) moment. Gravitational waves are perturbations of the background metric traveling at the speed of light.
A gravitational wave induces strain in spacetime which affects the relative distance
between two freely-falling test masses. Gravitational waves emitted from astrophysical
sources can be detected with modern day interferometric gravitational wave detectors.
One such detector is Virgo which is located in Italy near the city of Pisa. The Virgo
detector is a Michelson interferometer with 3 km long Fabry-Perot resonant cavities as
arms. The entire detection band runs from approximately 20 Hz to 10 kHz (sensitivities
about 10−20 Hz−1/2 and 10−21 Hz−1/2 respectively), and the best sensitivity of the Virgo
detector per 1-10-2009 was 6 × 10−23 Hz−1/2 for gravitational waves of 200 Hz.
Due to the losses in the mirrors of the arm cavities, the stored laser power will cause
these mirrors to heat up. This heating will cause variations in the optical path length
which change the properties of the Fabry-Perot cavities. This causes adverse effects in
the stable operation of the interferometer (e.g. loss of lock) and will limit the laser power
which can be injected in the interferometer.
Chapter 3 of this thesis describes a three dimensional finite element model of the
Virgo input mirror, including the recoil mass, which is used to simulate the heating
effects. It has been shown that in the steady state situation, the average temperature of
the mirrors increase by 8.52 mK. Furthermore, the simulation shows that two principal
eigenmodes of the input mirrosr (the drum and butterfly modes) increase in frequency
with 0.42 Hz/K and 0.29 Hz/K, respectively. These results can be used to monitor the
mirror temperature by extracting the resonance frequency of the modes from the detector output data. It has been shown that the mirrors have an even higher fractional
absorption value that expected. In the case of the West End input mirror the losses are
about an order of magnitude higher than previously thought. Also, the finite element
analysis shows that in the transient situation, the average temperature increase is governed by a time constant which is computed to be 4 hours. Finally, due to the shape of
the beam, the mirrors will heat up in a non-uniform way resulting in an effective radius
of curvature of 21 km.
One of the sources of gravitational waves are non-axisymmetric rotating neutron
127
Chapter 7. Summary
stars. Due to the high magnetic fields of these stars, cones of electromagnetic radiation
can be emitted from their magnetic poles. If such a neutron star is oriented favorably
with respect to Earth, these cones of radiation can be observed as pulses. These neutron stars are known as ‘pulsars’. From observing many of these pulsars and taking the
observational biases into account, about 105 active pulsars are expected to exist in our
Galaxy. Furthermore, it has been observed that the majority of the observed pulsars
with rotational frequencies above 20 Hz are in binary systems. Since the emission of
gravitational waves is not limited to just the magnetic poles, the unobserved neutron
stars could be seen as potentially observable gravitational wave sources. The amplitude
of the gravitational waves emitted by such systems is weak. However, they are also continuous in nature, meaning that observing such systems over longer stretches of time
will increase the signal-to-noise ratio with the square root of the observation time.
In order to take full advantage of the long integration times of a potential continuous
gravitational wave signal, despite the fact that the locations of the majority of the neutron stars in our Galaxy are unknown, so-called all-sky searches have been developed.
A fully-coherent search (i.e. the data are integrated over the entire observation time)
has been shown to be computationally limited. The most sensitive search for isolated
neutron stars required 104 CPU years and was limited to 30 hours of data. In order to
increase the observation time with limited computational resources, sub-optimal analysis methods have been developed. These so-called ‘semi-coherent’ analysis methods are
based on taking multiple short coherent stretches of data and combining them in an
incoherent way. What these searches lack in sensitivity, they partially make up in decreased computing requirements making it possible to consider longer stretches of data.
When attempting to apply such semi-coherent searches to signals from neutron stars in
binary systems, the initial coherent integration time is shorter by an order of magnitude
with respect to the isolated neutron star case. This short integration time will limit the
sensitivity of the semi-coherent analyses when applying them to binary systems.
A new data analysis algorithm, called ‘Polynomial Search’ has been developed and
is described in Chapter 5. Polynomial Search is designed to be an all-sky search for
gravitational waves from neutron stars in binary systems. Since Polynomial Search employs a bank of polynomial phase templates up to 4th order in time, the initial coherent
integration time can be increased by an order of magnitude with respect to applying
a traditional semi-coherent search to this signal. It has been shown that the increase
in coherence time is about 25 for neutron stars in a Keplerian orbit, where the orbital
period (P ) is larger than 2 hours for circular orbits (eccentricity e = 0). A filter bank has
been constructed which has been shown to cover all gravitational waveforms of which
the orbital parameters obey P > 2 hours for circular orbits and P > 6 hours for eccentric orbits with 0 < e < 0.6. These filters can be applied to the data with a coherence
time of 600 seconds (whereas the coherence time of a semi-coherent search for circular
orbits with P = 2 hours would be limited to 20 seconds). Furthermore, a strategy for
combining the results from the individual coherent stretches, called ‘the consistent filter
criterion’, has been developed. When applying Polynomial Search to white noise and to
simulated gravitational wave signals, the projected sensitivity of the new search method
has been presented.
128
Finally, in chapter 6 the developed framework is presented in which Polynomial
Search has been implemented. This so-called ‘AnaPulsar’ framework has been developed for implementing various types of gravitational wave data analysis algorithms.
Furthermore, the computing requirements of Polynomial Search have been discussed. It
has been shown that in order to perform Polynomial Search on all gravitational wave
signals originating from neutron stars in a Keplerian orbit with parameters P > 2 hours
and e = 0, or P > 6 hours and 0 < e < 0.6, for an analysis of 1 month of data
approximately 36 CPU years are required.
129
Chapter 7. Summary
130
Appendix
A
Fourier transform definitions
The Fourier transform of function s(t) and its inverse are defined as
Z ∞
F [s(t)] =
s(t)e2πif t dt = s̃(f )
Z−∞
∞
−1
F [s̃(f )] =
s̃(f )e−2πif t df = s(t),
(A.1)
(A.2)
−∞
where f represents the frequency in Hz in case t is measured in seconds, and s̃(f )
represents the Fourier transform of s(t). The correlation csg (τ ) between a function s(t)
and g(t) is defined as
Z
csg (τ ) ≡
∞
g(t)s(t + τ )dt,
(A.3)
−∞
where the correlation at a fixed lag, τ , measures the overlap of function g(t) with function
g(t + τ ). The parameter τ can be interpreted as a lag parameter with which the function
g is shifted in time.
Combining Eqs. (A.1), (A.2) and (A.3) results in the correlation theorem stating
that
csg (τ ) = F −1 [s̃(f )g̃(f )∗ ] .
(A.4)
When taking g(t) = s(t), the Wiener-Khinchin theorem [81] is obtained,
F(css (τ )) = |s̃(f )|2 ,
(A.5)
stating that the Fourier transform of the autocorrelation css is equal to the power spectrum of s(t). Using Eqs. (A.3), (A.4) and (A.5) leads to Parseval’s theorem
Z ∞
Z ∞
2
Ps ≡
|s(t)| dt =
|s̃(f )|2 df,
(A.6)
−∞
−∞
stating that the total power in s, Ps , is the same if computed in the frequency or the
time domain. When s(t) is a real function (e.g. s(t) represents the detector output),
then the total power in s becomes
Z ∞
Ps = 2
|s̃(f )|2 df.
(A.7)
0
131
Appendix A. Fourier transform definitions
The normalized power spectrum of s is then defined as
s̃(f )
ŝ(f ) = √ .
Ps
(A.8)
The spectral power of a function s between f and f + df can thus be written as
Ps (f ) = 2 |s̃(f )|2 ,
(A.9)
where Ps (f ) is the single sided power spectral density (PSD) of s.
If the function s(t) is not square integrable, which is the case for continuous signals,
then the power spectral density will become infinite. Consequently, a better measure
for the power spectral density is the power spectral density per unit time. The power
spectral density of random noise, Sn (f ) is defined as the ensemble average of many
possible realizations of the noise. When observing a noise source n(t) over a time period
T , the PSD is computed in Ref. [27] to be
Sn (f ) = 2
< |ñ(f )|2 >
,
T
(A.10)
where ñ is the Fourier transform of the noise source and the brackets denote the ensemble
average. In the case of Gaussian and white noise with mean µ = 0 and variance σ 2 , the
PSD reduces to
Sn (f ) = Sn = 2σ 2 .
(A.11)
Note that by defining the noise as white the PSD per definition does not depend on the
frequency.
When considering data as a series of discretely sampled points, the continuous signal
s(t) is replaced by a finite number of N samples. These N samples are the result of
sampling the signal over a time T with sample frequency1 fs . When defining
sk = s(tk ),
k = 0, 1, 2, ..., N − 1,
tk = k/fs ,
N = fs T,
the Fourier transforms defined in Eqs. (A.1) and (A.2) reduce to the Discrete Fourier
Transforms (DFT)
s̃n =
sk =
N
−1
X
sk e2πikn/N
k=0
N
−1
X
1
N
n=0
1
,
fs
s̃n e−2πikn/N
(A.12)
1
,
T
(A.13)
where fn and tk are the n-th and k-th frequency and time sample respectively. It has
been chosen to apply the normalization factor, 1/N , in the inverse Fourier transform.
1
When sampling data with a certain sample frequency, the Shannon-Nyquist theorem states that a
band limited analog signal can be perfectly reconstructed from an infinite sequence of samples when
the highest frequency component in the signal is at most fN = fsample /2, where fN is known as the
Nyquist frequency.
132
In practice, discrete Fourier transforms are computed by using a Fast Fourier Transform (FFT) algorithm [81]. Instead of requiring O(N 2 ) operations to compute a DFT,
the FFT algorithm enables the computation of a Fourier transform in O(N log2 N ) operations.
Furthermore, Eqs. (A.3) and (A.4) can be written in discrete form as
c(s, g)j =
N
−1
X
sj+k gk = F −1 [s̃k g̃k∗ ] ,
(A.14)
k=0
where j = τk fs . Now, the power spectral density per unit time is
Ss (fk ) =
2
|s̃k |2 ,
T
(A.15)
133
Appendix A. Fourier transform definitions
134
Appendix
B
The correlation statistic
When searching for a known signal in a noise background, a so-called matched filtering
approach is the most optimal [27]. A matched filter is obtained by correlating a signal
template to a data stream containing noise to detect the presence of the signal in the
data. The matched filter is the optimal linear filter for maximizing the signal to noise
ratio (SNR) in the presence of additive stochastic noise.
The detector output D(t) can be written as
D(t) = s(t) + n(t),
(B.1)
where s(t) and n(t) are the gravitational wave signal and the stochastic noise respectively. The correlation given in Eq. (A.4) between the detector output and a filter function F , is written in time and frequency domain for τ = 0 as
Z ∞
Z ∞
C=
D(t)F (t)dt =
D̃(f )F̃ ∗ (f )df.
(B.2)
−∞
−∞
When the shape of the gravitational wave, h(t) is known, the filter function that maximizes the signal-to-noise ratio for such a signal is known as the ‘optimal filter’. This
technique of matching the filter with the signal is called ‘matched filtering’.
The signal-to-noise ratio (in amplitude) is defined as S/N where S is the expected
value of C when the signal is present and N is the rms value of C when no signal is
present. The signal-to-noise ratio with filter function F is computed in Ref. [27] to be
R∞
s̃(f )F̃ ∗ (f )df
S
−∞
≡ R
(B.3)
1/2 ,
N
∞
2
1/2Sn (f )|F̃ (f )| df
−∞
where Sn is the single-sided power spectral density of the noise.
The optimal filter function is the function which maximizes Eq. (B.3). This maximization yields that the optimal filter has the form
F̃ (f ) = const.
s̃(f )
,
Sn (f )
(B.4)
135
Appendix B. The correlation statistic
where the constant is arbitrary since rescaling by an overall factor does not change the
signal-to-noise ratio. The signal-to-noise ratio for this optimal filter can be computed by
substituting Eq. (B.4) into Eq. (B.3) and yields.
2
Z ∞
S
|s̃(f )|2
=4
df.
(B.5)
N
Sn (f )
0
√
When taking a finite observation time T , the signal-to-noise ratio grows with T .
This can be seen by considering the fact that the power |s(f )|2 in Eq. (B.5) will grow
linearly with T while the power spectral density Sn will remain constant by construction
as defined in appendix A .
Polynomial Search is based on the assumption that shape of a GW signal can be
approximated by polynomial filters. These polynomial filters are per definition suboptimal thus the SNR shown in Eq. (B.5) does not apply. The SNR for Polynomial
Search can be computed from Eq. (B.3). Considering the fact that Polynomial Search
is applied to small frequency intervals of the data, the PSD in each interval can be
considered white (independent of frequency: Sn (f ) = Sn ). Under this assumption, the
denominator in Eq. (B.3) becomes
r
1/2 r
Z ∞
1
1
2
|F̃ (f )| df
=
Sn ×
Sn P F ,
2
2
−∞
where PF is defined in Eq. (A.6). With the notation defined in Eq. (A.8), Eq. (B.3)
becomes
R∞
s̃(f )F̂ ∗ (f )df
S
−∞
q
=
.
(B.6)
N
1
S
2
n
The numerator in Eq. (B.6) is the expectation value of C when signal is present. This
expectation value is obtained by computing the correlation of the data stream with the
filter as in Eq. (B.2). Furthermore, in order to accommodate for the sliding of the filter
function in time, the filter function is multiplied with a phase factor in the frequency
domain1 . Finally, defining a filter function parameter vector ξ and combining all the
aforementioned arguments yields the Polynomial Search correlation statistic, C(τ ; ξ), as
Z ∞
~
D̃(f )F̂ ∗ (f ; ξ)
~
q
C(τ ; ξ) =
e−2πif τ df.
(B.7)
1
−∞
S
2 n
Note that Eq. (B.7) is derived from Eq. (B.3) and thus is equivalent to a signal-to-noise
ratio. When considering the fact that the data are sampled during time T with sample
frequency fs , Eq. (B.7) becomes
Nf
1 X D̃(fk )F̂ ∗ (fk ) −2πifk τj
q
C(τj ) =
e
.
T k=0
1
S
2
(B.8)
n
In general one can show that F(g(t + τ )) = g̃(f )e−2iπf τ for any quadratically integrable function
g(t).
1
136
Here τj is the j th time sample, k denotes the frequency bin number, Nf = fs T /2 is the
number of frequency bins and fk is the k th frequency bin.
In order to distinguish between the correlation and the correlation statistic, the
correlation is denoted by a csg , where the sub indices denote the functions between which
the correlation is computed while the correlation statistic derived in Eq. (B.7) will be
~ However, since C(τ ; ξ)
~ can also be seen as a rescaled correlation, it
denoted by C(τ ; ξ).
will often be referred to as the correlation.
Assuming a coherent integration time T , Eq. (B.7) can be written in the time domain
by using the correlation theorem shown in Eqs. (A.3) and (A.4). The correlation statistic
in time-domain is
r Z T
2
~ =
~
C(τ ; ξ)
D(t)F̂ (t − τ ; ξ)dt.
(B.9)
Sn 0
And in the discrete case
r
~ =
C(τj ; ξ)
Ns
2 1 X
~
D(tk )F̂ (tk − τj ; ξ),
Sn fs k=0
When substituting Eq. (B.1) into Eq. (B.9), the correlation statistic becomes
r Z T
Z T
2
~
~
~
s(t)F̂ (t − τ ; ξ)dt +
n(t)F̂ (t − τ ; ξ)dt .
C(τ ; ξ) =
Sn
0
0
(B.10)
(B.11)
From Eq. (B.11) it can be seen that when the noise is Gaussian, C will also be Gaussian.
The mean and variance of C can be computed from Eq. (B.11). As the first term in Eq.
(B.11) does not contain any random variables (only n(t) is a random variable), the
variance can be computed to be
r Z T
2
~
Var(C) = Var
n(t)F̂ (t − τ ; ξ)dt
Sn 0
Z T
2
~
Var
n(t)F̂ (t − τ ; ξ)dt
=
Sn
0
Z T
2
~
=
Var(n(t))F̂ 2 (t − τ ; ξ)dt
Sn 0
2 2
=
σ = 1,
(B.12)
Sn
where the final equality arises from substituting the PSD of Gaussian noise (Eq. (A.11)).
When computing the mean of Eq. (B.11), it is important to realize that Mean(n(t)) =
0 so that
r Z T
2
~
Mean(C) = Mean
s(t)F̂ (t − τ ; ξ)dt
Sn 0
r Z T
2
~
=
s(t)F̂ (t − τ ; ξ)dt.
(B.13)
Sn 0
137
Appendix B. The correlation statistic
Polynomial Search operates under the assumption that the signal s(t) can be approximated by a polynomial during time T . Furthermore, T is assumed to be short enough
that amplitude modulation due to the beam-pattern functions do not affect the signal shape during the coherent integration time (e.g. the beam-pattern functions can be
taken constant). With these assumptions, the signal is of the form
s(t) = Asig sin (Φ(t)) ,
(B.14)
where Asig is the GW amplitude arriving at the detector (i.e. as given in Eq. (1.15))
times a constant due to the beam-pattern function and the phase Φ(t) is the Dopplershifted phase of the GW signal. With the signal shape shown in Eq. (B.14), the total
power in the signal, Ps can be computed to be
Z T
Z T
1 2
T
2
2 T
cos (2Φ(t)) dt ≈ A2sig ,
Ps =
s (t)dt = Asig − Asig
(B.15)
2
2
2
0
0
where the assumption that the integral over cos(2Φ(t)) term can be neglected as it is
a rapidly oscillating term and will be small compared
√ to the Asig T /2 which will grown
linearly with T . With the notation ŝ(t) = s(t)/ Ps (Eq. (A.8)), Eq. (B.13) can be
rewritten as
r Z T
T
~ = SNR × csF (τ ),
Mean (C) = Asig
ŝ(t)F̂ (t − τ ; ξ)dt
(B.16)
Sn 0
r
where
SNR ≡ Asig
T
Sn
(B.17)
and csF (τ ) is the correlation between normalized filter function F̂ and signal ŝ. Note
that Eq. (B.16) can be rewritten as
h
i
−1
∗
Mean (C) = SNR × ci = SNR × F
ŝ(f )F̂ (f ) .
(B.18)
Note that the functions ŝ and F̂ are periodic functions of time. This implies that −1 <
cŝF̂ (τ ) < 1; the correlation will be +1 if F̂ matches ŝ during T and −1 if F̂ matches ŝ
but is π out of phase during T . Furthermore, if ξ~ is such that the filter does not match
the signal, the correlation will lie in between these extremal values.
Note that the data are sampled with sample frequency fs,data . Since this means that
each second, fs,data samples are taken, the PSD of a Gaussian white noise process with
time-domain amplitude becomes
2A2noise
(B.19)
Sn =
fs
Inserting this expression in Eq. (B.17) yields
SNR =
Asig √
√ N,
Anoise 2
where N = T fs are the number of samples.
138
(B.20)
Appendix
C
Stepsize plots
In this appendix the slices of parameter space are shown for the remaining data sets
described in Table 5.4.
×10
1
-3
β (Hz/s^2) = 1e-06
0
600
×10-3
1
600.2
600.4
β (Hz/s^2) = 1.75e-06
0
600
×10
1
-3
600.2
600.4
β (Hz/s^2) = 2.5e-06
0
600
600.2
600.4
1
×10-3
1
0.5
0
0
600
1
×10-3
1
0.5
0
0
600
1
×10
1
0.5
0
-3
0
600
β (Hz/s^2) = 1.25e-06
600.2
600.4
β (Hz/s^2) = 2e-06
600.2
600.4
β (Hz/s^2) = 2.75e-06
600.2
600.4
1
×10-3
1
0.5
0
0
1
×10-3
1
0.5
0
600.2
600.4
β (Hz/s^2) = 2.25e-06
0
1
0.5
600
1
×10
1
0.5
0
-3
0
1
0.5
600
0
β (Hz/s^2) = 1.5e-06
600.2
600.4
β (Hz/s^2) = 3e-06
0
1
0.5
600
600.2
600.4
0
Figure. C.1: Step size plots for data set 2 (Table 5.6). The plots show α versus f0 for
different values for β and the color codes shows the correlation, normalized to unity.
139
Appendix C. Stepsize plots
3
×10-6
α (Hz/s) = -0.00016
3
0.5
2
0
1
1
3
0.5
2
0
1
1
3
0.5
2
0
1
1
3
2
1
3
600
×10-6
600.2
600.4
α (Hz/s) = -1e-05
2
1
600
×10
-6
3
600.2
600.4
3
600
×10-6
600.2
600.4
f0 (Hz) = 599.98
0.5
2
600
×10-6
600
×10
-6
α (Hz/s) = 0.00014
2
1
×10-6
1
600
×10-6
α (Hz/s) = -0.00011
600.2
600.4
α (Hz/s) = 4e-05
600.2
600.4
α (Hz/s) = 0.00019
600.2
600.4
f0 (Hz) = 599.985
0
1
3
0
×10-6
0.001
f0 (Hz) = 599.995
1
0.5
2
0
×10
-6
3
0.001
f0 (Hz) = 600.01
3
0
×10-6
0.5
2
0
0.001
0
1
1
3
0.5
2
0
1
1
3
0.5
2
0
1
1
3
1
600
×10-6
600.2
0
3
×10
0.001
f0 (Hz) = 600.015
α (Hz/s) = 9e-05
0.5
2
600
×10
600.2
600.4
α (Hz/s) = 0.00024
0
0.001
0
1
0.5
600
×10-6
600.2
600.4
f0 (Hz) = 599.99
0
1
0.5
2
0
1
3
0
×10-6
0.001
f0 (Hz) = 600.005
1
0.5
2
0
1
0
3
×10
0.001
f0 (Hz) = 600.02
1
0.5
2
0
0
1
1
0.5
-6
1
0
600.4
0
1
1
0.5
-6
0.5
0
1
2
0.001
f0 (Hz) = 600.0
2
-6
1
0.5
α (Hz/s) = -6e-05
0
1
0
1
3
0.5
2
×10-6
1
1
0
0.001
Figure. C.2: Step size plots for data set 2 (Table 5.6). The top nine plots show β versus f0 for different values for α. The bottom nine plots showα versus β
for different values for f0 . The color codes shows the correlation, normalized to unity.
140
×10-3
1.1
β (Hz/s^2) = 1e-06
0.5
1
0.9
599.9
600
×10-3
1.1
0
600.1
β (Hz/s^2) = 1.75e-06
1
0.5
1
0.9
599.9
600
×10
1.1
-3
0
600.1
β (Hz/s^2) = 2.5e-06
600
×10-3
1.1
0
600.1
β (Hz/s^2) = 1.25e-06
0.9
599.9
600
×10-3
1.1
0
600.1
β (Hz/s^2) = 2e-06
1
0.5
1
0.9
599.9
600
×10
1.1
0
600.1
β (Hz/s^2) = 2.75e-06
600
×10-3
1.1
0
600.1
β (Hz/s^2) = 1.5e-06
1
0.5
1
0.9
599.9
600
×10-3
1.1
0
600.1
β (Hz/s^2) = 2.25e-06
0.9
599.9
600
×10
1.1
0
600.1
β (Hz/s^2) = 3e-06
1
0.5
1
0.9
599.9
1
0.5
1
-3
1
0.5
1
0.9
599.9
1
0.5
1
-3
1
0.5
1
0.9
599.9
1
600
0
600.1
Figure. C.3: Step size plots for data set 3 (Table 5.6). The plots show α versus f0 for
different values for β and the color codes shows the correlation, normalized to unity.
141
Appendix C. Stepsize plots
3
×10-6
α (Hz/s) = 0.0009
2
1
3
600
×10-6
600.2
600.4
α (Hz/s) = 0.000975
2
1
600
×10
-6
3
600.2
600.4
α (Hz/s) = 0.00105
2
1
3
600
×10-6
600.2
600.4
f0 (Hz) = 599.64
3
0.5
2
0
1
1
3
0.5
2
0
1
1
3
0.5
2
0
1
1
3
600
×10-6
600
×10
-6
0.5
2
×10-6
1
600
×10-6
α (Hz/s) = 0.000925
600.2
600.4
α (Hz/s) = 0.001
600.2
600.4
α (Hz/s) = 0.001075
600.2
600.4
f0 (Hz) = 599.73
0
1
0.0009
3
0.001
×10-6
0.0011
f0 (Hz) = 599.91
1
0.5
2
0.001
×10
-6
3
0.0011
f0 (Hz) = 600.18
3
0.001
×10-6
0.5
2
0.001
0.0011
0
1
1
3
0.5
2
0
1
1
3
0.5
2
0
1
1
3
1
600
×10-6
600.2
600.4
α (Hz/s) = 0.001025
0.001
3
×10
0.0011
f0 (Hz) = 600.27
600
×10
0.5
2
600.2
600.4
α (Hz/s) = 0.0011
600
×10-6
600.2
600.4
f0 (Hz) = 599.82
0.001
0.0011
0
1
0
1
0.5
2
0
1
0.0009
3
0.001
×10-6
0.0011
f0 (Hz) = 600.09
1
0.5
2
0
1
0.0009
0.001
3
×10
0.0011
f0 (Hz) = 600.36
1
0.5
2
0
1
0.0009
1
0.5
-6
1
0
0.5
0
1
0.0009
1
0.5
-6
0.5
0
1
0.0009
2
0.0011
f0 (Hz) = 600.0
2
-6
1
0.5
α (Hz/s) = 0.00095
0
1
0.0009
0
1
0.0009
3
0.5
2
×10-6
1
0
1
0.0009
0.001
0.0011
Figure. C.4: Step size plots for data set 3 (Table 5.6). The top nine plots show β versus f0 for different values for α. The bottom nine plots showα versus β
for different values for f0 . The color codes shows the correlation, normalized to unity.
142
1.2
×10-3
β (Hz/s^2) = -1e-07
1
1.1
1.2
×10-3
β (Hz/s^2) = -5e-08
1.1
1
1.2
600.5
×10-3
β (Hz/s^2) = 5e-08
0
0.9
1
1.2
1.1
0.9
600
600.5
×10-3
β (Hz/s^2) = 1e-07
×10
-3
1.2
1
1.2
β (Hz/s^2) = 2e-07
0.9
1
1.2
600
×10
600.5
β (Hz/s^2) = 2.5e-07
0.9
1
1.2
600
×10
-3
600.5
β (Hz/s^2) = 3e-07
0.9
1
0
1
1.1
0.5
1
0
β (Hz/s^2) = 1.5e-07
0.5
0
1.1
600.5
×10-3
0
1
0.5
600
600.5
1.1
-3
1
600
0.5
0
1.1
0.9
0.9
1
600.5
1
0.5
0
1.1
600
β (Hz/s^2) = 0.0
1
0.5
1
×10-3
0.5
1
600
1.2
1.1
0.5
0.9
1
0.5
1
600
600.5
0
0.9
600
600.5
0
Figure. C.5: Step size plots for data set 4 (Table 5.6). The plots show α versus f0 for
different values for β and the color codes shows the correlation, normalized to unity.
143
Appendix C. Stepsize plots
0.4
×10-6
α (Hz/s) = 0.0008
1
0.2
0.4
×10-6
α (Hz/s) = 0.00085
0.2
0
0.4
600.5
×10-6
α (Hz/s) = 0.00095
0
-0.2
1
0.4
0.2
-0.2
600
600.5
×10-6
α (Hz/s) = 0.001
×10
-6
0.4
α (Hz/s) = 0.0011
-0.2
1
0.4
600
×10
600.5
α (Hz/s) = 0.00115
×10-6
f0 (Hz) = 599.28
0.2
1
0.4
-0.2
1
0.4
0.5
1
×10-6
1.2
f0 (Hz) = 599.82
0.2
0.5
0
600
600.5
1
×10
-6
1.2
f0 (Hz) = 600.36
×10-6
f0 (Hz) = 599.46
-0.2
1
0.4
0.5
1
×10-6
1.2
f0 (Hz) = 600.0
0.2
0.2
0.5
0
-0.2
0.8
1.2
600.5
×10-6
f0 (Hz) = 599.64
0.2
0
1
0.5
1
0.4
1
×10-6
1.2
f0 (Hz) = 600.18
0.2
0 -3
×10
1
0.5
0
1
×10
1.2
f0 (Hz) = 600.54
0.2
0 -3
×10 -0.2
0.8
0 -3
×10 -0.2
0.8
1
×10
-6
1
0.5
0
1
600
0 -3
×10 -0.2
0.8
0.5
-6
0.4
1
0
0 -3
×10 -0.2
0.8
1
α (Hz/s) = 0.0012
0
0.5
0
0
-0.2
0.8
600.5
0
0.2
0.4
1
0.2
0 -3
×10 -0.2
0.8
1
×10
-6
0
-0.2
0.8
600
0.5
0
0
0.4
-0.2
0
600.5
α (Hz/s) = 0.00105
0.5
0
0.2
600
×10-6
0
0
-6
0
0.4
0.4
600.5
0.2
0.5
0.4
1
600
0.5
0
0.2
-0.2
-0.2
0
600.5
1
0.5
0
0.2
600
α (Hz/s) = 0.0009
0
0.5
0
×10-6
0.5
0
600
0.4
0.2
0.5
-0.2
1
0.4
1.2
f0 (Hz) = 600.72
0.2
0 -3
×10
1
0.5
0
1
1.2
0 -3
×10 -0.2
0.8
1
1.2
0 -3
×10
Figure. C.6: Step size plots for data set 4 (Table 5.6). The top nine plots show β versus f0 for different values for α. The bottom nine plots showα versus β
for different values for f0 . The color codes shows the correlation, normalized to unity.
144
Appendix
D
AnaPulsar inheritance diagrams
In this appendix the inheritance and contents of the classes of the AnaPulsar framework
will be shown in more detail. The boxes in Figs. D.1 to D.8 have the following meaning:
ˆ A filled gray box represents the struct or class for which the graph is generated.
ˆ A box with a black border denotes a documented struct or class. This documentation is available with the source code distribution.
ˆ A box with a grey border denotes an undocumented struct or class.
Each class is denoted by the class name, a list of class data members and a list of class
methods. The − sign represents a private data member or method, the + sign represents
a public data member or method and the # sign represents a protected data member
or method.
The arrows have the following meaning:
ˆ A solid arrow is used to visualize a public inheritance relation between two classes.
ˆ A dashed arrow denotes a relation between a template instance and the template
class it was instantiated from. The arrow is labeled with the template parameters
of the instance.
145
Appendix D. AnaPulsar inheritance diagrams
Figure. D.1: The inheritance plot of the AnaPulsarBase class with all its derived
classes, data members and methods.
146
Figure. D.2: The inheritance plots of the AnaPulsarTime (left) and AnaPulsarTimeDuration (right) classes with all their derived classes, data members and
methods.
147
Appendix D. AnaPulsar inheritance diagrams
Figure. D.3: The inheritance plot of the DatObj class with all its derived classes, data
members and methods.
148
Figure. D.4: The inheritance plots of the DataStream (left) and Generator classes
with all their derived classes, data members and methods.
149
Appendix D. AnaPulsar inheritance diagrams
Figure. D.5: The inheritance plot of the GenericException class with all its derived
classes, data members and methods.
150
Figure. D.6: The inheritance plot of the IOobj class with all its derived classes, data
members and methods.
151
Appendix D. AnaPulsar inheritance diagrams
Figure. D.7: The inheritance plot of the OptionFileReader class with all its derived
classes, data members and methods.
152
Figure. D.8: The inheritance plot of the Options class with all its derived classes, data
members and methods.
153
Appendix D. AnaPulsar inheritance diagrams
154
References
[1] A. Einstein, “Die grundlage der allgemeinen relativitätstheorie,” Annalen der
Physik, 1916.
[2] H. Bondi, “Plane gravitational waves in general relativity,” Nature, 1957.
[3] J. Weber and J. A. Wheeler, “Reality of the cylindrical gravitational waves of Einstein and Rosen,” Rev. Mod. Phys., vol. 29, pp. 509–515, July 1957.
[4] J. Taylor and R. Hulse, “Discovery of a pulsar in a binary system,” AstroPhys. J.,
1975.
[5] J. Weisberg and J. Taylor, “The relativistic binary pulsar B1913+16: thirty years
of observations and analysis,” in Binary Radio Pulsars, vol. 328 of Astronomical
Society of the Pacific Conference Series, p. 25, July 2005.
[6] J. Abadie et al., “Predictions for the rates of compact binary coalescences observable by ground-based gravitational-wave detectors,” Classical and Quantum Gravity, vol. 27, p. 173001, September 2010.
[7] L. Santamaria, Coalescence of black-hole binaries: from theoretical source models
to applications in gravitational-wave astronomy. PhD thesis, Universität Potsdam,
2010.
[8] B. Allen, “χ2 time-frequency discriminator for gravitational wave detection,” Phys.
Rev. D, vol. 71, p. 62001, March 2005.
[9] C. Ott, “Probing the core-collapse supernova mechanism with gravitational waves,”
Classical and Quantum Gravity, vol. 26, p. 204015, October 2009.
[10] B. Abbott et al., “Search for gravitational wave ringdowns from perturbed black
holes in LIGO S4 data,” Phys. Rev. D, vol. 80, p. 62001, September 2009.
[11] R. Chapman, R. Priddey, and N. Tanvir, “Short gamma-ray bursts from SGR giant flares and neutron star mergers: two populations are better than one,” mnras,
vol. 395, pp. 1515–1522, May 2009.
155
References
[12] B. Abbott et al., “Search for gravitational-wave bursts associated with gamma-ray
bursts using data from LIGO Science Run 5 and Virgo Science Run 1,” AstroPhys.
J., vol. 715, pp. 1438–1452, June 2010.
[13] J. Abadie et al., “Search for gravitational-wave inspiral signals associated with short
gamma-ray bursts during LIGO’s fifth and Virgo’s first science run,” AstroPhys. J.,
vol. 715, pp. 1453–1461, June 2010.
[14] B. Abbott et al., “Search for gravitational waves associated with 39 gamma-ray
bursts using data from the second, third, and fourth LIGO runs,” Phys. Rev. D,
vol. 77, p. 62004, March 2008.
[15] P. Sutton et al., “X-Pipeline: an analysis package for autonomous gravitational-wave
burst searches,” New Journal of Physics, vol. 12, p. 53034, May 2010.
[16] B. Abbott et al., “Searching for a stochastic background of gravitational waves with
the Laser Interferometer Gravitational-Wave Observatory,” AstroPhys. J., vol. 659,
pp. 918–930, April 2007.
[17] M. Gasperini and G. Veneziano, “The pre-big bang scenario in string cosmology,”
Phys. Rev., vol. 373, p. 212, January 2003.
[18] D. Lorimer, “Binary and millisecond pulsars,” Living Reviews in Relativity, vol. 8,
no. 7, 2005. http://www.livingreviews.org/lrr-2005-7.
[19] D. Jones, “Gravitational wave emission from rotating superfluid neutron stars,”
mnras, vol. 402, pp. 2503–2519, March 2010.
[20] J. Middleditch et al., “Predicting the starquakes in PSR J0537-6910,” AstroPhys.
J., vol. 652, pp. 1531–1546, December 2006.
[21] R.
Manchester,
G.
Hobbs,
M.
Hobbs,
and
A.
ATNF
pulsar
catalog,”
AstroPhys.
J.,
vol.
http://www.atnf.csiro.au/research/pulsar/psrcat/.
Teoh,
129,
“The
2005.
[22] C. M. Espinoza, A. Lyne, M. Kramer, R. N. Manchester, and V. Kaspi,“The braking
index of PSR J1734-3333 and the magnetar population,” 2011.
[23] B. Abbott et al., “Coherent searches for periodic gravitational waves from unknown
isolated sources and Scorpius X-1: Results from the second LIGO science run,”
Phys. Rev. D, vol. 76, p. 82001, October 2007.
[24] B. Owen et al., “Gravitational waves from hot young rapidly rotating neutron stars,”
Phys. Rev. D, vol. 58, p. 84020, October 1998.
[25] D. Jones and N. Andersson, “Gravitational waves from freely precessing neutron
stars,” mnras, vol. 331, pp. 203–220, March 2002.
156
References
[26] L. Bildsten, “Gravitational radiation and rotation of accreting neutron stars,” AstroPhys. J., vol. 501, p. 89, July 1998.
[27] M. Maggiore, Gravitational waves volume 1: theory and experiments, ch. 7. Oxford
university press, 2008.
[28] S. Chandrasekhar, “Solutions of two problems in the theory of gravitational radiation,” Phys. Rev. Lett., vol. 24, pp. 611–615, March 1970.
[29] B. Schutz and J. Friedman, “Secular instability of rotating newtonian stars,” The
Astrophysical Journal, vol. 222, pp. 281–296, 1978.
[30] A. Melatos, “Gravitational waves from accreting neutron stars,” Advances in Space
Research, vol. 40, no. 10, pp. 1472–1479, 2007.
[31] A. Melatos and D. Payne, “Gravitational radiation from an accreting millisecond
pulsar with a magnetically confined mountain,” AstroPhys. J., vol. 623, pp. 1044–
1050, April 2005.
[32] A. Watts, B. Krishnan, L. Bildsten, and B. Schutz, “Detecting gravitational wave emission from the known accreting neutron stars,” Monthly Notices of the Royal Astronomical Society, vol. 389, no. 2, p. 839, 2008.
http://www3.interscience.wiley.com/journal/121376945/abstract.
[33] B. Abbott et al.,“LIGO: the Laser Interferometer Gravitational-Wave Observatory,”
Reports on Progress in Physics, vol. 72, p. 076901, July 2009.
[34] K. Danzmann et. al., “GEO600: Proposal for a 600 m Laser Interferometric Gravitational Wave Antenna,” Max Planck Institut fúr Quantenoptik internal report,
1994.
[35] B. Abbott et al., “Virgo final design report,” Internal Virgo note: VIR-TRE-100013, May 1997.
[36] C. Casciano, Seismic Isolation for the Test Masses of the VIRGO Gravitational
Wave Antenna. PhD thesis, University of Pisa, 2002.
[37] M. Punturo, “The VIRGO sensitivity curve,” Internal Virgo note: VIR-NOT-PER1390-51, October 2004.
[38] M. Beccaria et. al., “Relevance of Newtonian seismic noise for the VIRGO interferometer sensitivity,” Classical and Quantum Gravity, vol. 15, no. 11, p. 3339, 1998.
http://stacks.iop.org/0264-9381/15/i=11/a=004.
[39] X. Siemens, B. Allen, J. Creighton, M. Hewitson, and M. Landry, “Making h(t) for
LIGO,” Class. Quant. Grav., vol. 21, pp. S1723–S1736, 2004.
[40] T. Accadia, et. al., “Virgo calibration and reconstruction of the gravitational wave
strain during VSR1,” J. Phys. Conf. Ser., vol. 228, p. 012015, 2010.
157
References
[41] A. Królak, B. Schutz, and P. Jaranowski, “Data analysis of gravitational-wave signals from spinning neutron stars. I. The signal and its detection,” Phys. Rev. D,
vol. 58, p. 063001, Aug 1998.
[42] A. Nicolas et al., “Detection in coincidence of gravitational wave bursts with a
network of interferometric detectors: Geometric acceptance and timing,” Phys. Rev.
D, vol. 65, p. 42004, January 2002.
[43] R. Lawrence, Active wavefront correction in laser interferometric gravitational wave
detectors. PhD thesis, Massachusetts Institute of Technology, 2003. LIGO-P03000100-R.
[44] The COMSOL group, “COMSOL Multiphysics,” 2007. http://www.comsol.com.
[45] P. Hello, “Compensation for thermal effects in VIRGO mirrors,” Internal VIRGO
note, 2001. VIR-NOT-LAL-1390-170.
[46] P. Puppo, “A Finite Element Model of the VIRGO Mirrors,” Internal VIRGO note,
2004. VIR-NOT-ROM-1390-262.
[47] P. Puppo, “Recoil mass 2D CAD drawings,” Internal Virgo note, 2006.
[48] M. Punturo, “Thermal lensing analysis and simulation,” Internal Virgo note, 2006.
VIR-0489A-06.
[49] I. H. Malitson, “Inter specimen comparison of the refractive index of fused silica,”
J. Opt. Soc. Am., vol. 55, p. 1205, 1965.
[50] V. Sannibale and P. Willems, “Measurement of the dependence of the Young’s
modulus of fused silica on temperature,” LIGO internal note, 2000. LIGO-T00018800-R.
[51] S. Spinner, “Temperature dependence of elastic constants of vitreous silica,” Journal
of the American Ceramic Society, vol. 45:8, p. 394, 1962.
[52] COMSOL Multiphysics, “COMSOL material database,” 2006.
[53] F. Travasso, Analisi e modellizzazione dei picchi di rumore termico
dell’interferometro VIRGO. PhD thesis, Universitá degli Studi di Perugia, 2004.
[54] W. Winkler, K. Danzmann, A. Rüdiger, and R. Schilling, “Heating by optical absorption and the performance of interferometric gravitational-wave detectors,”Phys.
Rev. A, vol. 44, pp. 7022–7036, December 1991.
[55] M. Punturo, “Thermal lensing FEM with Matlab,” Internal Virgo note, 2006. VIR0114A-06.
[56] M. Punturo, “Sidebands behavior,” Internal Virgo note, 2006. VIR-0434A-06.
158
References
[57] A. Chiummo, J. Marque, and A. Rocchi, “Finesse Simulations for MSRCs with TCS
maps: preliminary results,” Internal Virgo note, April 2011. VIR-0231A-11.
[58] A. Rocchi and V. Fafone, “Thermal Compensation System,” Internal Virgo note,
2007. VIR-0065A-10.
[59] S. Dhurandhar and A. Vecchio, “Searching for continuous gravitational wave sources
in binary systems,” Phys. Rev. D, vol. 63, p. 122001, May 2001.
[60] R. Prix, “Search for continuous gravitational waves: Metric of the multidetector
F-statistic,” Phys. Rev. D, vol. 75, p. 023004, January 2007.
[61] B. Abbott et al., “Beating the spin-down limit on gravitational wave emission from
the Crab pulsar,” AstroPhys. J.l, vol. 683, pp. L45–L49, August 2008.
[62] B. Abbott et al., “Einstein@Home search for periodic gravitational waves in early
S5 LIGO data,” Phys. Rev. D, vol. 80, p. 42003, August 2009.
[63] B. Krishnan et al., “Hough transform search for continuous gravitational waves,”
Phys. Rev. D, vol. 70, p. 82001, October 2004
.
http://link.aps.org/doi/10.1103/PhysRevD.70.082001.
[64] B. Abbott et al., “All-sky search for periodic gravitational waves in LIGO S4 data,”
Phys. Rev. D, vol. 77, p. 22001, January 2008
.
http://Phys.Rev.D.aps.org/abstract/PRD/v77/i2/e022001.
[65] P. Hough, “Machine Analysis of Bubble Chamber Pictures,” Proc. Int. Conf. High
Energy Accelerators and Instrumentation, 1959.
[66] E. A. Goetz, Gravitational wave studies: Detector calibration and an all-sky search
for spinning neutron stars in binary systems. PhD thesis, University of Michigan,
2010.
[67] C. Messenger, “The incoherent metric for parameter space searches for continuous wave sources in binary systems,” Internal LIGO note, February 2010. LIGOP1000020-v1.
[68] E. Weisstein, “Order statistic.” From MathWorld A Wolfram Web Resource.
http://mathworld.wolfram.com/OrderStatistic.html.
[69] E. Weisstein, “Extreme value distribution.” From MathWorld A Wolfram Web
Resource. http://mathworld.wolfram.com/ExtremeValueDistribution.html.
[70] B. Caron et al., “SIESTA, a time domain, general purpose simulation program for
the VIRGO experiment,” Astroparticle Physics, vol. 10, pp. 369–386, May 1999.
[71] “LIGO analysis library
https://www.lsc-group.phys.uwm.edu/daswg/projects/lalsuite.html.
.”
159
References
[72] B. Abbott et al., “Searches for gravitational waves from known pulsars with S5
LIGO data,” Astrophys. J., vol. 713, pp. 671–685, 2010.
[73] R. Dupuis and G. Woan, “Bayesian estimation of pulsar parameters from gravitational wave data,” Phys. Rev. D, vol. 72, p. 102002, November 2005.
[74] “Boost C++ Librairies.” http://www.boost.org/.
[75] “The Frame library: A common data frame format for interferometric gravitational
wave detectors.” http://lappweb06.in2p3.fr/virgo/FrameL/.
[76] M. Frigo and S. Johnson, “The design and implementation of FFTW3,” Proceedings
of the IEEE, vol. 93, no. 2, pp. 216–231, 2005. Special issue on“Program Generation,
Optimization, and Platform Adaptation”, http://fftw.org/.
[77] R. Brun and F. Rademakers, “ROOT - An Object Oriented Data Analysis
Framework,” Nucl. Inst. & Meth. in Phys. Res. A, vol. 389, pp. 81–86, 1997.
http://root.cern.ch/.
[78] svn+ssh://login.nikhef.nl/public/svn/gravwav/Sipho/Sipho_Pulsar.
[79] “Egee - enabling grids for e-science.” http://eu-egee.org/.
[80] B. Abbott et al., “Einstein@Home search for periodic gravitational waves in LIGO
S4 data,” Phys. Rev. D, vol. 79, p. 22001, January 2009
.
http://Phys.Rev.D.aps.org/abstract/PRD/v79/i2/e022001.
[81] W. Press, S. Teulkolsky, W. Vetterling, and B.Flannery, Numerical recepies, the art
of scientific computing, ch. 12,13. Cambridge university press, 2007.
160
Samenvatting
De algemene relativiteitstheorie voorspelt dat een willekeurige massadistributie met een
tijdsafhankelijk quadrupool (of hoger) moment gravitatiegolven zal uitzenden. Deze gravitatiegolven zijn variaties op de achtergrondmetriek en planten zich voort met de lichtsnelheid. Een gravitatiegolf vervormt de ruimtetijd dusdanig dat de relatieve afstand
tussen twee testmassa’s tijdsafhankelijk wordt. Gravitatiegolven die worden uitgezonden
door astrofysische bronnen kunnen gedetecteerd worden met interferometrische gravitatiegolf detectoren. Eén van deze detectoren, Virgo genaamd, is gebouwd in Italië in de
buurt van Pisa. Virgo is een zogenaamde Michelson interferometer met twee Fabry-Perot
resonante armen van elk 3 km lang. De detectieband loopt van 10 Hz tot 10 kHz waarbij
de gevoeligheid tussen de 10−20 Hz−1/2 tot 10−21 Hz−1/2 ligt voor beide uitersten. The
beste gevoeligheid van de Virgo detector gemeten op 1-10-2009 was 6 × 10−23 Hz−1/2
voor gravitatiegolven van 200 Hz.
Omdat de zogenaamde ‘Input’ spiegels van de armen van de Michelson interferometer niet perfect reflecterend zijn, zal de laserbundel de spiegels opwarmen. Door
deze opwarming zullen de optische eigenschappen van spiegels veranderen, hetgeen de
eigenschappen van de Fabry-Perot armen beı̈nvloedt. Deze thermische effecten hebben
negatieve gevolgen voor de stabiele werking van de interferometer.
Hoofdstuk 3 van dit proefschrift beschrijft een volledig drie-dimensionaal eindige elementen model van de Virgo ‘Input’ spiegels samen met hun reactie massa’s. Dit model
wordt gebruikt om een analyse te doen naar de opwarming van de spiegels. Deze analyse
laat zien dat in de evenwichtssituatie de gemiddelde temperatuur met 8.52 mK stijgt.
Verder laat de analyse zien dat de frequentie van de twee prominente resonanties van
de input spiegels (de zogenaamde ‘drum’ en ‘butterfly’ modes) met respectievelijk 0.42
Hz/K en 0.29 Hz/K toenemen. Dit resultaat kan gebruikt worden om de spiegeltemperatuur te meten door de frequentie van de beide resonanties uit de data te extrapoleren.
Het is gebleken dat de spiegels een grotere fractie van het laserlicht absorberen dan
aanvankelijk werd voorspeld. In het geval van de ‘West End’ inputspiegel is gebleken
dat de absorptie een orde van grootte hoger is dan voorheen gedacht. In het geval
van een tijdsafhankelijke analyse is gevonden dat de gemiddelde temperatuursverhoging
bepaald wordt door een tijdconstante van ongeveer 4 uur. Tot slot is berekend dat het
intensiteitsprofiel van de laser leidt tot een niet-uniforme opwarming van de spiegel.
De opgewarmde spiegel heeft in de evenwichtssituatie een berekende kromtestraal van
161
Samenvatting
ongeveer 21 km.
Een van de bronnen van gravitatiegolven zijn niet axiaal-symmetrisch, roterende
neutronensterren. Doordat deze sterren een sterk magnetisch veld hebben, kunnen ze
bundels van elektromagnetische straling vanaf beide polen uitzenden. Als een neutronenster zó georiënteerd is ten opzichte van de Aarde dat de bundel de Aarde kruist, dan
wordt dit geobserveerd als een puls. Deze neutronensterren staan bekend als ‘pulsars’.
Door veel van deze pulsars te observeren en te compenseren voor het feit dat slechts een
deel van de neutronensterren als pulsars gezien kunnen worden, wordt geschat dat er
zich ongeveer 105 actieve pulsars in ons melkwegstelsel bevinden. Verder tonen waarnemingen aan dat het merendeel van de pulsars met rotatiefrequentie boven de 20 Hz
zich in binaire systemen bevindt. Omdat de emissie van gravitatiegolven niet beperkt is
tot de magnetische polen van een neutronenster kunnen de onbekende neutronensterren
gezien worden als potentiële bronnen van gravitatiegolven. De amplitude van de gravitatiegolven die uitgezonden worden door neutronensterren is zwak. Aan de andere kant
zijn de golven continue, dus als zulke systemen over langere tijd geobserveerd worden
zal de signaal-ruis verhouding toenemen met de wortel van de observatietijd.
Om volledig gebruik te maken van het continue karakter van deze gravitatiegolven en
zo de signaal-ruis verhouding te verbeteren, is het noodzakelijk om de precieze golfvorm
te kennen. Dit heeft als vereiste dat de positie van de neutronenster bekend is. Aangezien
we de meeste neutronensterren in ons melkwegstelsel niet hebben waargenomen, is het
niet mogelijk om gebruik te maken van een lange integratietijd. Er zijn zogenaamde ‘allsky’ analysemethodes ontwikkeld die in meer of mindere mate de relatie tussen observatietijd en signaal-ruis verhouding gebruiken. Het is gebleken dat een volledig coherente
analyse (d.w.z. data worden geı̈ntegreerd over de volledige observatietijd) computationeel gelimiteerd is. De meest gevoelige blinde analyse voor geı̈soleerde neutronensterren
kostte 104 CPU jaren aan computertijd en bestreek slechts 30 uur aan data. Om toch
de observatietijd te kunnen verlengen gegeven de gelimiteerde hoeveelheid aanwezige
computertijd, zijn er verschillende sub-optimale analyses ontwikkeld. Deze zogenaamde
semi-coherente analysemethodes zijn gebaseerd op het analyseren van meerdere korte
stukjes data die naderhand op een incoherente manier weer gecombineerd worden. Wat
deze individuele, korte stukjes, data missen aan gevoeligheid wordt deels gecompenseerd
door het feit dat het computationeel mogelijk is om langere stukken data te analyseren.
Het blijkt dat als zulke semi-coherente analyses toegepast worden op gesimuleerde gravitatiegolven afkomstig van neutronensterren in binaire systemen, de lengte van de korte
stukjes data met een orde van grootte gereduceerd moet worden. Deze korte integratietijd zal de gevoeligheid van de semi-coherente analyses extreem benadelen als ze zonder
meer toegepast worden op binaire systemen.
Een nieuwe analysemethode genaamd ‘Polynomial Search’ is ontwikkeld en omschreven in hoofdstuk 5. Polynomial Search is ontwikkeld als een ‘all-sky’ analyse voor
gravitatiegolven die worden uitgezonden door niet axiaal-symmetrisch roterende neutronensterren die zich in binaire systemen bevinden. Polynomial Search maakt gebruik van
een collectie van test-golfvormen gebaseerd op een polynomiaal verloop van de fase als
functie van de tijd. Omdat deze test-golfvormen (ook wel filters genoemd) tot de vierde
orde in tijd gaan, is het mogelijk om de initiële coherente tijd een orde van grootte langer
162
Samenvatting
te maken vergeleken met het toepassen van traditionele semi-coherente analysemethodes op dit soort systemen. Voor een neutronenster in een circulaire, Kepleriaanse, baan
(e = 0) met een periode (P ) van 2 uur is gebleken dat Polynomial Search de coherente
tijd ongeveer een factor 25 kan vergroten. Aan de hand van de verwachtte golfvormen
is er een geschikte collectie filters samengesteld. Alle golfvormen uit deze filtercollectie
beschrijven neutronensterren in Kepleriaanse banen met parameters P > 2 uur en e = 0
zowel als P > 6 uur en e < 0.6. Deze filters kunnen op de data toegepast worden met een
coherente tijd van 600 seconden (ter vergelijking: een traditionele semi-coherente analyse
toegepast op uitsluitend een circulaire baan van P = 2 uur kan maximaal 20 seconden
coherent blijven). Naast de collectie van filters is er ook een strategie ontwikkeld voor
het combineren van de resultaten van de verschillende stukjes coherente analyse. Dit
wordt het ‘consistente filter criterium’ genoemd. Door Polynomial Search op witte ruis
samen met gesimuleerde gravitatiegolf signalen toe te passen, is de voorspelde gevoeligheid berekend. Deze gevoeligheid samen met de volledige beschrijving van Polynomial
Search is gepresenteerd in dit proefschrift.
Tot slot wordt in hoofdstuk 6 het ontwikkelde ‘framework’ waarin Polynomial Search
is geı̈mplementeerd gepresenteerd. Dit framework, genaamd ‘AnaPulsar’, is ontwikkeld
om verschillende soorten gravitatiegolf analysealgoritmes te implementeren. Ook zijn de
computationele eisen voor Polynomial Search besproken. Het is gebleken dat om Polynomial Search met de voorheen genoemde filtercollectie toe te passen op een maand data,
er ongeveer 36 CPU jaren aan rekenkracht nodig is.
163
Samenvatting
164
Acknowledgements
The people without whom I could not have finished this thesis are too many to list here.
Still, I would like to thank a few people in particular.
First of all I would like to thank my promotors Jo van den Brand and Frank Linde for
giving me the opportunity to do this PhD research at Nikhef. As well Henk-Jan Bulten,
Thomas Bauer and Maurik Holtrop for their help with starting a brand new analysis
from scratch in a new field. Also I would like to thank as Eric Hennes for helping me
understand the intricacies of finite element analysis and for letting me give that first
talk at Virgo.
Also, I would like to thank the members of the ever growing gravitational waves group
for their support, help and companionship: Mark, Mathieu, David, Tjonnie, Michalis,
Walter, Salvatore and Chris. May you guys hunt down all unexplained noise resonances
and find many gravitational waves.
An honorable mention goes to my office mates past, present and future: Fabian,
Martin, Paul, Rob, Reinier and of course Gideon. You guys have always been a source
of entertainment and discussions on a variety of subjects. In particular I would like to
thank Gideon for all discussions on anything from thermodynamics to general relativity
to field theory to advanced trigonometry and of course the odd bad-taste joke.
I must of course mention my fellow brothers-in-arms with whom I have had the
pleasure of ‘putting on the penguin’ in the past: Aras, Patrick, Gabriel G., Gordon and
Alex. This last time I will finally be standing in the middle to hopefully join the ranks
of learned and productive members of society.
Of course, without my paranymphs Fabian and Chiara I could never have never
brought this scientific endeavor to the proper conclusion. I know that I will have little
to fear with both of you at my side during the defense. Special thanks to Chiara and
Hegoi for all the work they put into designing the cover of this thesis.
Then, I would like to thank some of the people inside and outside of Nikhef with
whom I was lucky enough to have spent various coffee breaks, ‘borrels’, PhD defenses,
conferences, dinners, regular Saturday nights and many other social occasions. In no
particular order: Hegoi, Giada, Gabriel Y., Eleonora, Ermes, Gianni, Gossie, Aart, Surya,
Monique, Solkin, Kazu, Joana, Daniel, Duncan and many others. Thanks, all of you for
making these past few years unforgettable!
En uiteraard ook de mensen van #WERK en omstanders: Lamp, Monn, DQ, Harvey,
165
Acknowledgements
Oel, Xik, Wok, Wies, Jona, Kees, Marjolein, Wouter, Maarten, Leon, Mark en Jasper.
Bedankt voor de ruige midweken, gave vakanties in de bergen, de bekende dinsdagavond
kansloosheid en uiteraard de jaarlijkse Hemelvaart traditie. Dat we deze activiteiten nog
lang vol mogen houden.
Tot slot mijn familie: ma, pa en Andi ik dank jullie vanuit de bodem van mijn
hart voor alle ondersteuning, interesse en luisterende oren die jullie mij gegeven hebben.
Zonder jullie zou ik nooit de energie hebben gehad om dit proefschrift tot een goed einde
te brengen.
Sipho van der Putten
Amsterdam, Decmber 2011
166