Survey

* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project

Document related concepts

Photon polarization wikipedia, lookup

Transcript
```Gravitational Wave
Data Analysis
Probability and Statistics
Junwei Cao (曹军威) and Junwei Li (李俊伟)
Tsinghua University
Gravitational Wave Summer School
Kunming, China, July 2009
Drawback of Matched Filtering
Two premises of matched filtering
 Premise 1: A given signal is present in data
stream
 Premise 2: The form of h(t) is known
Premises are impractical
Repeat matched filtering with many
different filters
A number of “events” are extracted
 “events” indicate that in the detector
happened something, which deserves further
scrutiny
Definition of Probability
Consider a set S with subsets A, B…
Define probability P as a real function
 For every A in S, P(A)≥0
 For disjoint subsets (i.e. A∩B= ),
P(A∪B)=P(A)+P(B)
 P(S)=1
Further more, conditional probability
P( A  B)
P( A | B) 
P( B)
Two Approaches of Probability
Frequentist (also called classical)
 A, B, … are the outcome of a repeatable
experiment, and P(A) is defined as the
frequency of occurrence of A
 In considering the conditional probability, such
as P(data | hypothesis), one is never allowed to
think about the probability that the parameters
take a given a value, nor of the probability that a
hypothesis is correct
Two Approaches of Probability(contd.)
Bayesian approach
 Bayes’ theorem
P( B | A) P( A)
P( A | B) 
P( B)
(3.1)
P( B)   P( B | Ai ) P( Ai )
(3.2)
i
P( A | B) 
P( B | A) P( A)
 P( B | A ) P( A )
i
i
i
(3.3)
Prior and Posterior Probability
P(hypothesis | data)  P(data | hypothesis) P( hypothesis)
Posterior probability
Likelihood function Prior probability
(3.4)
Which One to be Chosen
Depend on the type of experiment
 Elementary particle physics is suited for the
classical approach since it is the physicist that
controls the parameters of the experiment
 In astrophysics, the sources can be rare, and
each one is very interesting individually, e.g. a
single BH-BH binary coalesces
Before Parameter Estimation
A number of free parameters
A family of possible templates
 Denoted generically as h(t ; )
   1,..., N  is a collection of parameters
A family of optimal filters
 Denoted generically as K (t; )
~
~
 Determined by eq. 2.12, K ( f ; ) ~ h( f ; ) / Sn ( f )
 Must discretize the θ-space
For some template, the SNR exceeds
a predefined threshold, indicating a
detection
Parameter Estimation
How to reconstruct parameters of the
source
Assume that n(t) is stationary and
Gaussian
Corresponding Gaussian probability
distribution for n(t)
p(n0 )  N exp (n0 | n0 ) / 2
(3.5)
This is the probability that n(t) has a
given realization n0(t)
Likelihood Function
Assumption
 s(t )  h(t;t )  n0 (t )
 θt is the true value of the parameters θ
Likelihood function for s(t), by
plugging n0=s-h(θt) into 3.5
 1

( s | t )  N exp  ( s  h(t ) | s  h(t )) 
 2

(3.6)
Introduce ht≡h(θt)
1
1


( s | t )  N exp (ht | s )  (ht | ht )  ( s | s ) 
2
2


(3.7)
Posterior Probability Distribution


(t ) to
Introduce a prior probability
eq. 3.7
1


p(t | s )  Np (t ) exp ( ht | s)  (ht | ht ) 
2


(0)
p
(t ) can be an un-flat prior in θt

(0)

(3.8)
Estimator: a rule for assigning  , the
most probable value of θt
 Consistency

 The bias b  E ( )  t
 Efficiency
 Robustness
Maximum Likelihood Estimator
First, the prior probability is flat
 Max. posterior  Max. likelihood (s | t )
 Denote the maximum likelihood estimator by

 ML ( s)
Simpler to maximize log 
1
log ( s | t )  (ht | s)  (ht | ht )
2
(3.9)
Define  i   / ti
(i ht | s)  (i ht | ht )  0
(3.10)
Maximum Posterior Probability
Maximize the full posterior probability
 Takes into account the prior probability distribution
 Non-trivial prior information
Example
 Two-dimensional parameters space (θ1,θ2)
~
 Only interested in θ1
p(1 | s)   p(1 , 2 | s)
Drawback
 An ambiguity on the value of the most probable
value of θ1
 Unable to minimize the error on θ1 determination
Bayes Estimator
Most probable value of parameters

 Bi ( s)    i p( | s)d
(3.11)
Errors
 i  i  j  j 
 B      B (s)     B (s)  p( | s)d
ij
(3.12)
The “operational” meaning

i
i

(
s
)

 B Is the value of , averaged over an
ensemble of same outputs
Drawback: computational costs
Gravitational Wave
Data Analysis
Junwei Cao
[email protected]
http://ligo.org.cn
```
Related documents