Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
GravStat 2005 Statistical problems in network data analysis: burst searches by narrowband detectors L.Baggio and G.A.Prodi ICRR Tokyo Univ.Trento and INFN narrowband detectors & same directional sensitivity Cons: probing a smaller volume of the signal parameter space Pros: simpler problem IGEC time coincidence search is taking advantage of “a priori” information • template search: matched filters optimized for short and rare transient gw with flat Fourier transform over the detector frequency band • many trials at once: - different detector configurations (9 pairs + 7 triples + 2 four-fold) - many target thresholds on the searched gw amplitude (30) - directional / non directional searches GravStat 2005 … IGEC cont`d • data selection and time coincidence search: - control of false dismissal probability - balance between efficiency of detection and background fluctuations • background noise estimation - high statistics: 103 time lags for detector pairs 104 – 105 detector triples - goodness of fit tests with background model (Poisson) • blind analysis (“good will”): - tuning of procedures on time shifted data by looking at all the observation time (no playground) … what if evidence for a claim would appear ? “GW candidates will be given special attention …” - IGEC-2 agreed on a blind data exchange (secret time lag) GravStat 2005 Poisson statistics verified For each couple of detectors and amplitude selection, the resampled statistics allows to test Poisson hypothesis for accidental coincidences. Example: EX-NA background (one-tail 2 p-level 0.71) As for all two-fold combinations a fairly big number of tests are performed, the overall agreement of the histogram of p-levels with uniform distribution says the last word on the goodness-of-the-fit. GravStat 2005 physical unknown A few basics: confidence belts and coverage 1 coverage 0 x x x experimental data GravStat 2005 Freedom of choice of confidence belt Fixed frequentistic coverage Feldman & Cousins (1998) and variations (Giunti 1999, Roe & Woodroofe 1999, ...) C() CL I can be chosen arbitrarily within this “horizontal” constraint 1 coverage 0 GravStat 2005 Confidence intervals from likelihood integral • Let Nc Nb N Nb Tobs e Nb N Nc • Poisson pdf: f ( Nc ; N ) Nb N Nc ! • Likelihood: (N ; Nc ) f (Nc ; N ) (Ninf ; Nc ) (Nsup ; Nc ) 1 N • I fixed, solve for 0 Ninf Nsup: sup I ( N ; N ) dN (N ; Nc )dN 0 c N inf • Compute the coverage C(N ) Nc|Ninf N Nsup f (Nc ; N ) I Plot of the likelihood integral vs. minimum (conservative) coverage minN C(N ), with background counts Nb=0.01-10 Example: Poisson background Nb = 7.0 10 10 9 9 99.9% 8 8 7 99% 7 95% 6 6 N 5 50% 50% 5 95% 4 99% 4 99.9% 3 3 2 2 1 1 0 0 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 coincidence counts Nc 99% Nc Nb N 95% Likelihood integral 85% N GravStat 2005 Confidence intervals from likelihood integral • Let Nc Nb N Nb Tobs e Nb N Nc • Poisson pdf: f ( Nc ; N ) Nb N Nc ! • Likelihood: (N ; Nc ) f (Nc ; N ) (Ninf ; Nc ) (Nsup ; Nc ) 1 N • I fixed, solve for 0 Ninf Nsup: sup I ( N ; N ) dN (N ; Nc )dN 0 c N inf • Compute the coverage C(N ) Nc|Ninf N Nsup f (Nc ; N ) I Plot of the likelihood integral vs. minimum (conservative) coverage minN C(N ), with background counts Nb=0.01-10 GravStat 2005 Multiple configurations/selection/grouping within IGEC analysis GravStat 2005 Resampling statistics of accidental claims event time series 500 counts 400 300 200 100 0 expected found coverage “claims” 0.90 0.866 (0.555) [1] 0.95 0.404 (0.326) [1] 0 1 2 3 4 5 numer of false alarms Easy to set up a blind search GravStat 2005 Keep track of the number of trials (and their correlation) ! IGEC-1 final results consist of a few sets of tens of Confidence Intervals with min{C}=95% the “false positives” would hide true discoveries requiring more than 5 twosided C.I. to reach 0.1% confidence for rejecting H0 the procedure was good for Upper Limits, but NOT optimized for discoveries Need to decrease the “false alarm probability” (type I error) Freedom of choice of confidence belt Fine tune of the false alarm probability GWfanatic enthusiastic skeptical GravStat 2005 Example: confidence belt from likelihood integral Poisson background Nb = 7.0 Min{C}=95% P{false alarm} < 5 % P{false alarm} < 0.1% 1 - C(N ) Nc Nb N GravStat 2005 What false alarm threshold should be used to claim evidence for rejecting the null H0? • control the overall false detection probability: Familywise Error Rate < requires single C.I. with P{false alarm} < /m Pro: rare mistakes Con: high detection inefficiency • control the mean False Discovery Rate: R = total number of reported discoveries F+ = actual number of false positives F q R Benjamini & Hochberg (JRSS-B (1995) 57:289-300) Miller et. al. (A J 122: 3492-3505 Dec 2001; http://arxiv.org/abs/astro-ph/0107034) FDR control The p-values are uniformly distributed in [0,1] if the assumed hypothesis is true Usually, the alternative hypothesis is not known. However, the presence of a signal would contribute to bias the p-values distribution. pdf Typically, the measured values of p are biased toward 0. 1 signal background p-level Sketch of Benjamini & Hochberg FDR control procedure • compute p-values {p1, p2, … pm} for a set of tests, and sort them in creasing order; • choose your desired bound q on <FDR>; • determine the threshold T= pk by finding the index k such that pj<(q/m) j for every j>k; • OK if p-values are independent or positively correlated • in case NO signal is present (H0 is true), the procedure is equivalent to the control of the FamilyWise Error Rate at confidence < q m counts reject H0 q T p-value GravStat 2005 Open questions check the fluctuations of the random variable FDR with respect to the mean. check how the expected uniform distribution of p-values for the null H0 can be biased (systematics, …) would the colleagues agree that overcoming the threshold chosen to control FDR means & requires reporting a rejection of the null hypothesis ? To me rejection of the null is a claim for an excess correlation in the observatory at the true time, not taken into account in the measured noise background at different time lags. It could NOT be gws, but a paper reporting the H0 rejection is worthwhile and due.