Spectrum Reconstruction of Atmospheric Neutrinos with

Download Report

Transcript Spectrum Reconstruction of Atmospheric Neutrinos with

Spectrum
Reconstruction of
Atmospheric
Neutrinos with
Unfolding Techniques
Juande Zornoza
UW Madison
Introduction
 We will review different approaches for
the reconstruction of the energy
spectrum of atmospheric neutrinos:
 Blobel / Singular Value Decompostion
 actually, both methods are basically the same,
with only differences in issues not directly related
with the unfolding but with the regularization and
so on
 Singular Value Decomposition
Spectrum reconstruction
 In principle, energy spectra can be obtained using the
reconstructed energy of each event.
 However, this is not
efficient in our case
because of the
combination of two factors:
 Fast decrease (power-law) of
the flux.
 Large fluctuations in the
energy deposition.
 For this reason, an alternative way has to be used:
unfolding techniques.
Spectrum unfolding
•Quantity to obtain: y, which follows pdf→ftrue(y)
•Measured quantity: b, which follows pdf→fmeas(b)
•Both are related by the Fredholm integral equation of first kind:
f meas (b)   R(b | y ) ftrue ( y )dy
Matrix notation
ˆ b
Ay
•The response matrix takes into account three factors:
-Limited acceptance
-Finite resolution
-Transformation (in our case, b will be xlow)
•The response matrix inversion gives useless solutions, due to the
effect of statistical fluctuations, so two alternative methods
have been tried:
-Singular Value Decomposition
-Iterative Method based on Bayes’ Theorem
Dealing with instabilities
 Regularization:
 Solution with minimum curvature
 Solution with strictly positive curvature
 Principle of maximum entropy
 Iterative procedure, leading
asymptotically to the unfolded distribution
Single Value Decomposition1
 The response matrix is decomposed as:
Aˆ  USV T
U, V: orthogonal matrices
S: non-negative diagonal matrix (si, singular values)
 This can be also seen as a minimization
problem:
2


ˆ
  Aij xi  bi   min



i 1  j 1

nb
nx
 Or, introducing the covariance matrix to take
into account errors:

 

ˆAx  b T B 1 Aˆ x  b  min
1
A. Hoecker, Nucl. Inst. Meth. in Phys. Res. A 372:469 (1996)
SVD: normalization
 Actually, it is convenient to normalize the
unknowns
ˆ b
Ay
w j  y j / yini j
nx
A w
j 1
ij
j
 bi
 Aij contains the number of events, not the
probability.
 Advantages:
 the vector w should be smooth, with small bin-to-bin
variations
 avoid weighting too much the cases of 100%
probability when only one event is in the bin
SVD: Rescaling
 Rotation of matrices allows to rewrite the
system with a covariance matrix equal to
I, more convenient to work with:
B  QRQ
T
1
Aij   Qim Amj
ri m
~T ~
~
~
( Aw  b ) ( Aw  b )  min
bi 
1
Qimbm

ri m
Regularization
 Several methods have been proposed for the
regularization. The most common is to add a
curvature term add a curvature term
~T ~
~
~
( Aw  b ) ( Aw  b )   (Cw)T Cw  min
 1  

 1
C  0




1
0
2  
1
1
1
1
0
2  
1



0 

1 
1   
 Other option: principle of maximum entropy
Regularization
 We have transformed the problem in the
optimization of the value of , which tunes how
much regularization we include:
  too large: physical information lost
  too small: statistical fluctuations spoil the result
In order to optimize the
value of :
Evaluation using MC
information
Components of vector d  U T b
Maximum curvature of
the L-curve
L-curve
Solution to the system
A '  AC 1
w '  Cw
 Actually, the solution to the system with
the curvature term can be expressed as
a function of the solution without
curvature:
ny
w'  
i 1
di
f i vi
si
where
d
( )
i
si2
 di 2
si  
2
1

if
s
s
i  
fi  2
 2
si   si /  if si2  
2
i
(Tikhonov factors)
Tikhonov factors
 The non-zero tau is equivalet to change di by
 1 if si2  
si2
fi  2
 2
si   si /  if si2  
 And this allows to find a criteria to find a good tau
fun0
fun1
fun2
Components of d
k
 = sk2
Bayesian Iterative
2
Method
•If there are several causes (Ei) which can produce an effect Xj and we
know the initial probability of the causes P(Ei), the conditional probability of
the cause to be Ei when X is observed is:
P( Ei | X j ) 
smearing matrix: MC
P( X j | Ei ) P0 ( Ei )
 P( X
l 1
j
| El ) P0 ( El )
prior guess:
iterative
approach
•The expected number of events to be assigned to each of the
nX
causes is:
nˆ ( Ei )   n( X j ) P( Ei | X j )
j 1
experimental data (simulated)
•The dependence on the initial probability P0(Ei) can be overcome by
an iterative process.
nˆ ( Ei )
ˆ
P( Ei )  nE
 nˆ( Ei )
i 1
2
G. D'Agostini NIM A362(1995) 487-498
Iterative algorithm
1. Choose the initial distribution P0(E). For instance, a good guess
could be the atmospheric flux (without either prompt neutrinos or
signal).
2. Calculate nˆ ( E ) and Pˆ ( E ).
3. Compare nˆ ( E ) to n0 ( E ).
4. Replace P0 ( E ) by Pˆ ( E ) and n0 ( E ) by nˆ ( E ).
5. Go to step 2.
P(Xj|Ei)
Smearing matrix (MC)
no(E)
Initial
guess
P(Ei|Xj)
Reconsructed
spectrum
n(Ei)
Po(E)
n(Xj)
Experimental data
P(Ei)
Differences between SVD and Blobel
 Different curvature term
 Selection of optimum tau
 B-splines used in the standard Blobel
implementation
…
B-splines
 Spline: piecewise continuous
and differentiable function that
connects two neighbor points
by a cubic polynomial:
from H. Greene PhD.
 B-spline: spline functions can be expressed by a
finite superposition of base functions (B-spilines).
(first order)
(higher orders)
For IceCube
 Several parameters can be investigated:




Number of channels
Number of hits
Reconstructed energy
Neural network output…
 With IceCube, we will have much better
statistics than with AMANDA
 But first, reconstruction with 9 strings will be
the priority
Remarks
 First, a good agreement between data and MC
is necessary
 Different unfolding methods will be compared
(several internal parameters to tune in each
method)
 Several regularization techniques are also
available in the literature
 Also an investigation on the best variable for
unfolding has to be done
 Maybe several variables can be used in a
multi-D analysis