Skip to content
Snippets Groups Projects
Commit f4a8c678 authored by Jutta Schnabel's avatar Jutta Schnabel
Browse files

Adding ANTARES use case

parent 4cdf10bf
No related branches found
No related tags found
No related merge requests found
Pipeline #14796 passed
%% Cell type:markdown id: tags:
# Dummy point-source study with ANTARES data
The gammapy package is used to compute Feldman-Cousins statistics
## Prerequisits
`pip install gammapy, matplotlib, openkm3`
%% Cell type:code id: tags:
``` python
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import math
from scipy import stats
import gammapy.stats as gstats
import openkm3
```
%% Cell type:markdown id: tags:
## Helper functions
### Background evaluation
The function *get_number_background(decl, roi)* permits to evaluate the expected background for a give sky declination and a given region of interest ('half width' of the declination band around the source declination). The right ascension is not used here, as the background distribution depends only on the visibility of the given sky position from the location of ANTARES, and therefore on the declination only.
This function uses a polynomial interpolation to smoothen the background distibution otherwise obtained from the data. This function returns the number of expected background events that in the following will be indicated as nbkg.
%% Cell type:code id: tags:
``` python
def get_number_background(interpolation, decl, roi):
dec_low = decl - roi;
dec_high = decl + roi;
# estimation of background rate over the full declination band
bkg_low = interpolation(dec_low)
bkg_high = interpolation(dec_high)
bkg_band = bkg_high - bkg_low
# solid angle of declination band
solid_angle = abs(2*math.pi*(np.sin(math.radians(dec_high) - np.sin(math.radians(dec_low)))))
# background rate per unid solid angle
bkg_rate = bkg_band / solid_angle
# expected counts in the RoI
bkg_count = bkg_rate * math.pi * pow(math.radians(roi), 2)
return bkg_count
```
%% Cell type:markdown id: tags:
### Limit calcuation (Felman-Cousins
The functions *get_upper_limit(nbkg, nobs)* and *get_lower_limit(nbkg, nobs)* compute the upper and lower limits in presence of nbkg backgroud events, if the measurements is nobs observed events inside the region of interest.
%% Cell type:code id: tags:
``` python
# compute upper limit according to Feldamn-Cousins
def get_upper_limit(nbkg, nobs):
if nobs <= int(nbkg):
ul = Sens(nbkg) #if underfluctuation, UL set as the sensitivity
else:
x_bins = np.arange(0, 50)
mu_bins = np.linspace(0, 15, 300 + 1, endpoint=True) # 300 = 15 / 0.05
matrix = [stats.poisson(mu + nbkg).pmf(x_bins) for mu in mu_bins] # check nbkg!
acceptance_intervals = gstats.fc_construct_acceptance_intervals_pdfs(matrix, 0.9)
LowerLimitNum, UpperLimitNum, _ = gstats.fc_get_limits(mu_bins, x_bins, acceptance_intervals)
ul = UpperLimitNum
return ul
def get_lower_limit(nbkg, nobs):
x_bins = np.arange(0, 50)
mu_bins = np.linspace(0, 15, 300 + 1, endpoint=True) # 300 = 15 / 0.05
matrix = [stats.poisson(mu + nbkg).pmf(x_bins) for mu in mu_bins] # check nbkg!
acceptance_intervals = gstats.fc_construct_acceptance_intervals_pdfs(matrix, 0.9)
LowerLimitNum, UpperLimitNum, _ = gstats.fc_get_limits(mu_bins, x_bins, acceptance_intervals)
ll = LowerLimitNum
return ll
```
%% Cell type:markdown id: tags:
### Sensitivity calculation
The function *get_sensitivity(nbkg)* is used to calculate the sensitivity, defined as average upper limit, that is reached in presence of nbkg events in the declination band of the source.
%% Cell type:code id: tags:
``` python
confidence_level = 0.9 # set confidence level to 90%
def get_sensitivity(nbkg):
x_bins = np.arange(0, 20)
mu_bins = np.linspace(0, 200, 20000 + 1, endpoint=True) # 20000 = 200 / 0.01
matrix = [stats.poisson(mu + nbkg).pmf(x_bins) for mu in mu_bins] # check nbkg!
acceptance_intervals = gstats.fc_construct_acceptance_intervals_pdfs(matrix, confidence_level)
LowerLimitNum, UpperLimitNum, _ = gstats.fc_get_limits(mu_bins, x_bins, acceptance_intervals)
averageUL = gstats.fc_find_average_upper_limit(x_bins, matrix, UpperLimitNum, mu_bins)
return averageUL
```
%% Cell type:markdown id: tags:
### Picking number of observed events from full antares list
Retrieves the number of observed events from the eventlist for a given declination and region of interest.
-> This function could actually be replaced using the pyvo interface, as the VO offers the SCS protocol to do exactly this.
%% Cell type:code id: tags:
``` python
def get_number_observed_events(eventlist, decl, roi):
obs = 0
for counter in len(eventlist):
# fill vectors with RA and dec of each ANTARES track
event = make_vector(eventlist.RA[counter], eventlist.Decl[counter])
alpha = calc_angle(dir_source, event)
# count events whose angle from the source falls inside the RoI
if alpha < RoI:
obs = obs+1
return obs
```
%% Cell type:markdown id: tags:
### Math helpers
The two functions *make_vector(Decl, RA)* and *calc_angle(v1, v2)* are auxilary functions used later.
*make_vector* fills a 3-dimensional vector using the sky coordinates of an event (Decl, RA), and assigning length 1.
*calc_angle* computes the angular distance (in units degrees) between two points in the sky defined as the angle between their corresponding vectors v1 and v2.
%% Cell type:code id: tags:
``` python
def make_vector(Decl, RA):
x = math.cos(math.radians(Decl))*math.cos(math.radians(RA))
y = math.cos(math.radians(Decl))*math.sin(math.radians(RA))
z = math.sin(math.radians(Decl))
v =[x,y,z]
return v
def calc_angle(v1, v2):
unit_vector_1 = v1 / np.linalg.norm(v1)
unit_vector_2 = v2 / np.linalg.norm(v2)
dot_product = np.dot(unit_vector_1, unit_vector_2)
angle_between_vectors = math.degrees(np.arccos(dot_product))
return angle_between_vectors
```
%% Cell type:markdown id: tags:
## Calculating Limits
Here, the flux upper limits for the observation of ANTARES events from the 2007-2017 sample for a given direction in the sky is calculated.
### Setting the point source coordinates
The limits are calculated providing
* the source coordinates **right ascension** and **declination**, ra_source and dec_source, and
* the two additional source parameters **gamma** (spectral index) and
* **RoI** (region of interest, representing the source extension).
%% Cell type:code id: tags:
``` python
ra_source = 0.1 # between 0 and 360
dec_source = -29 # Declination must be between -80 and 50 degree
gamma = 2.0 # spectral index between 1.5 and 3.0
RoI = 2.6 # Region of interest > 0.1 and < 1000 (in degree)
```
%% Cell type:markdown id: tags:
### Getting ANTARES data and acceptance
This is a proposal of how retrieving the relevant ANTARES data should work.
The KM3Store should offer
* the event list
* the acceptance function
* the background estimate from interpolation
The following functions are now dummies, but they can be added to work accordingly in the first release of openkm3.
See http://pi1154.physik.uni-erlangen.de:82/data/collections/1/view for what is already provided
%% Cell type:code id: tags:
``` python
# read the ANTARES data tracks into a dataframe - PROPOSAL!
ks = openkm3.KM3Store()
ant07_17_collection = ks.get("ant2017")
ant07_17_collection.list()
```
%% Cell type:markdown id: tags:
should return a table e.g.
| identifier | name | description | type |
| --------| --------- | ------- | ------- |
| "events" | ANTARES 2007-2017 neutrino events | Event list of neutrino candidates | km3.data.l4.vo.scs |
| "acceptance" | ANTARES 2007-2017 detector acceptance | acceptance table (for different spectral indices and sin(decl) | km3.data.l4.table.d2 |
| "background_distribution" | ANTARES 2007-2017 background distribution | interpolation of expected background events for given declination and region of interest | km3.data.l4.function.d6 |
%% Cell type:code id: tags:
``` python
eventlist = ant07_17_collection.get_table("events", "pandas")
# would return the pandas eventlist, but perhaps better pyvo for this?
acceptance = ant07_17_collection.get_table("acceptance", "pandas")
# would return the acceptance as table.
bkgfit = ant07_17_collection.get_function("background_distribution")
# would return a function to be used for background calcuation
```
%% Cell type:markdown id: tags:
Could actually offer the acceptance table as lookup table with labeled rows and columns, so the right entry is easier to pick.
So more something like
`acceptance = ant07_17_collection.get_table("acceptance", "lookup")`
### Cut and count analysis
%% Cell type:code id: tags:
``` python
nevents_obs = get_number_observed_events(eventlist, decl, roi)
nevents_bkg = get_number_background(bkgfit, decl, roi)
print("estimated background: ", nevents_bkg)
print("number of observed events: ", nevents_obs)
n_ul = get_upper_limit(nevents_bkg, nevents_obs)
n_ll = get_lower_limit(nevents_bkg, nevents_obs)
# get bins declination and spectral index
bin_dec = int(math.sin(math.radians(dec_source))+1)*10;
bin_gamma = int((gamma*10)) - 15;
# convert number upper limits into flux upper limits dividing by acceptance
f_ul = n_ul/acceptance({"gamma": gamma, "declination": dec_source})
f_ll = n_ll/acceptance({"gamma": gamma, "declination": dec_source})
# output
print("Upper Limit: ", n_ul)
#print("LL ", n_ll)
print("flux Upper Limit (Normalized at GeV): ", f_ul , " (GeV cm s)^-1")
print("flux Lower Limit (Normalized at 100 TeV) ", f_ul*pow(1e-5, gamma), " (GeV cm s)^-1")
# if wished: plot
df.plot(kind='scatter', x='RA', y='Decl', alpha=0.1);
plt.show()
```
......@@ -40,5 +40,5 @@ The FAIR principles provide a solid set of requirements for the development of a
#### Reusable data
* **[Licensing standards](https://open-data.pages.km3net.de/licensing/)** for data, software and supplementary material have been introduced.
* Basic **provenance** information is provided with the data, which serves as a development start point to propagate provenance management through the complex the full data processing workflow in the future.
* Basic **provenance** information is provided with the data, which serves as a development start point to propagate provenance management through the complex data processing workflow in the future.
......@@ -10,24 +10,23 @@ status: unedited
# Example programs and Use cases
## ANTARES 2007-2017 Point source analysis
### Use case description
One of the primary goals for ANTARES is the identification of neutrino sources, whose signature would appear in ANTARES data (in form of neutrino arrival directions) as clusters of events at given sky coordinates.
The significance of a neutrino excess from a given sky position must be assessed over the expected background fluctuations using, for instance the Feldman-Cousins statistics.
This ANTARES use case allows to inspect a sample of neutrino arrival directions in equatorial coordinates (RA, dec), evaluate from it the expected background rate for a user-selected sky position, and finally assess the significance of a cluster defined as 'all arrival directions that fall inside a given radius, selected by the user and indicated here 'region of interest' (RoI).
Here follow the step-by-step description of the code provided at https://git.km3net.de/srgozzini/workflow_ps/-/blob/master/PS.ipynb
- The function *bkg_evaluation(decl, roi)* permits to evaluate the expected background for a give sky declination and a given region of interest ('half width' of the declination band around the source declination). The right ascension is not used here, as the background distribution depends only on the visibility of the given sky position from the location of ANTARES, and therefore on the declination only.
This function uses a polynomial interpolation to smoothen the background distibution otherwise obtained from the data. This function returns the number of expected background events that in the following will be indicated as nbkg.
- the functions *UL(nbkg, nobs)* and *LL(nbkg, nobs)* compute the upper and lower limits in presence of nbkg backgroud events, if the measurements is nobs observed events inside the region of interest.
- the function *Sens(nbkg)* is used to calculate the sensitivity, defined as average upper limit, that is reached in presence of nbkg events in the declination band of the source.
- the two functions *make_vector(Decl, RA)* and *calc_angle(v1, v2)* are auxilary functions used later. *make_vector* fills a 3-dimensional vector using the sky coordinates of an event (Decl, RA), and assigning length 1. *calc_angle* computes the angular distance (in units degrees) between two points in the sky defined as the angle between their corresponding vectors v1 and v2.
Here follow the step-by-step description of the code provided as [Jupyter notebook](notebooks/A01_recorded_rate)
- The function *compute_limits(ra_source, dec_source, gamma, RoI)* is the main function of this script and computes the flux upper limits for a given direction in the sky defined from the coordinates ra_source and dec_source, and for the two additional source parameters gamma (spectral index) and RoI (region of interest, representing the source extension).
### Data set
## Using ANTARES data
ANTARES data has already been published to the VO for two data sets by using the services of the German Astrophysical Virtual Observatory (GAVO) which run the DaCHS software. The most recent public data sample of the 2007-2017 point source search is available through the ANTARES website, however, it is thus not findable through VO and does not match the FAIR criteria. Including ANTARES data in the development of future VO data types allowes to increase the chance for a long-term availability of high-quality ANTARES data. On the other hand, the KM3NeT VO server could be registered to the VO and protocols be tested using the ANTARES 2007-2017 sample.
ANTARES data has already been published to the VO for two data sets by using the services of the German Astrophysical Virtual Observatory (GAVO) which run the DaCHS software. The most recent public data sample is available through the ANTARES website, however, it is thus not findable through VO and does not match the FAIR criteria. Including ANTARES data in the development of future VO data types allowes to increase the chance for a long-term availability of high-quality ANTARES data. On the other hand, the KM3NeT VO server could be registered to the VO and protocols be tested using the ANTARES 2007-2017 sample.
The provided data set includes
* The **full event list** of 2007-2017 selected astrophysics neutrino candidates, provided through the VO server,
* Supplementary distributions from simulations provided via the ODC including
* the **detector acceptance** for a given source spectral index and declination
* the interpolated **distribution of background events** for a given declination and region of interest
* the **effective area** for an E^-2 source spectrum in three different zenith bands
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment