This is a fixed-text formatted version of a Jupyter notebook
You can contribute with your own notebooks in this GitHub repository.
Source files: light_curve.ipynb | light_curve.py
Light curve estimation¶
Introduction¶
This tutorial presents how light curve extraction is performed in gammapy.
We will demonstrate how to compute a gammapy.time.LightCurve
from 3D reduced datasets (gammapy.cube.MapDataset
) as well as 1D ON-OFF spectral datasets (gammapy.spectrum.SpectrumDatasetOnOff
).
The data reduction will be performed with the high level interface for the data reduction. Then we will use the gammapy.time.LightCurveEstimator
class, which is able to extract a light curve independently of the dataset type.
We will compute two LCs: one per observing run and one per night.
We will use the four Crab nebula observations from the H.E.S.S. first public test data release and compute per-observation fluxes. The Crab nebula is not known to be variable at TeV energies, so we expect constant brightness within statistical and systematic errors.
Setup¶
As usual, we’ll start with some general imports…
[1]:
%matplotlib inline
import matplotlib.pyplot as plt
import astropy.units as u
from astropy.coordinates import SkyCoord
import logging
from astropy.time import Time
log = logging.getLogger(__name__)
Now let’s import gammapy specific classes and functions
[2]:
from gammapy.modeling.models import PowerLawSpectralModel
from gammapy.modeling.models import PointSpatialModel
from gammapy.modeling.models import SkyModel, SkyModels
from gammapy.time import LightCurveEstimator
from gammapy.analysis import Analysis, AnalysisConfig
Analysis configuration¶
For the 1D and 3D extraction, we will use the same CrabNebula configuration than in the notebook analysis_1.ipynb using the high level interface of Gammapy.
From the high level interface, the data reduction for those observations is performed as followed
Building the 3D analysis configuration¶
[3]:
conf_3d = AnalysisConfig()
Definition of the data selection¶
Here we use the Crab runs from the HESS DL3 data release 1
[4]:
conf_3d.observations.obs_ids = [23523, 23526, 23559, 23592]
Definition of the dataset geometry¶
[5]:
# We want a 3D analysis
conf_3d.datasets.type = "3d"
# We want to extract the data by observation and therefore to not stack them
conf_3d.datasets.stack = False
# Here is the WCS geometry of the Maps
conf_3d.datasets.geom.wcs.skydir = dict(
frame="icrs", lon=83.63308 * u.deg, lat=22.01450 * u.deg
)
conf_3d.datasets.geom.wcs.binsize = 0.02 * u.deg
conf_3d.datasets.geom.wcs.fov = dict(width=1 * u.deg, height=1 * u.deg)
# We define a value for the IRF Maps binsize
conf_3d.datasets.geom.wcs.binsize_irf = 0.2 * u.deg
# Define energy binning for the Maps
conf_3d.datasets.geom.axes.energy = dict(
min=0.7 * u.TeV, max=10 * u.TeV, nbins=5
)
conf_3d.datasets.geom.axes.energy_true = dict(
min=0.3 * u.TeV, max=20 * u.TeV, nbins=10
)
Run the 3D data reduction¶
[6]:
analysis_3d = Analysis(conf_3d)
analysis_3d.get_observations()
analysis_3d.get_datasets()
Setting logging config: {'level': 'INFO', 'filename': None, 'filemode': None, 'format': None, 'datefmt': None}
Fetching observations.
Number of selected observations: 4
Creating geometry.
Creating datasets.
Processing observation 23523
Processing observation 23526
Processing observation 23559
Processing observation 23592
Define the model to be used¶
Here we don’t try to fit the model parameters to the whole dataset, but we use predefined values instead.
[7]:
target_position = SkyCoord(ra=83.63308, dec=22.01450, unit="deg")
spatial_model = PointSpatialModel(
lon_0=target_position.ra, lat_0=target_position.dec, frame="icrs"
)
spectral_model = PowerLawSpectralModel(
index=2.702,
amplitude=4.712e-11 * u.Unit("1 / (cm2 s TeV)"),
reference=1 * u.TeV,
)
sky_model = SkyModel(
spatial_model=spatial_model, spectral_model=spectral_model, name="crab"
)
# Now we freeze these parameters that we don't want the light curve estimator to change
sky_model.parameters["index"].frozen = True
sky_model.parameters["lon_0"].frozen = True
sky_model.parameters["lat_0"].frozen = True
We assign them the model to be fitted to each dataset
[8]:
models = SkyModels([sky_model])
analysis_3d.set_models(models)
Reading model.
SkyModels
Component 0: SkyModel
name value error unit min max frozen
--------- --------- ----- -------------- ---------- --------- ------
lon_0 8.363e+01 nan deg nan nan True
lat_0 2.201e+01 nan deg -9.000e+01 9.000e+01 True
index 2.702e+00 nan nan nan True
amplitude 4.712e-11 nan cm-2 s-1 TeV-1 nan nan False
reference 1.000e+00 nan TeV nan nan True
Light Curve estimation: by observation¶
We can now create the light curve estimator.
We pass it the list of datasets and the name of the model component for which we want to build the light curve. We can optionally ask for parameters reoptimization during fit, that is most of the time to fit background normalization in each time bin.
If we don’t set any time interval, the gammapy.time.LightCurveEstimator
is determines the flux of each dataset and places it at the corresponding time in the light curve. Here one dataset equals to one observing run.
[9]:
lc_maker_3d = LightCurveEstimator(
analysis_3d.datasets, source="crab", reoptimize=False
)
lc_3d = lc_maker_3d.run(e_ref=1 * u.TeV, e_min=1.0 * u.TeV, e_max=10.0 * u.TeV)
The LightCurve object contains a table which we can explore.
[10]:
lc_3d.table["time_min", "time_max", "e_min", "e_max", "flux", "flux_err"]
[10]:
time_min | time_max | e_min | e_max | flux | flux_err |
---|---|---|---|---|---|
TeV | TeV | 1 / (cm2 s) | 1 / (cm2 s) | ||
float64 | float64 | float64 | float64 | float64 | float64 |
53343.92234009259 | 53343.94186555556 | 1.0 | 10.0 | 2.6538788850507653e-11 | 2.0923044822094324e-12 |
53343.95421509259 | 53343.97369425926 | 1.0 | 10.0 | 2.2999220249353757e-11 | 1.9916886558068632e-12 |
53345.96198129629 | 53345.98149518518 | 1.0 | 10.0 | 3.0354451809219976e-11 | 2.7380781806749725e-12 |
53347.913196574074 | 53347.93271046296 | 1.0 | 10.0 | 2.7716594082114026e-11 | 2.5951086629601995e-12 |
Running the light curve extraction in 1D¶
Building the 1D analysis configuration¶
[11]:
conf_1d = AnalysisConfig()
Definition of the data selection¶
Here we use the Crab runs from the HESS DL3 data release 1
[12]:
conf_1d.observations.obs_ids = [23523, 23526, 23559, 23592]
Definition of the dataset geometry¶
[13]:
# We want a 1D analysis
conf_1d.datasets.type = "1d"
# We want to extract the data by observation and therefore to not stack them
conf_1d.datasets.stack = False
# Here we define the ON region and make sure that PSF leakage is corrected
conf_1d.datasets.on_region = dict(
frame="icrs",
lon=83.63308 * u.deg,
lat=22.01450 * u.deg,
radius=0.1 * u.deg,
)
conf_1d.datasets.containment_correction = True
# Finally we define the energy binning for the spectra
conf_1d.datasets.geom.axes.energy = dict(
min=0.7 * u.TeV, max=10 * u.TeV, nbins=20
)
conf_1d.datasets.geom.axes.energy_true = dict(
min=0.3 * u.TeV, max=20 * u.TeV, nbins=40
)
Run the 1D data reduction¶
[14]:
analysis_1d = Analysis(conf_1d)
analysis_1d.get_observations()
analysis_1d.get_datasets()
Setting logging config: {'level': 'INFO', 'filename': None, 'filemode': None, 'format': None, 'datefmt': None}
Fetching observations.
Number of selected observations: 4
Reducing spectrum datasets.
Processing observation 23523
Processing observation 23526
Processing observation 23559
Processing observation 23592
Define the model to be used¶
Here we don’t try to fit the model parameters to the whole dataset, but we use predefined values instead.
[15]:
target_position = SkyCoord(ra=83.63308, dec=22.01450, unit="deg")
spatial_model = PointSpatialModel(
lon_0=target_position.ra, lat_0=target_position.dec, frame="icrs"
)
spectral_model = PowerLawSpectralModel(
index=2.702,
amplitude=4.712e-11 * u.Unit("1 / (cm2 s TeV)"),
reference=1 * u.TeV,
)
sky_model = SkyModel(
spatial_model=spatial_model, spectral_model=spectral_model, name="crab"
)
# Now we freeze these parameters that we don't want the light curve estimator to change
sky_model.parameters["index"].frozen = True
sky_model.parameters["lon_0"].frozen = True
sky_model.parameters["lat_0"].frozen = True
We assign the model to be fitted to each dataset. We can use the same gammapy.modeling.models.SkyModel
as before.
[16]:
models = SkyModels([sky_model])
analysis_1d.set_models(models)
Reading model.
SkyModels
Component 0: SkyModel
name value error unit min max frozen
--------- --------- ----- -------------- ---------- --------- ------
lon_0 8.363e+01 nan deg nan nan True
lat_0 2.201e+01 nan deg -9.000e+01 9.000e+01 True
index 2.702e+00 nan nan nan True
amplitude 4.712e-11 nan cm-2 s-1 TeV-1 nan nan False
reference 1.000e+00 nan TeV nan nan True
Extracting the light curve¶
[17]:
lc_maker_1d = LightCurveEstimator(
analysis_1d.datasets, source="crab", reoptimize=False
)
lc_1d = lc_maker_1d.run(e_ref=1 * u.TeV, e_min=1.0 * u.TeV, e_max=10.0 * u.TeV)
Compare results¶
Finally we compare the result for the 1D and 3D lightcurve in a single figure:
[18]:
ax = lc_1d.plot(marker="o", label="1D")
lc_3d.plot(ax=ax, marker="o", label="3D")
plt.legend()
[18]:
<matplotlib.legend.Legend at 0x11e0ad2b0>
Night-wise LC estimation¶
Here we want to extract a night curve per night. We define the time intervals that cover the three nights.
[19]:
time_intervals = [
Time([53343.5, 53344.5], format="mjd", scale="utc"),
Time([53345.5, 53346.5], format="mjd", scale="utc"),
Time([53347.5, 53348.5], format="mjd", scale="utc"),
]
To compute the LC on the time intervals defined above, we pass the LightCurveEstimator
the list of time intervals.
Internally, datasets are grouped per time interval and a flux extraction is performed for each group.
[20]:
lc_maker_1d = LightCurveEstimator(
analysis_1d.datasets,
time_intervals=time_intervals,
source="crab",
reoptimize=False,
)
nightwise_lc = lc_maker_1d.run(
e_ref=1 * u.TeV, e_min=1.0 * u.TeV, e_max=10.0 * u.TeV
)
nightwise_lc.plot()
[20]:
<matplotlib.axes._subplots.AxesSubplot at 0x11bdc0828>
[ ]: