Deep Learning for Imaging Cherenkov Telescopes Data Analysis
GammaLearn is a collaborative project to apply deep learning to the analysis of low-level Imaging Atmospheric Cherenkov Telescopes such as CTA. It provides a framework to easily train and apply models from a configuration file.
Gammalearn's CI/CD pipelines produce containers that can be used to quickly get a working environment for development or production. The containers are uploaded to the gammalearn's container registry. A set of containers is available for each branch in the repository, and the containers are re-build for any commit that involves modifications made to the environment or dockerfiles files (anything in the docker
directory).
Gammalearn containers include the pytorch and cuda dependencies to do computation on nvidia gpus, in addition to a big stack of scientific computing software. They are typically take several GBs.
We recommend the use of apptainer. To get the production image of the version of gammalearn you want to use, for instance to get gammalearn v0.13.0
, use apptainer pull
:
apptainer pull docker://gitlab-registry.in2p3.fr/gammalearn/gammalearn:v0.13.0
This will create a .sif
container file that contains a ready to use gammalearn installation. Warning: apptainer
can use several GB of disk space as cache when building the .sif
file. By default, the cache is located in your home folder ~/.apptainer/cache
. You can change this location by setting the APPTAINER_CACHEDIR
environment variable. Clean apptainer
's cache with apptainer cache clean
You can now get a shell inside the container to test gammalearn:
apptainer shell path_to_your_sif_file.sif
Apptainer> /opt/conda/bin/gammalearn --help
You can run an experiment using apptainer run
. Since apptainer containers are read-only by default, you will need to mount the paths to your input and output files. To use nvidia gpus, you will need to specify the --nv
option as well. A typical command example:
# Run the experiment in the container
# Parameters:
# --nv to use nvidia gpus from inside the container
# CUDA_VISIBLE_DEVICES env variable used by pytorch to discover the gpus
# NUMBA_CACHE_DIR a writable directory where numba can store its compiled functions
# (needs to be outside of the container, which is read-only)
# CTAPIPE_CACHE ctapipe needs a writable place, to store its downloaded files.
# Mounts: input (data and settings file) and output directories
#
# We call the gammalearn entrypoint directly in /opt/conda/bin, because micromamba is not initialized
# inside the container for a new user (and every user is new, since with apptainer the user remains the same
# as the user on the host system by default (only users defined in the containers are known))
apptainer run \
--nv \
--env "CUDA_VISIBLE_DEVICES=$CUDA_VISIBLE_DEVICES" \
--env "NUMBA_CACHE_DIR=/tmp/NUMBA" \
--env "CTAPIPE_CACHE=/tmp/CTAPIPE" \
--mount type=bind,source=/path/to/input/data_dir/,destination=/corresponding/path/in/container/ \
--mount type=bind,source=/path/to/output/data_dir/,destination=/corresponding/path/in/container/ \
path_to_your_sif_file.sif /opt/conda/bin/gammalearn path_to_your_experiment_settings_file.py
You can find examples of setting file in the examples and some sample data in example data
Developers are encouraged to use the development containers available in the repository container registry. The development containers contain all gammalearn run-time dependencies, and several tools required to develop the software (for instance to run tests, build the documentation, visualize results, etc.), but it does NOT contain gammalearn nor it source code. The goal is to provide a containerized development environment, but leave the source code in the host system. It is recommended to use the development containers with docker
, either directly or through an IDE.
To pull the container of the branch you use to develop:
docker pull gitlab-registry.in2p3.fr/gammalearn/gammalearn/dev:your_branch_name
Clone gammalearn locally and go in the gammalearn directory:
git clone https://gitlab.in2p3.fr/gammalearn/gammalearn.git && cd gammalearn
Enter the container while mounting the source code, and install gammalearn
# - get a shell with interactive mode
# - mount gammalearn sources in "bind" mode with --mount
# - use your host user uid and gid (-u $(id -u):$(id -g)) to have write permissions in the mounted gammalearn sources
# - set working directory to /src
# - optionally: set pull policy to always to always get the latest image for your branch
docker run --rm -it --mount type=bind,source=.,destination=/src -w /src --pull=always -u $(id -u):$(id -g) gitlab-registry.in2p3.fr/gammalearn/gammalearn/dev:your_branch_name
(base) mambauser@...:/src$ pip install --no-deps -e .
# Develop !
Many IDE's offer the possibility to start or interact with running docker containers. In VScode, this is handled by the "dev container" extension included in the remote development extension pack (see the extension documentation). The extension can start a container and automatically install a vs-code server inside the container, allowing to transparently develop the software while using the environment from inside the container. The following devcontainer.json
configuration allows to start a development container with the right permissions and build gammalearn from the mounted sources. As examples, the python extension and ruff linter are installed in the vs-code server running in the container.
// For format details, see https://aka.ms/devcontainer.json. For config options, see the
// README at: https://github.com/devcontainers/templates/tree/main/src/miniconda
{
"image": "gitlab-registry.in2p3.fr/gammalearn/gammalearn/dev:your_dev_branch",
"workspaceMount": "source=${localWorkspaceFolder},target=/src,type=bind",
"workspaceFolder": "/src",
"postAttachCommand": "micromamba run -n base pip install --no-deps -e /src",
"runArgs": [ "--network=host"],
"customizations": {
// Configure properties specific to VS Code.
"vscode": {
// Add the IDs of extensions you want installed when the container is created.
"extensions": [
"charliermarsh.ruff",
"ms-python.python",
"ms-python.vscode-pylance",
"njpwerner.autodocstring",
"tamasfe.even-better-toml",
"wmaurer.change-case"
]
}
}
}
Note: By default, the dev containers extension will not re-pull the specified image unless you rebuild the container in vscode. Make sure to pull the latest container for your branch before entering the container.
Once the container is running, you can enter it from another external terminal with
# Get the container ID of your dev container started with vscode
docker ps
# Get a shell in the container
docker exec -it -w /src container_ID bash
Gammalearn dependencies are managed using Anaconda. The dependencies are split into run-time, test and development dependencies available in the docker directory. According to your usage, install the dependencies you need. For instance:
conda env create -n glearn -f dependencies.yml
conda env update -n glearn -f test_dependencies.yml
conda activate glearn
Note for OSX and/or no-gpu users: please edit the environment file to remove cudatoolkit
from the dependencies.
Create the gammalearn release environment:
VERSION=0.12.0
wget https://gitlab.in2p3.fr/gammalearn/gammalearn/-/raw/v${VERSION}/environment.yml -O glearn_${VERSION}_env.yml
conda env create -f glearn_${VERSION}_env.yml
conda activate glearn
Note for OSX and/or no-gpu users: please edit the environment file to remove cudatoolkit
from the dependencies.
Contributions are very much welcome: please see CONTRIBUTING.
Please cite
Jacquemont M, Vuillaume T, Benoit A, Maurin G, Lambert P, Lamanna G, Brill A. GammaLearn: A Deep Learning Framework for IACT Data. In36th International Cosmic Ray Conference (ICRC2019) 2019 Jul (Vol. 36, p. 705). DOI: https://doi.org/10.22323/1.358.0705
For reproducibility purposes, please also cite the exact version of GammaLearn you used by citing the corresponding DOI on Zenodo:
GammaLearn is distributed under an MIT license.
## Back to top
Deep Learning For Imaging Cherenkov Telescopes Data Analysis