a scalable and extensible image analysis framework
The BIOMERO framework, for BioImage analysis in OMERO, allows you to run (FAIR) bioimage analysis workflows directly from OMERO on a high-performance compute (HPC) cluster, remotely through SSH.
The BIOMERO framework consists of this Python library biomero
, together with the BIOMERO scripts that can be run directly from the OMERO web interface.
The package includes the SlurmClient
class, which provides SSH-based connectivity and interaction with a Slurm (high-performance compute) cluster. The package enables users to submit jobs, monitor job status, retrieve job output, and perform other Slurm-related tasks. Additionally, the package offers functionality for configuring and managing paths to Slurm data and Singularity images (think Docker containers...), as well as specific FAIR image analysis workflows and their associated repositories.
Overall, the biomero
package simplifies the integration of HPC functionality within the OMERO platform for admins and provides an efficient and end-user-friendly interface towards both the HPC and FAIR workflows.
WARNING: Please note that default settings are for short/medium jobs. If you run long workflows (>45min), you will run into 2 lethal issues:
In the figure below we show our BIOMERO framework, for BioImage analysis in OMERO.
BIOMERO consists of this Python library (biomero
) and the integrations within OMERO, currently through our BIOMERO scripts.
For a quick overview of what this library can do for you, we can install an example setup locally with Docker:
localhost
server (... we are working on a web importer ... TBC)slurm/init/SLURM Init environment...
image.sc
of this software, or here on the Discussions
tab of Github.slurm/workflows/SLURM Run Workflow...
Select how to import your results
, e.g. change Import into NEW Dataset
text to hello world
cellpose
.
nuc channel
to the channel to segment (note that 0 is for grey, so 1,2,3 for RGB)use gpu
(step 2, our HPC cluster, doesn't come with GPU support built into the containers)Explore
tab to see your hello world
dataset with a mask image when the workflow is done!Note: This library has only been tested on Slurm versions 21.08.6 and 22.05.09 !
Your Slurm cluster/login node needs to have:
Your OMERO processing node needs to have:
python3 -m pip install biomero
python3 -m pip install 'git+https://github.com/NL-BioImaging/biomero'
/etc/slurm-.ini
python3 -m pip install ezomero==1.1.1 tifffile==2020.9.3
and the OMERO CLI Zarr plugin.Your OMERO server node needs to have:
https://github.com/NL-BioImaging/biomero-scripts
/opt/omero/server/OMERO.server/lib/scripts/slurm/
, e.g. git clone https://github.com/NL-BioImaging/biomero-scripts.git <path>/slurm
!!NOTE: Do not install Example Minimal Slurm Script if you do not trust your users with your Slurm cluster. It has literal Command Injection for the SSH user as a FEATURE.
To connect an OMERO processor to a Slurm cluster using the biomero
library, users can follow these steps:
Setup passwordless public key authentication between your OMERO processor
server and your HPC server. E.g. follow a SSH tutorial or this one.
processor
servers, and share the same private key to all of them.Create a SSH config file named config
in the .ssh
directory of (all) the OMERO processor
servers, within the omero
user's home directory (~/.ssh/config
). This file should specify the hostname, username, port, and private key path for the Slurm cluster, under some alias. This alias we will provide to the library. We provide an example in the resources directory.
This will allow a uniform SSH naming, and makes the connection headless; making it easy for the library.
Test the SSH connection manually! ssh slurm
(as the omero user) should connect you to the Slurm server (given that you named it slurm
in the config
).
Congratulations! Now the servers are connected. Next, we make sure to setup the connection between OMERO and Slurm.
At this point, ensure that the slurm-config.ini
file is correctly configured with the necessary SSH and Slurm settings, including the host, data path, images path, and model details. Customize the configuration according to the specific Slurm cluster setup. We provide an example in the resources section. To read it automatically, place this ini
file in one of the following locations (on the OMERO processor
server):
/etc/slurm-config.ini
~/slurm-config.ini
Note: Make sure to place the slurm-config.ini
in the target folder at build time of your docker container instead of mounting it at runtime. This is because the library reads the config file at import time, and if it is not found, it will not work.
Install OMERO scripts from OMERO Slurm Scripts, e.g.
cd /opt/omero/server/OMERO.server/lib/scripts
git clone https://github.com/NL-BioImaging/biomero-scripts.git slurm
!!NOTE: Do not install Example Minimal Slurm Script if you do not trust your users with your Slurm cluster. It has literal Command Injection for the SSH user as a FEATURE.
Install BIOMERO Scripts requirements, e.g.
python3 -m pip install ezomero==1.1.1 tifffile==2020.9.3
python3 -m pip install omero-cli-zarr==0.5.3
&& yum install -y blosc-devel
unzip -d /opt bioformats2raw-0.7.0.zip && export PATH="$PATH:/opt/bioformats2raw-0.7.0/bin"
To finish setting up your SlurmClient
and Slurm server, run it once with init_slurm=True
. This is provided in a OMERO script form at init/Slurm Init environment , which you just installed in previous step.
slurm-config.ini
, pull any described Singularity images to the server (note: might take a while), and generate (or clone from Git) any job scripts for these workflows:with SlurmClient.from_config(configfile=configfile,
init_slurm=True) as slurmClient:
slurmClient.validate(validate_slurm_setup=True)
With the configuration files in place, you can utilize the SlurmClient
class from the biomero
library to connect to the Slurm cluster over SSH, enabling the submission and management of Slurm jobs from an OMERO processor.
The easiest interaction from OMERO with this library currently is through our BIOMERO scripts, which are just a set of OMERO scripts using this library for all the steps one needs to run a image analysis workflow from OMERO on Slurm and retrieve the results back into OMERO.
!!NOTE: Do not install Example Minimal Slurm Script if you do not trust your users with your Slurm cluster. It has literal Command Injection for the SSH user as a FEATURE.
We have provided the BIOMERO scripts at https://github.com/NL-BioImaging/biomero-scripts (hopefully installed in a previous step).
For example, workflows/Slurm Run Workflow should provide an easy way to send data to Slurm, run the configured and chosen workflow, poll Slurm until jobs are done (or errors) and retrieve the results when the job is done. This workflow script uses some of the other scripts, like
data/Slurm Image Transfer
: to export your selected images / dataset / screen as TIFF files to a Slurm dir.data/Slurm Get Results
: to import your Slurm job results back into OMERO as a zip, dataset or attachment.Other example OMERO scripts are:
data/Slurm Get Update
: to run while you are waiting on a job to finish on Slurm; it will try to get a %
progress from your job's logfile. Depends on your job/workflow logging a %
of course.
workflows/Slurm Run Workflow Batched
: This will allow you to run several workflows/Slurm Run Workflow
in parallel, by batching your input images into smaller chunks (e.g. turn 64 images into 2 batches of 32 images each). It will then poll all these jobs.
workflows/Slurm CellPose Segmentation
: This is a more primitive script that only runs the actual workflow CellPose
(if correctly configured). You will need to manually transfer data first (with Slurm Image Transfer
) and manually retrieve data afterward (with Slurm Get Results
).
You are encouraged to create your own custom scripts. Do note the copy-left license enforced by OME.
We host BIOMERO container dockerfiles at NL-BIOMERO, which publishes container images to our public dockerhub cellularimagingcf. Specifically the cellularimagingcf/biomero image is an OMERO processor container with BIOMERO library installed. When we release a new version of BIOMERO, we will also release a new version of these containers (because we deploy these locally at our Core Facility - Cellular Imaging).
You can mount your specific configurations over those in the container, for example:
# Run the biomero container
echo "Starting BIOMERO..."
podman run -d --rm --name biomero \
-e CONFIG_omero_master_host=omeroserver \
-e OMERO_WORKER_NAME=biomero \
-e CONFIG_omero_logging_level=10 \
--network omero \
--volume /mnt/datadisk/omero:/OMERO \
--volume /mnt/data:/data \
--volume /my/slurm-config.ini:/etc/slurm-config.ini \
--secret ssh-config,target=/tmp/.ssh/config --secret ssh-key,target=/tmp/.ssh/id_rsa --secret ssh-pubkey,target=/tmp/.ssh/id_rsa.pub --secret ssh-known_hosts,target=/tmp/.ssh/known_hosts \
--userns=keep-id:uid=1000,gid=997 \
cellularimagingcf/biomero:0.2.3
This will spin up the docker container (in Podman) with omero config (-e CONFIG_omero_..
), mounting the required data drives (--volume /mnt/...
) and adding a new slurm config (--volume /my/slurm-config.ini:/etc/slurm-config.ini
) and the required SSH settings (--secret ...,target=/tmp/.ssh/...
) to access the remote HPC.
Note: the BIOMERO scripts are installed on the main server, not on the BIOMERO processor.
Note2: We will also update these containers with our own desired changes, so they will likely not be 1:1 copy with basic omero containers. Especially when we start making a nicer UI for BIOMERO. We will keep up-to-date with the OMERO releases when possible.
I have also provided tutorials on connecting to a Local or Cloud Slurm, and tutorials on how to add your FAIR workflows to this setup. Those can give some more insights as well.
Note: this library is built for SSH-based connections. If you could, it would be a lot easier to just have the OMERO processor
server and the slurm
client server be (on) the same machine: then you can just directly call sbatch
and other slurm
commands from OMERO scripts and Slurm would have better access to your data.
This is mainly for those cases where you already have an external HPC cluster and want to connect your OMERO instance.
Theoretically, you could extend the SlurmClient
class and change the run
commands to not use SSH, but just a subprocess
. We might implement this if we need it in the future.
But then you could also look at other Python libraries like submitit.
The SlurmClient class is the main entrypoint in using this library. It is a Python class that extends the Connection class from the Fabric library. It allows connecting to and interacting with a Slurm cluster over SSH.
It includes attributes for specifying paths to directories for Slurm data and Singularity images, as well as specific paths, repositories, and Dockerhub information for different Singularity image models.
The class provides methods for running commands on the remote Slurm host, submitting jobs, checking job status, retrieving job output, and tailing log files.
It also offers a from_config
class method to create a SlurmClient
object by reading configuration parameters from a file. Overall, the class provides a convenient way to work with Slurm clusters and manage job execution and monitoring.
The slurm-config.ini
file is a configuration file used by the biomero
Python package to specify various settings related to SSH and Slurm. Here is a brief description of its contents:
[SSH]: This section contains SSH settings, including the alias for the SLURM SSH connection (host). Additional SSH configuration can be specified in the user's SSH config file or in /etc/fabric.yml
.
[SLURM]: This section includes settings specific to Slurm. It defines the paths on the SLURM entrypoint for storing data files (slurm_data_path), container image files (slurm_images_path), and Slurm job scripts (slurm_script_path). It also specifies the repository (slurm_script_repo) from which to pull the Slurm scripts.
[MODELS]: This section is used to define different model settings. Each model has a unique key and requires corresponding values for <key>_repo
(repository containing the descriptor.json file, which will describe parameters and where to find the image), and <key>_job
(jobscript name and location in the slurm_script_repo
). The example shows settings for several segmentation models, including Cellpose, Stardist, CellProfiler, DeepCell, and ImageJ.
Note also that you can override the default Slurm job values using this model configuration, like memory, GPU, time limit, etc. All values for sbatch can be applied (see e.g. here) and will be forwarded to the job command.
For example
# Run CellPose Slurm with 10 GB GPU
cellpose_job_gres=gpu:1g.10gb:1
# Run CellPose Slurm with 15 GB CPU memory
cellpose_job_mem=15GB
The slurm-config.ini
file allows users to configure paths, repositories, and other settings specific to their Slurm cluster and the biomero
package, providing flexibility and customization options.
An important Slurm job config is the time limit: SBATCH --time=00:45:00
is the default in BIOMERO (max 45 minutes per job).
The format is d-hh:mm:ss
WARNING: After this time, the job will timeout and this scenario is not handled by BIOMERO (yet)! You will lose your processing progress.
You can change this timeout value:
#SBATCH --time=08:00:00
for 8 hours)cellpose_job_time=08:00:00
for 8 hours)Note that it might take longer for Slurm to schedule your job if you put the time very high, or possibly even make it wait indefinitely (see --time explanation in https://slurm.schedmd.com/sbatch.html). We will work on smart timing, but for now it is hardcoded and configurable.
To add an existing (containerized) workflow, add it to the slurm-config.ini
file like in our example:
# -------------------------------------
# CELLPOSE SEGMENTATION
# -------------------------------------
# The path to store the container on the slurm_images_path
cellpose=cellpose
# The (e.g. github) repository with the descriptor.json file
cellpose_repo=https://github.com/TorecLuik/W_NucleiSegmentation-Cellpose/tree/v1.2.7
# The jobscript in the 'slurm_script_repo'
cellpose_job=jobs/cellpose.sh
Here,
cellpose
<slurm_images_path>/cellpose
https://github.com/TorecLuik/W_NucleiSegmentation-Cellpose
v1.2.7
<slurm_script_repo>/jobs/cellpose.sh
.
A lot of the automation in this library is based on metadata of the workflow, provided in the source code of the workflow, specifically the descriptor.json.
For example, the OMERO script UI can be generated automatically, based on this descriptor. And also, the Slurm job script can be generated automatically, based on this descriptor.
This metadata scheme is (based on) Cytomine / BIAFLOWS, and you can find details of it and how to create one yourself on their website, e.g. this Cytomine dev-guide or this BIAFLOWS dev-guide.
NOTE! We do not require the cytomine_<...>
authentication parameters. They are not mandatory. In fact, we ignore them. But it might be beneficial to make your workflow compatible with Cytomine as well.
At this point, we are using the cytomine-0.1
schema, in the future we will also want to support other schemas, like Boutiques, commonwl or MLFlow.
We will try to stay compatible with all such schemas (perhaps with less functionality because of missing metadata).
At this point, we do not strictly validate the schema, we just read expected fields from the descriptor.json
.
Note that while it is possible to have multiple versions of the same workflow on Slurm (and select the desired one in OMERO), it is not possible to configure this yet. We assume for now you only want one version to start with. You can always update this config to download a new version to Slurm.
Unless you change the Slurm
job, the input is expected to be:
infolder
parameter
gtfolder
parameter (Optional)
ground-truth
input files, generally not needed for prediction / processing purposes.outfolder
parameter
Note that you can also use the wrapper.py setup from BIAFLOWS to handle the I/O for you:
with BiaflowsJob.from_cli(argv) as bj:
# Change following to the actual problem class of the workflow
...
# 1. Prepare data for workflow
in_imgs, gt_imgs, in_path, gt_path, out_path, tmp_path = prepare_data(problem_cls, bj, is_2d=True, **bj.flags)
# 2. Run image analysis workflow
bj.job.update(progress=25, statusComment="Launching workflow...")
# Add here the code for running the analysis script
# 3. Upload data to BIAFLOWS
...
# 4. Compute and upload metrics
...
# 5. Pipeline finished
...
This wrapper handles the input parameters for you, providing the input images as in_imgs
, et cetera. Then you add your commandline call between point 2 and 3, and possibly some preprocessing between point 1 and 2:
#add here the code for running the analysis script
For example, from Cellpose container workflow:
...
# 2. Run image analysis workflow
bj.job.update(progress=25, statusComment="Launching workflow...")
# Add here the code for running the analysis script
prob_thresh = bj.parameters.prob_threshold
diameter = bj.parameters.diameter
cp_model = bj.parameters.cp_model
use_gpu = bj.parameters.use_gpu
print(f"Chosen model: {cp_model} | Channel {nuc_channel} | Diameter {diameter} | Cell prob threshold {prob_thresh} | GPU {use_gpu}")
cmd = ["python", "-m", "cellpose", "--dir", tmp_path, "--pretrained_model", f"{cp_model}", "--save_tif", "--no_npy", "--chan", "{:d}".format(nuc_channel), "--diameter", "{:f}".format(diameter), "--cellprob_threshold", "{:f}".format(prob_thresh)]
if use_gpu:
print("Using GPU!")
cmd.append("--use_gpu")
status = subprocess.run(cmd)
if status.returncode != 0:
print("Running Cellpose failed, terminate")
sys.exit(1)
# Crop to original shape
for bimg in in_imgs:
shape = resized.get(bimg.filename, None)
if shape:
img = imageio.imread(os.path.join(tmp_path,bimg.filename_no_extension+"_cp_masks.tif"))
img = img[0:shape[0], 0:shape[1]]
imageio.imwrite(os.path.join(out_path,bimg.filename), img)
else:
shutil.copy(os.path.join(tmp_path,bimg.filename_no_extension+"_cp_masks.tif"), os.path.join(out_path,bimg.filename))
# 3. Upload data to BIAFLOWS
We get the commandline parameters from bj.parameters
(biaflows job) and provide that the cmd
commandline string. Then we run it with subprocess.run(cmd)
and check the status
.
We use a tmp_path
to store both input and output, then move the output to the out_path
after the processing is done.
Also note that some preprocessing is done in step 1:
# Make sure all images have at least 224x224 dimensions
# and that minshape / maxshape * minshape >= 224
# 0 = Grayscale (if input RGB, convert to grayscale)
# 1,2,3 = rgb channel
nuc_channel = bj.parameters.nuc_channel
resized = {}
for bfimg in in_imgs:
...
imageio.imwrite(os.path.join(tmp_path, bfimg.filename), img)
Another example is this imageJ
wrapper:
...
# 3. Call the image analysis workflow using the run script
nj.job.update(progress=25, statusComment="Launching workflow...")
command = "/usr/bin/xvfb-run java -Xmx6000m -cp /fiji/jars/ij.jar ij.ImageJ --headless --console " \
"-macro macro.ijm \"input={}, output={}, radius={}, min_threshold={}\"".format(in_path, out_path, nj.parameters.ij_radius, nj.parameters.ij_min_threshold)
return_code = call(command, shell=True, cwd="/fiji") # waits for the subprocess to return
if return_code != 0:
err_desc = "Failed to execute the ImageJ macro (return code: {})".format(return_code)
nj.job.update(progress=50, statusComment=err_desc)
raise ValueError(err_desc)
Once again, just a commandline --headless
call to ImageJ
, wrapped in this Python script and this container.
Building workflows like this will make them more FAIR (also for software) and uses best practices like code versioning and containerization!
Also take a look at our in-depth tutorial on adding a Cellprofiler pipeline as a workflow to BIOMERO.
Here is a shorter version: Say you have a script in Python and you want to make it available on OMERO and Slurm.
These are the steps required:
descriptor.json
(see previous chapter). E.g. like this.requirements.txt
, like this. Make sure to pin your versions for future reproducability!latest
. This improves reproducability!latest
if the code repository is given no version, but the master
branch.slurm-config.ini
SlurmClient.from_config(init_slurm=True)
(e.g. the init environment script.)By default, biomero
will generate basic slurm jobs for each workflow, based on the metadata provided in descriptor.json
and a job template.
It will replace $PARAMS
with the (non-cytomine_
) parameters given in descriptor.json
. See also the Parameters section below.
You could change the job template and generate new jobs, by running SlurmClient.from_config(init_slurm=True)
(or slurmClient.update_slurm_scripts(generate_jobs=True)
)
Or you could add your jobs to a Github repository and reference this in slurm-config.ini
, both in the field slurm_script_repo
and every <workflow>_job
:
# -------------------------------------
# REPOSITORIES
# -------------------------------------
# A (github) repository to pull the slurm scripts from.
#
# Note:
# If you provide no repository, we will generate scripts instead!
# Based on the job_template and the descriptor.json
#
slurm_script_repo=https://github.com/TorecLuik/slurm-scripts
[MODELS]
# -------------------------------------
# Model settings
# -------------------------------------
# ...
# -------------------------------------
# CELLPOSE SEGMENTATION
# -------------------------------------
# The path to store the container on the slurm_images_path
cellpose=cellpose
# The (e.g. github) repository with the descriptor.json file
cellpose_repo=https://github.com/TorecLuik/W_NucleiSegmentation-Cellpose/tree/v1.2.7
# The jobscript in the 'slurm_script_repo'
cellpose_job=jobs/cellpose.sh
You can update the jobs by calling slurmClient.update_slurm_scripts()
, which will pull the repository('s default branch).
This might be useful, for example if you have other hardware requirements for your workflow(s) than the default job asks for, or if you want to run more than just 1 singularity container.
The library will provide the parameters from your descriptor.json
as environment variables to the call. E.g. set DIAMETER=0; sbatch ...
.
Other environment variables provided are:
DATA_PATH
<slurm_data_path>/<input_folder>
. The base dir for data folders for this execution. We expect it to contain /data/in
, /data/in
and /data/in
folders in our template and data transfer setup.IMAGE_PATH
<slurm_images_path>/<model_path>
, as described in slurm-config.ini
IMAGE_VERSION
SINGULARITY_IMAGE
IMAGE_VERSION
above, as <container_name>_<IMAGE_VERSION>.sif
We (potentially) override the following Slurm job settings programmatically:
--mail-user={email}
(optional)--time={time}
(optional)--output=omero-%4j.log
(mandatory)We could add more overrides in the future, and perhaps make them available as global configuration variables in slurm-config.ini
.
We can simply use Slurm
for running your workflow 1:1, so 1 job to 1 workflow. This could speed up your workflow already, as Slurm
servers are likely equipped with strong CPU and GPU.
However, Slurm
is also built for parallel processing on multiple (or the same) servers. We can accomplish this by running multiple jobs for 1 workflow. This is simple for embarrassingly parallel tasks, like segmenting multiple images: just provide each job with a different set of input images. If you have 100 images, you could run 10 jobs on 10 images and (given enough resources available for you on Slurm) that could be 10x faster. In theory, you could run 1 job per image, but at some point you run into the overhead cost of Slurm (and OMERO) and it might actually slow down again (as you incur this cost a 100 times instead of 10 times).
Note, the default Slurm job script will not request any GPU resources.
This is because GPU resources are expensive and some programs do not work with GPU.
We can instead enable the use of GPU by either providing our own Slurm job scripts, or setting an override value in slurm-config.ini
:
# -------------------------------------
# CELLPOSE SEGMENTATION
# -------------------------------------
# The path to store the container on the slurm_images_path
cellpose=cellpose
# The (e.g. github) repository with the descriptor.json file
cellpose_repo=https://github.com/TorecLuik/W_NucleiSegmentation-Cellpose/tree/v1.2.7
# The jobscript in the 'slurm_script_repo'
cellpose_job=jobs/cellpose.sh
# Override the default job values for this workflow
# Or add a job value to this workflow
# If you don't want to override, comment out / delete the line.
# Run CellPose Slurm with 10 GB GPU
cellpose_job_gres=gpu:1g.10gb:1
In fact, any ..._job_...=...
configuration value will be forwarded to the Slurm commandline.
Slurm commandline parameters override those in the script, so the above one requests 1 10GB gpu for Cellpose.
E.g. you could also set the time limit higher:
# -------------------------------------
# CELLPOSE SEGMENTATION
# -------------------------------------
# The path to store the container on the slurm_images_path
cellpose=cellpose
# The (e.g. github) repository with the descriptor.json file
cellpose_repo=https://github.com/TorecLuik/W_NucleiSegmentation-Cellpose/tree/v1.2.7
# The jobscript in the 'slurm_script_repo'
cellpose_job=jobs/cellpose.sh
# Override the default job values for this workflow
# Or add a job value to this workflow
# If you don't want to override, comment out / delete the line.
# Run with longer time limit
cellpose_job_time=00:30:00
Now the CellPose job should run for maximum of 30 minutes, instead of the default.
We have added methods to this library to help with transferring data to the Slurm
cluster, using the same SSH connection (via SCP or SFTP).
slurmClient.transfer_data(...)
slurmClient.unpack_data(...)
slurmClient.zip_data_on_slurm_server(...)
slurmClient.copy_zip_locally(...)
slurmClient.get_logfile_from_slurm(...)
And more; see the docstring of SlurmClient
and example OMERO scripts.
You can test the library by installing the extra test dependencies:
Create a venv to isolate the python install:
python -m venv venvTest
Install OSC with test dependencies:
venvTest/Scripts/python -m pip install .[test]
Run pytest from this venv:
venvTest/Scripts/pytest
Debug logging can be enabled with the standard python logging module, for example with logging.basicConfig():
import logging
logging.basicConfig(level='DEBUG')
For example in (the __init__
of) a script:
if __name__ == '__main__':
logging.basicConfig(level=logging.INFO,
format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',
stream=sys.stdout)
runScript()