Events
30
min read

Reproducible Particle Physics Simulation Workflows on DECTRIS CLOUD

January 20, 2026
Share this post
Reproducible Particle Physics Simulation Workflows on DECTRIS CLOUD
Camilla Buhl Larsen
Camilla Buhl Larsen
Scientific Solution Architect

Across scientific disciplines, research workflows are becoming increasingly complex. Researchers must manage intricate software stacks, large datasets, and growing compute requirements, while still ensuring their work remains reproducible and easy to share.

DECTRIS CLOUD was built to support this reality. By combining data management, reproducible environments, and scalable compute in a single platform, it enables scientists from different domains to work efficiently while still supporting highly specialized workflows.

In this post, we highlight a particle physics use case from Philip Ploner, a student at ETH Zürich, who has used DECTRIS CLOUD to run large-scale simulation workflows for high-energy physics research. See his presentation from a recent DECTRIS CLOUD power user meeting below:

or follow the link here for a full recording, including a general introduction to DECTRIS CLOUD.

Particle Physics Simulation Workflows for Machine Learning Input

Experiments at CERN’s Large Hadron Collider generate enormous volumes of data, with hundreds of millions of proton–proton collisions occurring every second. Because only a tiny fraction of these events can be stored long term, experiments rely on triggering systems that decide, in real time, which data to keep and which to discard.

Simulation workflows play a critical role in this process. Large-scale simulations are used to generate labeled collision events that can be analyzed and used to train machine learning models. These models help make more refined trigger decisions, enabling researchers to identify subtle signatures that may be relevant for new physics.

Such simulation pipelines are complex and typically rely on multiple software tools, including MadGraph [1], Pythia [2], and Delphes [3], which are used together in Philip’s workflow. These tools must work in carefully configured environments, and ensuring that such workflows remain reproducible and easy to share across teams is a significant challenge, particularly as datasets and compute requirements continue to grow.

Reproducibility with Shared Environments

In DECTRIS CLOUD, Philip addresses this challenge by using containerized software environments to define an isolated and reproducible software stack for his simulations. An environment captures all required tools and dependencies and can be reused consistently across sessions and jobs.

Once created, environments can be shared with collaborators and are versioned automatically. This ensures that simulations can be rerun at any time using the exact same setup, removing common issues related to local installations or incompatible software versions.

Image: Particle physics simulation software was first configured inside a DECTRIS CLOUD virtual machine (session). A versioned snapshot of the virtual machine was saved as a shareable software container (environment), which could then be used as a basis for starting new pre-configured sessions as well as for running simulation jobs. 

From Code to Reusable Job Templates

With the environment in place, Philip created a job template to make the simulation workflow easy to run and reuse. Job templates expose key parameters, such as the number of simulated events, the physics process, and random seed, directly in the DECTRIS CLOUD interface.

This allows simulations to be launched with just a few clicks on scalable compute resources. Results, including validation plots and output datasets, are returned directly to the platform and remain fully reproducible, since the software environment and input data are preserved.

With the simulation results generated in DECTRIS CLOUD, Philip will continue his work with developing machine learning methods for improved triggering decisions while working under the supervision of Dr. Thea Aarrestad at ETH Zürich.

Image: Alongside the main event.parquet output used for machine-learning training, the DECTRIS CLOUD particle physics job template also generates validation plots. In the figure above, the jet multiplicity distribution compares generator-level, offline, and trigger-level jets and illustrates how PUPPI pileup mitigation reduces event complexity.

References

[1] Alwall, J., Frederix, R., Frixione, S., Hirschi, V., Maltoni, F., Mattelaer, O., Shao, H.S., Stelzer, T., Torrielli, P. and Zaro, M., 2014. The automated computation of tree-level and next-to-leading order differential cross sections, and their matching to parton shower simulations. Journal of High Energy Physics, 2014(7), pp.1-157.

[2] Bierlich, Christian, et al. "A comprehensive guide to the physics and usage of PYTHIA 8.3." SciPost Physics Codebases (2022): 008, arXiv:2203.11601 [hep-ph]

[3] De Favereau, J., Delaere, C., Demin, P., Giammanco, A., Lemaitre, V., Mertens, A. and Selvaggi, M., 2014. DELPHES 3: a modular framework for fast simulation of a generic collider experiment. Journal of High Energy Physics, 2014(2), pp.1-26.

Scientists
Tutorials
Share this post
Subscribe to newsletter

Subscribe to receive the latest news about features and releases.

By subscribing you agree to with our Privacy Policy.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.