SciPost Submission Page
Normalizing Flows for High-Dimensional Detector Simulations
by Florian Ernst, Luigi Favaro, Claudius Krause, Tilman Plehn, David Shih
Submission summary
Authors (as registered SciPost users): | Luigi Favaro · Claudius Krause · Tilman Plehn |
Submission information | |
---|---|
Preprint Link: | scipost_202312_00040v2 (pdf) |
Code repository: | https://github.com/heidelberg-hepml/CaloINN |
Data repository: | https://zenodo.org/records/14178546 |
Date submitted: | 2024-11-20 16:45 |
Submitted by: | Favaro, Luigi |
Submitted to: | SciPost Physics |
Ontological classification | |
---|---|
Academic field: | Physics |
Specialties: |
|
Abstract
Whenever invertible generative networks are needed for LHC physics, normalizing flows show excellent performance. In this work, we investigate their performance for fast calorimeter shower simulations with increasing phase space dimension. We use fast and expressive coupling spline transformations applied to the CaloChallenge datasets. In addition to the base flow architecture we also employ a VAE to compress the dimensionality and train a generative network in the latent space. We evaluate our networks on several metrics, including high-level features, classifiers, and generation timing. Our findings demonstrate that invertible neural networks have competitive performance when compared to autoregressive flows, while being substantially faster during generation.
Author indications on fulfilling journal expectations
- Provide a novel and synergetic link between different research areas.
- Open a new pathway in an existing or a new research direction, with clear potential for multi-pronged follow-up work
- Detail a groundbreaking theoretical/experimental/computational discovery
- Present a breakthrough on a previously-identified and long-standing research stumbling block
Author comments upon resubmission
We additionally publish the sample in a Zenodo repository with the full set of high-level observables studied.
In light of the extremely encouraging reports from the referees and the promising results, we believe our manuscript meets the criteria for the publication in Scipost Physics.
Main comments:
Regarding the main comments from the referees, section 4.5 now contains a comparison to similar normalizing flows and to a diffusion model in terms of shower generation time and accuracy.
We added a timing comparison and the classifier AUC for high- and low-level features.
We did not perform and in-depth ablation study because of resource constraints. We expanded the appendices with a discussion on the hyperparameters selection and the small grid search done.
List of changes
Report 1
See main comments.
Report 2
We adopted almost all the text improvements suggested by the referee. We kept the usage of "networks" when explicitly referring to a model defined by a neural network.
Abstract
We extended the abstract with the bottom line of our results.
Introduction
We added a paragraph on the effort done in ATLAS and CMS on fast detector simulations.
The claim is now supported by Sec.4.5 and the appendix on the CaloGAN dataset.
Dataset
We adopted all the suggested changes.
CaloINN
We expanded the text describing the INN, including the scaling aspect with input dimensionality. We also include a new schematic representation of the INN workflow.
We tagged the current state of the Github repo, which only contains refactorizations of the code at the time of submission.
VAE+INN
We added a VAE+INN flowchart and the motivation for a kernel-based encoding/decoding step.
Results
We clarified the first paragraph on photon showers.
We moved the inclusive set of high-level observables to the appendices. As suggested, we show and discuss three incident energies. The layers are chosen such that we discuss where there is the largest energy deposition.
We adopt the same scheme for the pion showers.
For dataset 2, we divided the high-level feature figures in the three E_inc windows (1-10), (10-100), (100-1000) GeV and selected only one layer to discuss the networks.
To avoid overcrowding the paper with figures, we have not done the same for dataset 3 but we included all the histograms in the published Zenodo repository.
We added a timing comparison in Table 3, the high-level classifier in Table 4, and the approximate training time on our cluster in the text.
We clarified the comparison between the INN and the VAE+INN throughout the text.
Conclusion
In the conclusions we expanded the discussion of the performance of our networks on all the studied metrics.
We comment on the observed differences between photon showers in dataset 1 and the electrons in dataset 2.
We give an outlook of what can be improved and directions for future works.
We updated all the references.
Report 3
VAE+INN
1-2 See main comment
3 We expanded the discussion on the selection of the beta parameter. The small value used ensures that the latent space is compact and therefore learnable by the latent INN.
4 We added citations to the beta-VAE and other latent models.
Current status:
Reports on this Submission
Report
The authors have addressed all my questions and comments satisfactorily. I find the manuscript suitable for publication in SciPost and recommend its acceptance.
Recommendation
Publish (meets expectations and criteria for this Journal)