Loading [MathJax]/extensions/Safe.js
SciPost logo

SciPost Submission Page

Systematically Constructing the Likelihood for Boosted $H\to gg$ Decays

by Andrew J. Larkoski

This is not the latest submitted version.

This Submission thread is now published as

Submission summary

Authors (as registered SciPost users): Andrew Larkoski
Submission information
Preprint Link: scipost_202503_00004v1  (pdf)
Date submitted: 2025-03-03 20:20
Submitted by: Larkoski, Andrew
Submitted to: SciPost Physics
Ontological classification
Academic field: Physics
Specialties:
  • High-Energy Physics - Phenomenology
Approaches: Theoretical, Phenomenological

Abstract

We study the binary discrimination problem of identification of boosted $H\to gg$ decays from massive QCD jets in a systematic expansion in the strong coupling. Though this decay mode of the Higgs is unlikely to be discovered at the LHC, we analytically demonstrate several features of the likelihood ratio for this problem through explicit analysis of signal and background matrix elements. Through leading-order, we prove that by imposing a constraint on the jet mass and measuring the energy fraction of the softer subjet an improvement of signal to background ratio that is independent of the kinematics of the jets at high boosts can be obtained, and is approximately equal to the inverse of the strong coupling evaluated at the Higgs mass. At next-to-leading order, we construct a powerful discrimination observable through a sort of anomaly detection approach by simply inverting the next-to-leading order $H\to gg$ matrix element with soft gluon emission, which is naturally infrared and collinear safe. Our analytic conclusions are validated in simulated data from all-purpose event generators and subsequent parton showering and demonstrate that the signal-to-background ratio can be improved by a factor of several hundred at high, but accessible, jet energies at the LHC.

Author indications on fulfilling journal expectations

  • Provide a novel and synergetic link between different research areas.
  • Open a new pathway in an existing or a new research direction, with clear potential for multi-pronged follow-up work
  • Detail a groundbreaking theoretical/experimental/computational discovery
  • Present a breakthrough on a previously-identified and long-standing research stumbling block

Author comments upon resubmission

I have addressed the comments by the referees.
Current status:
Has been resubmitted

Reports on this Submission

Report #2 by Anonymous (Referee 1) on 2025-3-15 (Invited Report)

Report

The authors has addressed the points raised in my first report and in the one of a second referee. As far as I am concerned, I am satisfied with the answers and with the modifications made, and I confirm my recommendation of acceptance of this paper.

Recommendation

Publish (easily meets expectations and criteria for this Journal; among top 50%)

  • validity: -
  • significance: -
  • originality: -
  • clarity: -
  • formatting: -
  • grammar: -

Report #1 by Anonymous (Referee 2) on 2025-3-12 (Invited Report)

  • Cite as: Anonymous, Report on arXiv:scipost_202503_00004v1, delivered 2025-03-12, doi: 10.21468/SciPost.Report.10819

Report

I thank the author for answering my questions and clarifying some points in the article. There is one point I would like to follow up on, please see my comment below.

Requested changes

On the previous point 9), the author's answer does not fully address my question. It's still not clear to me why the performance of the d2/z^2 observable is so much worse for low pT than for high pT. At a signal efficiency of about 0.2, the signal/background efficiency is about 1.1-1.2 for pt~500 GeV, but it is about 25 for pT~2000 GeV. Why is this variable not better than a random guess at low pT?

Connected to this, the author notes that Fig. 3 shows that d2/z^2 performs better than (1+O_NLO)/z and therefore only d2/z^2 is considered further. But Fig. 3 only shows the performance for pT>2 TeV. How does the comparison presented in Fig. 3 look at lower pT, or similarly, how does Fig. 4 look for (1+O_NLO)/z ? Can this information be added to the paper?

Recommendation

Ask for minor revision

  • validity: -
  • significance: -
  • originality: -
  • clarity: -
  • formatting: -
  • grammar: -

Author:  Andrew Larkoski  on 2025-03-24  [id 5309]

(in reply to Report 1 on 2025-03-12)

I thank the referee for their comments. At relatively low pT, around pT ~ 2m/R, the two hard prongs aren't necessarily completely captured in the jet. As such, the discrimination power of the softer subjet's energy fraction is a less powerful discriminant than at high pT. Further, if the hard prongs are less well-defined at lower pT, that has the consequence of making the soft dipole observable d_2 less efficient, as well. If the directions of the hard subjets are sculpted by the jet finding algorithm, then the angles of soft emissions from the hard subjets are smeared, and less correlated with the actual dipole that emitted them.

To address this, I have added the following paragraph to the end of section 5: "Note that at sufficiently low transverse momentum $p_\perp$, the discrimination power of $d_2/z^2$ is not much better than random guess. At lower $p_\perp$, especially close to the minimum at which both hard prongs from decay are captured in the jet, $p_{\perp,\min}\sim 2m_H/R$, the hard subjets may be significantly sculpted by the jet algorithm. Further, at smaller $p_T$, the range of the softer subjet's energy fraction $z$ decreases, and both effects reduce the efficacy of this observable for discrimination. Additionally, with sculpted subjets, their direction is smeared so that sensitivity to additional soft emissions through the observable $d_2$ is smeared, as $d_2$ is sensitive to the angles between soft emissions and the subjet directions. For these reasons, our simple perturbative calculations are most valid at the highest $p_T$ ranges."

Indeed, a comparison of the two NLO observables could be performed at lower pTs. However, with the perturbative calculations as a guide and as their validity decreases at lower pT, I felt that comparison of those distributions at lower pT was not necessary and somewhat distracting. That is, the goal of this paper is to use the perturbative analysis to understand the optimal observables where they are valid; i.e., at the highest pT. Then in Fig. 4, focusing on the observable d_2/z^2, this argument is validated by observation of significant reduction of discrimination power as pT decreases.

Login to report or comment