1 Introduction

The measurement of differential diphoton production cross sections offers an important test of both perturbative and non-perturbative quantum chromodynamics (QCD). At leading order (LO), diphotons are produced via quark–antiquark annihilation \(\mathrm{q}\) \(\overline{\mathrm{q}}\) \(\rightarrow \) \(\gamma \) \(\gamma \). At next-to-leading order (NLO), diphoton production also includes the quark–gluon channel, while next-to-next-to-leading order (NNLO) adds the gluon–gluon channel, which includes a box diagram and represents a non-negligible fraction of the total cross section. Diphoton production is sensitive to the emission of soft gluons in the initial state and to the non-perturbative fragmentation of quarks and gluons to photons in the final state. Due to this rich phenomenology, theoretical predictions are challenging especially in restricted regions of phase space.

Diphoton production constitutes the major source of background in the diphoton decay channel of the newly discovered Higgs boson [13], as well as to searches for physics beyond the standard model. New physics processes may also appear as non-resonant deviations from the predicted diphoton spectrum in events with large missing transverse energy, as in gauge-mediated SUSY breaking [4] or in models of universal extra dimensions [5]. Alternatively, some models predict narrow resonances, such as the graviton in the Randall–Sundrum model for warped extra dimensions [6, 7].

The most recent diphoton measurements were published by the CDF and D0 Collaborations [8, 9] at the Tevatron and by the ATLAS Collaboration [10] at the LHC. This paper presents an update of a previous CMS measurement at \(\sqrt{s}=7\,\text {TeV} \) [11] and is based on the full 2011 data sample of 5.0\(\,\text {fb}^{-1}\). It probes a phase space defined by a highly asymmetric selection for the transverse energy (\(E_{\mathrm {T}}\)) of the two photons. The leading-order kinematic configuration where photons are produced back-to-back in the transverse plane is suppressed, enhancing the sensitivity to higher-order diagrams. The ratio of the NNLO to the LO prediction is increased by 20 % with respect to the previous CMS measurement [11].

The main experimental challenge for the measurement of the diphoton cross section is distinguishing the “prompt” photon signal produced either directly or as a result of fragmentation from the background that arises mainly from energetic neutral mesons, predominantly \(\mathrm {\pi ^0}\) and \(\mathrm {\eta }\) mesons, inside jets. These mesons typically decay to two collimated photons that are reconstructed as a single photon candidate, which is referred to as “non-prompt” in this paper. The main features used to discriminate a prompt photon from a non-prompt one are the shape of the shower measured by the electromagnetic calorimeter (ECAL) [12] and the isolation energy in a cone around the photon direction [11, 13]. This information can be used to trigger on diphoton candidate events and, at the analysis level, to statistically evaluate the fraction of prompt diphoton candidates.

The particle flow (PF) event reconstruction [14] consists in reconstructing and identifying each particle with an optimal combination of all sub-detector information. In this process, the identification of the particle type (photon, electron, muon, charged hadron, neutral hadron) plays an important role in the determination of the particle direction and energy.

In this analysis, the photon component of the PF isolation is used as the discriminating variable. The distributions of signal and background components are built from data and used in a maximum likelihood fit to estimate the signal fraction. An increased separation power with respect to previous results [11] is achieved by improving the identification and subtraction of the photon energy deposit in the isolation cone.

After a brief description of the CMS detector in Sect. 2, data and simulated samples are presented in Sect. 3, and the photon reconstruction in Sect. 4. The diphoton signal is estimated as described in Sect. 5. The number of signal events is then corrected for inefficiencies and unfolded as described in Sect. 6. Systematic uncertainties are assessed in Sect. 7, and the differential cross sections are presented and compared to theoretical predictions in Sect. 8.

2 The CMS detector

A detailed description of the CMS detector can be found elsewhere [15]. Its central feature is a superconducting solenoid, 13\(\text {\,m}\) in length and 6\(\text {\,m}\) in diameter, which provides an axial magnetic field of 3.8\(\text {\,T}\). The bore of the solenoid is instrumented with both the tracker (TRK) and the calorimeters. The steel flux-return yoke outside the solenoid is instrumented with gas-ionisation detectors used to reconstruct and identify muons. Charged-particle trajectories are measured by the silicon pixel and strip tracker, with full azimuthal (\(\phi \)) coverage within \(|\eta | < 2.5\), where the pseudorapidity \(\eta \) is defined as \(\eta = -\ln [\tan (\theta /2)]\), with \(\theta \) being the polar angle of the trajectory of the particle with respect to the counterclockwise beam direction. A lead tungstate crystal electromagnetic calorimeter (ECAL) and a brass/scintillator hadron calorimeter (HCAL) surround the tracking volume and cover the region \(|\eta | < 3\). The ECAL barrel (EB) extends to \(|\eta | < 1.479\) while the ECAL endcaps (EE) cover the region \(1.479 < |\eta | < 3.0\). A lead/silicon-strip preshower detector (ES) is located in front of the ECAL endcap in the region \(1.653 < |\eta | < 2.6\). The preshower detector includes two planes of silicon sensors measuring the \(x\) and \(y\) coordinates of the impinging particles. In the \((\eta , \phi )\) plane, and for \(|\eta | < 1.48\), the HCAL cells map onto \(5\times 5\) ECAL crystal arrays to form calorimeter towers projecting radially outwards from points slightly offset from the nominal interaction point. In the endcap, the ECAL arrays matching the HCAL cells contain fewer crystals. A steel/quartz-fibre Cherenkov forward calorimeter extends the calorimetric coverage to \(|\eta | < 5.0\).

3 Data sample

The data sample consists of proton–proton (\(\mathrm {p}\) \(\mathrm {p}\)) collision events collected at the LHC with the CMS detector in the year 2011, at a centre-of-mass energy (\(\sqrt{s}\)) of 7\(\,\text {TeV}\) and corresponding to an integrated luminosity of 5.0\(\,\text {fb}^{-1}\).

Events are triggered [15] by requiring the presence of two photons with asymmetric transverse energy thresholds. The \(E_{\mathrm {T}}\) thresholds at trigger level are 26 (18) and 36 (22)\(\,\text {GeV}\) on the leading (sub-leading) photon, depending on the running period. Each candidate is required to satisfy either loose calorimetric identification requirements, based on the shape of the electromagnetic shower, or loose isolation conditions. The trigger efficiency is evaluated using a tag-and-probe technique on \(\mathrm{Z} \rightarrow \mathrm {e}^+\mathrm {e}^-\) events [16], with electrons treated as photons. The trigger efficiency for photons selected in this analysis is measured to be between 98.8 and 100 % depending on the pseudorapidity and the interaction with the material in front of the ECAL. The total trigger efficiency is found to be constant over the data taking period.

Several samples of simulated events are used in the analysis to model signal and background processes. Drell–Yan+jets and \(\gamma \) \(\gamma \)+jets signal events are generated with MadGraph 1.4.8 [17]. The \(\mathrm{g}\) \(\mathrm{g}\) \(\rightarrow \) \(\gamma \) \(\gamma \) box signal process, \(\gamma \)+jet, and QCD dijet background processes are generated with pythia 6.4.24 [18]. For all simulated samples the CTEQ6L1 [19] parton distribution functions (PDFs) are used. All generated events are then processed with pythia (Z2 tune) [20] for hadronization, showering of partons and the underlying event; a detailed simulation of the CMS detector based on Geant4 [21] is performed, and the simulated events are finally reconstructed using the same algorithms as used for the data.

The simulation includes the effects of in-time pileup (overlapping \(\mathrm {p}\) \(\mathrm {p}\) interactions within a bunch crossing) and out-of-time pileup (overlapping \(\mathrm {p}\) \(\mathrm {p}\) interactions from interactions happening in earlier and later bunch crossings) with a distribution matching that observed in data.

4 Photon reconstruction and selection

4.1 Photon reconstruction

Photon candidates are reconstructed from the energy deposits in the ECAL by grouping its channels into superclusters [22]. About half of the photons convert into an \(\mathrm {e}^+\) \(\mathrm {e}^-\) pair in the material in front of the ECAL. Conversion-track pairs are reconstructed from a combination of Gaussian-sum filter (GSF) electron tracks [23] and ECAL-seeded tracks fit to a common vertex and then matched to the photon candidate. The superclustering algorithms achieve an almost complete collection of the energy of such converted photons. In the barrel region, superclusters are formed from five-crystal-wide strips in \(\eta \), centred on the locally most energetic crystal (seed), and have a variable extension in the azimuthal direction (\(\phi \)). In the endcaps, where the crystals are arranged according to an \(x\)\(y\) rather than an \(\eta \)\(\phi \) geometry, matrices of \(5\times 5\) crystals (which may partially overlap) around the most energetic crystals are merged if they lie within a narrow \(\phi \) road. The photon candidates are reconstructed within the ECAL fiducial region \(|\eta |<2.5\) but excluding the barrel-endcap transition regions \(1.44 < |\eta | < 1.57\). This exclusion of the barrel-endcap transition regions ensures containment of the shower of the selected photon candidate in either the ECAL barrel or one of the ECAL endcaps. The fiducial region requirement is applied to the supercluster position (defined as the log-weighted barycentre of the supercluster’s active channels) in the ECAL.

The photon energy is computed starting from the raw crystal energies measured in the ECAL. In the region covered by the preshower detector the energy recorded in that sub-detector is added. The variation of the crystal transparency during the run is continuously monitored and corrected using a factor based on the change in response to light from a laser and light-emitting-diode based monitoring system. The single-channel response of the ECAL is equalised by exploiting the \(\phi \) symmetry of the energy flow, the mass constraint on the energy of the two photons in decays of \(\mathrm {\pi ^0}\) and \(\mathrm {\eta }\) mesons, and the momentum constraint on the energy of isolated electrons from \(\mathrm {W}\) and \(\mathrm{Z}\) decays. A correction factor compensates for the imperfect containment of the shower in the cluster crystals. The absolute energy scale and the residual long term drifts in the response are further corrected using \(\mathrm{Z} \rightarrow \mathrm {e}^+\mathrm {e}^-\) decays [22].

Interaction vertices are reconstructed from charged tracks and the vertex of the diphoton event is taken as the one with the largest sum of squared transverse momenta (\(\Sigma p_{\mathrm {T}} ^2\)) of the associated tracks. The photon four-momentum is recalculated with respect to this vertex.

4.2 Photon selection

The photon candidates are first required to pass a sequence of filters that aim to remove beam backgrounds or identified detector issues and to satisfy more stringent criteria than the trigger requirements. The preselection is based on the shape of the electromagnetic shower in the ECAL and on the degree of isolation of the photon (i.e. the amount of energy deposited in the vicinity of the photon). The variables used are:

  • Photon supercluster raw energy \(E^\mathrm{raw}_\mathrm{SC}\): the sum of the calibrated crystal energies;

  • Preshower energy \(E^\mathrm{ES}_\mathrm{SC}\): the sum of the energy deposits reconstructed in the preshower detector (ES) and associated with the supercluster;

  • \(R_\mathrm {9}\): the energy sum of \(3 \times 3\) crystals centred on the most energetic crystal in the supercluster divided by the raw energy of the supercluster;

  • \(H/E\): the ratio of the energy deposited in HCAL that is inside a cone of size \(\Delta R=\sqrt{{(\Delta \eta )^2 + (\Delta \phi )^2}}=0.15\) centred on the photon direction, to the supercluster energy;

  • \(\sigma _{\eta \eta }\): the shower transverse extension along \(\eta \) that is defined as:

    $$\begin{aligned} \sigma ^2_{\eta \eta }=\frac{\sum \left( \eta _i-{\bar{\eta }}\right) ^2 w_i }{\sum w_i}, \end{aligned}$$
    (1)

    where the sum runs over all elements of the \(5\times 5\) matrix around the most energetic crystal in the supercluster, and \(\eta _i=0.0174\ {\hat{\eta }}_i\) in EB, \(\eta _i=0.0447\ {\hat{\eta }}_i\) in EE with \({\hat{\eta }}_i\) denoting the index of the \(i\)th crystal along the \(\eta \) direction. The individual weights \(w_i\) are given by \(w_i = \max \left( 0, 4.7 + \ln (E_i/E_{5\times 5})\right) \), where \(E_i\) is the energy of the \(i\)th crystal and \({\bar{\eta }}=\sum \eta _iE_i/\sum E_i\) is the weighted average pseudorapidity;

  • \(\text {Iso}_\mathrm{ECAL}^{0.3}\) (ECAL isolation): the scalar sum of the \(E_{\mathrm {T}}\) of the deposits in the electromagnetic calorimeter lying inside a cone of size \(\Delta R= 0.3 \), centred on the direction of the supercluster but excluding an inner cone of size 3.5 crystals and an \(\eta \)-slice region of 2.5 crystals;

  • \(\text {Iso}_\mathrm{HCAL}^{0.3}\) (hadronic calorimeter isolation): the scalar sum of the \(E_{\mathrm {T}}\) of the deposits in the hadron calorimeter that lie inside a hollow cone of outer radius of size \(\Delta R = 0.3\) and inner radius of size \(\Delta R = 0.15\) in the \(\eta \)\(\phi \) plane, centred on the direction of the supercluster;

  • \(\text {Iso}_\mathrm{TRK}^{0.3}\) (tracker isolation): the scalar sum of the \(p_{\mathrm {T}}\) of the tracks that are consistent with originating from the primary vertex in the event, and lie inside a hollow cone of outer radius of size \(\Delta R = 0.3\) and inner radius of size \(\Delta R = 0.04\) in the \(\eta \)\(\phi \) plane, centred around a line connecting the primary vertex with the supercluster but excluding an \(\eta \)-slice region (\(\Delta \eta = 0.015\)).

The isolation requirements are kept loose because the isolation is used as the discriminating variable in the signal extraction procedure. The selection criteria are defined to be slightly tighter than the trigger selection. The shower shape variables in the simulation are corrected to compensate for their imperfect modeling, mainly connected with (a) the simulation of effective readout noise in ECAL channels, (b) the effect of overlapping energy deposits from collisions in adjacent bunch crossings, and (c) the description of the material budget in the detector geometry. The correction factors are extracted from a sample of photons in \(\mathrm{Z} \rightarrow \mathrm {\mu ^+}\mathrm {\mu ^-}\mathrm {\gamma }\) events, and validated as a function of \(E_{\mathrm {T}}\) and \(\eta \) in a sample of electrons from \(\mathrm{Z}\) boson decays. The list of preselection criteria is presented in Table 1.

Table 1 List of requirements that a candidate has to satisfy to pass the analysis preselection

The preselected photons must satisfy additional requirements to be considered as photon candidates. These consist of the absence of reconstructed electron track seeds in the pixel detector which match the candidate’s direction, and a tighter selection on the hadronic leakage of the shower and the \(\sigma _{\eta \eta }\) shower shape variable. The list of additional selection criteria is shown in Table 2.

Table 2 List of additional requirements applied in the photon candidate selection

In the simulation, prompt photons are defined as candidates satisfying the analysis selection requirements and geometrically matched to an isolated generator-level photon, either directly produced or originating from a fragmentation process. The generator-level isolation is defined as the \(p_{\mathrm {T}}\) sum of stable particles in a cone of size \(\Delta R= 0.4 \), and is required to be less than 5\(\,\text {GeV}\).

5 Signal yield determination

The diphoton signal is extracted from events containing two photon candidates with transverse energy greater than 40 (25)\(\,\text {GeV}\) for the leading (sub-leading) photon, and with a separation of \(\Delta R>0.45\). If more than two photon candidates are selected, the two with highest \(E_{\mathrm {T}}\) are retained. The minimum separation requirement ensures that the energy deposit of one photon does not enter the isolation cone centered on the other one. The signal fraction is statistically separated from jets misidentified as photons by means of a binned maximum likelihood fit that uses the photon component of the PF isolation as the discriminating variable.

The diphoton signal is then studied as a function of the diphoton invariant mass \(m_{\mathrm {\gamma }\mathrm {\gamma }}\), the diphoton transverse momentum \(p_{\mathrm {T}} ^{\mathrm {\gamma }\mathrm {\gamma }}\), the azimuthal angle difference \(\Delta \phi _{\mathrm {\gamma }\mathrm {\gamma }}\) between the two photons, and the cosine of the polar angle \(\theta ^*\) in the Collins–Soper frame of the diphoton system [24]. A maximum likelihood fit is performed for each bin of the distributions in the above variables.

5.1 Particle flow isolation

The photon component of the PF isolation (\(\text {Iso}\)) is used to discriminate signal from background. The choice of the isolation variable is optimized to obtain the smallest total uncertainty of the measured cross section. This variable is computed, in a cone of size \(\Delta R= 0.4 \) around each selected photon candidate, as the \(E_{\mathrm {T}}\) sum of photons reconstructed with the PF algorithm [14]. The PF isolation deals more effectively with cases of overlapping particles than the calorimetry-based isolation.

When calculating the isolation, the energy deposited by the selected photon candidate is subtracted by removing from the cone the area where the photon is expected to have deposited its energy (“footprint”), since photon energy leaking into the cone could bias the isolation sum. This is done on an event-by-event basis relying on simple geometrical considerations. The directions of the momenta of reconstructed photon candidates around the selected photon are extrapolated from the interaction vertex to the inner surface of the ECAL, and whenever they overlap with a crystal belonging to the supercluster these photon candidates are removed from the isolation sum. For the matching between the propagated trajectory and the crystal front width, a tolerance of 25 % of the face size is applied.

This procedure does not use any generator-level information and can therefore be applied in both data and simulated events.

The pile-up introduces a spurious correlation between the two candidate photons’ isolation sums. For this reason the PF isolation sums for both photons are corrected, event by event, for the presence of pile-up with a factor proportional to the average pile-up energy density (\(\rho \)) calculated with FastJet [25].

5.2 Template construction

The diphoton signal is extracted through a two-dimensional binned maximum likelihood fit that uses the isolation of the two selected photon candidates as discriminating variables. Different templates are built for the prompt–prompt (\(f_{pp}\)), prompt–non-prompt (\(f_{pn}\)), non-prompt–prompt (\(f_{np}\)), and non-prompt–non-prompt (\(f_{nn}\)) components in the (\(\text {Iso}_1\), \(\text {Iso}_2\)) plane, where \(\text {Iso}_1\) and \(\text {Iso}_2\) represent the isolation variables for the two selected photon candidates in the event. The probability distribution function has the following form:

$$\begin{aligned}&{\mathcal P}_{2D}(\text {Iso}_1,\text {Iso}_2)\nonumber \\&\quad =f_{pp} T_{pp}(\text {Iso}_1,\text {Iso}_2) + f_{pn} T_{pn}(\text {Iso}_1,\text {Iso}_2)\nonumber \\&\qquad +\, f_{np} T_{np}(\text {Iso}_1,\text {Iso}_2) + f_{nn}T_{nn}(\text {Iso}_1,\text {Iso}_2) \end{aligned}$$
(2)

where \(T_{kk}(\text {Iso}_1,\text {Iso}_2)\) is the function describing the isolation distribution (template) for the component \(f_{kk}\). Techniques have been developed to extract the templates from data to avoid possible biases coming from an imperfect modeling of the events in the simulation. Samples of events where at least one photon passes the photon selection are used to create prompt–prompt, prompt–non-prompt, non-prompt–prompt and non-prompt–non-prompt templates with high statistical precision, as described in the following.

The “random cone” technique is used to extract the prompt photon template with high statistical accuracy. In this procedure we compute the isolation energy in a region separated from the candidate photon. Starting from the photon \((\eta ,\phi )\) axis, a new axis is defined at the same pseudorapidity \(\eta \) but with a random separation in azimuthal angle \(\phi _{RC}\) between 0.8 and \(2\pi - 0.8\) radians from the photon \(\phi \). This new axis is used to define the random cone provided that no jet with \(p_{\mathrm {T}} > \text {20}\) \(\,\text {GeV}\) or photon or electron with \(p_{\mathrm {T}} > \text {10}\) \(\,\text {GeV}\) is reconstructed within \(\Delta R < \text {0.8}\) and no muon is reconstructed within \(\Delta R < \text {0.4}\) from this axis. In the case where the new axis does not meet these requirements, a new azimuthal angle is generated. The isolation energy, which is defined as the energy collected in a cone of size \(\Delta R < \text {0.4}\) about the new axis once the fraction corresponding to the area of the photon supercluster has been removed, is then used to populate the prompt photon template.

The distribution of the template variable has been studied in \(\mathrm{Z}\) \(\rightarrow \) \(\mathrm {e}^+\) \(\mathrm {e}^-\) events and found to be in agreement with the template built with the random cone technique.

The background (non-prompt) template cannot be defined by simply inverting the photon preselection, because the candidates entering the analysis, i.e. fulfilling the preselection requirements, have “photon-like” characteristics, while the set of candidates not fulfilling the photon preselection criteria includes a large number of genuine jets. To avoid this bias, the candidates selected to populate the non-prompt photon template are chosen from those that fulfil all the photon selection criteria, except the \(\sigma _{\eta \eta }\) shower shape, which is not strongly correlated with the isolation variable as a result of the footprint removal technique described in the previous section. The events in a “sideband” close to the photon selection criterium are used to populate the non-prompt photon template. The sideband is defined as \(0.011<\sigma _{\eta \eta }<0.014\) for candidates reconstructed in the ECAL barrel and \(0.030<\sigma _{\eta \eta }<0.034\) for candidates reconstructed in the ECAL endcaps.

The same procedure (Sect. 5.1) is used for subtracting the pile-up energy from the photon isolation sums. The templates obtained using the random cone and the sideband techniques in the simulation are compared with the one-dimensional PF isolation distribution for prompt and non-prompt photons in simulated events and with the templates obtained from data (Figs. 1, 2).

The residual differences in the simulation between the isolation distribution and the templates defined with the random cone and the sideband techniques are accounted for as systematic uncertainties on the template shapes.

Fig. 1
figure 1

Comparison of prompt photon templates in data and simulation: prompt photons in the simulation (squares), prompt photon templates extracted with the random cone technique from simulation (triangles) and from data (dots); (top) candidates in the ECAL barrel, (bottom) candidates in the ECAL endcaps. All histograms are normalized to unit area

Fig. 2
figure 2

Comparison of non-prompt photon templates in data and simulation: non-prompt photons in the simulation (squares), non-prompt photon templates extracted with the sideband technique from simulation (triangles) and from data (dots); (top) candidates in the ECAL barrel, (bottom) candidates in the ECAL endcaps. All histograms are normalized to unit area

The two-dimensional templates are built selecting candidate photons from data with the same kinematics as the diphoton events to be fitted. The procedure presented below correctly models the isolation distribution even in the case of overlap between the isolation cones of the two photon candidates.

The prompt–prompt template is built from events where the pileup energy density matches that of the event to be fitted, and where the two random cone directions are found having the same pseudorapidity and the same azimuthal angular separation as the selected photons.

The prompt–non-prompt template is built from events where a sideband photon is selected. The isolation sum around the sideband photon is used for the candidate to be fitted under the non-prompt-hypothesis. A direction satisfying the random cone criteria is then searched for in the same template event (oriented as the second candidate in the selected diphoton event) and used to calculate the isolation sum for the candidate to be fitted under the prompt hypothesis.

Fig. 3
figure 3

Result of the first step of the fitting procedure, for the \(90\,\text {GeV} <m_{\mathrm {\gamma }\mathrm {\gamma }} <95\,\text {GeV} \) bin in the EB-EE category: isolation distribution for the photon reconstructed in the (left) ECAL barrel, (right) ECAL endcaps

The non-prompt–non-prompt template is built selecting two events, each of which contains one sideband photon and such that their orientation matches the orientation of the candidate photons in the event to be fitted. Then, depending on the fraction of photon candidates with \(\Delta R_{\gamma \gamma } < 1.0\) present in the bin of the observable under analysis, a choice between two different strategies is made. If the fraction is below 10 %, the effect of the overlapping isolation cones can be neglected. The two-dimensional non-prompt–non-prompt template is then built by calculating each of the two isolation sums in the separate events. If the fraction is above 10 %, an additional requirement is imposed: the sum of the FastJet \(\rho \) of the two selected template events has to match the one of the diphoton event to be fitted. Then, the sets of reconstructed particles in the two template events are merged, and the isolation sums are calculated from this merged set of reconstructed particles along the direction of each sideband photon.

In this procedure, the pileup energy density of the template events is used to model the pileup energy density of the event to be fitted, and this allows us to describe the correlation between the isolation sums. The effect of the residual correlation mis-modeling is added to the template shape systematic uncertainty in the final result.

5.3 Fitting technique

The fit is performed separately for the cases where both candidates are reconstructed in the ECAL barrel, one in the ECAL barrel and one in the ECAL endcaps, or both in the ECAL endcaps. If both candidates are in the same detector region (EB-EB and EE-EE categories), the leading selected photon is assigned randomly to axis 1 or 2 of the two-dimensional plane, and the prompt–non-prompt (\(f_{pn}\)) and non-prompt–prompt (\(f_{np}\)) fractions are constrained to have the same value.

The fit, performed in each bin of the differential variables, is restricted to the region where the isolation of the photons is smaller than 9\(\,\text {GeV}\). To guarantee its stability even in the less populated bins, the fit is performed in steps. First the size of the bins in the two-dimensional plane \((\text {Iso}_1,\text {Iso}_2)\) is optimised to reduce statistical fluctuations of template shape in the tails; then a first fit is performed on the projections of the isolation distributions on the two axes of the plane using the one-dimensional templates described above. In a subsequent step, the fractions of prompt–prompt, prompt–non-prompt, non-prompt–prompt, and non-prompt–non-prompt, which are constrained to sum up to unity, are fit in the two-dimensional plane using as a constraint the results of the previous fit. The final likelihood maximisation is then performed after removing all constraints, and using as initial values of the parameters those found in the previous step.

An example of the first step of the procedure is obtained by fitting the one-dimensional projections of the isolation distributions as shown in Fig. 3. An example of the results of the final two-dimensional fit (projected on the axes for the sake of clarity) is shown in Fig. 4. The fractions of prompt–prompt, prompt–non-prompt, and non-prompt–non-prompt components are shown in Fig. 5 for the observables of the differential analysis. We fit about 69000 prompt diphoton events in the whole acceptance of the analysis.

Fig. 4
figure 4

Projections of the result of the final step of the fitting procedure, for the \(90\,\text {GeV} <m_{\mathrm {\gamma }\mathrm {\gamma }} <95\,\text {GeV} \) bin in the EB-EE category: isolation distribution for the photon reconstructed in the (left) ECAL barrel, (right) ECAL endcaps

Fig. 5
figure 5

Fractions of prompt–prompt, prompt–non-prompt and non-prompt–non-prompt components as a function of \(m_{\mathrm {\gamma }\mathrm {\gamma }}\), \(p_{\mathrm {T}} ^{\mathrm {\gamma }\mathrm {\gamma }}\), \(\Delta \phi _{\mathrm {\gamma }\mathrm {\gamma }}\), \(|\cos \theta ^* |\) in the whole acceptance of the analysis. Uncertainties are statistical only

The reported purity suffers from a contamination of electrons coming predominantly from Drell–Yan \(\mathrm {e}^+\) \(\mathrm {e}^-\)and incorrectly reconstructed as photons. The contamination is most significant in the \(\mathrm{Z}\) peak region, where it reaches about 25 % of the raw diphoton yield. The fraction of electron pairs passing the analysis selection and contributing to the prompt–prompt fitted fraction is estimated from simulation, where correction factors are applied to obtain the electron to photon mis-identification probability measured in data, and used to subtract the contamination.

6 Efficiencies and unfolding

Following the methodology presented in the previous sections, a “raw” diphoton production cross section is extracted. To obtain the final result, this cross section is corrected for inefficiencies and finally unfolded from the measured quantities to the corresponding particle-level quantities.

The total diphoton efficiency can be separated into the trigger efficiency and reconstruction/selection efficiency, and can be written as:

$$\begin{aligned} \epsilon _{\mathrm {\gamma }\mathrm {\gamma }}&= \epsilon _{\text {trig}} \times \epsilon _{\text {sel}} \times C_{\mathrm {\gamma }1}^{\mathrm{Z} \rightarrow \mathrm {e}^+\mathrm {e}^-} \times C_{\mathrm {\gamma }2}^{\mathrm{Z} \rightarrow \mathrm {e}^+\mathrm {e}^-}\\&\quad \times C_{\mathrm {\gamma }1}^{\mathrm{Z} \rightarrow \mathrm {\mu ^+}\mathrm {\mu ^-}\mathrm {\gamma }} \times C_{\mathrm {\gamma }2}^{\mathrm{Z} \rightarrow \mathrm {\mu ^+}\mathrm {\mu ^-}\mathrm {\gamma }}, \end{aligned}$$

where \(\epsilon _{\text {trig}}\) is the trigger efficiency and \(\epsilon _{\text {sel}}\) is the diphoton reconstruction/selection efficiency from simulation. The factors \(C_{\mathrm {\gamma }1}^{\mathrm{Z} \rightarrow \mathrm {e}^+\mathrm {e}^-}\) and \(C_{\mathrm {\gamma }2}^{\mathrm{Z} \rightarrow \mathrm {e}^+\mathrm {e}^-}\) are the corrections to the efficiency for each photon candidate to pass all the selection requirements except the electron veto; \(C_{\mathrm {\gamma }1}^{\mathrm{Z} \rightarrow \mathrm {\mu ^+}\mathrm {\mu ^-}\mathrm {\gamma }}\) and \(C_{\mathrm {\gamma }2}^{\mathrm{Z} \rightarrow \mathrm {\mu ^+}\mathrm {\mu ^-}\mathrm {\gamma }}\) are the corrections to the electron veto efficiency.

The values of the correction factors are determined from the ratio of the efficiency in data to that in the simulation, measured with a tag-and-probe method using (i) samples of \(\mathrm{Z} \rightarrow \mathrm {e}^+\mathrm {e}^-\) for the full selection except the electron-veto requirement, and (ii) samples of photons from the final-state-radiation of \(\mathrm{Z} \rightarrow \mathrm {\mu ^+}\mathrm {\mu ^-}\mathrm {\gamma }\) for the electron-veto requirement.

The diphoton reconstruction/selection efficiency \(\epsilon _{\text {sel}}\) is about 85 % when both photons are in the barrel, 75 % when one photon is in the barrel and the other in one endcap, and 64 % when both photons are in the endcaps. All these correction factors are estimated from data and range from 0.99 to 1.02, depending on the photon \(E_{\mathrm {T}}\) and \(\eta \).

The detector effects are unfolded from the measured yields for a direct comparison of experimental measurements with theoretical predictions. The number of unfolded diphoton events in each bin of the differential observables is obtained from the reconstructed diphoton events in the data, \({\varvec{N}}_\mathrm{GEN}^\mathrm{{data}}=M^{-1} \times {\varvec{N}}_\mathrm{RECO}^\mathrm{{data}}\), where the unfolding matrix \(M\) is obtained from simulation, \({\varvec{N}}_\mathrm{RECO}^\mathrm{MC}=M \times {\varvec{N}}_\mathrm{GEN}^\mathrm{MC}\). The unfolding matrix is calculated using the iterative Bayesian technique [26, 27]. The diphoton simulated sample from MadGraph hadronized with pythia is used. The distributions of diphoton candidates in the simulation are reweighted to the distributions of the raw diphoton yields from data as obtained from the fit procedure, for all the observables. The difference between the weighted and unweighted results is taken into account as a systematic uncertainty, and amounts to about 1 %. The unfolding correction amounts to 7 % of the raw yield at maximum, for the bins where the slope of the kinematic distributions is the steepest.

Table 3 Sources of systematic uncertainty in the measurement of integrated cross section

7 Systematic uncertainties

Table 3 summarises the main sources of systematic uncertainty in the measurement of the integrated cross section.

The dominant uncertainty in the template shapes arises from the difference in shape between the templates built with the techniques described in Sect. 5 and the distributions of the isolation variable for prompt or non-prompt isolated photons for simulated events. The latter are used to generate data samples for each bin of the differential variables, with the fractions measured in data. Then, each of these datasets is fitted with templates built in the simulation with the same techniques used on data, and the average difference between the fitted fractions and those used for the generation is quoted as a systematic uncertainty. It amounts to 3 % (barrel template) and 5 % (endcap template) for the prompt component, and between 5 % (barrel template) and 10 % (endcap template) for the non-prompt component. The uncertainty in the template shape for fragmentation photons is evaluated in the simulation by doubling the probability of the fragmentation process, and that yields an additional 1.5 % uncertainty in the measured cross section. In the case of the non-prompt–non-prompt template, and only for the bins where a significant fraction of the diphoton candidates are close in \(\Delta R_{\gamma \gamma }\), an additional uncertainty ranging from 3 to 5 % is introduced to account for the imperfections on the template shape description due to the effect of ECAL noise and PF thresholds on the combination of two different events to build the template.

The systematic uncertainty arising from the statistical uncertainty in the shape of the templates is evaluated generating modified templates, where the content of each bin is represented by a Gaussian distribution centred on the nominal bin value and with standard deviation equal to the statistical uncertainty of the bin. The root mean square of the distribution of the fitted purity values, divided by the purity measured with the original template, is used as systematic uncertainty in the purity measurement and amounts to about 3 %.

A possible bias associated with the fitting procedure is evaluated using pseudo-experiments. Pseudo-data samples are generated with given fractions of prompt–prompt, prompt–non-prompt, and non-prompt–non-prompt contributions, using the templates from simulation as generator probability density functions. Each data sample is then fitted with the same templates used for the generation. The average bias is negligible in all bins.

The systematic uncertainty associated with the subtraction of Drell–Yan \(\mathrm {e}^+\) \(\mathrm {e}^-\) events is evaluated by propagating the uncertainty in the electron to photon misidentification probability to the subtracted yield. The uncertainty in the fraction of such events that is fitted as prompt–prompt is also taken into account. This contribution is maximal for \(m_{\mathrm {\gamma }\mathrm {\gamma }}\) close to the \(\mathrm{Z}\)-boson mass. The relative contribution to the total systematic uncertainty is below 0.5 %.

The systematic uncertainty in the trigger efficiency is found to be below 0.5 %. The systematic uncertainty in the reconstruction and selection efficiencies is dominated by the uncertainty in the data-to-simulation corrections from the \(\mathrm{Z} \rightarrow \mathrm {e}^+\mathrm {e}^-\) and \(\mathrm{Z} \rightarrow \mathrm {\mu ^+}\mathrm {\mu ^-}\mathrm {\gamma }\) control samples, and it ranges from 2 % in the barrel to 4 % in the endcap.

The systematic uncertainty in the integrated luminosity that corresponds to our data sample is 2.2 % [28].

The total systematic uncertainty in the measurement amounts to approximately 8 % when both candidates are reconstructed within the ECAL barrel, and to 11 % for the full acceptance of the analysis.

Fig. 6
figure 6

The comparisons of the differential cross section between data and the sherpa, diphox + gamma2mc, resbos, and 2\(\mathrm {\gamma }\) nnlo predictions for \(m_{\mathrm {\gamma }\mathrm {\gamma }}\). Black dots correspond to data with error bars including all statistical and systematic uncertainties. Only the scale uncertainty is included for the sherpa prediction. Scale, PDF and \(\alpha _S\) uncertainties are included for diphox + gamma2mc and resbos. Only statistical and scale uncertainties are included for the 2\(\mathrm {\gamma }\) nnlo prediction

Fig. 7
figure 7

The comparisons of the differential cross section between data and the sherpa, diphox + gamma2mc, resbos, and 2\(\mathrm {\gamma }\) nnlo predictions for \(p_{\mathrm {T}} ^{\mathrm {\gamma }\mathrm {\gamma }}\). Black dots correspond to data with error bars including all statistical and systematic uncertainties. Only the scale uncertainty is included for the sherpa prediction. Scale, PDF and \(\alpha _S\) uncertainties are included for diphox + gamma2mc and resbos. Only statistical and scale uncertainties are included for the 2\(\mathrm {\gamma }\) nnlo prediction

Fig. 8
figure 8

The comparisons of the differential cross section between data and the sherpa, diphox + gamma2mc, resbos, and 2\(\mathrm {\gamma }\) nnlo predictions for \(\Delta \phi _{\mathrm {\gamma }\mathrm {\gamma }}\). Black dots correspond to data with error bars including all statistical and systematic uncertainties. Only the scale uncertainty is included for the sherpa prediction. Scale, PDF and \(\alpha _S\) uncertainties are included for diphox + gamma2mc and resbos. Only statistical and scale uncertainties are included for the 2\(\mathrm {\gamma }\) nnlo prediction

Fig. 9
figure 9

The comparisons of the differential cross section between data and the sherpa, diphox + gamma2mc, resbos, and 2\(\mathrm {\gamma }\) nnlo predictions for \(|\cos \theta ^* |\). Black dots correspond to data with error bars including all statistical and systematic uncertainties. Only the scale uncertainty is included for the sherpa prediction. Scale, PDF and \(\alpha _S\) uncertainties are included for diphox + gamma2mc and resbos. Only statistical and scale uncertainties are included for the 2\(\mathrm {\gamma }\) nnlo prediction

8 Results and comparison with theoretical predictions

The measured unfolded differential cross sections are compared with the following generators for QCD diphoton production: sherpa 1.4.0 [29], diphox 1.3.2 [30] supplemented with gamma2mc 1.1 [31], resbos [32, 33], and 2\(\mathrm {\gamma }\) nnlo [34]. Predictions with sherpa are computed at LO for the Born contribution with up to three additional real emissions (three extra jets) and with the box contribution at the matrix element level. The diphox NLO generator includes the direct and fragmentation contributions and uses a full fragmentation function for one or two partons into a photon at NLO. The direct box contribution, which is formally part of the NNLO corrections since it is initiated by gluon fusion through a quark loop, is computed at NLO with gamma2mc. The resbos NLO generator features resummation for Born and box contributions, and effectively includes fragmentation of one quark/gluon to a single photon at LO. The latter process is regulated to avoid divergences and does not include the full fragmentation function. The resbos \(p_{\mathrm {T}} ^{\mathrm {\gamma }\mathrm {\gamma }}\) spectrum benefits from a soft and collinear gluon resummation at next-to-next-to-leading-log accuracy. 2\(\mathrm {\gamma }\) nnlo predicts the direct \(\mathrm {\gamma }\mathrm {\gamma }\)+X processes at NNLO. The sherpa sample is used after hadronization while diphox + gamma2mc, resbos, and 2\(\mathrm {\gamma }\) nnlo are parton-level generators only and cannot be interfaced with parton shower generators.

The predictions have been computed for the phase space \(E_{\mathrm {T}}^{\mathrm {\gamma }\mathrm {1}} >40\) \(\,\text {GeV}\), \(E_{\mathrm {T}}^{\mathrm {\gamma }\mathrm {2}} >25\) \(\,\text {GeV}\), \(|\eta _{\gamma } |<1.44\) or \(1.57<|\eta _{\gamma } |<2.5\), \(\Delta R(\gamma _1,\gamma _2)>0.45\). An isolation requirement is applied at the generator level. In sherpa, the \(E_{\mathrm {T}}\) sum of stable particles in a cone of size \(\Delta R=0.4\) has to be less than 5\(\,\text {GeV}\) (after hadronization). In diphox, gamma2mc, and resbos the \(E_{\mathrm {T}}\) sum of partons in a cone of size \(\Delta R=0.4\) is required to be less than 5\(\,\text {GeV}\). In 2\(\mathrm {\gamma }\) nnlo, the smooth Frixione isolation [35] is applied to the photons to suppress the fragmentation component:

$$\begin{aligned} E_{\mathrm {T}} ^\mathrm{{Iso}} (\Delta R) < \epsilon \left( \frac{1-\cos (\Delta R)}{1-\cos (\Delta R_0)} \right) ^{n}, \end{aligned}$$
(3)

where \(E_{\mathrm {T}} ^\mathrm{{Iso}}\) is the \(E_{\mathrm {T}}\) sum of partons in a cone of size \(\Delta R\), \(\Delta R_0=0.4\), \(\epsilon =5\) \(\,\text {GeV}\), and \(n=0.05\). This criterion, tested with diphox, is found to have the same efficiency as that used for the other generators within a few percent. A non-perturbative correction is applied to diphox, gamma2mc, and 2\(\mathrm {\gamma }\) nnlo predictions to correct for the fact that those generators do not include parton shower or underlying event contributions to the isolation cone. The fraction of diphoton events not selected due to underlying hadronic activity falling inside the isolation cone is estimated using the pythia 6.4.22 [18] event generator with tunes Z2, D6T, P0, and DWT [20]. A factor of \(0.95\pm 0.04\) is applied to the parton-level cross section to correct for this effect.

Theoretical predictions are performed using the CT10 [36] NLO PDF set for sherpa, diphox + gamma2mc, and resbos, and the MSTW2008 [37] NNLO PDF set for 2\(\mathrm {\gamma }\) nnlo. The diphox and gamma2mc theoretical uncertainties are computed in the following way: the factorization and renormalization scales in gamma2mc are varied independently up and down by a factor of two around \(m_{\mathrm {\gamma }\mathrm {\gamma }}\) (configurations where one scale has a factor of four with respect to the other one are forbidden). In diphox, the factorization, renormalization and fragmentation scales are varied in the same way. In resbos, the factorization and renormalization scales are varied simultaneously by a factor of two. The maximum and minimum values in each bin are used to define the uncertainty. In diphox, gamma2mc, and resbos, the 52 CT10 eigenvector sets of PDFs are used to build the PDF uncertainty envelope, also considering the uncertainty in the strong coupling constant \(\alpha _S\), determined according to the CT10 \(\alpha _S\) PDF set. In 2\(\mathrm {\gamma }\) nnlo, a simplified and less computationally intensive estimate of the renormalization and factorization scale uncertainties is performed by varying these scales simultaneously by a factor of two up and down around \(m_{\mathrm {\gamma }\mathrm {\gamma }}\); no PDF uncertainty is computed. The same procedure is used in sherpa, using the internal METS scale, where scales are defined as the lowest invariant mass or negative virtuality in the core 2\(\rightarrow \)2 configuration clustered using a \(k_{\mathrm {T}}\)-type algorithm.

The total cross section measured in data for the phase space defined above is:

$$\begin{aligned} \sigma = 17.2 \pm 0.2\hbox {(stat.)} \pm 1.9\hbox {(syst.)} \pm 0.4\hbox {(lum.)}\, \hbox {pb}, \end{aligned}$$

compared with

$$\begin{aligned}&\sigma _{\text {NNLO}}({{2\mathrm {\gamma }\textsc {nnlo}}}) = 16.2^{+1.5}_{-1.3}\,\text {(scale)}\text {\,pb},\nonumber \\&\sigma _{\text {NLO}}({\textsc {diphox}} +{\textsc {gamma2mc}}) \nonumber \\&\quad =12.8^{+1.6}_{-1.5}\,\text {(scale)}^{+0.6}_{-0.8}\, \text {(pdf+}\alpha _S \text {)}\text {\,pb},\nonumber \\&\sigma _{\text {NLO}}({\textsc {resbos}}) =14.9^{+2.2}_{-1.7}\,\text {(scale)} \pm 0.6\,\text {(pdf+}\alpha _S \text {)}\text {\,pb},\nonumber \\&\sigma _{\text {LO}}({\textsc {sherpa}}) =13.8^{+2.8}_{-1.6}\,\text {(scale)}\text {\,pb}. \end{aligned}$$
(4)

Figures 6, 7, 8 and 9 show the comparisons of the differential cross section between data and the sherpa, diphox + gamma2mc, resbos, and 2\(\mathrm {\gamma }\) nnlo predictions for the four observables.

The NLO predictions of diphox + gamma2mc are known to underestimate the data [11], because of the missing higher-order contributions. Apart from an overall normalization factor, the phase space regions where the disagreement is the largest are at low \(m_{\mathrm {\gamma }\mathrm {\gamma }}\), low \(\Delta \phi _{\mathrm {\gamma }\mathrm {\gamma }}\). The resbos generator shows a similar trend, with a cross section closer to the data than diphox + gamma2mc; its prediction is improved at high \(\Delta \phi _{\mathrm {\gamma }\mathrm {\gamma }}\) due to soft gluon resummation. With higher-order diagrams included, 2\(\mathrm {\gamma }\) nnlo shows an improvement for the overall normalization. It also shows a better shape description, especially at low \(\Delta \phi _{\mathrm {\gamma }\mathrm {\gamma }}\), but it still underestimates the data in the same region. sherpa generally reproduces rather well the shape of the data, to a similar level as 2\(\mathrm {\gamma }\) nnlo. One can note that 2\(\mathrm {\gamma }\) nnlo and sherpa predict the \(p_{\mathrm {T}} ^{\mathrm {\gamma }\mathrm {\gamma }}\) shoulder near \(E_{\mathrm {T}}^{\mathrm {\gamma }\mathrm {1}} +E_{\mathrm {T}}^{\mathrm {\gamma }\mathrm {2}} \sim 65\) \(\,\text {GeV}\) observed in the data. This is expected since sherpa includes up to three extra jets at the matrix element level.

9 Summary

A measurement of differential cross sections for the production of a pair of isolated photons in \(\mathrm {p}\) \(\mathrm {p}\) collisions at \(\sqrt{s} = 7\) \(\,\text {TeV}\) has been presented. The data sample corresponds to an integrated luminosity of 5.0\(\,\text {fb}^{-1}\) recorded in 2011 with the CMS detector. To enhance the sensitivity to higher-order diagrams, this measurement covers a phase space defined by an asymmetric \(E_{\mathrm {T}}\) selection by requiring two isolated photons with \(E_{\mathrm {T}}\) above 40 and 25\(\,\text {GeV}\) respectively, in the pseudorapidity range \(|\eta |<2.5\), \(|\eta |\notin [1.44,1.57]\) and with an angular separation \(\Delta R > 0.45\).

A data-driven method based on the photon component of the particle flow isolation has been used to extract the prompt diphoton yield. The isolation is calculated so that the energy leakage from the photon deposit inside the isolation cone is effectively subtracted.

The measured total cross section is

$$\begin{aligned} \sigma = 17.2 \pm 0.2\,\text {(stat)} \pm 1.9\,\text {(syst)} \pm 0.4\,\text {(lumi)} \text {\,pb} \end{aligned}$$

in agreement with the 2\(\mathrm {\gamma }\) nnlo prediction. The sherpa and resbos predictions are compatible with the measurement within the uncertainties, while diphox + gamma2mc underestimates the total cross section.

Differential cross sections for prompt diphoton production have been measured as a function of the diphoton invariant mass \(m_{\mathrm {\gamma }\mathrm {\gamma }}\), the diphoton transverse momentum \(p_{\mathrm {T}} ^{\mathrm {\gamma }\mathrm {\gamma }}\), the azimuthal angular separation \(\Delta \phi _{\mathrm {\gamma }\mathrm {\gamma }}\) between the two photons, and the cosine of the polar angle \(\theta ^{*}\) in the Collins–Soper frame of the diphoton system. The 2\(\mathrm {\gamma }\) nnlo and sherpa predictions show an improved agreement in shape with the data for the kinematic distributions with respect to the diphox + gamma2mc and resbos predictions, especially in the low \(m_{\gamma \gamma }\), low \(\Delta \phi _{\gamma \gamma }\) regions, which are the most sensitive to higher-order corrections.