Search for extra dimensions in diphoton events from proton-proton collisions at √s = 7 TeV in the ATLAS detector at the LHC

. The large difference between the Planck scale and the electroweak scale, known as the hierarchy problem, is addressed in certain models through the postulate of extra spatial dimensions. A search for evidence of extra spatial dimensions in the diphoton channel has been performed using the full set of proton–proton collisions at √ s = 7 TeV recorded in 2011 with the ATLAS detector at the CERN Large Hadron Collider. This dataset corresponds to an integrated luminosity of 4 . 9 fb − 1 . The diphoton invariant mass spectrum is observed to be in good agreement with the Standard Model expectation. In the context of the model proposed by Arkani–Hamed, Dimopoulos and Dvali, 95% conﬁdence level lower limits of between 2.52 and 3.92 TeV are set on the ultraviolet cutoff scale M S depending on the number of extra dimensions and the theoretical formalism used. In the context of the Randall–Sundrum model, a lower limit of 2.06 (1.00) TeV at 95% conﬁdence level is set on the mass of the lightest graviton for couplings of k / M Pl = 0 . 1 ( 0 . 01 ) . Combining with the ATLAS dilepton searches based on the 2011 data, the 95% conﬁdence level lower limit on the Randall–Sundrum graviton mass is further tightened to 2.23 (1.03) TeV for k / M Pl = 0 . 1 ( 0 . 01 ) .


Introduction
One of the major goals of particle physicists in the past decades has been to solve the hierarchy problem: the fact that the electroweak scale is 16 orders of magnitude smaller than the Planck scale. One possible solution to the hierarchy problem is the existence of extra spatial dimensions. In this paradigm, gravity appears much weaker than the other interactions because it is diluted by the presence of extra spatial dimensions.
In the Randall-Sundrum (RS) model [1], a five-dimensional geometry is assumed, in which the fifth dimension is compactified with length r c . There are two four-dimensional branes sitting in a five-dimensional bulk with a "warped" geometry. The Standard Model (SM) fields are located on the so-called TeV brane, while gravity originates from the other brane, called the Planck brane. Gravitons are capable of propagating in the bulk. Mass scales on the TeV brane correspond to mass scales on the Planck brane (M P l ) as given by M D = M P l e −kπrc , where k is the curvature scale of the extra dimension. The observed hierarchy of scales can be naturally reproduced in this model if kr c ≈ 12 [2]. The compactification of the extra dimension gives rise to a Kaluza-Klein (KK) tower of graviton excitations G, a set of four-dimensional particles with increasing masses. The phenomenology can be described in terms of the mass of the lightest KK graviton excitation (m G ) and the dimensionless coupling to the SM fields, k/M P l , where M P l = M P l / √ 8π is the reduced Planck scale. From theoretical arguments [2], k/M P l values in the range [0.01, 0.1] are preferred. The lightest RS graviton is expected to be a fairly narrow resonance for k/M P l < 0.3. The most stringent experimental limits on RS gravitons have been obtained at the LHC: the current best 95 % confidence level (CL) lower limits on the graviton mass for k/M P l = 0.1 are 2.16 TeV [3] and 2.14 TeV [4]. The former limit has been obtained from an analysis of the G → ee/µµ channels in data corresponding to an integrated luminosity of 5 fb −1 collected by the ATLAS experiment. The latter limit has been obtained from an analysis of the same channels in data corresponding to an integrated luminosity of ≈5 fb −1 collected by the CMS experiment. The limits from earlier searches at the Tevatron can be found in Refs. [5,6].
Arkani-Hamed, Dimopoulos, and Dvali (ADD) [7] proposed a different scenario for extra dimensions. Motivated by the weakness of gravity, they postulated the existence of n flat additional spatial dimensions compactified with radius R, and proposed a model in which only gravity propagates in the extra dimensions. The fundamental Planck scale in the (4 + n)-dimensional spacetime, M D , is related to the apparent scale M P l by Gauss's law: M emission depend on M D , effects involving virtual gravitons depend on the ultraviolet cutoff of the KK spectrum, denoted M S . The strength of gravity in the presence of extra dimensions is typically parameterised by η G = F/M 4 S , where F is a dimensionless parameter of order unity reflecting the dependence of virtual KK graviton exchange on the number of extra dimensions. Several theoretical formalisms exist in the literature, using different definitions of F and, consequently, of M S : F = 1, (GRW) [8]; (1) F = log (M 2 S /ŝ) n = 2 2/(n − 2) n > 2 , (HLZ) [9]; (2) F = ±2/π, (Hewett) [10]; where √ŝ is the centre-of-mass energy of the parton-parton collision. The effect of ADD graviton exchange manifests itself as a non-resonant deviation from the SM background expectation. The effective diphoton cross section is the result of the SM and ADD amplitudes, as well as their interference. The interference term in the effective cross section is linear in η G and the pure graviton exchange term is quadratic in η G . In previous analyses by the ATLAS and CMS collaborations, 95 % CL lower limits on M S of 2.26-3.52 TeV [11] and 2.4-3.8 TeV [12] respectively were set, depending on the theoretical formalism used. Searches for ADD virtual graviton effects have also been performed at other colliders, e.g. at HERA [13,14], LEP [15,16,17] and the Tevatron [18,19].
This Letter reports the results of a search for extra dimensions in the diphoton channel using the full data sample recorded by the ATLAS detector in 2011, corresponding to a total integrated luminosity of 4.9 fb −1 of pp collisions at √ s = 7 TeV.
The results are interpreted in the context of both the ADD and RS scenarios. The dilepton and diphoton channels are particularly sensitive for this search due to the clean experimental signature, excellent mass resolution and modest backgrounds. The branching ratio for graviton decay to two photons is twice the value of that for graviton decay to any individual charged-lepton pair. This search, based on the full √ s = 7 TeV 4.9 fb −1 dataset, is an extension of an earlier study [11] performed using 1.0 to 2.1 fb −1 of data. The earlier data are included in the present study but they benefit from improvements in the reconstruction and calibration procedures. Moreover, the techniques used for the background estimation and their associated uncertainties have been also significantly improved. This results to an overall improvement of a factor 2 to 3 in the excluded production cross section of diphoton events predicted by the Randall-Sundrum model, corresponding to a graviton mass limit which is 10% to 15% larger.

The ATLAS detector
The ATLAS detector is a multipurpose particle physics instrument with a forwardbackward symmetric cylindrical geometry and near 4π solid angle coverage. ‡ A detailed description of the ATLAS detector can be found in Ref. [20]. The inner tracking detector (ID) consists of a silicon pixel detector, a silicon microstrip detector, and a transition radiation tracker. The ID is surrounded by a superconducting solenoid that provides a 2 T axial magnetic field. The ID allows an accurate reconstruction of tracks from the primary pp collision and also identifies tracks from secondary vertices, permitting the efficient identification of photon conversions. The electromagnetic calorimeter (ECAL) is a lead/liquid-argon (LAr) sampling calorimeter with an accordion geometry. It is divided into a barrel section, covering the pseudorapidity § region |η| < 1.475, and two endcap sections, covering the pseudorapidity regions 1.375 < |η| < 3.2. It consists of three longitudinal layers for |η| < 2.5 and two for 2.5 < |η| < 3.2. Up to |η| < 2.4, the first one uses highly granular "strips" segmented in the η direction for efficient event-by-event discrimination between single photon showers and two overlapping showers originating from π 0 decay. The second layer collects most of the energy deposited in the electromagnetic calorimeter by photon showers. Significant energy deposits in the third layer are an indication for leakage beyond the ECAL from a high energy shower. The measurements from the third layer are used to correct for this effect. A thin presampler layer in front of the accordion calorimeter, covering the pseudorapidity interval |η| < 1.8, is used to correct for energy loss upstream of the calorimeter.
The hadronic calorimeter (HCAL), surrounding the ECAL, includes a central (|η| < 1.7) iron/scintillator tile calorimeter, two endcap (1.5 < |η| < 3.2) copper/LAr calorimeters and two forward calorimeters that extend the coverage to |η| < 4.9, using copper and tungsten as absorber. The muon spectrometer, located beyond the calorimeters, consists of three large air-core superconducting toroid systems, instrumented with precision tracking chambers as well as fast detectors for triggering.

Trigger and event selection
The analysis uses data recorded by ATLAS between March and October 2011 during stable-beam periods of pp collisions at √ s = 7 TeV. A three-level trigger system is used to select events containing two photon candidates. The first level trigger is hardwarebased and relies on a readout with coarse granularity. The second-and third-level triggers, collectively referred to as the high-level trigger (HLT), are implemented in software and exploit the full granularity and energy calibration of the calorimeter.
Selected events have to satisfy a diphoton trigger where each photon is required to satisfy, at the HLT level, a transverse energy requirement E γ T > 20 GeV and a set of requirements [21,22] on the shape of the energy deposit. This set includes criteria for the energy leakage into the hadronic calorimeter and for the width of the shower in the second layer of the ECAL. This trigger is nearly 100 % efficient for diphoton events passing the final selection requirements.
Only events satisfying the standard ATLAS data quality requirements and having at least one primary collision vertex with at least three tracks are kept. Photon candidates reconstructed in the precision regions of the barrel (|η| < 1.37) or the endcaps (1.52 < |η| < 2.37), with E γ T > 25 GeV and satisfying the same requirements on the shower shape as the trigger are preselected. These requirements on the photon candidates are referred to as the loose selection. The two highest-E γ T photon candidates each have to satisfy a set of stricter requirements, referred to as the tight photon definition, which includes a more stringent selection on the shower width in the second layer of the ECAL and additional requirements on the energy distribution in the first layer of the ECAL. In addition, the two photon candidates of interest have to satisfy a calorimetric isolation requirement, E iso T < 5 GeV. This isolation is computed by summing the transverse energy over the cells of the calorimeters in a cone of radius ∆R = (η − η γ ) 2 + (φ − φ γ ) 2 < 0.4 surrounding the photon candidate. Here (η, φ) denotes the cell position and (η γ , φ γ ) denotes the position of the photon cluster [23]. The transverse energy contribution in the centre of the cone, which contains the photon cluster, is subtracted. Corrections, based on simulations, are applied to account for the expected energy leakage from the photon into the isolation region. An ambientenergy correction, based on the measurement of low-transverse-momentum jets [24] is also applied, on an event-by-event basis, to remove the contributions from the underlying event and from pileup which results from the presence of multiple pp collisions within the same or nearby bunch crossings.

Event simulation
Monte Carlo (MC) simulations are used to study the response of the detector to various RS and ADD scenarios as well as SM diphoton processes. The MC generators used in this analysis are listed below. For samples generated using Pythia [25], the ATLAS parameter tunes [26] are used in the event generation. For samples generated using Pythia or Sherpa [27], the generated events are processed through the Geant4 [28] detector simulation [29]. Then the simulated events are reconstructed with the same procedure as is used for the data. To reproduce the pileup conditions of the data, the simulated events are reweighted to match the observed distribution of the average number of interactions per bunch crossing. The determination of the number of interactions per bunch crossing is discussed in Ref. [30].
Pythia 6.424 with MRST2007LOMOD [31] parton distribution functions (PDFs) is used to simulate the SM diphoton production at leading order (LO). In addition, the next-to-leading (NLO) correction and the contribution of the fragmentation processes are evaluated using Diphox [32] 1.3.2 with MSTW2008NLO [33] PDFs and lead to an invariant-mass-dependent correction of the Pythia prediction. The so-called box contribution gg → γγ through a quark loop is included in both the Pythia and the Diphox predictions. From the point of view of power counting, this diagram is a nextto-next-to-leading (NNLO) contribution, but the gluon luminosity at the LHC is so large compared to the quark-antiquark one that this process represents a contribution that is comparable to the qq → γγ process. Samples of RS events are produced using the same generator and PDFs as for the SM diphoton samples. They are used to study the selection efficiency and the shape of the reconstructed invariant mass spectrum for various values of the graviton mass m G between 400 GeV and 3000 GeV and values of

Background estimate
The dominant background in this analysis is the irreducible background due to SM γγ production. Another significant background component is the reducible background which arises from events in which one or both of the photon candidates result from a different object being misidentified as a photon. This background is dominated by γ + jet (j) and j + j events, with one or both photon candidates arising from jets. Backgrounds with electrons faking photons, such as the Drell-Yan production of electron-positron pairs as well as W/Z + γ and tt processes, have been verified to be negligible after the event selection.
We first describe the tools and samples that are used to predict the contributions from the SM γγ background to the high-mass signal region, as well as their main limitation. We then describe in detail the two-step procedure that is used to overcome this limitation and to obtain our background estimate.
The estimate of the irreducible background makes use of the NLO calculations [32] that are available for the SM γγ production processes. The uncertainty in the absolute normalisation of the γγ cross section is, however, quite substantial: ≃ 20 % (≃ 25 %) uncertainty on the cross section for γγ events with an invariant mass m γγ > 200 GeV (m γγ > 1200 GeV) with both photons in the detector acceptance. The background estimates, which are discussed in detail below, are therefore normalised to the data in the low-mass control region ([142, 409] GeV) where the presence of any signal beyond the SM has been excluded by previous searches. Compared to the use of NLO calculations to The exact values of the upper and lower bound of the low-mass control region have been chosen to coincide with the boundaries of the first bin in Table 1. The results of this search are not altered by small variations of the control region edges nor a change of the binning.
predict the absolute rate of SM γγ production in the high-mass signal region, the use of NLO calculations for the extrapolation of the SM γγ production rates from the low-mass control region to the high-mass signal region results in significantly smaller uncertainties. The extrapolation method is described in detail below, along with quantitative estimates of the final uncertainties.
Following these considerations, the estimate the m γγ spectrum of the background events is done in two steps. First the shape of the m γγ spectrum is determined, separately for each background component. Then the normalisation for each component is determined using the low-mass control region.
The shape of the m γγ spectrum from SM γγ production is estimated using simulated events, reweighting the Pythia samples to the differential (in m γγ ) cross section predicted using Diphox.
The shape of the reducible background is estimated using data-driven techniques. It is split into three components: γ + j events (the leading-E T photon candidate is due to a real photon), j + γ (fake leading photon candidate) and j + j. The shapes of the j +γ and γ +j differ significantly because of the strong p T -dependence of the probability to mis-identify a jet as a photon. Separately for each component, several data control samples that are enriched in the given component are defined. The shape of the γ + j background is estimated from a data control sample that is selected in the same way as the signal sample, except that the subleading γ candidate is required to pass loose identification criteria and to fail tight identification criteria. The control samples used to extract the shapes of the m γγ spectra of j +γ and j +j backgrounds are defined following the same approach, i.e. by requiring the leading photon candidate (j + γ background) or both candidates (j + j background) to pass loose and to fail tight identification criteria. Since the data control samples contain relatively few events at high m γγ , a fit to a smooth function of the form f (m γγ ) = p 1 × (m γγ ) p 2 +p 3 log mγγ , where the p i are free parameters, is used to extrapolate the reducible background shapes to higher masses. This functional form has been used in the CDF dijet search [37] and in previous ATLAS searches in the dilepton [38], dijet [39], photon-jet [40] and diphoton [11] channels, and it describes the shapes of the data control samples well. In contrast to the earlier version of this analysis [11], the shapes of the three components of the reducible background are modelled separately.
To determine the contributions of each of the four sources (the irreducible γγ component plus the three reducible components, γ + j, j + γ and j + j) to the low-mass control region, a two-dimensional template fit to the distributions of the calorimetric isolation (E iso T ) of the two photon candidates is used. For the purpose of this fit, the isolation requirement in the event selection is relaxed to E iso T < 25 GeV. This method has been used previously in Refs. [41,42] and in an earlier version of this analysis [11]. Templates for the E iso T distribution of true photons and of fake photons from jets are both determined from data. The shape for fake photons is determined using a sample of photon candidates that fail the tight requirement but pass a set of requirements referred as the non-tight selection. This non-tight selection Table 1. Expected numbers of events from the irreducible and reducible background components, as well as the total background prediction and observed numbers of events in m γγ bins. The table quotes the sum of the statistical and systematic uncertainties in the background expectations. The boundaries of all bins in this table have been chosen to coincide with bin boundaries in Fig. 1. As discussed in the text, the background estimate is normalised to the data in the low-mass control region [142,409] GeV. Since the uncertainties in the normalisation are strongly anticorrelated between the reducible and irreducible components, their impact on the estimate of the total background is smaller than their impact on the individual components. By construction, the uncertainties in the control region are 100% anticorrelated and the total background expectation is identical to the number of observed events.

Mass Window
Background 0.39 ± 0.08 0.013 ± 0.009 0.40 ± 0.07 0 includes the same requirements as the loose selection plus additional requirements that reduce the correlation between the identification and the isolation. The shape for true photons is found from the sample of tight photon candidates, after subtracting the fake photon shape normalised to match the number of candidates with large values of E iso T (E iso T > 10 GeV; this control region is dominated by fake photons). Both the signal templates and fake templates are constructed separately for leading and subleading photon candidates. The significant correlation (∼ 20 %) between the E iso T values of the two fake photons in the j + j background is included in the two-dimensional template for this background component. The background expectation as a function of m γγ is summarised in Fig. 1 and Table 1. Fig. 2 details the different contributions to the uncertainty in the background expectation.
The uncertainties in the shape of the irreducible background are dominated by the uncertainties in the PDFs that are used with Diphox. Those uncertainties are evaluated using the MSTW2008 NLO eigenvector PDF sets and using CTEQ6.6 and MRST2007LOMOD PDF sets for comparisons. The spread of the variations includes the difference between the central values obtained with the different PDF sets. Smaller contributions arise from residual imperfections in the simulation of the isolation variable E iso T and from higher-order contributions that are not included in the Diphox model. The latter contribution is evaluated by varying in a coherent and an incoherent way the renormalisation, the initial factorisation and the final factorisation scales by a factor two around their nominal value, which is the invariant mass of the diphoton system. The shape predictions from Diphox are found to be in good agreement with those obtained using an alternative NLO generator, MCFM [44] (with the same PDF set as Diphox). The small difference between the two generators is assigned as an additional systematic uncertainty. It represents a negligible contribution for m γγ > 1 TeV. The uncertainties in the shape of the reducible background arise from the finite size of the background control samples, and from the extrapolation of the background shapes from the control sample to the signal region. The latter are assessed by varying the definition of the loose selection requirement which is used to define the control samples. The shape uncertainty of the reducible background has a small impact on the total background (Fig. 2). This is mainly because the reducible background is a small contribution to the total ( Table 1). The uncertainties in the normalisation of the background components are dominated by the uncertainties in the templates used in the fit to the E iso T distributions. The uncertainties in the templates are assessed by varying the definition of the non-tight selection requirement used in the extraction of the templates. They are then propagated to the normalisations by repeating the template fits with the varied templates.

Systematic uncertainties in the signal models
Experimental systematic uncertainties on the signal yields are evaluated for both the RS and the ADD signals. The main systematic uncertainties arise from the non-perfect simulation of the quantities used for the identification of photons and in the isolation requirements. For the former, the data-driven estimates from Ref. [22] are used, together with an extrapolation to larger values of p T , which are relevant in the present search. In the RS scenario, these uncertainties on the photon identification translate into a relative uncertainty of 4.7% to 3.3% on the signal yield for m G values of 0.5 TeV to 2.5 TeV. In the ADD scenario, the relative uncertainty on the signal yield is 4.5% with a weak dependence on M S . For the latter, a study comparing the data sample and simulation has shown that a 1 GeV shift of the isolation distribution in the simulation covers possible discrepancies between data and simulation in the signal region. In the RS scenario, these uncertainties translate into a relative uncertainty of 4% to 7% on the signal yield for m G values of 0.5 TeV to 2.5 TeV. In the ADD scenario the relative uncertainty on the signal yield is of ∼ 6% with a weak dependence on M S . The uncertainty on the integrated luminosity for the complete 2011 dataset is 3.9%, based on the calibration described in Refs. [30,45] plus an additional uncertainty for the extrapolation to the later data-taking period with higher instantaneous luminosity. An uncertainty of 2 % on the trigger efficiency arises from differences between the measured trigger efficiency in data and simulated samples. Given that the sum in quadrature of all experimental systematic uncertainties depends only weakly on the signal scenario, a common relative uncertainty value of 9% on signal yield is used for all scenarios.
Uncertainties from limited Monte Carlo sample sizes, as well as uncertainties in the RS resonant shape due to the current knowledge of the electromagnetic energy scale of the calorimeter, the resolution of the detector and the pileup conditions were verified to have a negligible impact on the result. Theoretical uncertainties due to the PDFs and due to uncomputed higher perturbative orders were considered. The uncertainty in the PDFs translates into an uncertainty of +10 −5 % in the signal yields predicted by the ADD models. In the RS scenario the PDF uncertainty in the signal yield is between 6 % and 25 % for m G values between 0.5 TeV and 2.5 TeV. As detailed in Ref. [11], an uncertainty of ±0.1 is considered on the NLO K-factor value. The theory uncertainties are not included in the limit calculation, but their size is indicated in Figs. 3 and 5, where graphical representations of the limits discussed in Section 7 are shown.

Results and interpretation
A comparison of the observed invariant mass spectrum of diphoton events and the background expectation is shown in Fig. 1, along with the statistical significance of the bin-by-bin difference [43] between data and the expected background. This significance is plotted as positive (negative) for bins with an excess (deficit) of data. Also shown are the expected signals for two examples of RS and ADD models. The geometrical acceptance for the RS scenario with m G = 1.5 TeV and k/M P l = 0.1 shown in Fig. 1 is 88 % and the selection efficiency for events within the acceptance is (54±5)%. Both Fig. 1 and Table 1 show that, over the entire m γγ range, the data are in good agreement with the expectations from SM backgrounds. The BumpHunter algorithm [46] is used to further quantify the level of agreement between data and the SM background expectation. The BumpHunter performs a scan of the mass spectrum in windows of progressively increasing width. It identifies the window with the most significant excess of data events over the background expectation anywhere in a predefined search region. In this analysis, the binned mass spectrum of Fig. 1 is used, the search region is defined as 142 < m γγ < 3000 GeV and the window size is allowed to vary from one bin to half the number of bins in the search region ¶. The most significant excess is found in the region 1312 < m γγ < 1644 GeV. The probability of observing, due to fluctuations in the background only, an excess that is at least as significant as the one in data is 0.86. This confirms the absence of a significant excess in the mass spectrum.
Given the absence of evidence for a signal, 95 % CL upper limits were determined on the RS signal cross section. The limits are computed using a Bayesian approach [47] assuming a flat prior for the RS cross section. The likelihood function is defined as the product of the Poisson probabilities over all mass bins in the search region defined as m γγ > 409 GeV. In each bin the Poisson probability is evaluated for the observed number of data events given the model prediction. The model prediction is the sum of the expected background and the expected signal yield. The expected signal yield is a function of the graviton mass m G and the signal cross section times branching ratio σ × B(G → γγ). Systematic uncertainties are incorporated as nuisance parameters with Gaussian priors; the computation of the likelihood function includes integration over the variation in each of these. The RS model results can be combined with the results from the dilepton channel [3] to obtain limits on σ × B(G → γγ/ee/µµ). Relative branching ratios of B(G → ee)/B(G → µµ) = 1 and B(G → γγ)/B(G → ee) = 2 [48] are assumed for the purpose of this combination. Correlations between the systematic uncertainties for different channels are taken into account in the combination, as discussed in Ref. [11]. Specifically, the systematic uncertainty in the QCD dijet background is treated as fully correlated across the ee and γγ channels. The PDF and scale uncertainties are treated as correlated across all three channels, and affect the irreducible background in the γγ channel as well as the Drell-Yan background in the ee/µµ channels. The resulting limits are shown in Fig. 3. The limit can be interpreted in the plane (m G , k/M P l ) using the theoretical dependence of the cross section on these parameters (Fig. 4). Alternatively, for a given value of k/M P l , the limit can be translated into a limit on m G . The limits ¶ While the background is normalised to the data in the low-mass control region, the presence of a bump inside this region would still be visible. This motivates the choice of the lower limit of the search region.
on m G are summarised in Table 2. Using a constant K-factor of 1.75, the 95 % CL lower limit from the diphoton channel is 1.00 (2.06) TeV for k/M P l = 0.01 (0.1), and the combined 95 % CL lower limit is 1.03 (2.23) TeV for k/M P l =0.01 (0.1).
[TeV] G m 0.5 1 1.5 2 2.5 3 [pb]  . Expected and observed 95 % CL limits from the combination of G → γγ/ee/µµ channels on σ × B, the product of the RS graviton production cross section and the branching ratio for graviton decay via G → γγ/ee/µµ, as a function of the graviton mass. The -1σ and -2σ variations of the expected limit exhibit a tendency to be particularly close to the expected limit at large m G . This behaviour is expected as signals with large m G would manifest themselves in regions of m γγ where the SM background is small and the poissonian fluctuations around the mean expected background are highly asymmetric. The theory curves are obtained using the Pythia generator, which implements the calculations from Ref. [48]. A K-factor of 1.75 is applied on top of these predictions to account for NLO corrections. The thickness of the theory curve for k/M P l = 0.1 illustrates the theoretical uncertainties due to the PDFs expressed at 90% CL.
A counting experiment is performed to set limits on the ADD model. Specifically, the number of diphoton events is counted in a search region above a given threshold in m γγ . The mass threshold is chosen to optimise the expected limit on the difference in the diphoton cross section for m γγ > 500 GeV between the ADD model and the SM-only hypothesis. For the purpose of this optimisation, a specific implementation of the ADD model and specific values of the parameters have to be chosen. For M S = 2500 GeV    (2.23) in the GRW convention, an optimal search region is obtained at m γγ > 1217 GeV. In the data, 4 events are observed in this search region, with a background expectation of 2.32 ± 0.37 events. The expected and observed 95 % CL upper limits on the number of signal events above the SM background in the search region are given in Table 3. Table 3. Expected and observed limits at 95 % CL on the number of signal events in the search region at m γγ > 1217 GeV.

Expected limit
Observed limit −2σ −1σ mean +1σ +2σ 3.08 3.08 5. 18 5.96 8.53 7.21 These limits on the event yield can be translated into limits on the parameter η G of the ADD model using a prediction of the expected excess of events over the SM-only background as a function of η G . This translation is shown, for one specific implementation of the ADD model, in Fig. 5.  Figure 5. Number of signal events as a function of η G . The solid horizontal line corresponds to the observed limit, the dashed line corresponds to the expected limit and the bands to the ±1σ and ±2σ uncertainty on the expected limit. The black curve corresponds to MC predictions from various samples without applying the Kfactor of 1.70. The band around it illustrates the theoretical uncertainties due to the PDFs expressed at 90% CL. The samples at positive (negative) η G have been simulated using the GRW (Hewett) formalism. When the prediction is greater than the observed limit, the corresponding value for η G is excluded.
The limits + for different ADD scenarios are summarised in Table 4. Using a constant K-factor of 1.70, the 95 % CL upper (lower) limit on η G is 0.0085 (−0.0159) TeV −4 for constructive (destructive) interference. In the RS scenario, a lower limit of 1.00 (2.06) TeV at 95 % CL is set on the mass of the lightest RS graviton, for RS couplings of k/M P l = 0.01 (0.1). These new limits are a significant improvement over the previous best limit in the diphoton channel of 0.80 (1.85) TeV for k/M P l = 0.01 (0.1) [11]. The combination with the latest ATLAS dilepton searches [3] yields an improvement by a factor of 2-3 in the cross-section limits compared to Ref. [11] and further tightens the RS graviton mass limits to 1.03 (2.23) TeV for k/M P l = 0.01 (0.1).

Acknowledgements
We thank CERN for the very successful operation of the LHC, as well as the support staff from our institutions without whom ATLAS could not be operated efficiently.
We acknowledge the support of ANPCyT, Argentina; YerPhI, Armenia; ARC, Australia; BMWF and FWF, Austria; ANAS, Azerbaijan; SSTC, Belarus; CNPq and FAPESP, Brazil; NSERC, NRC and CFI, Canada; CERN; CONICYT, Chile; CAS, + The limit in the HLZ convention with n=2 may appear weak compared to the corresponding limit reported in Ref. [12]. When adopting the same search region as in Ref. [12], namely m γγ > 901 GeV instead of m γγ > 1217 GeV, our HLZ, n=2 limit for a K-factor=1 becomes M s =3.57 TeV, which is more stringent thant the result obtained with our default search region.

The ATLAS Collaboration
Hiroshima, Japan