Measurement of the production cross-section of a single top quark in association with a Z boson in proton–proton collisions at 13 TeV with the ATLAS detector

The production of a top quark in association with a Z boson is investigated. The proton–proton collision data collected by the ATLAS experiment at the LHC in 2015 and 2016 at a centre-of-mass energy of √ s = 13 TeV are used, corresponding to an integrated luminosity of 36 . 1 fb − 1 . Events containing three identiﬁed leptons (electrons and/or muons) and two jets, one of which is identiﬁed as a b -quark jet are selected. The major backgrounds are diboson, t ¯ t and Z + jets production. A neural network is used to improve the background rejection and extract the signal. The resulting signiﬁcance is 4 . 2 σ in the data and the expected signiﬁcance is 5 . 4 σ . The measured cross-section for tZq production is 600 ± 170 (stat.) ± 140 (syst.)fb. 3 .


Introduction
At hadron colliders, the top quark is typically produced in tt pairs through the strong interaction or as a single top or antitop quark through the electroweak interaction. The top quark was first observed via tt production at the Tevatron [1,2]. This was followed by the observation of single top-quark production [3][4][5] in the tand s-channels, also at the Tevatron. The associated tW production was first observed in 8 TeV proton-proton collisions at the Large Hadron Collider (LHC) [6,7]. These single-top-quark channels allow a direct determination of the dominant tW b vertex and of the magnitude of the CKM matrix element |V tb | [8] using their measured cross-sections.
With increasing energy and integrated luminosity, the ability to study rare Standard Model (SM) phenomena becomes possible. In the case of single top-quark production, examples include pp → t Zq [9] and pp → t H [10]. The pp → t Zq process involves W W Z and t Z couplings and has not been observed so far [11]. Fig. 1 shows typical lowest-order Feynman diagrams for the process. This channel probes two SM couplings in a single process, whereas the similar final state tt Z only probes the t Z coupling. The tt Z process has been measured by the ATLAS [12,13] and CMS [14] collaborations. In addition, the production of pp → t Zq is a SM background to the t H final state [10].
This Letter presents evidence of the production of a single top quark in association with a Z boson in the t-channel process pp → E-mail address: atlas .publications @cern .ch.
t Zq, where the Z boson decays into electrons or muons and the W boson from the top quark decays leptonically.

ATLAS detector
The ATLAS experiment [15] at the LHC is a multi-purpose particle detector with a forward-backward symmetric cylindrical geometry and a near 4π coverage in solid angle. 1 It consists of an inner detector (ID) surrounded by a thin superconducting solenoid providing a 2 T axial magnetic field, electromagnetic and hadron calorimeters, and a muon spectrometer. The inner detector covers the pseudorapidity range |η| < 2.5. It consists of silicon pixel, silicon micro-strip and transition radiation tracking detectors. The innermost pixel layer, the insertable B-layer, was added between Run 1 and Run 2 of the LHC, at a radius of 33 mm around a new, thinner, beam pipe [16]. Lead/liquid-argon (LAr) sampling calorimeters provide electromagnetic (EM) energy measurements with high granularity. A hadron (steel/scintillator-tile) calorimeter covers the central pseudorapidity range (|η| < 1.7). The end-cap and forward regions are instrumented with LAr calorimeters for both the EM and hadronic energy measurements up to |η| = 4.9. 1 ATLAS uses a right-handed coordinate system with its origin at the nominal interaction point (IP) in the centre of the detector and the z-axis along the beam pipe. The x-axis points from the IP to the centre of the LHC ring, and the y-axis points upwards. Cylindrical coordinates (r, φ) are used in the transverse plane, φ being the azimuthal angle around the z-axis. The pseudorapidity is defined in terms of the polar angle θ as η = − ln tan(θ/2). Distances in the η-φ plane are measured in units of R ≡ ( η) 2 + ( φ) 2   The muon spectrometer surrounds the calorimeters and is based on three large air-core toroidal superconducting magnets with eight coils each. The field integral of the toroids ranges between 2.0 and 6.0 T m across most of the detector. The muon spectrometer includes a system of precision tracking chambers and fast detectors for triggering. A two-level trigger system is used to select events. The first-level trigger is implemented in hardware and uses a subset of the detector information to reduce the accepted rate to at most 100 kHz. This is followed by a software-based trigger level that reduces the accepted event rate to 1 kHz on average.

Data and simulation samples
The pp collision data sample used in this measurement was collected with the ATLAS detector at the LHC during the 2015 and 2016 data-taking periods, corresponding to integrated luminosities of 3.3 fb −1 and 32.8 fb −1 , respectively, for a total of 36.1 fb −1 , after requiring that the detector is fully operational. Events are considered if they were accepted by at least one of the single-muon or single-electron triggers [17,18]. The electron triggers select a cluster in the calorimeter matched to a track. Electrons must then satisfy identification criteria based on a multivariate technique using a likelihood discriminant. In 2015, electrons had to satisfy a 'medium' identification requirement and have a transverse energy of E T > 24 GeV. In 2016, electrons had to satisfy a 'tight' identification together with an isolation criterion and have E T > 26 GeV.
To avoid efficiency loss due to isolation at high E T , an additional trigger was used, selecting 'medium' electrons with E T > 60 GeV.
Muons are triggered on by matching tracks reconstructed in the muon spectrometer and in the inner detector. In 2015, muons had to satisfy a 'loose' isolation requirement and have a transverse momentum of p T > 20 GeV. In 2016, the isolation criteria were tightened and the threshold increased to p T = 26 GeV. In both years, another muon trigger without any isolation requirement was used, selecting muons with p T > 50 GeV.
In order to evaluate the effects of the detector resolution and acceptance on signal and background and to estimate the SM background, a full Geant4-based detector simulation was used [19,20]. Event generators were used to estimate the expected signal and background contributions and their uncertainties. The topquark mass in the event generators described below was set to 172.5 GeV. Multiple inelastic pp collisions (referred to as pile-up) are simulated with Pythia 8.186 [21], and are overlaid on each Monte Carlo (MC) event. Weights are assigned to the simulated events such that the distribution of the number of pile-up interactions in the simulation matches the corresponding distribution in the data. All simulation samples are processed through the same reconstruction algorithms as the data.
Monte Carlo t Zq signal samples were generated at leading order (LO) in QCD using MG5_aMC@NLO 2.2.1 [22] in the four-flavour scheme, treating the b-quark as massive, with the CTEQ6L1 [23] LO parton distribution functions (PDFs). The Z boson was simulated to be on-shell and off-shell Z /γ * contributions and their interference are not taken into account. Following the discussion in Ref. [24], the renormalisation and factorisation scales (μ r and where the b-quark is the external one produced from gluon splitting in the event. This choice is motivated by the total scale dependence being dominated by this external b-quark, shown in Fig. 1. The parton shower and the hadronisation of signal events were simulated with Pythia 6 [25] using the Perugia2012 set of tuned parameters [26]. The t Zq total cross-section, calculated at next-to-leading order (NLO) using MG5_aMC@NLO 2.3.3 with the NNPDF3.0_nlo_as_0118 [27] PDF, is 800 fb, with an uncertainty of +6.1 −7.4 %. The uncertainty is computed by varying the renormalisation and factorisation scales by a factor of two and by a factor of 0.5.
A comparison of the event kinematics before parton showering between the LO MG5_aMC@NLO 2.2.1 sample and a sample generated using NLO MG5_aMC@NLO 2.3.3 showed agreement within 10%, justifying the use of a LO sample for the detector simulation.
Monte Carlo simulated events are used to estimate the SM background that can produce three leptons and at least two jets in the final state. In tt production, if both W bosons decay into leptons (referred to as 'prompt') and either a b-or c-hadron decays into a lepton (referred to as 'non-prompt') that is isolated, the final state can mimic the t Zq final state. The nominal tt simulated sample was generated at NLO with the Powheg-Box [28][29][30] event generator using the CT10 PDFs [31]. The cut-off parameter, h damp , for the first emission of gluons was set to the top-quark mass. The events were then processed using Pythia 6 to perform the fragmentation and hadronisation, and to generate the underlying event.
Events from the associated production of a tt pair and a boson (W / Z /H ) provide additional modes for the production of leptons in the final state. For tt + W the MC simulated events were generated using MG5_aMC@NLO 2.2.2 [22], while the tt + H and tt + Z MC simulated events were generated using MG5_aMC@NLO 2.2.3. The generated events were then processed with Pythia 8 [21] to perform the fragmentation and hadronisation, and to generate the underlying event, using the NNPDF2.3LO PDF set and the A14 tune [32].
Processes that include the production of W W , W Z and Z Z events were simulated using Sherpa 2.1.1 at LO with up to three additional partons and the CT10 PDF set. In the trilepton topology, the diboson background consists mainly of W Z events, while the contribution to the background from W W final states, corresponding to the case where a jet is misidentified as a lepton, is negligible. The Z Z background gives a small contribution of 9% of all diboson events. The gluon-induced diboson production, which amounts to about 10% of the quark-induced diboson production, is therefore negligible in the t Zq signal region, and is not included in the diboson samples. In order to estimate the systematic uncertainty, additional diboson samples were simulated using the Powheg-Box generator in combination with Pythia 8 and the CTEQ6L1 PDF sets.
Of the aforementioned single-top-quark production channels, only the t W channel contributes to the trilepton final state. This sample was produced using the NLO Powheg-Box event generator with the CT10 PDF set. The events were then processed with Pythia 6 to perform the fragmentation and hadronisation, and produce the underlying event. A sample of tW Z events was produced using the MG5_aMC@NLO 2.2.3 generator and showered with Pythia 8, using the NNPDF3.0_NLO PDF set and the A14 tune.

Object reconstruction
The reconstruction of the basic physics objects used in this analysis is described in the following. The primary vertex is chosen as the proton-proton vertex candidate with the highest sum of the squared transverse momenta of all associated tracks with p T > 400 MeV.
Electron candidates are reconstructed from energy deposits in the electromagnetic calorimeter that match a reconstructed track [33][34][35][36]. The clusters are required to be within |η| < 2.47 excluding the transition region between the barrel and end-cap calorimeters at 1.37 < |η| < 1.52. Electron candidates must also satisfy a transverse energy requirement of E T > 15 GeV. A likelihood-based discriminant is constructed from a set of variables that enhance the electron selection, while rejecting photon conversions and hadrons misidentified as electrons [34]. An |η|and p T -dependent selection on the likelihood discriminant is applied, such that it has an 80% efficiency when used to identify electrons from the Z -boson decay. This working point corresponds to an approximate rejection factor against jets of 700 at a p T of 40 GeV. Electrons are further required to be isolated using criteria based on ID tracks and topological clusters in the calorimeter, with an isolation efficiency of 90(99)% for p T = 25(60) GeV. Correction factors are applied to simulated electrons to take into account the small differences in reconstruction, identification and isolation efficiencies between data and MC simulation.
Muon candidates are required to have |η|<2.5 and p T >15 GeV, and are reconstructed by combining a reconstructed track from the inner detector with one from the muon spectrometer [37]. To reject misidentified muon candidates, primarily from pion and kaon decays, several quality requirements are imposed on the muon candidate. An isolation requirement based on ID tracks and topological clusters in the calorimeter is imposed, and results in an isolation efficiency of 90(99)% for p T = 25(60) GeV. The overall efficiency obtained for muons from W -boson decays in simulated pp → tt events is 96% and the rejection factor for non-prompt muons with p T > 20 GeV is approximately 600. As for electrons, correction factors are applied to muons to account for the small differences between data and simulation.
Jets are reconstructed from topological clusters using the anti-k t algorithm [38,39] with the radius parameter set to R = 0.4. They are reconstructed for p T > 30 GeV in the region with |η| < 4.5. To account for inhomogeneities and the non-compensating response of the calorimeter, the reconstructed jet energies are corrected using p T -and η-dependent factors that are derived in MC simulation and validated in data. Any remaining differences in the jet energy scale are corrected using in situ techniques, where a well-defined reference object is momentum-balanced with a jet [40]. To suppress pile-up, a discriminant called the jet-vertex-tagger (JVT) is constructed using a two-dimensional likelihood method [41]. For jets with p T < 60 GeV and |η| < 2.4 a JVT requirement corresponding to a 92% efficiency, while rejecting 98% of jets from pile-up and noise, is imposed. To identify jets containing a b-hadron (b-tagging), a multivariate algorithm is employed [42]. This algorithm uses the impact parameter and reconstructed secondary vertex information of the tracks contained in the jet as input for a neural network. Due to its use of the inner detectors, the reconstruction of b-jets is done in the region with |η| < 2.5. Jets initiated by b-quarks are selected by setting the algorithm's output threshold such that a 77% b-jet selection efficiency is achieved in simulated tt events. With this setting, the misidentification rate for jets initiated by light-flavour quarks or gluons is 1%, while it is 17% for jets initiated by c-quarks [43]. Correction factors are derived and applied to correct for the small differences in b-quark selection efficiency between data and MC simulation [42].
The missing transverse momentum, with magnitude E miss T , is calculated as the negative of the vector sum of the transverse momenta of all reconstructed objects, p miss T . In addition to the identified jets, electrons and muons, a track-based 'soft' term is included in the p miss T calculation, by considering tracks associated with the primary vertex in the event but not with an identified jet, electron, or muon [44,45].
To avoid cases where the detector response to a single physical object is reconstructed as two separate final-state objects, several steps are followed to remove such overlaps, following Ref. [46].

Signal, control and validation regions
The reconstructed t Zq final state consists of three charged leptons (electron and/or muon), a b-tagged jet, an additional jet and E miss T . Reconstructing the Z boson and the top quark is important in order to identify specific features that help to separate the signal from the background. For example, the Z -boson mass distributions can contribute to the reduction of top-quark backgrounds, as these do not include a Z boson in the final state, while the untaggedjet pseudorapidity distribution differs in shape between t Zq signal events and diboson and tt Z events, which constitute some of the largest backgrounds.
The signal region (SR) definition reflects the t Zq final state by selecting only events that have exactly three charged leptons, one b-tagged jet and one additional jet, referred to as the untagged jet as no b-tagging requirement is applied. In order to better separate the t Zq signal from background, additional requirements are imposed on the properties of the selected objects. The three leptons are sorted by their p T , irrespective of flavour, and required to have transverse momenta of at least 28, 25 and 15 GeV, respectively.
Both jets are required to have p T > 30 GeV.
An opposite-sign, same-flavour (OSSF) lepton pair is required in order to reconstruct the Z boson. In the μee and eμμ channels, the pair is uniquely identified. For the eee and μμμ events, both possible combinations are considered and the pair that has the invariant mass closest to the Z -boson mass is chosen. The W boson is reconstructed from the remaining lepton and the missing transverse momentum, using as constraint the W -boson mass to evaluate the z component of the neutrino momentum. 2 The top quark is reconstructed from the reconstructed W boson and the b-tagged jet.
To suppress background sources that do not contain a Z boson, the invariant mass of the leptons is required to be between 81 and 101 GeV. Because a W boson is expected in the final state, Table 1 Overview of the requirements applied for selecting events in the signal, validation and control regions.

Common Selections
Exactly 3 leptons with |η| < 2.5 and p T > 15 GeV No OSSF pair 2 jets, |η| < 4.5 1 j e t , |η| < 4.5 2 j e t s , |η| < 4.5 2 j e t s , |η| < 4.5 The selection criteria that define the SR are summarised in Table 1. In total, 141 events are selected using these criteria. The criteria are modified to define validation regions, which are used to check the modelling of the main background contributions. Two validation regions (VR) are defined as follows: the diboson VR uses the same event selection as the SR, except that only one jet is required in the event and no b-tagging requirement is applied. The tt VR also uses the same selection as the SR, except that the invariant mass of the OSSF pair must be outside the Z -mass window (m < 81 GeV or m > 101 GeV). In addition, two control regions (CR) are defined, from which the normalisations of the diboson and the tt background sources are computed, as explained in Section 6. The diboson CR is defined in the same way as the diboson VR, except with a tighter requirement on m T ( W , ν). The tt CR instead has the same selection as the SR but it requires an oppositesign, different-flavour (OSDF) lepton pair and rejects events with an OSSF pair.

Background estimation
Different SM processes are considered as background sources for this analysis. These are either processes such as diboson or tt V + tt H production, in which three or more prompt leptons are produced, or processes with only two prompt leptons in the final state (such as Z + jets and tt production) and one additional non-prompt or 'fake' lepton that meets the selection criteria. Such non-prompt or fake leptons can originate from decays of bottom or charm hadrons, a jet that is misidentified as an electron, leptons from kaon or pion decays, or electrons from photon conversions.
The dominant source of background originates from diboson production. This consists mainly of W Z events with a small fraction of Z Z events in which the fourth lepton is missed (roughly 9% of the total number of diboson events). Studies in the diboson VR indicated that the number of events predicted by the Sherpa MC samples is lower than the number observed. The kinematic distributions are otherwise well described. Hence, the total number of diboson events predicted by the Sherpa samples is scaled by a factor of 1.47, leading to an expected number of diboson events in the SR of 53. The scale factor is derived from the diboson CR, defined in Section 5, by computing the data-to-MC ratio for events that satisfy the condition m T (W ) > 60 GeV. This selection is applied in order to reduce the Z + jets contamination and ensure a dibosondominated region. The uncertainty in the scale factor is estimated 3 The transverse mass is calculated using the momentum of the lepton associated with the W boson, p miss T and the azimuthal angular difference between the two: In order to take into account a possible difference between data and MC simulation for tt events, the number of events containing a non-prompt or fake lepton in the MC simulation is scaled by a data/MC factor that is derived in the tt CR defined in Section 5. This tt control region and the signal region have very similar non-prompt lepton compositions. Requiring a pair of oppositesign, different-flavour leptons, and rejecting events with an OSSF pair, ensures that there is no contamination from Z + jets events and from the SR. Different electron-muon invariant mass windows around the Z mass, with widths ranging from 20 GeV to 60 GeV, were investigated and the average of the obtained factors is used for scaling the tt background in the signal region. The total uncertainty in the scaling factor is calculated taking into account this variation and the statistical uncertainty of the sample. This leads to a data/MC scale factor of 1.21 ± 0.51. Deriving separate factors depending on the fake lepton's flavour or on the lepton p T was also investigated. All approaches are consistent with each other within the assigned uncertainties. The expected number of tt events in the SR is 18 ± 9. According to the MC prediction, the t W contribution is found to be less than one event.
A data-driven technique called the fake-factor method is used to estimate the Z + jets background contribution. A region defined by selecting events with m T (W ) < 20 GeV is used for deriving the fake factors. Since it is observed that the number of non-prompt or fake electrons and muons can be very different, the estimation is done separately for the electron and muon channel. Fake factors are defined as the ratio of data events that have three isolated leptons to events in which one of the leptons fails the isolation requirement. They are derived in bins of the p T of the lepton not associated with the Z boson. According to MC simulation, this lepton is in over 95% of the cases the non-prompt or fake lepton. These factors are then applied to events passing the signal region selection (including a m T (W ) > 20 GeV cut) that have one of the three leptons failing the isolation requirement. Contamination from other background sources, which is about 50% and mainly coming from tt, is taken into account and subtracted before making the final Z + jets estimate. The expected number of Z + jets events in the SR is 37. Different sources of uncertainty are investigated, including consistency checks of the fake-factor method using MC Z + jets samples, the effect of changing the diboson scale factor and the statistical uncertainties in the estimated and observed number of events. All these amount to a total uncertainty of 40%. The expected tt V , tt H and t W Z contributions are evaluated from the MC samples normalised to their predicted NLO crosssections [22]. The tt V + tt H contribution is approximately 10% of the total background estimate, while t W Z events amount to 3%.
The expected number of tt V + tt H + tW Z events is 20 ± 3. The uncertainty in the predictions is taken to be 13% [22].

Multivariate analysis
A multivariate analysis is used to separate the signal from the large number of background events. The neural-network package NeuroBayes [47,48] is used, which combines a three-layer feedforward neural network with a complex robust preprocessing. Several variables are combined into one discriminant, then mapped onto the interval [0, 1], such that background-like events have an output value, O NN , closer to 0 and signal-like events have an output closer to 1. All background processes are considered in the training except tt production, due to the very small number of available MC events that meet the selection criteria. Only variables that provide separation power and are well modelled are taken into account in the final neural network (NN). For the NN training, the ten variables with the highest separation power are used. These variables are explained in the order of their importance in Table 2 Variables used as input to the neural network, ordered by their separation power.
Absolute value of η of the lepton coming from the W -boson decay Table 2. They include simple variables, such as the p T and η of jets and of the lepton not associated with the Z boson. Information about the reconstructed W boson, Z boson and top quark, such as their p T as well as their masses, is also used. In addition, the R between the untagged jet and the Z boson is employed as an input. The modelling of the input variables is checked both in the validation regions defined in Table 1 and in the signal region. The distributions of some input variables in the signal region are shown in Fig. 2, normalised to the expected number of events, including the scale factors determined in Section 6. Good agreement between data and the prediction is observed.
The output of the NN is checked in the validation regions, shown in Fig. 3. Good agreement between the expected and ob-  served numbers of events and in the shape of the NN output distribution are seen, demonstrating reliable background modelling. NeuroBayes includes extensive protection against overtraining and several further checks confirm that it functions well.

Systematic uncertainties
Systematic uncertainties in the normalisation of the individual backgrounds and in the signal acceptance, as well as uncertainties in the shape of the NN distributions, are taken into account when determining the t Zq cross-section. For uncertainties where variations as a function of the NN distribution are consistent with being due to statistical fluctuations, only the normalisation difference is taken into account. The uncertainties are split into the following categories: Reconstruction efficiency and calibration uncertainties Systematic uncertainties affecting the reconstruction and energy calibration of jets, electrons and muons are propagated through the analysis. The dominant sources of uncertainty for this measurement are the jet energy scale (JES) calibration, including the modelling of pile-up, and the b-jet tagging efficiencies.
The uncertainties due to lepton reconstruction, identification, isolation requirements and trigger efficiencies are estimated using tag-and-probe methods in Z → events. Correction factors are derived to match the simulation to observed distributions in collision data and associated uncertainties are estimated. Uncertainties in the lepton momentum scale and resolution are also assessed using Z → events [34,37,49].
Several components of the JES uncertainty are considered [40,50]. Uncertainties derived from different dijet-p T -balance measurements as well as uncertainties associated with other in situ calibration techniques are considered. Furthermore, the presence of nearby jets and the modelling of pile-up affect the jet calibration. The uncertainty in the flavour composition covers effects due to the difference in quark-gluon composition between the jets used in the calibration and the jets used in this analysis. Also an uncertainty due to the different calorimeter responses to light-quark and gluon jets is taken into account. Finally, the JES uncertainty is Signal PDF and radiation The systematic effects due to uncertainties in the parton distribution functions are taken into account for the signal. As it was generated at LO, the uncertainty is evaluated using the 30 eigenvectors of the NNPDF3.0_lo_as_0118 [27] PDF set, in the four-flavour scheme. The events are reweighted according to each of the PDF uncertainty eigenvectors. As a crosscheck, the PDF uncertainty is also evaluated following the updated PDF4LHC recommendation [54] by using the PDF4LHC15 NLO PDF set. This has a smaller effect; hence the uncertainty from the LO PDF set is used.
Variations of the amount of additional radiation are studied by changing the hard-scatter scales and the scales in the parton shower simultaneously in the t Zq sample. A variation of the factorisation and renormalisation scale by a factor of two is combined with the Perugia2012 set of tuned parameters with lower radiation (P2012radLo) than the nominal set; while a variation of both scales by a factor of 0.5 is combined with the Perugia2012 set of tuned parameters with higher radiation (P2012radHi).
Luminosity The uncertainty in the combined 2015 + 2016 integrated luminosity is 2.1%. It is derived, following a methodology similar to that detailed in Ref.
[55], from a calibration of the luminosity scale using x-y beam-separation scans performed in August 2015 and May 2016.
The effects of the above uncertainties on the number of signal events are summarised in Table 3. This does not include the impact of the background uncertainties.
Background The uncertainties in the normalisation of the various background processes use the uncertainty estimated in Section 6. For the tt sample, the systematic effects due to uncertainties in the scale and the amount of radiation are included. Table 3 Breakdown of the impact of the systematic uncertainties on the number of t Zq signal events in order of decreasing effect. Details of the systematic uncertainties are provided in the text. MC statistics refers to the effect of the limited size of the MC samples used.

Results
Using the 141 selected events, a maximum-likelihood fit is performed to extract the t Zq signal strength, μ, defined as the ratio of the measured signal yield to the NLO Standard Model prediction. The statistical analysis of the data employs a binned likelihood function L(μ, θ), constructed as the product of Poisson probability terms, to estimate μ [56]. The likelihood is maximised on the NN output distribution in the signal region. The background normalisations are allowed to vary within the uncertainties given in Section 6. The impact of systematic uncertainties on the expected numbers of signal and background events is described by nuisance parameters, θ , which are each parameterised by a Gaussian or log-normal constraint for each bin of the NN output distribution. If the variation of the uncertainty in each bin is consistent with being due to statistical fluctuations, only the overall change in normalisation is included as a nuisance parameter. The uncertainties are set to be symmetric in the fit, using the average of the variations up and down. The expected numbers of signal and background events in each bin are functions of θ . The test statistic, q μ , is constructed according to the profile likelihood ratio: , where μ and ˆ θ are the parameters that maximise the likelihood, and ˆ θ are the nuisance parameter values that maximise the likelihood for a given μ. This test statistic is used to determine a probability for accepting the backgroundonly hypothesis for the observed data. Fig. 4 shows the NN discriminant in the signal region with background normalisations, signal normalisation and nuisance parameters adjusted by the profile likelihood fit.
The results for the numbers of fitted signal and background events are summarised in Table 4. The table also shows the result of a fit to the Asimov dataset [56]. The total uncertainty in the number of fitted events includes the effect of correlations, which are large among the background sources, as the O NN distributions have a similar shape. The strongest correlation is found to be between the diboson and the Z + jets contributions and it is about −0.5 for both the Asimov dataset and the data.
After performing the binned maximum-likelihood fit and estimating the total uncertainty, the fitted value for μ is 0.75 ± 0.21 (stat.) ± 0.17 (syst.) ± 0.05 (th.). The quoted theory (th.) uncertainty in μ includes the t Zq NLO cross-section uncertainty given in Section 3. This is not taken into account when evaluating the cross-section. The statistical uncertainty in the crosssection is determined by performing a fit to the data, including only the statistical uncertainties. The total systematic uncertainty is determined by subtracting this value in quadrature from the total uncertainty. The cross-section for t Zq production is measured

Table 4
Fitted yields in the signal region for the Asimov dataset and the data. The fitted numbers of events contain the statistical plus systematic uncertainties.
The probability p 0 of obtaining a result at least as signal-like as observed in the data if no signal were present is calculated using the test statistic q μ=0 in the asymptotic approximation [56]. The observed p 0 value is 1.3 × 10 −5 . The resulting significance is 4.2σ , to be compared with the expected significance of 5.4σ .

Conclusion
The cross-section for t Zq production has been measured using 36.1 fb −1 of proton-proton collision data collected by the ATLAS experiment at the LHC in 2015 and 2016 at a centre-of-mass energy of √ s = 13 TeV. Evidence for the signal is obtained with a measured (expected) significance of 4.2σ (5.4σ