Measurement of the CP properties of Higgs boson interactions with τ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tau $$\end{document}-leptons with the ATLAS detector

A study of the charge conjugation and parity (CP\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\textit{CP}$$\end{document}) properties of the interaction between the Higgs boson and τ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tau $$\end{document}-leptons is presented. The study is based on a measurement of CP\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\textit{CP}$$\end{document}-sensitive angular observables defined by the visible decay products of τ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tau $$\end{document}-leptons produced in Higgs boson decays. The analysis uses 139 fb-1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$^{-1}$$\end{document} of proton–proton collision data recorded at a centre-of-mass energy of s=13\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sqrt{s}= 13$$\end{document} TeV with the ATLAS detector at the Large Hadron Collider. Contributions from CP\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\textit{CP}$$\end{document}-violating interactions between the Higgs boson and τ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tau $$\end{document}-leptons are described by a single mixing angle parameter ϕτ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\phi _{\tau }$$\end{document} in the generalised Yukawa interaction. Without constraining the H→ττ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$H\rightarrow \tau \tau $$\end{document} signal strength to its expected value under the Standard Model hypothesis, the mixing angle ϕτ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\phi _{\tau }$$\end{document} is measured to be 9∘±16∘\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$9^{\circ } \pm 16^{\circ }$$\end{document}, with an expected value of 0∘±28∘\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$0^{\circ } \pm 28^{\circ }$$\end{document} at the 68% confidence level. The pure CP\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\textit{CP}$$\end{document}-odd hypothesis is disfavoured at a level of 3.4 standard deviations. The results are compatible with the predictions for the Higgs boson in the Standard Model.


Introduction
A detailed analysis of Higgs boson ( ) decays into -lepton pairs observed at the LHC [1-3] allows a direct probe of the charge conjugation and parity ( ) properties of the Yukawa coupling of the Higgs boson to the -lepton. The Standard Model (SM) of particle physics predicts the Higgs boson to be a -even (scalar) particle. The presence of a -odd (pseudoscalar) admixture has not yet been excluded, and any observed -odd contribution to the coupling properties would be a sign of physics beyond the SM.
Studies of the properties of Higgs boson interactions with gauge bosons performed by the ATLAS and CMS experiments [4][5][6][7][8][9] have shown no deviation from the SM predictions. However, these measurements probe the bosonic couplings in which -odd contributions enter only via higher-order operators that are suppressed by powers of 1/Λ 2 [10], where Λ is the scale of the new physics in an effective field theory. In contrast, a -odd contribution to Yukawa couplings can be present at tree level. Recently, measurements of the properties of the interaction between the Higgs boson and top quarks were performed by the ATLAS [11] and CMS [12] Collaborations, and excluded a pure -odd structure for the top-quark Yukawa coupling at 3.9 and 3.2 , respectively. This paper presents a measurement of the properties of the Higgs boson interaction with -leptons. The measurement is based on -sensitive angular observables defined using the visible -lepton decay products. Ideas about how to probe a -odd and -even admixture in the -lepton Yukawa coupling in → decay were initially developed in the context of + − colliders [13][14][15][16][17]. Originally, hadronic decays of the -leptons into ± and ± were used, and observables sensitive to the transverse spin correlations between the -lepton decay products were constructed. These methods, extended to ℓ ± (= ± , ± ) and ± 1 decays and re-evaluated in the context of collisions at the LHC [18][19][20][21][22][23], are adopted in this analysis. Recently, a similar study was also performed by the CMS Collaboration [24].
The general effective Yukawa interaction between the Higgs boson and -leptons can be parameterised as in Refs. [21,23]: where = 246 GeV is the vacuum expectation value of the Higgs field, is the reduced Yukawa coupling strength, and (where ∈ [−90 • , 90 • ]) is the -mixing angle that parameterises the relative contributions of the -even and -odd components to the coupling. The SM -even hypothesis is realised for = 0 • , while the pure -odd scenario corresponds to = ±90 • . Other values of represent admixtures of the two components and would indicate a -violating scenario.
The -mixing angle is encoded in the correlations between the transverse spin components of the -leptons in the → decays, which are then reflected in the directions of the -lepton decay products. The signed acoplanarity angle * CP between the decay planes (described in Section 3 and illustrated in Figure 1) is sensitive to the transverse spin correlations impacted by the -mixing angle of the Yukawa coupling. Such correlations are usually calculated by contracting the polarimeter vectors of the decayed -leptons 1 and the spin density matrix of the -lepton-pair spin state, , , which depends on the -lepton pair-production process [26][27][28]. In the case of Higgs boson decays, the spin density matrix , has only transverse components with respect to the -lepton direction, and these are first-order trigonometric Table 1: Notation for the dominant leptonic and hadronic -lepton decay modes used and their branching fractions. The symbol 'ℓ ± ' stands for ± or ± , and 'ℎ ± ' includes ± and ± . The parentheses show the hadronic decays involving ± and their corresponding branching fractions.
The analysis is performed using 139 fb −1 of Figure 1: Illustration of the -lepton decay planes for constructing the * CP observable in (a) → + − → + − + 2 decay using the impact parameter method, (b) → + − → + 0 − 0 using the -decay plane method, and (c) → + − → + 0 − using the combined impact parameter and -decay plane method. The decay planes are spanned by the spatial momentum vector of the charged decay particle of the -lepton ( ± ) and either its impact parameter n * ± or the spatial momentum vector of the neutral decay particle of the -lepton ( 0 ).  Figure 2: Normalised * CP distributions in simulated → + − → + − + 2 events at the generator level for different hypotheses. The predictions for a pure -even SM Higgs boson (scalar, red circle), a pure -odd hypothesis (pseudoscalar, green square), and -mix hypothesis ( = 45 • , blue triangle) are shown. The transverse momentum of the simulated leptons is required to be larger than 30 GeV (20 GeV) for the leading (sub-leading) lepton during the event generation.

Impact parameter (IP) method
The IP method is applied to -lepton decays with only one charged particle in the final state, specifically the direct hadronic decay ± → ± or leptonic decays ± → ℓ ± . This refers to the 1p0n-1p0n and ℓ-1p0n decay mode combinations. In this case, the -lepton decay plane is formed from the spatial momentum vector q ± of the charged particle ( ± , ℓ ± ) and the three-dimensional (3D) impact parameter vector n ± of the charged particle, defined as the directional distance of closest approach of the charged particle's track to the reconstructed primary vertex (PV) of the event. The four-vectors of the track momentum ± and the impact parameter ± = (0, n ± ), initially defined and measured in the laboratory frame, are boosted to the rest frame of the two charged decay particles (the visible di-ZMF, denoted by * ). The boosted and normalised impact parameter vectorn * ± is then decomposed into components which are parallel and transverse (n * ± ⊥ ) to the direction of the associated normalised spatial momentum vectorq * ± . Using these vectors, an angle * and a CP-odd triple correlation O * are defined as * = arccos(n * + ⊥ ·n * − ⊥ ) and O * =q * − · (n * + ⊥ ×n * − ⊥ ), and both are incorporated in a single observable * (1) In the case of leptonic decay, due to a different sign in the spectral function for the leptonic decays [20,29], an additional shift by 180 • is applied to synchronise the phase in * CP with the other decays.

-decay plane ( ) method
The method is applied to construct * CP in events with 1p1n-1p1n or 1p1n-1pXn decay mode combinations. In the case of consecutive decays ± → ± , ± → ± 0 , the -lepton decay plane can be formed from the spatial momentum vectors of the charged pion (q ± ) and neutral pion (q 0± ). The four-momentum vectors of the ± and 0 are boosted to the rest frame of the -meson pair (the visible -lepton-pair ZMF). The angle * and triple correlation O * are then defined in the same way as in the IP method using the unit spatial vectors, but replacing the impact parameter component with the neutral-pion vector, * = arccos(q * 0+ ⊥ ·q * 0− ⊥ ) and O * =q * − · (q * 0+ ⊥ ×q * 0− ⊥ ), whereq * 0+ ⊥ andq * 0− ⊥ are the normalised vectors transverse to the direction of the associated charged pion for each neutral pion. A signed observable * is defined similarly to Eq. (1), An additional requirement that depends on the sign of the product of -lepton spin-analysing functions ± = ( ± − 0 )/( ± + 0 ), where ±,0 is the pion energy in the laboratory frame, is needed to define the observable * CP sensitive to the -mixing angle as * (2) In the case of 1pXn decays (e.g. ± → ± 1 → ± 2 0 ), the sum of the four-momenta of all neutral pions is taken as the neutral component in the method.

Combined IP and (IP-) method
For events with the combinations 1p0n-1p1n, 1p0n-1pXn, ℓ-1p1n, and ℓ-1pXn, the IP method and the method are combined to compute the * CP angle. In the case of → → ∓ ± , ± → ± 0 (1p0n-1p1n) events, * CP is defined in the visible -lepton-pair ZMF by using the rest frame. One of the decay planes is defined using the IP method and the other with the method. The quantities * , O * and * are calculated in a way analogous to that in Section 3.2, but with the one-neutral-pion component replaced with the impact parameter vector defined in Section 3.1. The * CP observable is defined as * with a phase shift of 180 • depending on the sign of .

1 -decay ( 1 ) method
The 1 method is an extension of the method discussed in Section 3.2, and is used for ± → ± 1 , ± 1 → ± + − . The -lepton decay plane is defined by the charged pion ( ± 1 ) with the highest transverse momentum and the vector sum of the other two momenta. The observable ± used in the method is modified to take the effect of the masses into account, and is defined by 1 ± adopting the convention in Ref. [36] as where 3 is the invariant mass of the three charged pions from the ± 1 decay, and 2 ( 2 ) refers to the invariant mass (energy) of the system of the two in the decay that do not have the highest transverse momentum.
Similarly to Section 3.3, the 1 method is combined with the IP method for the ℓ-3p0n events and with the method for the 1p1n-3p0n events in the lep had and had had channels, respectively. In the ℓ-3p0n case, 1 ∓ is used to determine whether the 180 • shift is applied.

Data and simulated event samples
This analysis uses 139 fb −1 of √ = 13 TeV collision data recorded by ATLAS with good operating conditions [37] in 2015-2018. The data were collected using single-lepton or di-hadronic triggers [38][39][40][41]. Events used in the lep had channel were accepted by single-lepton triggers with T thresholds of 24 GeV (26 GeV) and 20 GeV (26 GeV) for electron and muon candidates, respectively, in the 2015 (2016-2018) dataset. Events used in the had had analysis channel were accepted by di-hadronic -lepton triggers, with a T threshold of 35 GeV for the leading candidate and 25 GeV for the sub-leading candidate. Due to rising instantaneous luminosity, the di-hadronic triggers used in the 2016-2018 data-taking period required an additional first-level triggered jet with T > 25 GeV and | | < 3.2.
The analysis considers the four main Higgs boson production processes at the LHC: gluon-gluon fusion (ggF), vector-boson fusion (VBF), associated production with a vector boson ( ), and associated production with top-quark pair (¯). The P NNLOPS program [42-46] was used to model ggF Higgs boson production with next-to-next-leading-order (NNLO) accuracy. The VBF and production processes were simulated with P at NLO accuracy in QCD. The production of¯events was simulated using P B v2 [44][45][46][47][48]  All samples of simulated events were passed through a full simulation of the ATLAS detector response [56] using G 4 [57]. The effects of multiple interactions in the same or neighbouring bunch crossings (pile-up) were modelled by overlaying each hard-scatter event with minimum-bias events, simulated using the soft QCD processes of P 8. The simulated events were then weighted such that the distribution of the average number of interactions per bunch crossing matches the one observed in data.

Event selection and background estimation
The analysis of -mixing in the → channel requires the selection of signal events characterised by the presence of isolated leptons, visible decay products in hadronic decays, jets and missing transverse momentum. Events with an isolated lepton (electron or muon) and a hadronic decay are used for the lep had analysis channel, whereas events with two hadronic decays are used for the had had channel. Additional jets, the invariant mass and the transverse momenta of the lepton pairs are used for event categorisation (Section 5.2). Various kinematic and identification requirements are used to suppress the background.

Objects and decay mode reconstruction
The analysis shares the same object reconstruction and identification algorithms as the → cross-section measurement [55]. Here the most important features specific to this measurement are recalled: • Tracks measured in the ID are used to reconstruct interaction vertices [58], of which the one with the highest sum of squared transverse momenta 2 T of the associated tracks is selected as the primary vertex of the hard interaction.
• Electrons are reconstructed from topological clusters of energy deposits in the electromagnetic calorimeter which are matched to a track reconstructed in the ID [59] and are required to satisfy 'loose' isolation and 'medium' identification criteria.
• Muons are reconstructed from signals in the muon spectrometer matched with tracks inside the ID. They are required to satisfy 'loose' identification and 'tight' isolation criteria based on track information [60].
• Jets are reconstructed using a particle-flow algorithm [61]. It applies an anti-algorithm [62,63] with distance parameter = 0.4 to noise-suppressed positive-energy topological clusters in the calorimeter after subtracting energy deposits associated with primary-vertex-matched tracks, and including the track momenta instead in the clustering, thereby improving the jet energy measurement. Cleaning criteria are used to identify jets arising from non-collision backgrounds or noise in the calorimeters [64]. A dedicated jet-vertex-tagger algorithm [65] is used to remove jets that are identified as not being associated with the primary vertex of the hard interaction. Similarly, a dedicated algorithm is used to suppress pile-up jets in the forward region [66].
• Hadronic -lepton decays are reconstructed from the visible decay products (neutral pions or charged pions/kaons). The reconstruction is seeded by jets reconstructed with an anti-algorithm using calibrated topological clusters [67] as inputs and a distance parameter of = 0.4 [68]. Reconstructed nearby tracks are matched to a hadronic candidate ( had-vis ) if they exceed the value required for a multivariate discriminant determining the likelihood that the tracks are produced from the had-vis decay. A recurrent neural network identification algorithm [69] is trained to separate the had-vis candidates from jets initiated by quarks or gluons, and boosted decision tree (BDT) discriminants are used to help reject misidentified hadronic decays due to electrons. In both cases, the had-vis is required to satisfy 'medium' criteria. Reconstructed had-vis objects are required to have one or three associated tracks, and have T > 20 GeV and | | < 2.47, excluding the transition region between the barrel and endcap electromagnetic calorimeters (1.37 < | | < 1.52).
• The missing transverse momentum (with magnitude miss T ) is reconstructed as the negative vector sum of the transverse momenta of leptons, had-vis objects, jets, and a 'soft-term'. The 'soft-term' is calculated as the vector sum of the T of tracks matched to the primary vertex but not associated with reconstructed leptons, had-vis objects, or jets [70]. The transverse (longitudinal) impact parameter 0 ( 0 ) of a charged-particle track is defined as the distance in the transverse plane ( direction) from the primary vertex to the track's point of closest approach in the transverse plane. The impact parameters are used in the calculation of the * CP observable in the cases involving single-pion or leptonic decays (Section 3.1), as well as in one of the variables defining the signal regions (Section 5.2).
An important addition in the analysis is that it requires the -lepton reconstruction algorithms to categorise the hadronic -lepton decays by using the number of reconstructed tracks and the capability to distinguish between single-0 and multi-0 clusters. This allows the had candidates to be classified as being from the 1p0n, 1p1n, 1pXn or 3p0n decay modes. For this measurement it is crucial to identify each decay mode with good efficiency and purity, and to measure the momenta of the charged and neutral pions. This is achieved through the ' had Particle Flow' reconstruction algorithm [72], in which the four-momenta of the charged and neutral pions from the decay are determined by combining measurements from the tracking detector and the calorimeter. The decay mode classification is performed by counting the number of charged and neutral pions, exploiting the kinematic properties of the -lepton decay products and the number of reconstructed photons by using BDTs. Three BDTs are built respectively for the 1p0n vs 1p1n, 1p1n vs 1pXn and 3p0n vs 3pXn decay modes to improve the determination of the number of neutral pions in each case. The efficiency (purity) of the classification is about 80% (70%-80%) for the dominant decay modes 1p0n and 1p1n; for the 3p0n decay mode the efficiency is over 90%, with 90% purity.

Event selection
Events are selected by requiring at least one hadronic -lepton decay in the signature. For events in the lep had channel, a had-vis with T > 30 GeV and an electron or muon with T > 21.0-27.3 GeV are required, with the latter T cut depending on the data-taking period. In the had had channel, two had-vis objects are required with T above 40 and 30 GeV. The T requirements are chosen to be above the T thresholds of the respective single-lepton and hadronic triggers to ensure trigger operation at the plateau efficiency. The had-vis and the lepton are required to match geometrically with their trigger counterparts in the had had and lep had channels, respectively. In both channels the two -lepton candidates are required to have opposite electric charges. The angular distances between the -lepton candidates are required to be Δ < 2.5 and |Δ | < 1.5 (0.6 < Δ < 2.5 and |Δ | < 1.5) to reject non-resonant events in the lep had ( had had ) channel. All events require at least one jet with T > 40 GeV. Due to triggering conditions, this selection is tightened to T > 70 GeV and | | < 3.2 in the had had channel. The miss T is required to be greater than 20 GeV, and requirements on the fraction of the -lepton momentum carried by its visible decay products are applied to further improve the invariant mass estimation. In the lep had channel, the transverse mass T of the lepton-plus-miss T system is required to be less than 70 GeV in order to efficiently reject + jets processes. These preselection criteria are the same as in the → cross-section measurement [55].
Events are further categorised to target the vector-boson fusion (VBF category) and gluon-gluon fusion (Boost category) Higgs boson production modes. The VBF category contains events with a second high-T jet with 2 T > 30 GeV. The two leading jets are required to satisfy the following kinematics: |Δ | > 3.0, 1 · 2 < 0, and > 400 GeV, with the pseudorapidity values of lep or had lying between those of the two leading jets (yielding 'central -leptons'). This kinematic selection is enhanced by splitting the VBF category into two regions, VBF_1 and VBF_0, based on the output of a BDT-based VBF tagger [55], with the VBF_1 region having an enhanced fraction of VBF Higgs boson production events. The Boost category targets events where the Higgs boson recoils against jets. The T of the Higgs boson ( T ) is computed as the magnitude of the vector sum of the transverse momenta of the -leptons' visible decay products and the missing transverse momentum. Events with T > 100 GeV that do not pass the VBF selection form the Boost category. These events are further separated into those with Δ < 1.5 and T > 140 GeV (Boost_1) and those not passing these selections (Boost_0). The event categorisation is applied to both the lep had and had had channels, as summarised in Table 3. For each category, a Higgs-enriched signal region is defined with 110 < MMC < 150 GeV, and a → control region (CR) is defined with 60 < MMC < 110 GeV, with the other selection criteria for each category remaining the same. Two additional control regions are defined using events with a ± → ± → ± 0 decay in the 60 < MMC < 110 GeV range: one region with ℓ-1p1n events for the lep had channel, and the other with 1p1n-1p1n events for the had had channel. The events in these regions are defined so as to be statistically independent of the four → control regions in Table 3. The use of the control regions is described in Section 7.
Depending on the decay mode combination, different additional selection criteria are applied to enhance the sensitivity of the * CP construction method. The * CP from the IP method is less effective in discriminating between different values when the impact parameter vector has a magnitude smaller than, or similar to, its resolution. The sensitivity of this method can be enhanced by using events with high significance of the track impact parameter in the transverse plane, sig 0 , defined as the transverse impact parameter 0 divided by its resolution ( 0 ). Events are therefore separated into two groups based on the value of | sig 0 | of the lepton in lep or the pion in had . In the method, events with larger absolute values of the product | + − | are more sensitive to . This quantity is also used to separate the events into two groups. In the case of the combined IP-method, the sensitivity of * CP is enhanced by separating the events based on the values | sig 0 | and | |. For the IP-1 and -1 methods the separation is based on | 1 ± |. Details of the additional selection criteria are summarised in Tables 4 and 5.
Among the -lepton pair decays used in this analysis, the most sensitive ('dominant') decay mode combinations are 1p0n-1p1n, 1p1n-1p1n, and 1p0n-1p0n in the had had decay channel, and ℓ-1p1n and ℓ-1p0n in the lep had decay channel, while the other combinations involving 1pXn and 3p0n are subdominant due to their weaker spin analysing power and smaller decay fractions. The events are therefore divided into three groups, the 'High', 'Medium' and 'Low' signal regions, which characterise the different levels of sensitivity. Events satisfying the additional selection criteria for the dominant and subdominant decay mode combinations define the High and Medium signal regions, respectively, while the rest are grouped into the Low signal region. This allows the * CP distributions from the decay mode combinations with similar sensitivity to -mixing to be merged to increase the statistical precision of the distribution templates within each signal region, allowing the use of finer binning in the * CP distributions for the regions with better sensitivity. The splitting of the signal regions summarised in Tables 4 and 5 is applied in each of the VBF and Boost categories (VBF_1, VBF_0, Boost_1 and Boost_0).
In this configuration, there are 12 signal regions in each of the lep had and had had decay channels, leading to 24 signal regions in total. Each signal or control region is orthogonal to the others.
All above Not satisfying selection criteria Low All above Not satisfying selection criteria

Background estimation
Expected SM processes other than the → signal are evaluated using a mixture of simulation and data-driven techniques. Processes with true -leptons and prompt light leptons are estimated through simulation. Among those the (→ ) + jets process is the dominant one, and dedicated control regions with 60 < MMC < 110 GeV (Section 5.2) are used to extract its normalisation (Section 7).
The second most significant background contribution arises from jets misidentified as hadronically decaying -leptons ( had ), referred to as 'misidentified-background'. It is determined by a data-driven approach using fake factors, as described in detail in Ref. [55]. The misidentified-background consists mostly of + jets, QCD multĳet, and top-quark events in the lep had channel, while in the had had channel it originates mainly from multĳet events. The fake factors are calculated as the ratio of the number of events passing the identification requirements to the number failing them in dedicated QCD and + jets-enriched regions, defined by inverting the lepton isolation or T requirement, respectively. The fake factors are estimated in each hadronic decay mode considered in the analysis, separately for VBF and Boost region events. To estimate both the shape and the normalisation of the misidentified-background, distributions in the regions failing the identification are multiplied by the fake factors to correct for the different efficiencies to pass or fail the identification selection. In the 'failed identification' regions, events not corresponding to misidentified-background from jets are subtracted using simulated event samples. For the had had final state the fake-factor method is modified slightly: the had objects are matched to their trigger counterparts and the estimate covers processes with one or two jets misidentified as had . The misidentified-background contribution is estimated for each di-decay combination considered (Section 3) in the preselection region for the had had events.
Smaller background contributions are due to diboson, (→ ℓℓ) + jets and → * processes in both decay channels, while + jets and top production also make small contributions in the had had channel. These are estimated from simulation and are normalised to their theoretical expectations.

Systematic uncertainties
Systematic uncertainties generally affect the yield in the signal and control regions as well as the shape of the * CP distribution, on which the fit is performed. They are grouped into three types: the experimental uncertainties, the theoretical uncertainties, and the -lepton decay reconstruction uncertainties.
The experimental uncertainties include those from the trigger, reconstruction, identification and isolation requirements of the final-state particle candidates (electrons, muons, -leptons, jets, miss T , -tagging), as well as uncertainties from misidentified-background estimation and the luminosity measurement [73,74].
The uncertainties on the jet energy resolution and scale [75] affect the acceptance of the signal and background contributions through their impact on miss T , and thus MMC [76]. They also directly affect the jet selection for the event categorisation in the VBF and Boost regions [76]. The jet energy scale uncertainty for central jets (| | < 1.2) varies from 1% for a wide range of jet T (250 GeV < T < 2000 GeV) to 5% for very low T jets ( T < 20 GeV), and 3.5% for very high T jets ( T > 2.5 TeV), and forward jets exhibit uncertainties of similar size. The relative jet energy resolution ranges from (24 ± 1.5)% to (6 ± 0.5)% for jets with T of 20 GeV to 300 GeV, respectively [75]. Other jet-related uncertainties have smaller impacts.
Uncertainties from the misidentified-background estimation arise from sample statistics of the events used for estimating the fake factors, subtraction of residual contributions from processes without misidentified-, and uncertainties in the flavour composition taken from comparisons between the predicted and observed backgrounds in a dedicated validation region. The uncertainties are assigned per -lepton-pair decay combination in the lep had and had had channels. The total expected uncertainties in the yield of the misidentified-events in the signal regions are typically 20% (20%-40%) in the Boost (VBF) region in the lep had channel, and 15%-30% in the had had channel.
Theoretical uncertainties are applied to + jets background and signal processes. For the + jets background, uncertainties are considered for the renormalisation and factorisation scales, the resummation scale, the jet-to-parton matching scheme, the choice of value for the strong coupling constant s , and the choice of PDF. Only distribution shapes and migrations between analysis regions are considered for these uncertainties since the absolute normalisation of → is determined at the fit level. For the signal, uncertainties are considered for the QCD scale because of missing higher orders in the matrix element calculation, for the parton shower and hadronisation model, and for renormalisation and factorisation scales and the PDF. These are applied to the production cross-sections for the ggF, VBF and processes, and do not have large impacts on the measurement as they only affect the signal normalisation, which is also determined in the fit, and not the shape of the * CP distribution. The -lepton decay reconstruction uncertainties primarily affect the shape of the * CP distributions. These concern the classification of the hadronic -lepton decay modes as well as measurement uncertainties of the track impact parameters, pion track momentum, and neutral pion momentum. The uncertainties in the classification of the hadronic -lepton decay modes are derived from a -lepton decay mode classification efficiency and correction factor measurement, using a tag-and-probe analysis in thehad final state performed on part of the Run 2 dataset. These uncertainties include those in the decay mode reconstruction efficiencies and in the event migration between decay modes. The size of the uncertainties ranges from 6% to 20% depending on the decay modes and their cross migrations. The uncertainties affecting the impact parameters and track measurements include those in alignment effects and in impact parameter resolution, and they account for differences between data and simulation.
The uncertainties in the 0 angular resolution and energy scale are estimated in situ in the analysis. The 0 energy and momentum direction are initially varied in accordance with the energy response of the calorimeter to charged pions [77] and the measured 0 angular resolution [72], respectively. These variations in the 0 energy scale and angular resolution are found to affect the shape of the distribution of the reconstructed invariant mass of the ± 0 system, ( ± , 0 ), in ± → ± → ± 0 decays. Dedicated → control regions with exclusive 1p1n decays, namely the ℓ-1p1n and 1p1n-1p1n events, are defined in the lep had and had had decay channels, respectively, as described in Section 5.2. The ( ± , 0 ) distribution is used as the observable in these control regions in the combined fit, so that the final size of the 0 angular resolution and energy scale uncertainties in the measurement are determined from data. Figure 3 shows the post-fit distributions of ( ± , 0 ) in the → control regions in the lep had and had had channels using ℓ-1p1n and 1p1n-1p1n events, respectively. For the 1p1n-1p1n events in the had had channel, only the leading had is selected for the ( ± , 0 ) distribution. The ( ± , 0 ) data distributions show good agreement with the prediction.

Statistical analysis
To measure the -mixing angle , a simultaneous fit to the data is performed using a likelihood function that depends on the -mixing angle as the parameter-of-interest, and the nuisance parameters that account for the systematic uncertainties and the floating normalisations for the Higgs boson signal and for the background. The likelihood function is constructed as a product of Poisson probability terms over the bins of the input distributions, and the parameter-of-interest is estimated by maximising the likelihood. The likelihood comprises 24 signal regions and 10 control regions. Constraints on the nuisance parameters are assigned with a Gaussian term, and bin-by-bin statistical fluctuations in the simulated background samples are included in the fit with a Poisson probability term.
In the fit the normalisation of the → events is left to float freely in the eight control regions (described in Section 5.2) to account for the → modelling in the different signal regions. Four normalisation  factors (NF) are defined to control the → background events in the Boost_0, Boost_1, VBF_0, VBF_1 categories, respectively. Each factor is shared between the lep had and had had channels, and is also shared between a control region and the corresponding signal region in the same category. Other backgrounds are normalised to their expected number of events estimated from the MC simulation.
Two control regions using events with a ± → ± → ± 0 decay (ℓ-1p1n in lep had and 1p1n-1p1n in had had ) are defined from the → control region events (Section 5.2). The distribution of the ± 0 invariant mass, ( ± , 0 ), is employed in these control regions to control the 0 -related uncertainties by using the data, as described in Section 6.
The Higgs boson signal strength (defined as the ratio of the measured signal yield to the SM expectation) is also left unconstrained in the fit, such that the signal normalisation does not depend on the SM assumption and only the shape of the * CP distribution is exploited in the estimation of . Model-dependence of the cross-section on -mixing scenarios is not exploited. The * CP distributions in the signal regions are binned to maximise the measurement sensitivity, taking into account the associated uncertainties. In general, finer binnings are used in the signal regions associated with higher sensitivity (High and Medium SRs), with coarser binning in the Low signal region. A smoothing procedure is applied in the signal regions to remove potentially large local fluctuations in the systematic variations of the * CP distributions because of the limited size of the MC samples used to build the templates. The test statistic is based on a profile likelihood ratio and the asymptotic approximation [78] is used for statistical interpretations. Figure 4 shows the post-fit * CP distributions for the data as well as the prediction in the High, Medium and Low signal regions in the lep had and had had channels, respectively. In each distribution, the * CP bins are counted incrementally through all VBF and Boost categories and cover the range [0, 360] • for each category. The observed and expected negative log-likelihood (Δ ln ) scans in are shown in Figure 5. The observed (expected) value of is 9 • ± 16 • (0 • ± 28 • ) at the 68% confidence level (CL), and ±34 • ( +75 • −70 • ) at the 2 level. The data disfavours the pure -odd hypothesis at the 3.4 level, while the expected exclusion level is 2.1 . The fitted signal and background normalisations are shown in Table 6. The results are compatible with the SM expectation within the measured uncertainties.

Results
The total uncertainty is dominated by the statistical uncertainties of the data sample. The dominant contributions to the systematic uncertainties are from jets, followed by limited sample size of the background simulations, uncertainties from the free-floating normalisation factors, and theory uncertainties. The uncertainties in the -lepton decay reconstruction have small impacts of less than 1 • on . The effects from the 0 uncertainties measured in situ are compared with those from a set of simulated event samples with systematically varied modelling of the hadronic response in the calorimeter, and the latter are found to have a similar impact on the measurement. Effects from other sources are negligible. The impact of the uncertainties is summarised in Table 7.
The difference between the observed and expected sensitivities of the measurement can be attributed to a statistical fluctuation in data. The uncertainties of the measurement are highly dependent on the size of the modulation amplitude of the * CP distributions. In the measured data, there are more distortions of the * CP shape due to bins fluctuating from their expected values, resulting in overall larger modulation amplitudes than the predicted shape. A set of pseudo-experiments is performed (with the condition = 1) that shows that the probability of obtaining the observed distribution, given the expectation, is about 4%.
The present measurement is compatible with the recent measurement of the same mixing-angle parameter by the CMS Collaboration [24], for which an observed (expected) mixing-angle value of −1 • ±19 • (0 • ±21 • ) at the 68% CL was reported.
The expected sensitivities of the had had and lep had channels in excluding a pure -odd coupling are 1.7 and 1.1 , respectively. The 'High' signal regions contribute the most, with 1.4 and 1.0 in the had had and lep had channels, respectively. Other signal regions have expected sensitivities below 1 .
A 2D scan of Δ ln as a function of the signal strength versus is shown in Figure 6. The 1 and 2 2D confidence levels correspond to Δ ln values of 1.15 and 3.09, respectively. No strong correlation between and is observed. The SM prediction of = 1 and = 0 is compatible with the measurement within the 1 confidence region. Figure 7 shows the data distribution of * CP with all signal regions in both the lep had and had had channels combined, together with the best-fit → signal, pure -even and pure -odd hypotheses. The events in each signal region are weighted with ln(1 + / ), where and are the event yields for the signal and the total background, respectively. The background is subtracted from data in the figure. The distribution illustrates that the data disfavours the pure -odd scenario. (best-fit)  (best-fit) (best-fit)  (best-fit)  (best-fit)  (best-fit)       CP from all signal regions in both the lep had and had had channels. Events are weighted with ln(1 + / ) for the corresponding signal region. The background is subtracted from data. The best-fit → signal is shown in solid pink, while the red and green lines indicate the predictions for the pure -even (scalar, SM) and pure CP-odd (pseudoscalar) hypotheses, respectively, all scaled to the best-fit → signal yield. The hatched uncertainty band includes all sources of uncertainty after the fit to data, and represents the same uncertainty in the total signal and background predictions as in Figure 4.

Conclusion
A measurement of the properties of the interaction between the Higgs boson and -leptons is presented. In the generalised Yukawa interaction, a single mixing parameter parameterises -violating interactions between the Higgs boson and -leptons. The measurement is performed using 139 fb −1 of proton-proton collision data recorded at a centre-of-mass energy of √ = 13 TeV with the ATLAS detector at the Large Hadron Collider. The measurement is based on a maximum-likelihood fit to the -sensitive angular observable * CP defined by the visible decay products in the lep had and had had decay channels. Depending on the decay channels, different methods are used to reconstruct the * CP distributions. The sensitivity is optimised by applying decay-mode-dependent kinematic selections. The observed (expected) value of is 9 • ± 16 • (0 • ± 28 • ) at the 68% confidence level, and ±34 • ( +75 • −70 • ) at the 2 level. The pure -odd hypothesis is disfavoured at a level of 3.4 standard deviations. The analysis precision is limited by the statistical uncertainty of the data sample. The measurement is consistent with the Standard Model expectation.    The ATLAS Collaboration