1 Introduction

At this unique juncture in particle physics, a multitude of theories once believed to be ideal complements to the Standard Model (SM) have been either disqualified or so significantly confined in their scope that they no longer hold the same prominence. However, the unresolved issues of the SM that were prevalent prior to the initiation of the Large Hadron Collider (LHC) remain unresolved. These circumstances have led some to consider this to be a period of stagnation. Historically, such periods have often served as precursors to significant breakthroughs. The challenge we face is due to our inability to predict if this impending revolution will occur in the coming decade or necessitate a more extended timeline. The initial two runs of the LHC have largely corroborated the SM, exceeding our own expectations. In this scenario, where there is no decisive conception of the scale and structure of New Physics (NP), indirect searches and precision measurements emerge as highly promising avenues. Among these, precise assessments of Flavour Physics hold particular intrigue as they could potentially elucidate the Flavour puzzle - the distinctive pattern of masses and couplings across different families that, although described, remain inexplicable within the SM. In more technical terms, the intriguing aspect is that the gauge portion of the SM Langrangian has an extensive discrete symmetry, \(U(3)^5\), which is broken by the Higgs mechanism via Yukawa couplings and is suggestive of a concealed dynamics at work.

Among the various flavour-physics experiments, LHCb has been one of the most prominent for the last 15 years. The reason for this is due to the unique sensitivity, meaning that LHCb results from both the original detector and its upgrades have the potential to significantly influence the future direction of particle physics. The LHCb experiment [1] is one of the four large detectors located at the LHC and is specifically designed to study the decays of beauty and charm quarks. Unlike the ATLAS and CMS detectors, LHCb is designed more like a fixed target experiment, with the collision point situated on one side of the detector. This single-arm forward spectrometer geometry allows the maximum amount of space in which to instrument in a single direction, which is crucial to the performance of particle ID and momentum resolution. This is at the expense of having a smaller acceptance, where only the pseudo-rapidity range \(2< \eta < 5.0\) is instrumented. However, due to the particular nature of \(b\bar{b}\) production at the LHC, most beauty quarks are produced in the forward direction, and therefore around 20% of which decay within the LHCb acceptance.

Being situated at the LHC, the LHCb detector operates in a hadronic environment, with hundreds of particles produced in every hard pp interaction. This is in contrast to the B-factories, where two B-mesons are much more isolated and kinematically constrained. The reason why LHCb can compete in this environment is due to the huge production rate at the LHC. For example, starting from 2035, the high-luminosity LHC will produce O(100 M) beauty quark pairs every second [2]. This is equivalent to producing a B-factory dataset every 10 s, albiet in much harsher conditions. The exploitation of this huge dataset for the purposes of heavy flavour physics is a formidable, yet as we will discuss later, rewarding challenge.

2 Status of the LHCb experiment

Aside from its geometry, the LHCb detector has a few features which distinguish it when compared to the other experiments at the LHC. Firstly, the vertex detector (VELO) is placed only 5.1 mm from the beam, allowing for an excellent impact parameter resolution of roughly 35 \(\mu \hbox {m}\) depending on the transverse momentum of the particle. This is vitally important for timing resolution and to reject particles that originate from the primary pp collision. Another feature of LHCb crucially important for heavy flavour physics is the ability to identify hadrons. This is provided by ring imaging Cherenkov detectors (RICH), situated upstream and downstream of a dipole magnet. The tracking system, comprised of the VELO, a silicon tracking system situated upstream of the magnet and further tracking stations downstream. These provide an excellent momentum resolution resulting in \(\sim \) 20 MeV mass resolution for a fully reconstructed b-hadron decay.

The LHCb experiment does not operate at the maximum luminosity of \(2\times 10^{34}\) cm\(^{-2}\) s\(^{-1}\) currently accesable at the LHC. The reason for this choice is firstly due to the very large occupancies in the forward region of the LHC, which would make particle reconstruction prohibitively complicated. The second reason is due to the extremely high signal rates, resulting in unfeasible requirements for any trigger or DAQ system realisable with current technology. The luminosity is therefore ‘levelled’ by defocusing the beam at the interaction point, allowing for a more manageable data rate and simpler to analyse. During runs I and II, the luminosity at LHCb was levelled to a maximum rate of \(4\times 10^{32}\) cm\(^{-2}\) s\(^{1}\), which resulted in around 1–2 hard pp interactions per bunch crossing. In this environment the first level of the trigger system, based on partial readout such as high transerve energy clusters in the calorimeters, was able to achieve good efficiency for a wide variety of signatures.

For the ongoing LHC run III, the luminosity will be levelled to a rate of \(2\times 10^{33}\) cm\(^{-2}\) s\(^{-1}\), a five-fold increase from the previous runs. At this luminosity, there are around 7 pp interactions per bunch crossing, which is a significantly more challenging environment for both triggering and reconstruction. The LHCb detector was therefore upgraded [3] to meet this challenge: The tracking systems were replaced, with the strips in the VELO replaced by a pixel detector; the Tracker Turicensis was replaced by the Upstream Tracker improving to material budget and increased acceptance and finally the Outer and Inner Trackers were replaced with Scintilating Fibres. These upgrades improved the occupancy and radiation hardness, which are mandatory upgrades to deal with the increased data.

The second important aspect of the LHCb upgrade 1 is the near-complete replacement of the readout electronics, allowing for the full detector to be readout at the bunch crossing rate of 30MHz. This allows for a trigger system that is fully implemented in software. The main advantage of this is to utilise the large impact parameter of particles that originate from beauty and charm hadron decays. This allows to have softer transverse momentum requirements compared to the original trigger system, which substantially improves the efficiency for fully hadronic decays which, among other things, are of crucial importance for LHCb’s CP-violation programme discussed later on in this review. The LHCb upgrade 1 has now been completed, and the detector is currently being commissioned. A total of 50 fb\(^{-1}\) are expected to be taken during the next two run periods until 2032, which will allow for significantly more precise measurements in addition to a broader programme enabled by the new trigger system.

The LHCb physics programme includes a large and diverse set of measurements, including studies of charm decays, electroweak physics and analyses of heavy-ion collisions. In this review, we focus on measurements performed with b-hadron decays, which remain a core part of the experiment’s physics potential. The b physics programme is summarised by highlighting three areas: Studies of rare b decays, measurements of charged-current semileptonic decays and searches for new sources of CP violation.

3 Rare decays

Rare decays of b-hadrons serve as promising probes for NP, as they are suppressed and can therefore introduce contributions from NP that are competitive with the SM decay amplitude. A particularly sensitive set of rare decays are those proceeding via flavour-changing-neutral-currents (FCNC), where a b-quark transitions into either a strange or down quark. FCNCs have had a significant historical impact in the construction of the SM, for example which led Glashow, Iliopoulos, and Maiani to predict the existence of the charm quark (c) [4]. Interestingly, when operators of dimensions greater than four are added to the Standard Model (SM), FCNC transitions naturally emerge, a key reason why rare decays are often targeted in the search for NP. Rare decays play an outsized role in the physics programme of LHCb as the experiment has a unique sensitivity, thanks to the vast number of b-quarks produced at the LHC and its exceptional performance in particle identification, momentum, and vertex resolution which is crucial for background control.

In a scenario where we postulate that there are no new particles below the electroweak scale, we can characterise \(b\rightarrow s\ell ^+\ell ^-\) transitions using an effective Lagrangian encompassing only light SM fields. The main distinction between the SM and effective Lagrangians, renormalized at a scale \(\mu \sim m_b\), is the number of effective operators, which can be more extensive in the NP case, in addition to the value of the corresponding coefficients.

$$\begin{aligned} \Delta {{\mathcal {L}}}^{b\rightarrow s\ell \ell }_{\textrm{NP}} = \frac{4 G_F}{\sqrt{2}}\, \sum _{i } C_i {{\mathcal {O}}}_i + \mathrm{h.c.}, \end{aligned}$$
(1)

where \(G_F\) denotes the Fermi constant, and where the index i indicates the following set of dimension-six operators (treated independently for \(\ell =e\) and \(\mu \)):

$$\begin{aligned}{} & {} {{\mathcal {O}}}^\ell _{9}= (\bar{s}_L\gamma _\mu b_L)(\bar{\ell }\gamma ^\mu \ell ) , \quad \ {{\mathcal {O}}}^\ell _{10}= (\bar{s}_L\gamma _\mu b_L)(\bar{\ell }\gamma ^\mu \gamma _5\ell ) \,, \nonumber \\{} & {} {{\mathcal {O}}}^{\ell \prime }_{9}= (\bar{s}_R\gamma _\mu b_R)(\bar{\ell }\gamma ^\mu \ell ) , \quad {{\mathcal {O}}}^{\ell \prime }_{10}= (\bar{s}_R\gamma _\mu b_R)(\bar{\ell }\gamma ^\mu \gamma _5\ell ) \,, \nonumber \\{} & {} {{\mathcal {O}}}^\ell _{{\hat{S}}}= (\bar{s}_L b_R)(\bar{\ell }_R \ell _L) , \qquad \ {{\mathcal {O}}}^{\ell \prime }_{{\hat{S}}}= (\bar{s}_R b_L)(\bar{\ell }_L \ell _R) \,. \end{aligned}$$
(2)

This methodological approach is highly effective in correlating different decays, not only facilitating a combined interpretation of different measurements for enhanced sensitivity but also leveraging the coherence of various modes to mitigate the impact of statistical and systematic effects.

In recent years, several discrepancies in rare decays have emerged, which form a large part of the so-called flavour anomalies. The anomalies related to rare decays comprised of three types of measurements: lower than predicted branching fractions for various \(b\rightarrow s\mu \mu \) decays [5,6,7], a deviation in the angular distrubution of \(B\rightarrow K^*\mu ^+\mu ^-\) decays [8, 9], and previous deviations in tests of lepton flavour universality. Furthermore, discrepancies with respect to the SM in tree-level semiltauonic decays, proceeding through \(b\rightarrow c\tau \nu \) processes, have been identified and will be elaborated upon in the next section.

The first anomaly in rare decays observed was a discrepancy in the angular distribution of the decay \(B\rightarrow K^*\mu ^+\mu ^-\) [10], specifically in the observable \(P_5'\) [11, 12], which was later corroborated by further measurements [8, 9]. An effective field theory analysis of the deviation points towards a lower than expected value of the Wilson coefficient, \({\mathcal {C}}_{9}\), which is the coupling associated with the vector current to the di-lepton pair at the level of around 3\(\sigma \) standard deviations. One of the striking features of this interpretation is that such a shift is numerically consistent with the lower branching fractions seen in other decay modes, which dampens the possibility of a statistical fluctation. Despite this intriguing coherence, it appears that a theoretical breakthrough is necessary for definitive conclusions, as they indicate a shift to \(C_9\) which can also be affected by long-distance contributions from decays \(b\rightarrow s \bar{c} c\), where the \(c\bar{c}\) annihilates to form the two leptons. As \(c\bar{c}\) decay proceeds via a photon, the contribution interferes with the vector contribution of pure semileptonic \(b\rightarrow s \mu \mu \) decay and can mimic a NP effect. As this \(c\bar{c}\) contribution is fully hadronic, it is very difficult to compute and is therefore controlling it is the main avenues to clarify the situation. Approaches using data [13,14,15,16,17] have been proposed and implemented in order to help constrain the theoretical predictions. At the current moment there is not yet a consensus on the size of this effect, but progress is being steadily made (see for example Ref. [18]).

Fortunately, this issue is not present for fully leptonic rare decays, of the type \(B_{(s)}\rightarrow \ell ^+\ell ^-\), as they can proceed only in the SM with an axial-vector coupling to the two leptons. A famous example of this type is the decay \(B_s\rightarrow \mu ^+\mu ^-\), which has long been considered a critical channel for the search for NP due to its helicity suppression as well as that from the GIM mechanism. This makes it significantly enhanced in models with an extended Higgs sector, such as the MSSM, where the branching ratio is proportional to \(\tan ^6\beta \) [19,20,21,22]. In terms of effective operators, NP can contribute new scalar/pseudoscalar operators, yielding non-zero values for the Wilson coefficients \(C_{S,P}^{(\prime )}\) or by modifying the value of the axial-vector coefficient \(C_{10}\). The main challenge in the experiment for this decay is to distinguish it from background sources which have much larger branching fractions compared to the ultra rare decay rate.

The latest LHCb measurement has measured the value of \({{{\mathcal {B}}}}(B_{s}\rightarrow \mu \mu )=3.09^{+0.48}_{-0.44} \times 10^{-9}\) [23], the value of which is consistent with the SM prediction [24, 25]. More recently, the CMS experiment updated their result to \({{{\mathcal {B}}}}(B_{s}\rightarrow \mu \mu )=(3.66\pm 0.14) \times 10^{-9}\) [26]. A combination of various measurements of these branching ratios including one from ATLAS [27] limits potential NP contributions to \(C_{10}\).

In the absence of substantial theoretical advancement for rare semileptonic decays, one viable approach to address the anomalies observed in \(b\rightarrow s\mu ^{+}\mu ^{-}\) processes, assuming they are a result of NP, would be to relate them to other phenomena and look for predicted signatures in those systems. Consequently, when tests on lepton universality in \(b\rightarrow s\ell ^{+}\ell ^{-}\) processes at LHCb displayed discrepancies with respect to the Standard Model (SM), it was reasonable to consider these discrepancies as potentially related (see Refs. [28, 29] and references therein). The following ratio of branching ratios, denoted by \(R(X_s, q_{min}^2, q_{ma}^2)\), are considered:

$$\begin{aligned} R(X_s, q_{min}^2, q_{max}^2)=\frac{\int _{q_{min}^2}^{q_{max}^2} \frac{d{{{\mathcal {B}}}} (X_s\mu ^+\mu ^-)}{dq^2}}{\int _{q_{min}^2}^{q_{max}^2} \frac{d{{{\mathcal {B}}}} (X_s e^+e^-)}{dq^2}}. \end{aligned}$$
(3)

Lepton Universality is an accidental symmetry of the SM. Hence, it would not be surprising for extensions of the SM to break it. Although the quantity in Eq. (3) is theoretically clean (see Refs. [30, 31] for discussions on theory uncertainty), its experimental measurement is more challenging, particularly at LHCb.

These experimental challenges arise from the significantly larger bremsstrahlung radiation of electrons compared to muons, which results in poorer momentum resolution. This, in turn, leads to lower efficiency, poorer invariant mass resolution, and ultimately, more complex handling of various backgrounds.

LHCb had initially measured the observables \(R_K\) [32], \(R_{K^*}\) [33], \(R_{K_S}\) [34] and \(R_{pK}\) [35] to be all below unity, with a local combined significance [36] exceeding four standard deviations. This brought considerable attention to the flavour anomalies, as they appeared to present a coherent pattern involving a non-lepton universal NP coupling \(C_V=C9-C_{10}\).

The general analysis strategy comprises the following steps. The primary trigger for electron events utilizes the electromagnetic calorimeter, a method that is less efficient than triggering muons, which is accomplished through the muon detector. The remainder of the event is also utilized as an additional trigger line. To alleviate the issue of poor resolution caused by bremsstrahlung radiation, a bremsstrahlung recovery algorithm is implemented for the electrons. This involves adding to the electron momentum the momentum of bremsstrahlung photons that are compatible with the electron’s trajectory.

The primary difference between the latest LHCb analysis [37, 38] and previous ones lies in the treatment of misidentified \(B\rightarrow hhh X\) type backgrounds that contaminate the electron channel. To account for all potential hadron misidentified backgrounds, a data-driven procedure was developed. This involves creating a hadron background-enriched sample by inverting the PID criteria. The candidates are weighted according to their misidentification probability, obtained with control channels. After subtracting the residual electron component, a high-purity background sample for single and double hadron misidentification is obtained. This provides a template for fitting the invariant mass distributions, as shown in Fig. 1.

Fig. 1
figure 1

Invariant mass distribution of \(B^+\rightarrow K^+\ell ^+\ell ^-\) and \(B^0\rightarrow K^{*0}\ell ^+\ell ^-\) for the muon mode (top four figures) and electron mode (bottom four figures) with the fit to data overimposed. Reproduced from Ref. [37]

The latest LHCb measurements of \(R_K\) and \(R_{K^*}\) perform the analysis in two different regions of \(q^2\), resulting in four different measurements. These measurements are compatible with SM predictions, significantly altering the broader picture. Figure 2 compares the results previous analyses [33, 34] with the latest results [37, 38]in the four bins of \(q^2\).Footnote 1 It’s worth emphasising that while these measurements align with the SM, they can still accommodate a 5-10% lepton universality (LU) violation. The new results have strongly influenced global fits for the flavour anomalies, which were approaching a significance of five standard deviations prior to these analyses (see e.g. Ref. [39]). While different groups yield compatible values under the same assumptions, papers obtaining different significance values mainly differ in how they handle non-local contributions from charm. Given the updated results and the current uncertainty of the \(c\bar{c}\) contribution, the situation for the rare decays anomalies remains unresolved.

Fig. 2
figure 2

Summary of the measurements for \(R_{K}\) and \(R_{K^*}\) across different \(q^2\) regions, as reported by LHCb. Previous measurements refer to [32, 33]. The more recent measurements [37, 38] have introduced improved background handling techniques and thus supersede the previous ones. Please note that the measurements of \(R_{K^*}\) in the low-\(q^2\) region, as specified in the cited references, were carried out within two distinct \(q^2\) ranges: \(0.045<q^2<1.1\) \(\hbox {GeV}^2\) and \(0.01<q^2<1.1\) \(\hbox {GeV}^2\), respectively

Importantly, the uncertainty of the measurements is statistical, thus the LHCb Upgrade(s) will remain the primary experiment for LU tests in rare decays. This is significant because any LU violation would be a clear indication of NP. The anomalies have inspired innovative theoretical models that provide an intriguing explanation for the flavour puzzle. Many of these compelling models regard the third family as unique. For these models, LFU tests with the tau in the final states are critical, while electron/muon LU violation can be minimal without causing issues for the models. Consequently, a promising way to test these models, besides semi-tauonic measurements, would be rare decays with taus in the final state of the type \(b\rightarrow s\tau \tau \).

Most models that account for the flavour anomalies through the addition of NP also predict potentially significant effects in \(b\rightarrow s\tau \tau \) transitions due to a hierarchical coupling to the different generations. The current best limits are as follows:

$$\begin{aligned} {{{\mathcal {B}}}}(B_{s}^{0}\rightarrow \tau \tau )< 2.1\times 10^{-3} \, (95\%\, ~\mathrm CL)\\ {{{\mathcal {B}}}}(B^+\rightarrow K^+\tau \tau )< 2.25\times 10^{-3} \, (90\%\,~ \mathrm CL)\\ {{{\mathcal {B}}}}(B^0\rightarrow K^{*0}\tau \tau ) < 3.1\times 10^{-3} \, (90\%\, ~ \mathrm CL). \end{aligned}$$

These were established by the LHCb [40], BaBar [41] and Belle [42] experiments. Evidently, rare decays involving tau particles in the final state pose greater challenges than their counterparts involving electrons and muons, as tau particles decay before reaching the detector. The upgrade of the Belle experiment, Belle II, is expected to achieve a limit of \(5.4\times 10^{-4}\) for \(B^0\rightarrow K^{*0}\tau \tau \) with a luminosity of 50 \(\hbox {ab}^{-1}\) [43].

Such decays prove even more experimentally challenging at the LHCb experiment, where there is a high density of particles and the momentum of the decaying B-meson is unknown, in contrast to the Belle II experiment. Regardless, LHCb established the world’s first limit on \(B_s \rightarrow \tau \tau \) and (assuming no contribution of \(B_s\)), the best limit on \(B^0 \rightarrow \tau \tau \) [40]. Studies on \(B\rightarrow X_s \tau \tau \) at LHCb are ongoing.

Moreover, a significant \(b\rightarrow s\tau \tau \) NP contribution could be investigated in the dimuon spectrum of \(B^+\rightarrow K^+\mu \mu \) by fitting the dimuon spectrum, as suggested in Ref. [44]. While this technique necessitates assumptions on the theoretical model of the dimuon spectrum, it has the potential to yield results competitive with direct searches. Thus, such indirect searches could hold particular relevance for the LHCb upgrade II.

Finally, the search for lepton flavour violation (LFV) represents a fascinating area of study, as this accidental symmetry is already violated in the Standard Model (SM) due to neutrino oscillations. However, the LFV induced by the SM is at an unmeasurably low level. Many extensions of the SM, including models aimed at explaining flavour anomalies, predict LFV at a level that upcoming particle physics experiments may be capable of measuring. Searches for LFV in the LHCb Upgrades complement those of dedicated experiments. Though it’s inherently more challenging to determine LFV limits for the third family of quarks and leptons compared to the second family, the potential preferential coupling of NP to the third family might offset these experimental constraints. A summary of LFV limits set by the LHCb experiment can be found in Table 1.

Table 1 Table of decay modes, analysed data, and limits at 90% CL for various lepton flavour violating modes

Limits on \(\tau \rightarrow 3 \mu \) are currently dominated by the Belle experiment, with Belle II anticipated to reach levels of a few parts in \(10^{-10}\) assuming no background [43]. For LHCb, surpassing the limits set by Belle II will pose a considerable challenge. However, sensitivity studies indicate that LHCb Upgrade II could also explore the level of \(10^{-9}\). It’s worth mentioning that LHCb might be the only experiment capable of potentially confirming a discovery by Belle II.

LHCb is leading in setting limits on \(b\rightarrow s\mu e\) across all channels, as well as on \(b\rightarrow s\tau \mu \). All these limits are statistically limited and are expected to remain so after Run 3 and Run 4. It’s important to note that LHCb has begun to explore an intriguing region for \(b\rightarrow s\tau \mu \), making future measurements especially interesting.

4 Tree-level semileptonic decays

In the SM, tree-level semileptonic decays of the type \(b\rightarrow c,u \ell ^{-}\bar{\nu }_{\ell }\) proceed via a charged-current interaction mediatated by the W boson. The presence of the weakly interacting neutrino in the final state is an unambiguous signal of a short-distance interaction, substantially simplifying the theoretical treatment as no long-distance contributions are present. The price to pay for this theoretical clarity is large, however, as the signal is necessarily partially reconstructed. Therefore, it was a surprise that LHCb contributed as significantly as it has in this system.

Given their theoretical appeal, tree-level semileptonic decays are the ideal system to measure the magnitudes of the CKM elements \(V_{ub}\) and \(V_{cb}\). There are two main ways to determine these parameters. The first is to study inclusive \(b\rightarrow c (u) \ell \nu _{\ell }\) decays, where all hadronic final states of the charm or the up quark are included in the signal. The second is to reconstruct an exclusive hadronic final state such as \(B\rightarrow \pi \ell \nu \). Measurements based on exclusive and inclusive methods have had a long-standing disagreement for more than a decade for both \(|V_{ub}|\) and \(|V_{cb}|\) [45]. This disagreement is one of the issues that should be resolved to maximise the sensitivity for the test of CKM unitarity.

Determining a CKM magnitude requires control over the absolute normalisation of the production and efficiency, which is difficult at a hadron collider. Therefore, the strategy at LHCb is to normalise the \(b\rightarrow u\) transition by a corresponding \(b \rightarrow c\) transition, thus performing a measurement of \(|V_{ub}|/|V_{cb}|\). This has been done with \(\Lambda _b^{0} \rightarrow p \mu \nu \) and \(B_{s}^{0} \rightarrow K^{+}\mu ^{-}\nu \) decays, which are easier signatures than the canonical \(B\rightarrow \pi \mu \nu \) decay used at the B-factories due to the large amount of pions that originate from background sources. The main strategy of the analysis is to take advantage of the high signal yields at LHCb to apply very tight selection criteria which improves the purity and consequently reduces systematic uncertainties associated with the background. An example of this is to select candidates which have the best vertex resolution, as are more easily distinguished from background as well as having better kinematic resolution.

When using lattice QCD, the results from LHCb agree with the B-factory results based on \(B\rightarrow \pi \mu \nu \). This agreement significantly strengthens the confidence in the exclusive determinations of \(|V_{ub}|\) which will be crucial to finally solving the inclusive-exclusive puzzle. The uncertainties are also competitive with those from the B-factories and are expected to reach an ultimate precision of 1% after the second LHCb upgrade [46], which relies on further progress on lattice QCD techniques.

On the \(|V_{cb}|\) side, the determination is strongly affected by the treatment of the hadronic form factors and how they are combined with the experimental data [47,48,49,50]. The differential shape of the branching fraction as a function of the decay kinematics such as the squared di-lepton mass \(q^{2}\) provides a large amount of experimental information in which to constrain the theory parameters. Here is where LHCb could provide a complementary measurement, as the large signal yields will result in very precise measurements in wider kinematic bins due to the lower resolution compared to the B-factories. These measurements will have a significant impact, provided that systematic uncertainties remain below the very high statistical precision.

Due to their large branching fractions, semileptonic decays are a unique place to study decays involving the notoriously difficult \(\tau \)-lepton. Despite its large SM amplitude, the decay \(b\rightarrow c\tau ^{-}\bar{\nu }_{\tau }\) is highly sensitive to NP due to the large number of third-generation fermions involved in the decay. The branching fractions of these decays are tested with respect to the SM by normalising them to their muonicFootnote 2 counterparts to form lepton universality ratios

$$\begin{aligned} R(X_{c}) = \frac{{\mathcal {B}}(X_{b}\rightarrow X_{c}\tau ^{-}\bar{\nu }_{\tau })}{{\mathcal {B}}(X_{b}\rightarrow X_{c}\mu ^{-}\bar{\nu }_{\mu })} \end{aligned}$$
(4)

which are predicted to be around 30% in the SM due to the kinematic suppression with the presence of the \(\tau \)-lepton. \(\tau \)-leptons are notoriously difficult to reconstruct as they decay with at least one neutrino with a low visible branching fraction. For semitauonic decays, this means at least two neutrinos missing from the decay, resulting in a broad signal shape that is highly susceptable to backgrounds.

There are two decay modes of the \(\tau \)-lepton that are considered. The first is the leptonic decay \(\tau ^{-} \rightarrow \mu ^{-}\bar{\nu }_{\mu }\nu _{\tau }\), which suffers from three missing neutrinos but has a large signal yield and a well understood \(\tau \) decay. The second is the three-prong decay \(\tau ^{-}\rightarrow \pi ^{-}\pi ^{+}\pi ^{-}\nu _{\tau } (\pi ^{0})\) which allows for a better kinematic reconstruction and higher purity at the expense of a smaller signal yield. For both modes, the most dangerous background is from \(X_{b}\rightarrow X_{c}\bar{X}_{c} (X)\) decays, where one of the \(X_{c}\) hadrons mimics a \(\tau \) decay. Due to the similarity in the mass and lifetime, the kinematics of the \(X_{c}\) decay products can be fiendishly difficult to disentangle from the signal.

The signal yields are determined by three-dimensional fits to variables that discriminate between signal and the various backgrounds. Examples include the \(q^{2}\) distribution or the \(\tau ^{-}\) decay time. The latest LHCb measurements [51, 52] of \(R(D^{(*)})\) are shown in Fig. 3 together with those of the B factories [53,54,55,56,57,58]. The fact that the hadron collider measurements are competitive with those from the B-factories is a huge success story of the LHCb experiment. The SM prediction is also depicted, which is 3.3 \(\sigma \) from the world average. Given the complexity of the measurements, caution should be taken when interpreting this discrepancy as a sign of NP, but it is interesting to note that three seperate experiments consistency measure higher semitauonic yields than predicted and in more than one \(\tau \) decay channel. Further measurements, including those only accessible at the LHC like those using \(\Lambda _b\) baryons [59] will be crucial to clarify the situation in the future.

Fig. 3
figure 3

Measurements and SM predictions (black) of the lepton universality ratios R(D) and \(R(D^*)\) [45]

5 Searches for CP violation

The observed level of baryon asymmetry in the universe mandates sources of CP violation beyond the SM [60]. The LHCb experiment is contributing to the search for new sources of CP violation in several different ways. The first that we will discuss is the contribution from LHCb in testing the unitarity of the CKM matrix. Here, LHCb dominates the precision of the CKM angle \(\gamma \), providing world-leading results on the CKM parameter \(sin(2\beta )\) and providing important contributions to the CKM magnitude ratio \(|V_{ub}|/|V_{cb}|\).

When LHCb started collecting data in 2009, the angle \(\gamma \) was the least precisely known of the CKM unitarity triangle, with a precision of \(\sim \)25% [61]. Over a decade of LHCb measurements sensitive to the interference between \(b\rightarrow c\) and \(b\rightarrow u\) transitions have improved the precision by over a factor of six. Unlike the CKM parameter \(sin(2\beta )\), there is no golden mode for \(\gamma \) measurements, which means that the ultimate precision is obtained by combining several different \(B\rightarrow D h\) analyses. Therefore, the ability to reconstruct a wide variety of fully hadronic final states is a key experimental requirement for the experiment. Of particular importance is \(K-\pi \) separation as the main sensitivity arises from CKM-suppressed decays \(B^{+}\rightarrow \bar{D^{0}} K^{+}\), which would be swamped by the more abundant decay \(B^{+}\rightarrow \bar{D^{0}}\pi ^{+}\) if no particle ID requirements were applied. More recently, decay modes involving \(\pi ^{0}\) mesons were included [62] to further boost the sensitivity using charm decays such as \(D\rightarrow h^{+} h^{'-}\pi ^{0}\). The ability to reconstruct neutral pions and obtain physics sensitivity from decays involving them is a surprise given the large occupancy in the calorimeters. Other novel reconstruction techniques involve determining the signal yield of \(B\rightarrow D^{*0} h\) decays without reconstructing the \(\pi ^{0}\) from the \(D^{*0}\rightarrow D^{0} \pi ^{0}\) decay. This is possible due to the low Q-value of the \(D^{*0}\) decay. All the different \(B\rightarrow Dh\) decays have a complementary sensitivity to \(\gamma \) and associated hadronic nuisance parameters. The combination of which results in \(\gamma = (65.4^{+3.8}_{-4.2})^\circ \), where the charm mixing parameters are also measured simultaneously [63]. Most of the measurements that contribute to this average use the full LHCb dataset of 9 \(fb^{-1}\), although there are few which are yet to be updated such as that from \(B_{s}\rightarrow D_{s} K^{+}\) [64]. Being based on fully hadronic decays, the improved trigger system in run III will particularly benefit these measurements in the future.

The CKM parameter sin(2\(\beta \)) is another place where LHCb is currently world leading. Here, the golden mode is the decay \(B^{0}\rightarrow J/\psi K^{0}_{s}\), where the decay into a CP eigenstate allows for interference between the mixing and decay amplitudes. There are two key experimental challenges associated with this analysis. The first is the ability to reconstruct the long-lived \(K_{s}^{0}\) meson, which has an average decay length of around one metre in the LHCb detector. This is significantly longer than for the B-factories due to the larger boost at the LHC centre of mass energy which results in a comparatively lower reconstruction efficiency. This is one of the reasons why the precision on \(\gamma \) has been dominated by LHCb measurements for years whereas the world’s most precise measurement of sin(2\(\beta \)) has only just become a measurement from LHCb.

Fig. 4
figure 4

Tagging power of hadronic and leptonic B decays reported in LHCb publications as a function of the pre-print date

The other challenge for this measurement is flavour tagging, as the final state does not determine the flavour of the \(B^{0}\) meson at production. Flavour tagging at a hadron collider is much more difficult compared to the B-factories, where the two B mesons are quantum entangled and are produced in isolation. Flavour tagging at LHCb relies on two aspects of \(b\bar{b}\) events in LHC collisions. The first is to exploit correlations of accompanying particles in the hadronisation of the b-hadron of interest, known as same-side tagging. The second is to exploit anticorrelations in the decay products of the other b-hadron produced in the event, known as opposite-side tagging. The performance of flavour tagging is quantified by the tagging power, which represents the size of a perfectly tagged dataset that would have equal precision. The tagging power is a result of the exploitation of several different weak correlations and has steadily improved as more sophisticated techniques have been employed to utilise more inclusive decay pathways. This can be seen in Fig. 4, which shows the tagging power reported in LHCb publications as a function of the preprint date. This improvement is one of the reasons why the latest precision of sin(2\(\beta \)) has improved beyond the niave scaling of luminosity.

Beyond the unitarity of the CKM matrix, CP violation in \(B_{s}^{0}\) mixing is also highly sensitive to many NP models which favour second and third generation couplings [65]. This means that the measurement of the CP violating phase in the \(B_{s}^{0}\) system, \(\phi _{s}\) has unique NP potential. The golden modes for this observable is \(B_{s}^{0} \rightarrow J/\psi hh\), where hh is either a pair of pions or kaons. In addition to flavour tagging and particle ID, a key aspect of the LHCb performance is the time resolution needed due to the very fast oscillations of the \(B_{s}^{0}\) meson. This is where the large boost provided in high energy LHC environment is highly advantageous, leading to a time resolution of 35–45 fs. This performance is demonstrated by an ultra precise measurement of the \(B_{s}^{0}\) oscillation frequency [66] using \(B_{s}^{0}\rightarrow D_{s}^{-} \pi ^{+}\) decays. The decay time of candidates that have undergone an oscillation or not is shown in Fig. 5, showing that the remarkably fast oscillations can be resolved by the LHCb detector. This level of fidelity is crucial for the measurement of \(\phi _{s}\), for which has recently been performed with the full LHCb run I–II dataset of 9 \(\mathrm fb^{-1}\) [67]. The result is \(-0.039\pm 0.022\pm 0.006\) rad which is consistent with the SM, and is two times more precise than any other measurement. Despite the remarkable precision, the comparison to the SM is still limited by the statistical uncertainty of the measurement, meaning that there are excellent prospects for the upgraded LHCb detector in the future. Such progress will rely on control of the penguin pollution, which will be helped to be controlled using the SU(3) related decay \(B_{s}^{0}\rightarrow J/\psi K^{*0}\) [68].

Fig. 5
figure 5

Decay time distribution of \(B_{s}^{0}\rightarrow D_{s}^{-} \pi ^{+}\) decays, split between cases in which the \(B_{s}^{0}\) has undergone oscillation or not [66]

In order to observe CP violation directly in the decay of a particle, one needs two decay amplitudes with different weak and strong phases. The size of the CP asymmetry is maximised when the magnitudes of the two amplitudes are of similar size. Fully hadronic charmless B decays are perfect laboratories for studying such large effects, as both tree- and loop-level diagrams contribute to the final state. Furthermore, the CKM suppression of the tree-level \(b\rightarrow u\) amplitude allows for a large CP asymmetry as it places the magnitude of that amplitude to a similar level of the loop diagram. The loop diagram typically occurs through a gluonic penguin, which ensures a strong phase difference with respect to the tree-level amplitude. Indeed, for three-body \(B^{+}\rightarrow h^{+}h^{-}h^{+}\) decays, a 80% CP asymmetry was observed in certain kinematic regions [69].

The complication with interpreting CP asymmetries in charmless B decays is that its size depends on the amplitude magnitudes and the strong phase, which are difficult to calculate. Fully hadronic decays cannot be factorised in the same way that semileptonic decays can, which makes theoretical calculations much more challenging. For this reason, cancelations between similar decays can be useful to bring uncertainties from QCD under control. A well-known example of this is to compare the CP asymmetry of \(B^{0} \rightarrow K^{+}\pi ^{-}\) decays with its isospin partner, \(B^{+} \rightarrow K^{+} \pi ^{0}\). Given that only the spectator quark is different between the two decays, one would expect the CP asymmetry to be very similar if not the same. However, the CP asymmetries of the two decays, \(A_{CP}(B^{0}\rightarrow K^{+}\pi ^{-}) = -0.084 \pm 0.004\) and \(A_{CP}(B^{+}\rightarrow K^{+}\pi ^{0}) = 0.040 \pm 0.021\), are over five standard deviations from each other, indicating either unexpectedly large hadronic effects or even contributions from NP. This difference is surprising and is known in the field as the \(K\pi \) puzzle. Given the difficult signature of the decay \(\pi ^{0}\rightarrow \gamma \gamma \), the uncertainty is dominated by the \(B^{+}\) transition. As there is only one charged particle in the signature, there is no B vertex to reconstruct which is normally a vital signature to identify b-hadron decays. A measurement of the \(B^{+}\rightarrow K^{+}\pi ^{0}\) decay was therefore considered impossible at a hadron collider. However, due to the very large production at the LHC, a tight selection could be employed using the four-momentum of the \(\pi ^{0}\) which allowed to reduce the background enough to determine the signal for a CP measurement [70]. The signal is not as clean as fully charged b-hadron decays, is still of remarkable purity given the challenging signature. The LHCb analysis allowed for the most precise measurement of the CP asymmetry in the world, which significantly increased the \(K\pi \) puzzle. The success of this measurement gives encouragement to other related decays such as \(B^{+}\rightarrow pi^{+} K^{0}\), which should be of a similar challenge and are crucial to understand the \(K\pi \) puzzle further via a sum-rule relation [71].

Finally, all these searches described in this section so far have been using B meson decays. Despite the connection to baryeogenesis, no observation of CP violation has been made yet in baryons themselves. There was hope with evidence reported in \(\Lambda _b^{0}\rightarrow p \pi ^{+}\pi ^{-}\pi ^{+}\) decays [72] but unfortunately faded when analysed with an increased dataset [73]. Searches are on-going in several different decay modes of both \(\Lambda _b\) and \(\Xi _{b}\) baryons. Given the lack of baryon oscillations in the SM, the main area of focus is charmless decays. Given the amplitudes involved it is inevitible that CP violation will be eventually seen in baryons, which will mark another important milestone in the history of CP violation attained by LHCb after the observation of CP violation in charm decays [74].

6 The case for a second LHCb upgrade

The second LHCb upgrade [46, 75] aims for an unprecedented luminosity of \(1.5\times 10^{34}\) cm\(^{-2}\) s\(^{-1}\), which is expected to cause approximately 40 visible proton–proton interactions per crossing, yielding about 2000 charged particles within the LHCb’s area of coverage. To handle this substantial increase, changes are envisioned for the existing spectrometer components, including increasing their granularity, lowering the material within the detector, and incorporating advanced timing precision. This initiative will strive to preserve or enhance the current performance of the detector, particularly in areas such as track-finding efficiency and momentum resolution.

The LHCb Upgrade II will fully harness the HL-LHC potential as a flavour physics machine while maintaining a rich and diverse research programme. A glimpse of the potential available at the LHC can be found in Fig. 6, which shows the vast increase of the number of b-hadrons that will be produced. Some of the measurements that LHCb Upgrade II will undertake will constrain theoretical models for decades to come. From the conclusion of the HL-LHC to the commencement of the future circular collider (FCC), it will be crucial to have a robust set of measurements to guide future theoretical work. Neglecting to fully exploit the third family’s flavour potential could result in a missed opportunity.

Fig. 6
figure 6

Number of b-hadrons produced or to be produced at the B-factories, run II of the LHC, the Belle II experiment and at the high luminosity LHC

Although the LHCb Upgrade II presents great opportunities, it also poses substantial technological challenges. The key points of the upgrade for tracking, particle identification, and data processing are summarised below.

The VELO is a critical component of the LHCb upgrade physics agenda, designed to perform real-time track reconstruction from all LHC bunch crossings within the software trigger system. The enhanced luminosity projected for Upgrade II necessitates a new and more capable VELO. This includes the ability to manage an enlarged data output, higher radiation levels, and an increased occupancy. Innovative techniques are being devised to facilitate real-time pattern recognition and correct association of each b-hadron to its originating primary vertex (PV). This effort will involve the development of a novel pixel detector with superior rate and timing capabilities, and a more efficient mechanical design.

To better associate the decay products of heavy flavor with their parent PV, effective utilization of timing information is critical. Research indicates that through a timing precision of 50-100 ps, mis-association levels of PVs can be curtailed from about 20% to approximately 5%. Further, reducing the pixel pitch from its current size of \(55\mu \)m at Upgrade II is considered advantageous, especially for the VELO’s innermost region. Timing will not only aid in the track reconstruction process but also help conserve computational resources. R &D for the VELO is currently ongoing with the major candidate solutions for the sensors including Planar Sensors, Low Gain Avalanche Diodes (LGADs) and 3D sensors.

Upgrade II also plans changes to the downstream tracking system, which currently comprises a silicon strip detector (UT) and three tracking stations (T-stations). The redesigned system will need to accommodate higher occupancies by enhancing detector granularity and minimizing incorrect matches between upstream and downstream track segments.

For the inner portions of the T-stations, as well as the UT, Monolithic Active Pixel Sensors (MAPS) are being considered. Optimization efforts for the UT are ongoing, and both HV- and LV-CMOS solutions are under consideration. Regarding the inner part of the T-stations, a preliminary prototype of the MightyPix silicon detector, based on HV-CMOS technology, has been tested. Significant synergies exist with R &D work conducted in the context of both the Mu3e experiment and the ATLAS Upgrade. The requisite number of detector layers is currently being investigated. As with the VELO, the integration of timing information into the tracking system could substantially improve the accuracy of matching upstream and downstream track segments.

Quality particle identification (PID) is a vital component in precision flavor measurements. The PID subdetectors in the ongoing experiment will be improved and, in some cases, enhanced for the upcoming Upgrade II. The primary enhancements include improved granularity and for select subdetectors, rapid timing of a few tens of picoseconds. The improved timing helps to associate signals with a few proton–proton interactions in the bunch crossing.

The RICH system for Upgrade II builds on the current detectors, consisting of an upstream RICH 1 for lower momentum tracks, and a downstream RICH 2. To manage increased track multiplicity, it will be necessary to replace the current MaPMTs with more granular photodetectors. Multiple technologies are under consideration, with SiPMs being a prime candidate.

An opportunity to enhance low-momentum hadron-identification capabilities is under consideration by installing a TORCH detector. This detector measures time-of-flight by detecting internally reflected Cherenkov light produced in a thin quartz plane. The TORCH can help identify kaons in the region below 10 GeV/c.

The electron, photon, and pi-zero identification provided by the current electromagnetic calorimeter (ECAL) has proved essential. A suitably designed new ECAL will be necessary for achieving high performance in key modes and for ensuring a broad physics program at Upgrade II. One of the challenges for the ECAL at Upgrade II will be the harsh radiation environment. Solutions to these challenges include reducing the Molière radius of the converter, moving to a smaller cell size, and using fast timing information.

Upgrade II aims to provide ECAL performance at least as good as the current detector in Run 1 and Run 2 conditions. An active R &D program has begun to investigate potential technologies for the Upgrade II ECAL. It’s possible that the final detector might use multiple solutions due to the rapid variation of multiplicity and radiation flux with position.

The current muon system in Upgrade I conditions would result in a degraded performance due to the higher background flux foreseen for Upgrade II. Enhanced shielding will be necessary in front of the muon detector to mitigate the incidence of punch-through particles. This could involve substituting the HCAL with up to 1.7 ms of iron, thereby adding an extra four interaction lengths compared to the existing setup. As the HCAL’s main function of contributing to the hardware trigger will become obsolete starting from Run 3, this change is feasible. In terms of the muon system, upgrades are needed for the detectors in the most central areas of all stations, aiming for a design with both superior granularity and rate capabilities.

Potential solutions are the micro-resistive WELL detector (\(\mu \)-RWELL) for the higher occupancy region, and MWPC or RPC detectors in the lower flux region. R &D is ongoing to study all of these possibilities.

The LHCb Upgrade II detector is expected to generate data in the range of 200 Tb per second. This vast amount of data necessitates real-time processing and a substantial reduction before it can be stored. Advances in radiation-resistant optical links and commercial networking technology, along with the specific design of the LHCb that positions the readout cables outside of the detector acceptance, should enable the transmission of this massive data volume to a computational farm by the time of Upgrade II.

Data processing in Upgrade II presents a significant challenge. Traditional trigger methods will not suffice for substantial data rate reduction at even the initial trigger stages. Instead, the data processing approach for Upgrade II will focus on pile-up suppression, prioritizing the early-stage discarding of detector hits not associated with the specific proton–proton interaction of interest. Accurate timing data is vital for rapidly selecting reconstructed objects based on their originating proton–proton interaction. Therefore, access to detailed timing information across multiple subdetectors is imperative. It’s equally significant to carry out an early-stage reconstruction of charged particle trajectories and neutral particle clusters created within the LHCb acceptance, to optimally associate specific heavy flavour indicators with proton–proton interactions. LHCb has already showcased its proficiency in performing a comprehensive offline-quality detector alignment, calibration, and reconstruction in a near-real-time environment during Run 2, as well as its capacity to conduct precision physics assessments using this real-time data processing.

Adapting this capability to the more rigorous conditions of Upgrade II will require the construction of a processor farm leveraging whichever technology or technologies are deemed most commercially sustainable over a decade from now. Given the increasing trend towards more diverse computing architectures, with CPU server farms being augmented with GPU or FPGA accelerators, it becomes crucial to sustain and expand collaborations with computing institutions that specialize in the design of these hybrid architectures. Research has been carried out on FPGA-based downstream tracking. Collaborations with industry partners will be instrumental in ensuring that LHCb’s trigger and reconstruction algorithms are optimally calibrated to make the best use of the most cost-effective architecture.

In conclusion, despite the technological challenges of the LHCb Upgrade II, it also offers a unique opportunity to develop novel detector technologies during an exciting period of anticipation while we await the FCC. The combination of achieving cutting-edge physics goals and advancing technological capabilities underscores the significance of the LHCb Upgrade II.

7 Summary

In 2008, shortly before the LHCb experiment was due to start collecting data, the experiment published a roadmap [76] of six key measurements to be performed with high priority in the upcoming run. These were dominated by decays with fully charged final states, where the only leptons considered were muons. In this review we highlighted three areas of b physics, but even within that highlight one can see a huge diversity of signatures, involving electrons, \(\tau \)-leptons, neutrinos and neutral pions. The last 15 years has brought about a huge expansion of the LHCb programme beyond what was considered possible at a hadron collider and is one of the great successes of the LHCb experiment. The next 15 years will certainly contain more surprises in addition to the expected improvements to the measurements already established within the programme. The motivation for the LHCb upgrades is clear.

This is not to mention the huge amount of physics potential for which we could not cover in this review. We did not cover the pentaquark discovery [77] and associated measurements into exotic spectrospopy, which have a huge interest both in and outside the field. Beyond b-physics the LHCb physics programme includes a large and diverse set of measurements, which have steadily expanded over the last decade. In flavour physics, measurements of a wide variety of charm-quark measurements from oscillations to spectroscopy. In particular, one of LHCb’s landmark measurements has been the discovery of CP violation in \(D\rightarrow hh\) decays [74], which is a milestone in the history of the field. Beyond flavour, there is a broad programme of electroweak physics measurements, from drell-yan to top quark measurements. Highlights here include the W boson mass measurement [78] and ion-proton cross-section measurements [79], which inform background estimates in astrophysics.

Given the unique opportunity of HL-HLC, pushing the LHCb experiment to its maximal capability is mandatory. There is a possibility that the current discrepancies in \(b\rightarrow s\mu ^{+}\mu ^{-}\) and \(b\rightarrow c\tau ^{-}\bar{\nu }_{\tau }\) could be resolved or new anomalies could emerge. In both cases, the LHCb Upgrade II will be a fundamental resource for future theoretical work and, in turn, for upcoming experiments. It also provides an important bridge to design and implement new detector technologies in the time period between the final upgrades of ATLAS and CMS and the new generation accelerator complex, the FCC. It is clear to us that neither LHCb nor b-hadron physics will lose any relevance over the next decade, whether or not NP is discovered in that time.