Customer sojourn time in GI/G/1 feedback queue in the presence of heavy tails

We consider a single-server GI/GI/1 queueing system with feedback. We assume the service times distribution to be (intermediate) regularly varying. We find the tail asymptotics for a customer's sojourn time in two regimes: the customer arrives in an empty system, and the customer arrives in the system in the stationary regime. In particular, in the case of Poisson input we use the branching processes structure and provide more precise formulae. As auxiliary results, we find the tail asymptotics for the busy period distribution in a single-server queue with an intermediate varying service times distribution and establish the principle-of-a-single-big-jump equivalences that characterise the asymptotics.


Introduction
In queueing theory, the sojourn time U of a customer in a queueing system is one of important characteristics, this is the time from its arrival instant to departure instant. In general, the distribution of U is hard to find analytically, and research interest is directed to the asymptotics of the tail probability, P(U > x), as x → ∞ under various stochastic assumptions. Among them, the following assumption is typically used. probability p 20 , or joins the queue to server 1, with probability p 21 = 1 − p 20 . Customers are server in the order of their (external and internal) arrival to the servers.
If we let σ (2) n ≡ 0 and p 12 p 21 = p, we obtain our model as a particular case indeed. So the study of our model is not only of interest itself, but also opens a window to analysing a broad class of more general models.
In the G I /G I /1 feedback queue, one may change the service order in such a way that each customer continuously gets service without interruption when it completes service and returns to the queue. Then such a system is nothing else than the standard G I /G I /1 queue with "new" i.i.d. service times K i j=1 σ ( j) i , and the sojourn time is again the sum of the waiting time and of the (new) service time. Although the busy period, which is the time from the moment when the system becomes non-empty to the moment when it is again empty, is unchanged by this modification, the sojourn time does change. We will use this modified system to study the tail asymptotic of the busy period.
Thus, our analysis is connected to the standard G I /G I /1 queue. In this case, there is no feedback, and the monotonicity is satisfied. Hence, the waiting time of a tagged customer is a key characteristic because the sojourn time U is the sum of the waiting and service times, which are independent. In particular, the stationary waiting time is a major target for the tail asymptotic analysis. Let u 0 be the unfinished work found by the "initial" customer 1 that arrives at the system at time 0, and let W n be the waiting times of the nth arriving customer. Then, W 1 = u 0 and we have the Lindley recursion: We assume both inter-arrival and service times to have finite means, a = Et n and b = Eσ n . Here W n forms a Markov chain which is stable (i.e. converges in distribution to the limiting/stationary random variable W = W ∞ ) if the traffic intensity ρ := b/a is less than 1. It is well-known (see e.g. [1]) that if u 0 = 0, then W coincides in distribution with the supremum M = sup n≥0 n i=1 (σ n − t n ) of a random walk with increments σ n − t n . The tail asymptotics for P(M > x) as x → ∞ is known under the light-tail and heavy-tail regimes. In the case of light tails, there are three types of the tail asymptotics, depending on properties of the moment generating function ϕ(s) = E exp(sσ ) -see e.g. [6] and references therein. In the case of heavy tails, the tail asymptotics are known in the class of so-called subexponential distributions and are based on the principle of a single big jump (PSBJ): M takes a large value if one of the service times is large. This PSBJ has been used for the asymptotic analysis in several other (relatively simple) stable queueing models, for a number of characteristics (waiting time, sojourn time, queue length, busy cycle/period, maximal data, etc.) that possess the monotonicity property (1) (see e.g. [2]).
Our proofs rely on the tail asymptotics for the first and stationary busy periods of the system. We establish the PSBJ for the busy period first. This allows us to establish the principle for the sojourn time since the tail distribution asymptotics of the busy period is of the same order with that of the sojourn time. Then insensitivity properties of the intermediate varying distributions (see Appendix A again) allow us to compute the exact tail asymptotics for the sojourn time. The main result from [7] is a key tool in our analysis.
The paper is organised as follows. Section 2 formally introduces the model and presents main results. Section 3 states the tail asymptotic of the busy period and the PSBJ. All theorems from Sects. 2 and 3 are proved in Sect. 4. The Appendix consists of three parts. Part A contains an overview on basic properties of heavy-tailed distributions and part B the proof of Corollary 3.1. In part C, we propose an alternative approach to the proof of Corollary 3.2.
Throughout the paper, we use the following notation: 1(·) is the indicator function of the event "·". For two positive functions f and g, we write f ( if lim sup x→∞ f (x)/g(x) ≤ 1. For a distribution function F, its tail F is defined as For random variables X, Y with distributions F, G, respectively, X = st Y if F = G, and X ≤ st Y if F(x) ≤ G(x) for all real x. Two families of events A x and B x of non-zero probabilities are equivalent, where Note also that A x B x is stronger than equivalence P(A x ) ∼ P(B x ). We complete the Introduction by a short

Summary of main classes of heavy tail distributions
In this paper, we are concerned with several classes of heavy tail distributions. We list their definitions below. Their basic properties are discussed in Appendix A.
In all definitions below, we assume that F(x) > 0 for all x.
1. Distribution F on the real line belongs to the class L of long-tailed distributions if, for some y > 0 and as x → ∞, (we may write equivalently F(x + y) ∼ F(x)). 2. Distribution F on the positive half-line belongs to the class S of subexponential distributions if  (1.6) where L(x) is a slowly varying function, i.e. L(cx) ∼ L(x) as x → ∞, for any c > 0.

Distribution
The following relations between the classes introduced above may be found, say, in the books [3] or [9]: in the class of distributions F with finite m + (F), (1.7)

The Modelling Assumptions and Main Results
In this section, we describe the dynamics of the sojourn time of a tagged customer (customer 1), and present main results on the tail asymptotics of its sojourn time. In Sect.

G I/ G I/1 Feedback Queue
Let K be the number of services of the tagged customer until its departure. By the feedback assumption, K is geometrically distributed with parameter p, that is, and independent of everything else. Throughout the paper, we make the following assumptions: (i) The exogenous arrival process is a renewal process with a finite mean interarrival time a > 0. (ii) All the service times that start after time 0 are i.i.d. with finite mean b > 0, they are jointly independent of the arrival process. (iii) The system is stable, that is, We denote the counting process of the exogenous arrivals by N e (·) ≡ {N e (t); t ≥ 0}. We use the notation G for the service time distribution, and use σ for a random variable subject to G.
Let (X 0 , R s 0 ) be the pair of the number of earlier customers and the remaining service time of a customer being served at time 0, where R s 0 = 0 if there is no customer in the system. Let u 0 be the waiting time of the tagged customer before the start of its first service, Then where σ 0,i 's for i ≥ 2 are i.i.d. random variables each of which has the same distribution as σ . There are two typical scenarios for the initial distribution, that is, the distribution of (X 0 , R s 0 ). (2a) A tagged arriving customer finds the system empty. That is, (2b) A tagged arriving customer finds X 0 customers and the remaining service time R s 0 of the customer being served. Thus, the initial state (X 0 , R s 0 ) = (0, 0). In this paper, we assume that the service time distribution is heavy tailed, and mainly consider the tail asymptotic of the sojourn time distribution of the G I /G I /1 feedback queue under the scenario (2a). The case (2b) when X 0 and R s 0 are bounded by a constant may be studied very similarly to the case (2a), therefore we do not analyse it. We consider the case (2b) when (X 0 , R s 0 ) is subject to the stationary distribution embedded at the arrival instants. For given (X 0 , R s 0 ), we have defined u 0 . Let X k be the queue length behind the tagged customer when it finished its kth service for k ≥ 1 when the tagged customer gets service at least k times. Similarly, let U k be the sojourn time of the tagged customer measured from its (k − 1)th service completion to its kth service completion, and let T k be the sojourn time of the tagged customer just after its kth service completion.
We now formally define random variables X k , U k and T k by induction. Let T 0 = 0. Denote the kth service time of the tagged customer by σ k,0 , while σ k,0 , i = 1, . . . , X k−1 are the service times of the customers waiting before the tagged one on its kth return. Note that σ k,i 's for k ≥ 1, i ≥ 0 are i.i.d. random variables subject to the same distribution as σ . Then, X k , U k and T k for k ≥ 1 are defined as

4)
where u 0 is given by (2.3), and N B k (n)'s are i.i.d. random variables each of which is subject to the Binomial distribution with parameters n, p. The dynamics of the sojourn time is depicted above when X 0 = 0, that is, a tagged customer finds the system empty ( Fig. 1).
To make clear the dependence of X k , U k , T k , we introduce a filtration {F t ; t ≥ 0} as where N s (t) and N r (t) are the numbers of customers who completed service and who return to the queue, respectively, up to time t. Clearly, T k is a F t -stopping time, and X k and U k are F T k -measurable. Furthermore, σ k,0 and σ k,i for i ≥ 1 are independent of F T k−1 . Then U , the sojourn time of the tagged customer, may be represented as For k ≥ 0, let Y k = k =0 X for k ≥ 0, which is the total number of external and internal arrivals to the queue up to time T k plus the number of customers in system at time 0. Then Hence, under scenario (2a), we have u 0 = X 0 = 0, so while, under scenario (2b), where σ i 's are i.i.d. random variables each of which has the same distribution as σ . Note that K + Y K −1 is F T K −1 -measurable that depends, in general, on all σ i 's of customers who arrive before T K −1 . This causes considerable difficulty in the asymptotic analysis of U . Thus, we need to consider dependence structure in the representation of U . Furthermore, {(U k , X k ); k ≥ 0} is generally not a Markov chain for a general renewal process.
On the other hand, if the arrival process N e (·) is Poisson, then not only {(U k , X k ); k ≥ 0} but also {X k ; k ≥ 0} is a Markov chain with respect to the filtration {F T k ; k ≥ 0}. In this case, we may obtain exact expressions for EX k and then an explicit form for the tail asymptotics.

M/ G I/1 Feedback Queue and Branching Process
In this subsection, we assume that the exogenous arrival process is Poisson with rate λ > 0. This model is analytically studied using Laplace transforms in [12], but no asymptotic results are given there. Note that we may consider {X k ; k ≥ 0} as a branching process and directly compute E(X k ), which then will be used for the general renewal input case.
Since the Poisson process N e (·) has independent increments, (2.6) is simplified to using independent Poisson processes N e k and independent Binomial random variables N B k (n). Furthermore, (2.11) can be written as 12) where N e k,i (·)'s are independent Poisson processes with rate λ. Hence, {X k ; k ≥ 1} is a branching process with immigration.
Due to the branching structure, we can compute the moments of X k explicitly. We are particularly interested in their means. From (2.12), we have where r = λb + p. By the stability condition (2.2), r < 1, and we have Hence, we have a uniform bound: (2.14) Furthermore, we have Under the scenario (2a), E(X k ) of the M/G I /1 feedback queue will be used for the tail asymptotic of the sojourn time in the G I /G I /1 feedback queue. Thus, we introduce notations for them. Let X We will use m

Main Results
We are ready to present the main results of this paper. They are proved in Sect. 4.

Theorem 2.1
For the stable G I /G I /1 feedback queue, assume that its service time distribution is intermediate regularly varying (IRV). If the tagged customer finds the system empty, then This corollary is easily obtained from arguments used in the proof of Theorem 2.1. On the other hand, if we take the geometrically weighted sum of (2.18) and if the interchange of this sum and the asymptotic limit are allowed, then we have (2.17). This interchange of the limits is legitimated by Theorem 3.2. However, Corollary 2.1 itself can be directly proved. We provide such a proof for a slightly extended version of Corollary 2.1 in Appendix C.
We next present the tail asymptotic for a tagged customer that arrives in the stationary system. By "stationary" we mean stationary in discrete time, i.e. at embedded arrival epochs, this is detailed in Sect. 4.4.

Theorem 2.2 Let U 0 be the sojourn time of a typical customer in the stationary G I /G I /1 feedback queue with IRV distribution G of service times with mean b, i.i.d. inter-arrival times with mean a and probability of feedback p
a slowly varying function, then (2.20) for sufficiently large x. In this case, let σ e be a random variable subject to G e , then we can replace σ I by σ e in (2.19), multiplying its right-hand side by b.
K −1 for the renewal arrivals is involved in (2.19). This is different from (2.17), and may come from averaging in the steady state. (c) It may be interesting to compare the asymptotics in (2.19) with those without feedback, which is well known (e.g., see [2]). Namely, let the stationary sojourn time U 0 in the standard G I /G I /1 queue with inter-arrival times {t n } and with service times {σ H n }. where σ H n has the same distribution as K i=1 σ i . If σ I has a subexponential distribution, then

Busy Period and the Principle of a Single Big Jump
In this section, we present the Principle of a Single Big Jump (PSJB) in Theorem 3.2 below, which will be used for a proof Theorem 2.1. For that, we first provide an auxiliary result on the tail asymptotics of the busy period in the G I /G I /1 queue without feedback. Denote its service time distribution by H and let σ H i be the ith service time. It is assumed that the arrivals are subject to the renewal process N e with interarrival times t i with mean a, and H has a finite and positive mean b H > 0. Denote the traffic intensity by ρ ≡ b H /a < 1. Let B be the (duration of the) first busy period in this G I /G I /1 queue, which is the time from the instant when the system becomes non-empty to the instant when it again becomes empty. We here omit the subscript H for ρ, B, because they will be unchanged for the G I /G I /1 feedback queue. We finally let τ H be the number of customers served in the first busy period. We Recall the definitions of classes of heavy-tailed distributions L, S * , IRV and RV at the end of Sect. 1. The following theorem is proved in Sect. 4.1.

Theorem 3.1 Consider a stable G I
If, in addition, H ∈ S * , then, for any 0 < c < 1, Finally, if H ∈ IRV, then, as x → ∞, Remark 3.1 For the class of regularly varying tails, the equivalence (3.3) was proved by Zwart in [13]. We provide a different proof which is shorter and works for a broader class of distributions. Our proof is based on probabilistic intuition related to the principle of a single big jump. A similar result holds for another class of distributions that overlaps with the IRV class but does not contain it, see e.g. [11].
Recall the equivalence A x B x for two families of events A x and B x with variable x. We have the following corollary, which is proved in Appendix B.
and, for any ε > 0, one can choose N = N e (ε) ≥ 1 such that, as x → ∞, Furthermore, the following PSBJ holds: We now return to the G I /G I /1 feedback queue with the service time distribution G. Assume that the first customer arrives at the system at time instant T 0 = 0 and finds it empty. Recall that K i is the number of services ith customer has in the system, K i 's are independent of everything else and i.i.d with the same geometric distribution as K [see and denote its distribution by H . Since the length of the busy period, B, does not depend on the order of services, we may allow the server to proceed with services of lengths σ j i , like in the queue with feedback, and conclude that the (the lengths of) the busy periods are the same in both queues. Similarly, the traffic intensity ρ in the new queue without feedback coincides with that in the G I /G I /1 queue with feedback. Furthermore, let τ be the number of service times in the first busy period of this feedback queue. Then, τ = τ H i=1 K i , and therefore we have We now consider the G I /G I /1 feedback queue introduced in Sect. 2.1. We establish the PSBJ, i.e. show that, for large x, the rare event {U > x} occurs mostly due to a big value of one of the service times. Our proof of Theorem 3.2 is based on Theorem 3.1 and is given in Sect. 4.2.

Theorem 3.2 Consider a stable single-server queue G I /G I /1 with feedback. Assume that the service times distribution is intermediate regularly varying. Denote by U be the sojourn time of the first customer, and let
If there exists a collection of positive functions {g k, ,i, j (x)} such that, as x → ∞,

8)
and constants C k, ,i, j such that, for any k

Proof of Theorem 3.1
We will prove Theorem 3.1 for the tail asymptotics of the busy period B only. The proof for τ H , the number of arriving customers in the busy period, is similar. It is enough to prove the lower and upper bounds in (3.1) and (3.2). Then the equivalences in (3.3) follow by letting c tend to 1 and using the property of IRV distributions.
Then, we have Here, the first inequality in (4.1) holds since S H τ H is non-positive, and the second inequality comes from the following facts. Events D i are disjoint and, given the event D i , we have Thus, (4.1) holds.
The events {A i } form a stationary sequence. Due to the SLLN, for any ε > 0, one can choose R so large that P(A i ) ≥ 1 − ε. For this ε and any N ≥ 1, we can choose sufficiently large C such that Hence, (4.1) implies that, as x → ∞, and therefore the long-tailedness of distribution H and (iii) of Remark A.1 yield Letting first N to infinity and then ε to zero completes the proof of the first inequality of (3.1).
where the equivalence follows from Theorem A.1. Further, where the first equivalence follows from the long-tailedness of the distribution of ψ H 1 and the second from Remark A.1. Letting ε tend to zero, we have Recall that we consider the scenario where the initial customer 1 arrives at the empty system.
Further, by Corollary 3.1, the PSBJ for B holds: Here τ is the number of customers served within the first busy period. Combining To derive the exact asymptotics for P(U > x), we recall that, for 1 ≤ k < K ≡ K 1 , X k ≥ 0 is the total number of services of other customers between the kth and the (k + 1)st services of customer 1, and let σ k,i be the service time of the i service there, 1 ≤ i ≤ X k . Further, under the scenario (2a), X 0 = 0. Then let ν ≥ 0 be the total number of services of other customers after the departure of the first customer within the busy period, and let σ * i be the ith service time there, 1 ≤ i ≤ ν. Then random variables σ k,.i and σ * i are i.i.d. with the same distribution as σ and U is given by (2.9). From (4.5), we get, On the other hand, we have Proof Indeed, the term in the right-hand side of (4.6) is bigger than P(D N (x)) and smaller than the sum P( Consider again the auxiliary G I /G I /1 queue with service times σ H i = K i j=1 σ ( j) i and the first-come-first-served service discipline. Consider the following majorant: assume that at the beginning of the first cycle, in addition to customer 1, an extra K − 1 new customers arrive, so there are K arrivals in total. Here K is a geometric random variable with parameter p that does not depend on service times. Then the first busy period in this queue has the same distribution as K i=1 B i where B i are i.i.d. random variables that have the same distribution as B and do not depend on K . By monotonicity, Due to (A.4), the latter probability is equivalent, as x → ∞, to We can go further and obtain the following result.

10)
Then, for any ε > 0, one can choose a positive integer R such that where the event D N (x) was defined in (4.8). Further, Proof Indeed, where the term E((τ +1)1(τ > R)) may be made as small as possible by taking a sufficiently large R. Then (4.12) follows since the probability of a union of events is always smaller than the sum of their probabilities, and is bigger than the sum of probabilities of events minus the sum of probabilities of pairwise intersections of events. Each probability of intersection of two independent events is smaller than therefore their finite sum is o(G(x)) and (4.12) follows.
We are now in a final step of the proof of Theorem 3.2. For k ≥ 1, , j ≥ 0, define D k, , j as where the second equality holds because K is geometrically distributed. Then, Lemma 4.5 implies (3.11) for g k, ,i, j (x) = P k, ,i, j (x) since, for any k, , j ≥ i, where, recall, τ is the total number of customers served in the first busy period. Clearly, (3.11) is also valid for a general {g k, ,i, j (x)} because of the conditions (3.9) and (3.10). This completes the proof of Theorem 3.2.

Proof of Theorem 2.1
We first recall the notation: U 1 , U 2 , . . . and X 0 , X 1 , . . . are the service cycles and the number of customers other than the tagged customer served in the cycles, respectively. Here u 0 = X 0 = 0. In general, the sojourn time is a randomly stopped sum of i.i.d. positive random variables, and both the summands and the counting random variable have heavy-tailed distributions. It is known that it is hard to study the tail asymptotics for general heavy-tailed distributions (see, e.g., [10]) in this case. We proceed under the assumption that the service time distribution is intermediate regularly varying.
Recall that σ k,0 is the kth service time of the tagged customer and, for i = 1, . . . , X k , σ k,i is the ith service time in the queue X k . Further, T k = k =1 U be the time instant when the kth service of the tagged customer is completed, where U 1 = σ 1,0 . Introduce the notation which is the remaining time the tagged customer spends in the system after the completion of the kth service, and let v k be the residual inter-arrival time of the input when the kth service of the tagged customer ends.
In what follows, we will say that an event involving some constants and functions/sequences occurs "with high probability" if, for any ε > 0, there exists constants and functions/sequences (that depend on ε) with the desired properties such that the event occurs with probability at least 1 − ε.
For example, let S σ n = n 1 σ i be the sum of i.i.d. random variables with finite mean b. Then the phrase "with high probability (WHP), for all n = 1, 2, . . ., with C > 0 and δ n ↓ 0" means that "for any ε > 0, there exist a constant C ≡ C ε > 0 and a sequence δ n ≡ δ n (ε) ↓ 0 such that the probability of the event {S σ n ∈ (n(a − δ n ) − C, n(a + δ n ) + C), for all n ≥ 1} is at least 1 − ε". We can say equivalently that "WHP, for all n = 1, 2, . . ., S σ n ∈ (na − o(n), na + o(n)) ", or, simply, "WHP, S σ n ∼ an", and this means that "for any ε > 0, there exists a positive function h(n) = h ε (n) which is an o(n)-function (it may tend to infinity, but slower than n) and is such that the probability of the event Namely, we show that, for all k ≥ 1, ≥ 0, 0 ≤ i ≤ j, We prove (4.16) by induction on ≥ 0, for each fixed k ≥ 1, 0 ≤ i ≤ j. Lower bound = 0.
Since σ k,i > x implies that U > x and σ k,i > (1 − ρ)x, the lower bound for the LHS of (4.16) is There is a constant w > 0 such that T k−1 ≤ w and 0≤i ≤ j,i =i σ k,i ≤ w WHP. Then U ≤ 2w + σ k,i , so the upper bound for the LHS of (4.16) is Letting ε tend to zero in this upper bounds yields that the lower and upper bounds are asymptotically identical. Since m (0) 0 = 0, they are further identical to g k,0,i, j (x) of (4.15). Thus, (4.16) is verified.
Turn to the case = 1. Lower bound = 1. Like in the case = 0, replace all other service times σ k,i , i = i by zero. Assume that all j customers from the group X k−1 leave the system after their service completions. WHP, v k−1 ≤ w. Given y = σ k,i is large and much bigger than w, we have that at least N e (y − w) customers arrive during time U k ≥ σ k,i = y. Again WHP, and, again WHP, their total service time is within the time interval (λby − o(y), λby + o(y)). Therefore, and the RHS is bigger than x if y > x/(1 + λb) + o(x). Therefore, the lower bound for the LHS of (4.16) is Then, WHP, U k ≤ y + 2w and the number of external arrivals within U k is bounded above by 1 + N e (y + 2w) = λy + o(y), again WHP. Assume that all X k−1 = j customers stay in the system after their services. Then again j + 1 + N e (y + w) = λy + o(y), WHP. Therefore, U k+1 = bλy + o(y). Then we arrive at the upper bound that meets the lower bound.

Induction step
We can provide induction for any finite number of steps. Here is the induction base. Assume that σ k,i = y 1 and that, after ≥ 1 steps, T k+ ∼ (1 + m (0) )y for 0 ≤ ≤ , and there are X k+ −1 customers in the queue and that X k+ −1 = wy +o(y), WHP, where w > 0. Then, combining upper and lower bounds, we may conclude that, again WHP, U k+ = bwy + o(y) and then  Hence, by (4.18), +1 )y. This completes the induction step for + 1.

PSBJ for the Stationary Queue
We now consider the case where customer 1 arrives to the stationary queue and denote by U 0 its sojourn time. In this section, we frequently use the following notation: for a distribution F having a finite mean, By "stationarity" we mean stationarity in discrete time, i.e. at embedded arrival epochs. So we assume that the system has started from time −∞ and that customer 1 arrives at time t 1 ≡ 0, customers with indices k ≤ 0 enter the system at time instants t k = − 0 j=k t j and customers with indices k ≥ 2 at time instants Then the stationary busy cycle covering 0 starts at t k , k ≤ 0 if So, if B 0 is the remaining duration of the busy period viewed at time 0, then where B k is the duration of the period that starts at time t k given that customer k arrives in the empty system (then, in particular, B = B 0 ). See Fig. 2. Let which is the number of customers arriving at or after time 0 in the busy period when it starts at time t k , and let  (4.20) and therefore, applying the PSBJ of Corollary 3.1 to each busy period B −k , where σ H −k+i , i ≥ 0 is the service time of the ith customer arriving in the busy period that starts at time t −k .
Hence, letting We first consider the event A 0 + (x), which is a contribution of big jumps at or after time 0, and show that its probability is negligible with respect to H I (x), as x → ∞. Clearly, for any positive function h(x) and for any ε ∈ (0, a), Then if one takes, say, h(x) = x c for some c < 1. Here the second inequality follows since τ H −k+i ≥ i = τ H −k+i ≤ i − 1 c is independent of σ H −k+i , the third inequality from Chernoff's inequality, for a small α > 0, and the final conclusion from property (A.7) in the Appendix.
Thus, we only need to evaluate the contribution of big jumps that occur before time 0. Namely, we analyse A 0 − (x). Note that, for any k 0 > 0, the probability of the event (1 − ρ))) which is negligible with respect to G I (x (1 − ρ)). Therefore, one can choose an integer-valued h(x) → ∞ such that P A x(1 − ρ))). So we may apply again the SLLN, t −k ∼ ak for sufficiently large k, to get On the other hand, for h(x) ↑ ∞ sufficiently slowly and for an appropriate sequence ε ↓ 0 (that comes from the SLLN), we have Since B 0 > x on the event E(x), we arrive at the following PSBJ for the stationary busy period.

Lemma 4.6 If the G I /G I /1 feedback queue is stable and its service time distribution has an IRV distribution with a finite mean, then
The lemma implies that since the sum of the probabilities of pairwise intersections is of order O(G I 2 (x)) =

o(G I (x)).
Then we may conclude that the principle of a single big jump can be applied to the stationary sojourn time too: where the second equivalence is valid for any integer-valued function h(x) ↑ ∞, h(x) = o(x) and follows from (4.19) and from the properties of IRV and integrated tail distributions, see Appendix A.

Proof of Theorem 2.2
First, we comment that it is easy to obtain the logarithmic asymptotics for the stationary sojourn time. Since the sojourn time of the customer entering the stationary queue at time 0 is not bigger than the stationary busy period and is not smaller than the stationary sojourn time in the auxiliary queue without feedback, and since both bounds have tail distributions that are proportional to the integrated tail distribution of a single service time (see the Appendix for definitions), we immediately get the logarithmic tail asymptotics: Now we provide highlights for obtaining the exact tail asymptotics for the stationary sojourn time distribution and give the final answer. For this, we use the following simplifications, which are made rigorous in the "WHP" terminology and due to the o(x)-insensitivity of the service-time distribution.
(1) We observe that the order of services prior to time 0 is not important for the customer that enters the stationary queue at time 0: the joint distribution of the residual service time and of the queue length at time 0 stays the same for all reasonable service disciplines (that do not allow processor sharing). So we may assume that, up to time 0, all arriving customers are served in order of their external arrival: the system serves the "oldest" customer a geometric number of times and then turns to the service of the next customer. (2) We simplify the model by assuming that all inter-arrival times are deterministic and equal to a = λ −1 . (3) We further assume that all service times of all customers but one are equal to b, so every customer but one has a geometric number of services of length b. The "exceptional" customer may be any customer −n ≤ 0, it has a geometric number of services, one of those is random and large and all others equal to b. So the total service time of the "exceptional" customer has the tail distribution equivalent to (4) We assume that the "exceptional" customer arrives at an empty queue, that is, the workload found by this customer is negligible compared with his exceptional service time.
Due to the arguments explained above, we can show that the tail asymptotics of the sojourn time of customer 1 in the original and in the auxiliary system are equivalent. We start by repeating our calculations from the proof of Theorem 2.1, but in two slightly different settings.
Assume all service times but the very first one are equal to b for the exceptional customer arriving at or before time 0. Assume that, if customer 1 arriving at time 0 is not exceptional, then it finds X 0 = N customers in the queue, and otherwise it finds a negligible number of customers compared with N while its first service time is N b. Assume customer 1 leaves the system after K = k services. Denote, as before, by U i the time between its (i − 1)st and ith services and by X i the queue behind customer 1 after its ith service completion. How large should N be for the sojourn time of customer 1 to be bigger than x where x is large?
(A) Assume that the (residual) service time, z, of the very first customer in the queue is not bigger than b (so we may neglect it). When N is large, we get that U 1 ∼ N b. Then we have Hence, X i ∼ X i−1 p + λU i ∼ Nr i and U i ∼ N br i−1 where r = p + λb < 1. Then r ). Thus, we may conclude that where (B) Assume now that both X 0 = N and z are large. Then U 1 ∼ z + N b and X 1 ∼ N p+λ(z + N b) and, further, Let W (t) be the total work in the system at time t. We illustrate W (t) below to see how the cases (A) and (B) occur.
We will see now that if K = k and if there is a big service time of the (−n)th "exceptional" customer, then the case (A) occurs if n > x k /b and the case (B) if n < x k /b (Fig. 3).
Let the big service time take value y 1. Recall from (4.24) that it is enough to consider values of For any k ≥ 1, assume K = k and y ≤ na, then the exceptional service is completed before or at time 0, and the situation (A) occurs. Hence, X 0 ≡ N = n − j for some nonnegative j ≤ n, and y + jb/q ≈ na because approximately j further customers leave the system prior to time 0. Then U ∼ N b(1 − r k )(1 − r ), and U > x is asymptotically equivalent to Since na ≥ y, this further implies that n x k /b. We next assume K = k and n < x k /b. Then, the contraposition of the above implication implies that y > na, and the situation (B) occurs. Therefore, we should take y = z + na, Combining together both cases, we obtain the following result: (4.28) Clearly, the second sum in the parentheses is equivalent to while the first sum in the parentheses is Hence, we have where we recall that If F ∈ L (long tailed), that is, (1.3) holds for some y > 0, then it holds for all y and, moreover, uniformly in |y| ≤ C, for any fixed C. Therefore, if F ∈ L, then there exists a positive function h(x)). In this case we say that the tail distribution F is h-insensitive.
In what follows, we make use of the following characteristic result (see Theorem 2.47 in [9]): We also use another characteristic result which is a straightforward minor extension of Theorem 2.48 from [9]: for any sequence of non-negative random variables V n with corresponding means v n = EV n satisfying V n → ∞ and V n /v n → 1 in probability.
Here is another good property of IRV distributions. Let random variables X and Y have arbitrary joint distribution, with the distribution of X being IRV and P(|Y | > x) = o(P(X > x)). Then If F is an IRV distribution with finite mean, then the distribution with the integrated tail F I (x) = min(1, ∞ x F(y)dy) is also IRV and F(x) = o (F I (x)) and, moreover, We use the following well-known result: if {σ 1, j } is an i.i.d. sequence of random variables with common subexponential distribution F and if the counting random variable K does not depend on the sequence and has a light-tailed distribution, then Here is the principle of a single big jump again: the sum is large when one of the summands is large.
.v.'s with negative mean −m and with common distribution function F such that F I is subexponential. Then Further, if F I is subexponential, then, for any sequence m n → m > 0 and any function and, for any sequence c n → 0, (ii) if the random variable ξ has distribution F ∈ K and c 1 > 0 and c 2 are any constants, then the distribution of the random variable η = c 1 ξ + c 2 also belongs to K; (iii) if the random variable ξ may be represented as ξ = σ − t where σ and t are mutually independent random variables and t is non-negative (or, slightly more generally, bounded from below), and if the distribution of σ belongs to class K, then P(ξ > x) ∼ P(σ > x), so the distribution of ξ belongs to K too.
The following result is a part of Theorem 1 in [7], see also [8] for a more general statement.  ρ)), and therefore the equivalence (3.4) is immediate from (3.3) of Theorem 3.1, while (3.5) easily follows from (3.4).
Thus, it remains to prove (3.6). For this, we introduce some notation.
Define a sequence of events E n , n = 0, 1, . . ., as which is stationary in n (here, by convention, S H 0 = S σ H 0 = 0). Due to the SLLN, there exists a sequence δ ↓ 0 such that as n → ∞. Therefore, for any ε > 0, there exists C = C ε > 0, for which E n is denoted by E n,ε , such that where [x/a] is the integer part of the ratio x/a. Then, for this ε and n ≥ 1, define J n,ε (x) as Then, on the event J n,ε (x) ∩ E n,ε , we have S H n−1 > 0, S H n > ξ H n > x(1 − ρ) + h ε (x), and therefore Hence, letting 0 = [x/a], we have, on the same event, For any integer N ≥ 1, let  by (B.11), and, for any N , by (3.3) of Theorem 3.1, Choosing N such that ∞ n=N +1 P(τ H > n) ≤ εEτ H , we get For x > 0, define events J (x) and J ε (x) as (1 − ρ)).

C Alternative Proof of Corollary 2.1
In this section, we give an alternative proof of Corollary 2.1, which is based on the result from [7], not using PSBJ. Instead of it, our basic tools are Theorem A.1 and the law of large numbers. We also slightly generalise Corollary 2.1. (C.14) Remark C.1 If η = σ 1,0 , then the conditions (I)-(III) are satisfied, and this theorem is just Corollary 2.1.
We prove (C.18) deriving upper and lower bounds. We first consider the case that = 1. Since U j ≤ y j for 1 ≤ j ≤ k − 1, we have that T j ≤ j j =1 y j for 1 ≤ j ≤ k − 1, and This inductively shows that X The corresponding lower bound is obvious. That is, Hence, letting y j → ∞ for j = 1, 2, . . . , k, we obtain (C.18) for = 1.
Similar to (C.19), we have On the other hand, by the law of large numbers, Hence, this term is asymptotically negligible, and therefore we have the asymptotic lower bound for I (2) k ( y, x), which agrees with the upper bound, by letting ε ↓ 0. Thus, we have proved (C.18) for = 2. For = 3, . . . , k, (C.18) is similarly proved (we omit the details). Then the proof of the corollary is completed.