8 Appendix
1.1 8.1 LTE Basic Terms
In this study we generally follow the conventional frequency division duplex (FDD) cellular resource units, in which the time is slotted into frames, and each frame is divided into constant 1 ms intervals, denoted as sub-frames. Each subframe is divided into parts termed physical resource resource blocks, which we shall refer to as simply resource blocks. Each such resource block comprises a bandwidth and time duration, e.g., in LTE each resource block comprises 12 sub-carriers in the frequency domain and 14 OFDM symbols in the time domain.
1.2 8.2 Downlink Link Adaptation (DLLA)
As mentioned earlier, opportunistic scheduling, e.g., proportional fairness, takes into consideration the users channel quality reports for better scheduling decisions. In particular, note that in the algorithm presented above, in order for the scheduler to select the user according to Eq. \(\underset{k}{\arg \max } \frac{R_k(t)}{T_k(t)}\) it needs to know the instantaneous rates of all users. In wireless networks, these channel states of users are attained via reports indicating the users’ supported rates for transmission. Furthermore, each practical system supports only a finite set of rates. Link Adaptation is the mechanism where the users’ transmission code rates and modulation schemes are selected based on the channel conditions.
In this section, we briefly explain the concepts and processes of DLLA that is utilized in simulations and experimental results for scheduling using RL. since in the evaluation part both in the simulations and experimental results we follow a typical LTE DLLA, in the following subsection we will provide a technical description of the DLLA we utilized. Our description follows the common terminology and the accepted acronyms hence it is somewhat cumbersome.
The DLLA process is a crucial part of current wireless communication systems. Such technique increases the data rate that can be reliably transmitted [4] and has been adopted as a core feature in cellular standards such as LTE. The LA role in the MAC layer of the base station (BS) is to suggest the scheduler an appropriate modulation and coding scheme (MCS) to be used in the next transmissions to a certain user equipment (UE) in order to keep the block error rate (BLER) below a target. The proposed MCS is signaled from the UE by means of channel quality indicator (CQI) in the form of reports it sends to the BS, [14]. Afterwards, the BS uses a pre-calculated table for the mapping of CQI to a transport block size index (ITBS), an integer ranging from 1–26, which is used in the decision of the transport block (TB) size to be transmitted to the UE. The TB size is also determined by the number of physical resource blocks (PRBs) which can be allocated to the UE. In LTE the radio resources are allocated in the time/frequency domain. In particular, the time is slotted into intervals of 1 ms corresponding to 14 OFDM symbols. and in the frequency domain, the total bandwidth is divided into sub-channels of 180 kHz, each one with twelve consecutive and equally spaced OFDM sub-carriers. A time/frequency radio resource spanning over 1 ms time slot/14 OFDM symbols and twelve consecutive sub-carriers is called a physical resource block(PRB), or just RB, and corresponds to the smallest radio resource unit that can be assigned to a user for transmission. As the sub-channel size is fixed, the number of RBs varies according to the system bandwidth configuration, and it is the scheduler’s decision to divide the total number of RBs to each scheduled UE in the time slot. The ITBS, together with the number of RBs that are allocated to the UE are mapped to the size of the TB.
The CQI reported by the UE on a per transmission time interval (TTI) basis, delivers information on how good/bad the downlink communication channel is. The UE’s measurement of CQI depends solely on the chipset vendors and is derived from UE’s measurement of the reference signals transmitted by the BS. The reference signals received power (RSRP) that is measured by the UE is than used to calculate the link quality metric (LQM) which quantifies the quality of the downlink and is used to determine the CQI. The LQM that is mostly used in LTE is the exponential effective SNR mapping (EESM) [5]. The process of selecting the most suitable MCS based on the link quality measurements is called inner loop link adaptation (ILLA) [6].
Due to various errors in the CQI measurements of the UE, the delay in the reporting process and deviations from the assumed channel conditions, e.g., multi-path environment, UE speed [11], a compensation process is needed and called outer loop link adaptation (OLLA). The correction of OLLA is based on the hybrid automatic repeat request (HARQ) feedback and is depicted as follows, the mapped ITBS from the UE’s CQI report, defined as, ITBS(CQI), is updated by a margin, \(ITBS_{margin}\), for each received positive/negative acknowledgment (ACK/NACK) from the UE. When an ACK is received, \(ITBS_{margin}\) is decreased by \(\varDelta _{down}\), and when a NACK is received, the margin is increased by, \(\varDelta _{up}\). The ratio \(\frac{\varDelta _{down}}{\varDelta _{up}}\) is controlled by the target BLER that OLLA is designed to converge to, given by
$$\begin{aligned} \frac{\varDelta {down}}{\varDelta _{up}} = \frac{BLER_{T}}{100-BLER_{T}} \end{aligned}$$
Intuitively, if \(BLER_{T}\) is set to \(10\%\), this means that the user should receive at least \(90\%\) successful downlink transmissions. As explained the OLLA process is formulated as such,
$$\begin{aligned} ITBS = ITBS(CQI) - ITBS_{margin} \end{aligned}$$
$$\begin{aligned} ITBS_{margin} = {\left\{ \begin{array}{ll} ITBS_{margin} - \varDelta _{down} &{} \text {if ACK}\\ ITBS_{margin} + \varDelta _{up} &{} \text {if NACK}\\ \end{array}\right. } \end{aligned}$$
(5)