Key Points

  • Endurance athletes show training seasonal fluctuations in TEE, energy intake, and body composition.

  • Dietary recommendations should consider the actual training load, TEE, and body composition goals.

Background

Total energy expenditure (TEE) is composed of the energy costs of the processes essential for life (basal metabolic rate (BMR), 60–80% of TEE), of the energy expended in order to digest, absorb, and convert food (diet-induced thermogenesis, ~10%), and the energy expended during physical activities (activity energy expenditure, ~15–30%) [1, 2]. Elite endurance athletes are characterized by high fluctuations of TEE, mainly due to the variability of the energy expended during sporting activities. Among elite senior endurance athletes, training loads from 500 h/year [3, 4] up to 1000 h/year [57] have been reported, depending on the specific muscular loading characteristic of the sport. During heavy sustained exercise (e.g., during the Tour de France), TEE can be as high as fivefold the BMR over several weeks [8]. On the other hand, during recovery days, pre-competition tapers, or during the off-season, the energy expended in activities is far less. Therefore, TEE is expected to be much lower and may even reach levels comparable to that of sedentary behavior.

An appropriate energy intake supports optimal body function, determines the capacity for intake of macronutrients and micronutrients, and assists in manipulating body composition in athletes [9]. It is a challenge for each endurance athlete to appropriately match energy intake and TEE in order to achieve energy balance and thus, weight stability, both on a micro level (i.e., over 1 day or several days) and through the training and competitive season. Furthermore, endurance athletes in general strive for a low body mass and/or body fat level for various advantages in their sports, specifically during the competition season [10]. This allows runners and cyclists to reach greater economy of movement and better thermoregulatory capacity from a favorable ratio of weight to surface area and less insulation from subcutaneous fat tissue. Elite endurance athletes are therefore characterized by low body mass and body fat content. For example, in elite Kenyan endurance runners, the body fat percentage was 7.1% [11], which is only marginally above the recommended 5% minimum for males [12]. In the same athletes, body mass index (BMI) was 18.3 kg/m2 [11], which is generally classified as being underweight [13]. However, these athletes were in peak physical conditions as the investigations were undertaken and a low body fat percentage and body weight might be an advantage for competition. Achieving a negative energy balance and a concomitant loss of body and fat masses in preparation for competition can be accomplished in phases with high daily TEE solely by the reduction of energy intake, since any further training load increases could cause overtraining [12]. Therefore, the nutritional goals and requirements of endurance athletes are not static over the training year. Since endurance athletes undertake a periodized training program and follow periodized body composition goals, the nutritional support also needs to be periodized [9].

Usually, the annual training schedule of an elite endurance athlete is divided into distinct phases, each with very specific objectives. This is necessary to maximize physiological adaptations for improved performance, usually scheduled to peak around the main competitions of the year [14]. The principle of training periodization was first introduced in the 1960s by the Soviet trainer Leo Matveyev [15] and has not fundamentally changed since then [14]. The basis of this model is to prepare the athlete for one or more major competitions during the year by separating the training into the following three main phases (macrocycles): preparatory, competitive, and transition phases [15]. An example for a “one-peak annual plan” for a runner is shown in Fig. 1. The preparatory phase is characterized by predominantly high-volume training at moderate intensities, which improves endurance capacity and provides a more efficient use of fuel substrates. During the late preparatory phase, training volume is reduced while intensity is gradually increased. The goal of this phase is to reach peak performance and to transfer the training effects into the competitive phase, where exercise intensity is the highest. In the week before an important competition, volume and intensity are typically decreased (taper phase) to allow the body to optimally recover for competition. The days and weeks after a main competition are characterized by low-intensity and low-volume training, with goals to induce regeneration and to prepare the athlete mentally and physically for the next training cycle (transition phase) [14, 16].

Fig. 1
figure 1

Periodization of the training year for a “one-peak annual year” of an elite runner. Adapted from Bompa & Haff [16]

Although the concept of training periodization in elite endurance sports has been established for a long time, the coupling of periodized training with nutrition and body composition has gained scientific awareness only recently [17]. Stellingwerff’s group was one of the first to publish periodized nutrition guidelines for middle-distance athletes [17], they then expanded these recommendations for a multitude of power sports [18]. Nowadays, there are guidelines for carbohydrate, protein, and fat intake during training and competition phases, not exclusively focusing on endurance sports [1921]. Meanwhile, for endurance athletes, sport-specific dietary intake recommendations were developed only for a few endurance disciplines (e.g., swimmers [2225], distance runners [26], marathon/triathlon/road cycling [27]). But it remains unclear whether endurance athletes are actually following these nutrient guidelines across all seasonal training phases.

The validity of either body composition, energy intake, or TEE-determination in athletes strongly depends on the methods used. The measurement of body composition in general is prone to error. It has been shown that acute food or fluid ingestion [28], subject positioning [29], previous physical activity [30], and hydration status [31] have an impact on reliability of body composition measurement. Since endurance athletes often train several times per day, it might be difficult to assure best conditions for body composition assessment. According to a recent methodology review performed by Nana et al., only few of the studies, where body composition of athletes was measured with dual X-ray absorptiometry (DXA), provided details about their subject and device standardization [30]. However, other methods like skinfold measurements require highly experienced investigators [32] and strongly depend on the number of measurement sites and the formula used to calculate the percentage of body fat [33]. Therefore, it is important to report standardization protocols in order to evaluate the quality of data assessment. One main issue in assessing energy intake in athletes is the magnitude of under-reporting, which can amount to 10–45% of TEE [34]. It was shown that the magnitude of under-reporting increases as energy requirements increase [34]. Since endurance athletes are often characterized by high TEE, we must assume that these athletes are very prone to a high percentage of under-reporting. For determination of TEE objective methods such as doubly labelled water (DLW) or heart frequency measurements are available. However, in many studies subjective methods such as activity records and activity questionnaires are used in order to assess the activity level and TEE of subjects. These methods estimate TEE or activity level and their validity strongly depends on the breadth of the activity dimensions analyzed.

There exist some longitudinal studies that have assessed fluctuations in body composition, dietary intake, and/or TEE of endurance athletes across the training seasons [3552], but no systematic reviews have been performed. Therefore, the purpose of this study was to (1) systematically analyze TEE, energy intake, and body composition in highly trained athletes of various endurance disciplines and of both sexes with focusing on objective assessment methods and (2) analyze fluctuations in these parameters across the training season. We hypothesized that endurance athletes show large fluctuations of TEE during different seasonal training phases due to differing exercise loads, and concomitant alterations in energy intake and body composition.

Methods

The review protocol was developed according to the Meta-analysis of Observational Studies in Epidemiology Guidelines for meta-analyses and systematic reviews of observational studies [53].

Search Strategy

A systematic literature search was performed to retrieve articles pertaining to body composition, energy intake, and TEE in endurance athletes across the training season. One researcher (JH) conducted the search for publications on 31 January 2015 in the electronic databases MEDLINE (via PubMed) and SPORTDiscus with Full Text (via EBSCOHost). A hand search of relevant reviews was performed to obtain additional articles missed by the database search. No individual or organization was contacted to receive further publications. To identify the population of endurance athletes, the following keywords connected with the Boolean operator “OR” were searched: endurance athletes, endurance-trained, endurance trained, aerobically trained, runners, swimmers, triathletes, skiers, cyclists, and rowers. To identify the outcome of body composition, TEE, and energy intake, the following keywords connected with the Boolean operator “OR” were searched: body composition, fat mass, fat-mass, fat free mass, fat-free mass, body fat, metabolic rate, energy expenditure, dietary intake, food intake, energy intake, food consumption, and macronutrient*. Terms for the study population and outcomes were combined by the use of the Boolean operator “AND”. Limits included articles published in the English language, human studies, and publishing date limits between 1990 and January 2015. Keywords were searched as free text in the title, abstract, and subject heading. A detailed overview of search strategies in the two databases can be obtained in Additional file 1: Table S1.

Literature Selection

Two researchers independently assessed the eligibility of the records by screening the title, abstract, and keywords for inclusion and exclusion criteria. An agreement between the two researchers was quantified by kappa statistics [54]. The full texts of all abstracts meeting the eligibility criteria were retrieved and subjected to a second assessment for relevance performed by one author (JH).

The inclusion criteria included (1) articles reporting original data in peer-reviewed journals; (2) in vivo, human analyses; (3) adult endurance athletes (highly aerobically trained individuals who were engaged in a competitive endurance sport) with a mean age of 18–40 years; (4) reporting of training seasonal phase of data assessment; and (5) assessment of body composition and/or ad libitum daily energy intake and/or daily TEE. Articles were excluded from the review if (1) the article was only in abstract form or a case report, (2) data could not be split between the sexes (where both male and female subjects were analyzed), (3) body composition was assessed by skinfold measurements, (4) daily TEE was assessed by the use of questionnaires, and (5) descriptive quantitative results were not reported in a text or tabular form. Any difference in assessments between the two researchers was discussed in the first instance or resolved by a third author (KM).

Methodological Quality Assessment

All relevant articles were examined for full methodological quality using a modified version of the Downs and Black [55] checklist for the assessment of the methodological quality of randomized and non-randomized studies of health care interventions. According to Fox et al. [56], 10 of the 27 criteria that logically applied to all of the types of studies included in this review were used. The maximum possible total score was 10. Two researchers assessed the study quality independently, with differences resolved by consensus or by a third author (KM). The agreement between the two researchers was quantified by kappa statistics [54]. Based on the assessment of the methodological study quality, no studies were excluded and no additional analyses were undertaken. The methodological quality of the included studies is shown in Additional file 2: Table S2.

Data Extraction

Body composition, energy intake, and/or TEE data were extracted from all studies included in the review by the first author (JH). Demographic and methodological data were also extracted for the following confounding factors: age, sex, sports discipline, competition level, seasonal phase, and methods for assessing body composition, energy intake, and/or TEE. If the same subjects were analyzed during different time points in the same seasonal phase (e.g., energy intake before three different races, or assessment of energy intake at three time points during the training period), the first time point was chosen for data analysis to facilitate data entry and to avoid selection bias. If studies reported any intervention leading to a non-habitual behavior of athletes’ nutrient intakes (e.g., dietary supplementation), the baseline and/or control group data were used. To enable comparisons between studies, reported units were converted into standard units. These conversions were performed by using the reported mean values of the outcomes. Energy intake and TEE were reported in either absolute (kcal/day) or relative values (energy intake or TEE in relation to body weight [kcal/kg·day]). Body composition was converted into fat mass (%, kg) and fat-free mass (kg). According to the definition by Wang et al. [57], the terms lean body mass and fat-free mass (FFM) were considered synonymous. Duplicate publications from the same data set were identified according to the criteria published in the Cochrane Handbook for Systematic Reviews of Intervention [58]. The most complete record was then used for data extraction.

According to the traditional periodization model, the reported seasonal training phases of data assessment were clustered into three groups that included the preparation phase, the competition phase, and the transition phase [1416]. A detailed overview of the clustering can be obtained in Table 1.

Table 1 Clustering of seasonal training phases for body composition, energy intake, and total energy expenditure

Statistical Analysis

The main outcome measures were body composition (fat mass, FFM), energy intake, and TEE of endurance athletes across the season. Once all of the relevant data were extracted, the weighted mean and standard deviation of the weighted mean were calculated for the main outcome variables. Based on the number of subjects examined within the study, relative to the total number of subjects examined for the specific variable, a percentage weight (w) was allocated to each result within each outcome variable and used for the calculation of the overall weighted mean ( w ) and standard deviation of the weighted mean (SD w ) for each variable [59]. A capital “N” denotes the number of separate studies, while a small “n” denotes the number of included individual subjects.

Statistical analyses were performed using the statistical software SPSS statistics version 22 for Windows (IBM Corp., Chicago, IL, USA). p values < 0.05 were considered statistically significant. Kolmogorov-Smirnov tests were performed to check for normal distributions. All parameters were normally distributed except body mass, fat mass, and FFM. To test for comparisons of subgroups, one-factorial analyses of variance (ANOVAs) with Scheffé post hoc tests (parametric) and Kruskal-Wallis tests (H-test) with Mann-Whitney U post hoc tests (non-parametric) were performed. When multiple non-parametric post hoc tests were applied, Bonferroni-adjusted alpha levels were applied. Since parameters for body composition were not normally distributed, we abstained from multiple statistical comparisons between seasonal training phases and endurance disciplines to reduce the risk of type I errors. For comparisons of energy intake and TEE during different seasonal training phases, paired t-tests were used. The separate analysis of studies, where energy intake and TEE were assessed in parallel, and longitudinal studies that reported energy intake during different training season phases, were performed using the free software for meta-analysis Review Manager 5 version 5.3.5 for Windows (Cochrane Collaboration, Copenhagen, Denmark). The results were then presented as means and 95% confidence intervals (95% CI).

Results

Description of Studies and Assessment Methods

The flow chart for the study selection process is shown in Fig. 2. Data were extracted from 82 studies in endurance athletes, with 53 studies assessing body composition, 48 energy intake, and 14 TEE. The kappa value of 0.47 for the agreement between the two researchers who assessed the eligibility of records was considered to reflect a “fair agreement”, whereas “excellent agreement” (kappa value of 0.96) was obtained for the assessment of the methodological quality of included studies [54].

Fig. 2
figure 2

Flow chart for the present systematic review. NR = not reported. *Sum of studies not equal to total as multiple parameters were assessed in certain studies. N = number of studies

The characteristics of the included studies for body composition, energy intake, and TEE are shown in Table 2. In Additional file 3: Table S3, an overview of excluded studies and the reasons for their exclusion can be found.

Table 2 Characteristics of the studies included in the review of body composition (BC), energy intake (EI), and total energy expenditure (TEE)

The cumulative number of subjects included in the analysis was 1674 (71.4% male). Runners (27.8%), cyclists (18.7%), and swimmers (16.4%) comprised the largest proportion of subjects. All athletes for whom an endurance sports discipline was not described or for whom multiple endurance disciplines were mentioned were grouped into “other endurance athletes” (13.5%). On average, the mean age, VO2max, and training volume of study estimates were 26.3 ± 6.7 years, 61.8 ± 6.0 mL/kg min, and 12.0 ± 6.9 h/week, respectively ( w  ± SD w ). A detailed overview of physical characteristics of included study estimates is shown in Table 3.

Table 3 Physical characteristics of included study estimates

Body composition was assessed by DXA in 32.1% of studies, by bioelectrical impedance analysis (BIA) in 25.6% of studies, and by hydrostatic weighing in 25.6% of studies. In 71.7% of the studies, where body composition was measured, no details of standardization were provided. Ten studies (18.9%) reported some standardization details, whereas only three studies (5.7%) reported satisfactory details about their standardization. For determination of energy intake, dietary records (95.1%) with a mean observation time of 4.7 ± 4.1 days were most often utilized. Dietary recall (3.3%) and food frequency questionnaires (FFQs, 1.6%) played secondary roles in energy intake assessments. Half of the studies (50.0%) used DLW for determination of TEE. Other methods included heart rate monitoring (33.3%) and accelerometers (16.7%). The studies using heart rate monitoring for estimation of TEE used individual derived linear relationships between heart rate and oxygen consumption (HR–VO2) during different tasks to estimate the oxygen cost and energy expenditure during the observation period. Two third of the studies used the 24-h heart rate recordings and the individual HR–VO2 relationship to estimate TEE (gross calculation). Two studies calculated TEE by summation of activity energy expenditure (based on individual HR–VO2 relationship) and resting metabolic rate (RMR; net calculation).

Total Energy Expenditure and Energy Intake

In total, 14 studies where TEE was assessed during various seasonal training phases were identified by the literature search. Since no study assessed TEE during the transition phase, only data during the preparation phase (N = 8) and the competition phase (N = 6) are shown. In addition, due to limited data, no separations between the sexes and endurance disciplines of TEE were performed.

Absolute and relative TEE were significantly higher during the competition phase than during the preparation phase (9869 ± 4129 vs. 4345 ± 1062 kcal/day, and 98.9 ± 46.5 vs. 68.5 ± 11.4 kcal/kg·day, respectively, all p < 0.001). Most of the studies assessing TEE during the competitive phase were conducted during an ultra-endurance competition (N = 5), such as during a 24-h team relay cycling race [60], during a 6-day cycling stage race [61], or during a 4851-km team relay cycling race [62]. The maximum TEE amounted to 13,862 kcal/day and 156.0 kcal/kg·day, respectively, observed in male ultra-endurance runners during a 24-h ultra-marathon [63]. The absolute and relative TEE were significantly higher than the energy intake in the preparation phase (4345 ± 1062 vs. 2915 ± 761 kcal/day, and 68.5 ± 11.4 vs. 42.8 ± 10.5 kcal/kg·day, respectively, all p < 0.001) and competition phase (9869 ± 4129 vs. 3156 ± 967 kcal/day, and 98.9 ± 46.5 vs. 43.5 ± 11.3 kcal/kg·day, respectively, all p < 0.001).

Absolute and relative energy intake was higher in males compared to females in the preparation phase (3111 ± 717 vs. 2291 ± 525 kcal/day, and 44.0 ± 10.6 vs. 39.0 ± 9.1 kcal/kg·day, respectively, all p < 0.001) and competition phase (3405 ± 940 vs. 2337 ± 483 kcal/day, and 44.8 ± 11.9 vs. 39.3 ± 7.9 kcal/kg·day, respectively, all p < 0.001, Figs. 3 and 4).

Fig. 3
figure 3

Energy intake (EI) and total energy expenditure (TEE) in kcal/day of endurance athletes. Data are shown in weighted mean and standard deviation of the weighted mean (X̅w ± SDw). n = number of cumulative subjects

Fig. 4
figure 4

Energy intake (EI) and total energy expenditure (TEE) in kcal/kg·day of endurance athletes. Data are shown in weighted mean and standard deviation of the weighted mean (X̅w ± SDw). n = number of cumulative subjects. No data for TEE of females during competition phase available

In males, the absolute energy intake was higher during the competition phase compared to the preparation phase (p < 0.001), whereas relative energy intake was unchanged (p = 0.553). In females, neither the absolute (p = 0.735) nor relative (p = 0.951) energy intake was different between the two seasonal training phases.

Table 4 provides a detailed overview of the absolute and relative energy intakes differentiated by sex, endurance discipline, and seasonal training phase. Energy intake was significantly higher in male runners, swimmers, and rowers compared to their female counterparts during both the preparation and competition phases (all p < 0.01). In male and female runners, male endurance athletes, and combined male and female rowers and cross-country skiers, the energy intake was higher during the competition phase compared to the preparation phase, whereas for male and female swimmers, energy intake was higher during the preparation phase (all p < 0.01). The energy intake of female runners and rowers during the preparation phase was significantly lower than that of all other endurance athletes (all p < 0.05). Reasons for the lower energy intake in female rowers might be that during preparation phase the athletes often reduce their energy intake in order to reduce concomitantly their body weight to start in the lightweight category. During pre-season, body mass may reduce by as much as 8% among lightweight rowers [64]. Runners, in general, profit from a low body mass since greater economy of movement and better thermoregulatory capacity from a favorable ratio of weight to surface area and less insulation from subcutaneous fat tissue is reached [10].

Table 4 Energy intake in kcal/day and kcal/kg/day of endurance athletes in preparation and competition phase

A separate analysis of energy balance was performed by including only studies where both energy intake and expenditure were assessed in parallel. Male endurance athletes showed a significant energy deficit of 304 kcal/day (95% CI −549, −58, p = 0.02) during the preparation phase and 2177 kcal/day (95% CI −2772, −1582, p < 0.0001) during the competition phase (Fig. 5). In female endurance athletes, a negative energy balance was also observed during the preparation phase (−1145 kcal/day, 95% CI −1404, −887, p < 0.0001) and the competition phase (−1252 kcal/day, 95% CI −1778, −727, p < 0.0001, Fig. 6). The relative energy deficit was 6.6% of TEE during the preparation phase and 18.9% during the competition phase in males, and 29.0% of TEE during the preparation phase and 22.0% during the competition phase in females. When comparing energy intake during the preparation and competition phases by solely including studies where energy intake was assessed in both phases (N = 8), the energy intake was higher during the competition phase, being significant in males (+106 kcal/day, p = 0.03), but not in female endurance athletes (+134 kcal/day, p = 0.20, Fig. 7).

Fig. 5
figure 5

Energy balance (EB) of male endurance athletes during preparation and competition phase

Fig. 6
figure 6

Energy balance (EB) of female endurance athletes during preparation and competition phase

Fig. 7
figure 7

Forest plot for comparison of energy intake during preparation and competition phase in endurance athletes

In more than half (53.7%) of the female study populations, where TEE was assessed, the menstrual status was not reported. 24.4% of the female study populations were eumenorrheic, whereas in 22.0% menstrual irregularities were reported. However, a separate statistical analysis assessing seasonal training phase differences of TEE between eumenorrheic and amenorrheic athletes could not be performed, since the cumulative number of subjects was too low in the single training phases.

Body Composition

For the total sample during the competition phase, both body mass and FFM were significantly higher compared to the preparation and transition phases (p < 0.05, Table 5). For the percentage of fat mass, no differences were detected between the seasonal training phases (p > 0.05). Since the percentage of female data on total data varies between the seasonal training phases, we further split the data by sex. In males, the body mass was lowest during the transition phase (p < 0.05) and absolute and relative fat mass were highest during the competition phase (all p < 0.05). FFM was lowest during the transition phase (p < 0.001, Fig. 8). For females, absolute and relative body fat were higher during the preparation phase compared to those during the transition phase (p < 0.01, Fig. 8). Neither body mass nor FFM differences between seasonal training phases were observed (all p > 0.05). When separately analyzing the few studies where body mass and composition were assessed during both the preparation and competition phases (N = 5), male and female endurance athletes showed a significantly lower percentage of body fat and higher absolute FFM during the competition phase compared to the preparation phase (18.2 ± 5.0% vs. 19.6 ± 5.0%, and 56.6 ± 8.7 kg vs. 54.0 ± 8.7 kg, respectively, all p < 0.0001).

Table 5 Body composition of included study estimates across the season
Fig. 8
figure 8

Fat-free mass and fat mass of endurance athletes during preparation, competition, and transition phase. Data are shown in weighted mean and standard deviation of the weighted mean (X̅w ± SDw). n = number of cumulative subjects

In more than one third (34.5%) of the female study populations, where body composition was assessed, the menstrual status was not reported. 39.7% of the female study populations were eumenorrheic, whereas 16.4% menstrual irregularities were reported. However, a separate analysis between eumenorrheic and amenorrheic athletes could not be performed, since the cumulative number of subjects during the different seasonal training phases was too low.

Discussion

In this systematic review, we examined fluctuations in TEE, energy intake, and/or body composition in endurance athletes across the training season. We found that some, but not all, of the investigated outcomes depended on the time point of data assessment during seasonal training. TEE was highest during the competition phase and higher than energy intake in all seasonal training phases. Alterations in TEE did not lead to adaptations of energy intake in females, whereas in males, a higher absolute energy intake during the competition phase was observed. The finding that male endurance athletes demonstrated the highest fat mass values during the competition phase and the lowest FFM during the transition phase seems to be an anomaly from the pooling of data.

Our systematic search initially yielded many studies where TEE, energy intake, or body composition in endurance athletes were investigated. Only a few (2%) reported the time point of data collection with regard to the training season and could thus be included in this review. This is unfortunate since our analysis clearly illustrates how training volume and related TEE vary importantly with seasonal training phases. Specifically and expectedly, both absolute and relative TEEs were significantly higher during the competition phase compared to the preparation phase. Interestingly, these differences were only partly in agreement with alterations in energy intake and/or body composition of endurance athletes.

During the transition phase, limited data for TEE and energy intake of endurance athletes was available. Only for body composition, it was possible to compare with other seasonal training phases, although the number of study estimates and therefore, explanatory power, was weak. Future research on elite athletes should focus on the effects of a sudden stop or reduction in TEE on body composition (e.g., because of injury). There exist only a few studies (with conflicting results) where this question has been examined. Ormsbee and Arciero investigated the effects of 5 weeks of detraining on body composition and RMR in eight male and female swimmers [65]. RMR decreased, whereas fat mass and body weight increased with detraining. In contrast, LaForgia et al. showed that after 3 weeks of detraining, no differences in RMR and percentage of fat mass occurred in male endurance athletes [38]. Unfortunately, energy intake was not reported in either of these studies. Thus, it remains unclear when, whether, and to what extent the body adapts (through changes in energy intake and/or body composition) for the decrease in TEE caused by detraining.

Our analysis highlights an important apparent negative energy balance in endurance athletes, both in the preparation and competition phases, when separately examining the energy balance in articles where both energy intake and TEE were assessed (N = 11). Negative energy balance was reported during the preparation phase in male [66, 67] and female [67] cross-country skiers, male [11] and female [68] runners, and female lightweight rowers [69] and swimmers [70], and amounted to a mean of 304 kcal/day (4.7% of TEE) for males and 1145 kcal/day (27.8%) for females. During the competition phase, a negative energy balance was reported in male cyclists and triathletes [60], male [63] and female [63, 71] runners, and male cyclists [61, 62], averaging 2177 kcal/day (32.5%) for male and 1252 kcal/day (47.9%) for female endurance athletes. The most obvious explanation for these energy deficits is likely the classical issue of under-reporting energy intake through self-assessment in human studies. A review of nine studies using DLW to validate self-reported energy intake in athletes revealed that under-reporting can amount to 10–45% of TEE [34]. Since under-reporting increases in magnitude as energy requirements increase [34], we must assume that under-reporting in the present study estimates was more important during the competition phase. Even when 45% was added to the energy intake of all athletes included in our review, there still remained a negative energy balance of 118 kcal (2.7% of TEE) in the preparation and 5293 kcal (53.6%) in the competition phase. Another explanation for the negative energy balance might be the low accuracy and precision of methods used to estimate energy intake in athletes in the articles included in our review. For example, mostly dietary records with a mean observation time of 4.7 ± 4.1 days were used. According to Magkos and Yannakoulia, for athletes, a 3–7-day diet-monitoring period would be enough for reasonably accurate and precise estimations of habitual energy and macronutrient consumption [34]. However, other methods like FFQs and dietary recalls were also used for energy intake estimations. These methods are both memory-dependent and show lower accuracy and precision than prospective methods like dietary records [72]. However, even when only articles were considered where energy intake was assessed by the use of dietary records, the error remained high (2.5% of TEE during the preparation phase and 54.9% during the competition phase). Finally, the high negative energy balance during the competition phase may also be explained by the fact that, apart from one study, all included studies investigated the TEE during the days with actual competition and not during habitual training days in the competition phase. Thus, it is likely that the TEE during this phase was over-estimated. During the preparation phase, a negative energy balance leading to increased energy store utilization might be desirable by coaches and athletes to reach a sport-specific body composition, but during the competition phase, body composition should not be modified anymore since it is typically already at its optimum. There was one study in which dietary intake was strictly controlled since the subjects were in confinement. Brouns et al. simulated a Tour de France race in a metabolic chamber and calculated the daily energy balance from the energy expended and energy intake as calculated from daily food and fluid consumption [73]. They found a positive energy balance during active rest days whereas during the exercise days, a significant negative energy balance was observed. The authors concluded that if prolonged intensive cycling increases energy expenditure to levels above a certain threshold (probably around 20 MJ or 4780 kcal), athletes are unable to consume enough conventional food to provide adequate energy to compensate for the increased energy expenditure. The authors of a recent review addressing the criticisms regarding the value of self-reported dietary intake data reasoned that these should not be used as a measure of energy intake [74]. Our analysis supports this statement since, for athletes, relative energy deficits amounted up to 48% of TEE in female athletes and 33% in male athletes during the competition phase. Thus, there is an urgent need for better methods of dietary intake quantification, such as dietary biomarkers and automated image analysis of food and drink consumption [74]. The classical concept of energy balance, defined as dietary energy intake minus TEE, has been criticized, since according to this definition energy balance is the amount of dietary energy added to or lost from the body’s energy stores after the body’s physiological systems have done their work for the day [75]. Thus, energy balance is an output from those systems. In contrast, energy availability, defined as the dietary energy intake minus the energy expended during exercise, is an input to the body’s physiological systems, since energy availability is the amount of dietary energy remaining for all other metabolic processes [75]. Endurance athletes, especially female athletes, show low energy availability (<30 kcal/kg FFM/day) [76] and increased risk for changes of the endocrine system affecting energy and bone metabolism, as well as in the cardiovascular and reproductive systems [77]. In healthy young adults, energy balance = 0 kcal/day when energy availability = 45 kcal/kg FFM/day [75]. Since the results of the present study indicate a high negative energy balance in endurance athletes, we must assume that the athletes also demonstrate low energy availability. However, due to the limited data, it was not possible to account for other clinical markers (e.g., bone mineral density), menstrual status, or prevalence of eating disorders in the athletes. We recommend that energy balance-related studies in endurance athletes should also assess and report clinical markers, such as bone mineral density and menstrual status, in order to assess the clinical consequences of the mismatch of TEE and energy intake.

The aggregate analysis yielded a surprising finding. In male endurance athletes, the absolute and relative fat mass was highest during the competition phase. In contrast, during the transition phase, FFM was lowest, which goes along with our expectations with a decrease in exercise volume and intensity. For the female athletes, we did not find these fluctuations in body composition, except for a higher body fat content during the preparation phase compared to the transition phase. We believe that these findings are due to the paucity of data and to the fact that the number and type of athletes varied between seasonal training phases. Indeed, when separately analyzing the few studies where body mass and composition were assessed during both the preparation and competition phases (N = 5), both male and female endurance athletes showed a significantly lower percentage of body fat and higher FFM during the competition phase. Further studies with longitudinal assessments of body composition are required to support these findings. However, in only 5.7% of the studies, where body composition was assessed, satisfactory details about standardization were provided. According to Nana et al., studies involving DXA scans of body composition should report details of the DXA machine and software, subject presentation and positioning protocols, and analysis protocols [30]. It has been shown that the use of a non-standardized protocol increased the variability for total and fat-free soft tissue mass compared to a standard protocol, which might include a loss in ability to detect an effect of an intervention that might have relevance for sports performance [78]. The use of non-standardized protocols and the concomitant higher variability might explain some of the unexpected findings of body composition changes in athletes of the present study.

In male endurance athletes, absolute energy intake was higher during the competition phase compared to the preparation phase. The relative energy intake was not different, which can be explained by the apparent significant increase of body mass during the competition phase, and is likely an artifact of the aggregation of data from various studies. In female athletes, neither absolute nor relative energy intake was different between seasonal phases. When focusing on longitudinal studies that assessed energy intake during different training seasons in the same cohort, there was a tendency for male athletes to show greater fluctuations in energy intake. In female cross-country skiers, the energy intake was higher during the preparation phase [50], whereas in female runners and swimmers, the energy intake was higher during the competition phase [47]. However, summing up both studies, no significant differences between training season phases were found. In contrast, male endurance athletes showed a significantly higher energy intake during the competition phase, as seen in male runners [44], cross-country skiers [50], swimmers [43], and triathletes [49]. Although some of the included studies showed greater energy intake in male endurance athletes during the preparation phase (cyclists [46, 48], swimmers [43]), the power of these studies was too low to change the results. However, since energy intake varies in male endurance athletes depending on the training season phase, it indeed seems appropriate to adapt dietary recommendations according to the different training season phases, as proposed by Stellingwerff et al. [17, 18].

Strengths and Limitations

This is, to our knowledge, the first systematic review focusing on fluctuations in TEE, energy intake, and body composition in endurance athletes. To increase the robustness of the outcomes of our systematic review, we excluded articles where body composition was estimated by skinfold measurements and equations. The accuracy of skinfold measurements depends on the number of measurement sites and the formula used to calculate the percentage of body fat [33]. Since there are many different techniques [79], it is impossible to compare results accurately between studies. Furthermore, skinfold measurements cannot be used to assess intra-abdominal adipose tissue and are highly variable when assessors with limited training and experience perform the measurements [32]. Of course, since skinfolds are very often used for body composition assessments, the exclusion of these articles reduced the total number of articles measuring body composition, which were included in the present systematic review. The inclusion of articles with skinfold body composition determination would have led to a higher number of study estimates and comparisons of different seasonal training phases would have a higher explanatory power. The same is true for estimations of TEE. We included only articles measuring TEE in a more objective way (such as DLW) and excluded articles where TEE was assessed by questionnaires or activity records. This led to the inclusion of a limited number of high-quality studies.

Limitations of the present study relate to the limited cumulative number of subjects, which provided a low explanatory power, and the classification of the different seasonal training phases. In the literature, several similar-sounding terms have been used to describe time points of data collection in athletes. However, assigning the appropriate classification into one of the three seasonal training phases is essential and has a great impact on the final analysis. Furthermore, if articles reported several time points of data collection within one seasonal training phase, we included only the first time point into the analysis in order to assure standardization and avoid selection bias. The exclusion of other time points might have led to the loss of interesting data.

Conclusions

Our analysis highlights the important seasonal fluctuations in TEE, energy intake, and body composition in male and female endurance athletes across the training season. Therefore, dietary intake recommendations should take into consideration other factors including the actual training load, TEE, and body composition goals of the athlete. The present review supports the statement of the current position stand of the American College of Sports Medicine (ACSM) that energy and nutrient requirements are not static and that periodized dietary recommendations should be developed [9]. Importantly, our analysis again shows the uselessness of self-reported dietary intake, a well-known limitation to energy balance studies, in endurance athletes. The important underreporting suggested by our analysis again raises the question of whether self-reported energy intake data should be used for the determination of energy intake and illustrates the need for more valid and applicable energy intake assessment methods in free-living humans [74]. Since we observed a lack of data during the transition phase, future research should focus on the assessment of TEE, energy intake, and body composition on a reduction in training intensity and volume, such as at the end of the competitive season. In addition, future studies dealing with energy balance and nutrient intake in elite endurance athletes should always mention the time point of data assessments (e.g., seasonal training phase).