Introduction

According to reports from the Global Cancer Observatory 2020, colorectal cancer (CRC) is the third most common cancer worldwide [1]. It is estimated that there will be approximately 1.9 million new CRC cases in 2020, and that number is predicted to increase to 3.2 million CRC new cases in 2040 [1]. Colorectal carcinogenesis is strongly promoted by oxidative stress and chronic inflammation via reactive oxygen species and proinflammatory cytokines [2]. In addition, diets rich in antioxidants and anti-inflammatory factors have shown inverse associations with CRC development [3, 4]. Furthermore, diets may indirectly affect colorectal carcinogenesis risk via CRC risk factors (e.g., obesity) and the gut microbiota [5].

Given the contribution of genetic factors to the development of CRC [6], previous studies investigated the effect of dietary intake on CRC risks according to susceptibility loci [7, 8]. In the Genetics of Colorectal Cancer Consortium and Colon Cancer Family Registry, the interaction between 10 CRC susceptibility SNPs was evaluated with several dietary factors, including red meat, processed meat, fruit, vegetables, and alcohol consumption [7]. Among 10 genetic variants, only rs16892766 was found to interact with vegetable intake [7]. However, one single nucleotide polymorphism (SNP) may be limited in reflecting the overall genetic risk of CRC. With widespread genome-wide association studies (GWASs), subsequent research has examined whether the association between diet consumption and CRC risk differed according to CRC susceptibility status [9]. Genetic variants identified from GWAS vary from low penetrance (common variants) to moderate and high penetrance (rare variants) [10, 11]. Thus, combining multiple SNPs into a single polygenic risk score (PRS) is an alternative approach to reflect the overall genetic predisposition to CRC [12, 13].

In the concept of nutrigenetics, genetic factors may impact the effect of diets on health outcomes by altering the bioactivity of metabolic pathways and mediators [14]. To date, the extent to which the CRC risk of individuals with a high overall genetic risk can be improved by adherence to a healthy dietary habit remains unclear. Previous studies have reported interindividual variability in responses to the same dietary factors [15]. Identical meals were shown to largely contribute to postprandial responses of triglycerides, whereas overall genetic factors did not explain any variances in triglyceride responses [16]. However, a recent study revealed a relationship between genetically predicted polyunsaturated fatty acids (PUFAs) and CRC risk [17]. Thus, we hypothesized that the effect of dietary intake on CRC risks may differ by genetic factors associated with CRC.

Furthermore, recent dietary guidelines have shifted the focus from single food items to the overall diet, which considers the complex interrelationships among different foods and reflects individuals’ actual dietary habits [18]. Therefore, while hypothesizing that pleiotropic pathways, rather than the individual effects of each exposure, may be the underlying cause of the increased or reduced effect of dietary and genetic factors, we conducted this study to explore the associations of dietary intake with CRC risk and explore the joint effect of dietary factors and genetic factors contributing to the incidence of CRC by constructing a PRS using data from the largest-to-date GWAS [10, 11]. Additionally, by hypothesizing that genetic predisposition may alter the association of diets with CRC risk, we identified dietary factors that may attenuate CRC incidence in individuals at high genetic risk.

Materials and methods

Study design and data collection

We carried out a prospective cohort study of participants recruited from UK Biobank. A detailed description of the study design is available elsewhere [19,20,21]. Overall, eligible participants were recruited from 22 assessment centers across England, Wales, and Scotland between 2006 and 2010. All the study participants provided electronically signed consent using a signature-capture device. Information on demographics, lifestyles, and medication history was collected via a touchscreen questionnaire. This touchscreen questionnaire was also used to inquire about habitual diet consumption in the preceding year [22]. Anthropometric factors were measured following standardized procedures.

Of the 502,389 participants recruited, we excluded individuals with no genetic information (N = 15,208). We further excluded those with sex discordance (N = 367), putative sex chromosome aneuploidy (N = 651), and ethnic backgrounds other than White British (N = 78,378). Of 408,093 participants remaining after quality control, 374,004 participants who were free of any cancers at baseline and did not withdraw during the study were eligible for the final analysis (Fig. 1).

Fig. 1
figure 1

Flowchart for the study participants

Dietary information

In this study, habitual food intake was assessed via a touchscreen questionnaire [22]. Participants were asked about the frequency of consumption of oily and nonoily fish, processed meat, beef, lamb/mutton, pork, poultry, and cheese. The frequency was specified as follows: less than once a week, once a week, 2–4 times a week, 5–6 times a week, and once or more daily. Participants were also asked about the intake of cooked and salad/raw vegetables (tablespoons/day), fresh and dried fruit (pieces/day), and coffee and tea (cups/day). For alcohol consumption, participants were asked to choose one of the following intake frequencies: daily or almost daily, three or four times a week, once or twice a week, one to three times a month, special occasions only, and never. Any answers with ‘do not know’ or ‘prefer not to answer’ were considered missing. We grouped single items to obtain the consumption of red meat (times/week), total fish (times/week), total fruit (servings/day), and total vegetables (servings/day) [23]. Daily milk consumption (mL/day) was estimated based on information on the type of milk, and the numbers of bowls of breakfast cereal, cups of coffee, and cups of tea consumed [23]. The cutoffs for categories of food groups and food items were chosen based on the distribution of food frequencies. A summary of the touchscreen questionnaire for food items included in the analysis is available in Additional file 1: Table S1.

To capture the overall dietary habits, we calculated a World Cancer Research Fund (WCRF) dietary score based on the extent to which participants adhered to the dietary recommendations. This score was determined by counting the number of dietary components followed by each participant. The simplified-WCRF/American Institute for Cancer Research (AICR) 2018 score and the American Cancer Society Guidelines on Nutrition and Physical Activity for Cancer Prevention were obtained. Accordingly, the WCRF dietary score was calculated based on the adherence to three dietary recommendations, including consumption of red and processed meat less than 4 times/week, fruit and vegetable intake greater than 5 servings/day, and alcohol consumption less than 3 times/month [24, 25]. Each dietary recommendation was treated as a binary variable with a value of 1 assigned to adherence and a value of 0 assigned to nonadherence. We also created an IV-weighted dietary score by adding the natural logarithm of hazard ratio (HR) divided by the corresponding standard error for the estimate of each dietary factor in association with CRC. This approach may reflect both the strength and variation of the effect of dietary factors. Any dietary factors (red meat, processed meat, poultry, fish, milk, cheese, fruit, vegetables, coffee, tea, and alcohol), which were reported by the WCRF/AICR, regardless of the level of evidence, were included in the calculation [26]. The distribution of participants at different levels of adherence to dietary recommendations was as follows: score = 0 (N = 42,599), score = 1 (N = 127,264), score = 2 (153,729), and score = 3 (N = 47,310). Thus, we divided participants into two groups by the WCRF dietary score (0–1 and 2–3) and three groups by the IV-weighted dietary score (tertiles: -13.22 to < -2.88, -2.88 to < 1.09, and 1.09 to 13.02) to have similar numbers of participants in each group. Accordingly, higher WCRF and lower IV-weighted dietary scores indicated healthier diets, and vice versa.

Genotyping and polygenic risk score

Individuals were genotyped using the custom UK Biobank Axiom Array and the Affymetrix Axiom Array, which capture 805,426 markers, as described elsewhere [20]. Both the UK10K and 1000 Genomes Phase 3 and the Haplotype Reference Consortium reference panel were used to imputed genotyping data, which resulted in a total of 93,095,623 markers [20].

Susceptibility loci for CRC risk were derived from two large meta-analyses of GWAS in European populations [10, 11]. Utilizing whole-genome sequencing data for 1,439 CRC and 720 controls from five studies and GWAS data for 58,131 CRC or advanced adenoma cases and 67,347 controls from 45 studies in GECCO, CORECT, and CCFR, Huyghe discovered 40 novel independent SNPs [10]. Using GWAS data for 31,197 cases and 61,770 controls from 15 studies from CCFR1, CCFR2, COIN, CORSA, Croatia, DACHS, FIN, NSCCG-OncoArray, SCOT, Scotland1, SOCCS/GS, SOCCS/LBC, UK1, and UK Biobank, Law identified 31 additional independent SNPs [11]. Taken together with known genetic variants, a total of 221 SNPs were identified [27]. After excluding duplicate, missing, ambiguous, and high linkage disequilibrium variants, 127 SNPs remained [27]. By excluding UK Biobank data, beta-coefficients and standard errors for the effect estimate of these variants on CRC were recalculated to avoid bias and overlap [27]. Of the 127 variants for CRC susceptibility, 78 SNPs were available in the imputed UK Biobank data. We identified the closest SNPs determined to be in linkage disequilibrium (LD) with 49 unavailable SNPs. Of these, 20 SNPs with r2 greater than 0.8 for LD coefficients were considered good proxies and included in the calculation of the PRS. The proxy SNP selection was performed using the LDproxy tool, with the reference data panel of European ancestry [28, 29]. Thus, a total of 98 SNPs were used to calculate the PRS (Additional file 1:Table S2).

The PRS calculation was considered according to three approaches [30]. First, we used an unweighted PRS \(({PRS}_{unw})\), which corresponds to the sum of the number of effect alleles. Second, we used a standard weighted PRS \(({PRS}_{\beta })\), which added the log odds ratio (β) for each effect allele as weights. Third, we used an inverse variance (IV)-weighted PRS \(({PRS}_{IV})\), in which the weights incorporated both the log odds ratio (β) and standard error (SE) of effect alleles.

$${PRS}_{unw}={\sum }_{i=1}^{98}{SNP}_{i};{PRS}_{\beta }={\sum }_{i=1}^{98}{\beta }_{i}*{SNP}_{i};{PRS}_{IV}={\sum }_{i=1}^{98}\frac{{\beta }_{i}}{{SE}_{i}}*{SNP}_{i}$$

Outcome ascertainment

The primary outcome was CRC incidence, in which the diagnoses were defined by the International Classification of Disease 10 codes. CRC was defined as either colon cancer (C18.0-C18.9) or rectal cancer (C19 and C20). The follow-up time was defined from the date of study participation to the date of CRC diagnosis, death, loss-to-follow-up, or end of follow-up (June 25, 2021), whichever came first.

Statistical analysis

The characteristics and diet consumption are presented as the mean ± standard deviation for continuous variables and counts (percentage) for categorical variables. Univariate analysis was performed for demographic and lifestyle factors using the Cox regression model, and factors associated with CRC risk were identified (Additional file 1: Table S3). Accordingly, sex, first-degree family history of CRC, household income, smoking status, alcohol consumption, body mass index (BMI), and physical activity were adjusted in the multivariable analysis of dietary intake and CRC risk.

To estimate the effect of the PRS, the association between each tertile and decile of the PRS with CRC was estimated using the Cox regression model adjusting for sex and family history of CRC. The interactions between food consumption and the IV-weighted PRS, which upweighted the contribution of variants with more precisely estimated effects, were assessed. We modeled the dietary factor and the PRS independently and tested their interactions by comparing the model with a model additionally adjusted for the PRS*diet interaction term, using the likelihood ratio test. Since UK Biobank participants were aged 39 to 73 years, we calculated the adjusted cumulative risk of developing CRC at the age of 80 years for individuals of each combination of dietary and PRS categories. The cumulative risk was defined as the complement of the cumulative survival adjusted for covariates. We further examined the association between dietary intake and CRC risk stratified by PRS tertiles to explore the benefit of adherence to a healthy diet with CRC risk among individuals of different genetic risk profiles.

To quantify the contribution of dietary intake or genetic risk to CRC incidence, we calculated the attributable fraction, which is the estimated proportional reduction in CRC incidence that would occur if all had been unexposed to the risk factor for interest [31]. We assumed that the prevalence of exposure in our study would reflect the prevalence of exposure in the general European population.

Subgroup analyses were conducted according to sex (men and women) and anatomical cancer subsites (colon and rectal cancer). Missing data were handled in a complete-case analysis approach, where the analysis restricted to participants with complete information on all variables included in the model [32]. Quality controls for genotyping data were performed in PLINK [33], and statistical analyses were implemented using R version 4.1.2 (Foundation for Statistical Computing, Vienna, Austria).

Results

Demographic and lifestyle characteristics of study participants

After a median follow-up of 12.4 years (interquartile range 11.6–13.0), 4,686 participants were diagnosed with CRC (3,131 colon cancer and 1,555 rectal cancer incident cases). Participants had a mean (standard deviation) age of 56.6 (8.0) years, and 199,428 (53.3%) were women. Information on other characteristics, including family history of CRC, household income, smoking status, alcohol consumption, BMI, and physical activity, is also shown in Table 1.

Table 1 General characteristics of study participants

Dietary intakes and associations with colorectal cancer

Table 2 presents the habitual intake in association with CRC risk. Having a low adherence to the WCRF (HR = 1.12, 95% CI = 1.05–1.09) and a high IV-weighted dietary score (HR = 1.27, 95% CI = 1.18–1.37) was significantly associated with increased risks of CRC. Regarding dietary components, those who consumed ≥ 3 times per week had a 16% higher risk of CRC (HR = 1.16, 95% CI = 1.08–1.25) than individuals who consumed red meat < 2 times per week. Positive associations were also observed for intakes of processed meat (≥ 2 times per week vs. < 1 time/week, HR = 1.13, 95% CI = 1.05–1.21). The risk of CRC in individuals consuming alcohol ≥ 3 times/week was observed to be 8% higher than that in those consuming < 1 time/week (HR = 1.08, 95% CI = 1.01–1.17). In contrast, compared with individuals consuming < 200 mL/day of milk, those who consumed ≥ 300 mL/day had a 15% CRC risk reduction (HR = 0.85, 95% CI = 0.79–0.92). Compared with individuals who drank < 3 cups/day of tea, those who reported consumption ≥ 5 cups per day had a 12% lower risk of CRC (HR = 0.88, 95% CI = 0.82–0.94). Dose–response associations of red meat (ptrend < 0.001), milk (ptrend < 0.001), tea (ptrend < 0.001), and alcohol (ptrend = 0.01) consumption with CRC risk were observed.

Table 2 Dietary intakes and colorectal cancer risk in total study population

In the stratification analysis by sex, unfavorable diets of both dietary scores and habitual intakes of red meat, processed meat, and alcohol were associated with an increased risk of CRC in men only. In addition, the inverse associations of the highest categories of milk and tea consumption with CRC risk occurred in both men and women (Additional file 1: Table S4).

In the subgroup analysis by cancer subsites, we found an increased risk of both colon and rectal cancer for those with unhealthy diets and high intakes of red and processed meats. Additionally, inverse associations of milk and tea consumption with colon cancer and fruit intake with rectal cancer were observed. Alcohol consumption was shown to be related to rectal but not colon cancer risks (Additional file 1: Table S5).

Polygenic risk scores and associations with colorectal cancer

The association between the PRS and CRC risk is summarized in Additional file 1: Table S6. Compared to the lowest tertile, individuals in the highest tertile of the PRS had an increased risk of CRC by 98% (HR = 1.98, 95% CI = 1.84–2.13) for unweighted PRS, 109% (HR = 2.09, 95% CI = 1.94–2.24) for standardized weighted PRS, and 112% (HR = 2.12, 95% CI = 1.97–2.29) for IV-weighted PRS. In the subgroups, the highest tertile PRS exerted an approximately twofold higher risk of CRC incidence compared to the lowest tertile PRS.

All deciles showed significantly increased risks of CRC in a dose–response manner (ptrend < 0.001). The adjusted HRs (95% CIs) of the highest decile compared to the lowest decile were 3.23 (2.81–3.71) for unweighted PRS, 3.66 (3.16–4.24) for standardized weighted PRS, and 3.87 (3.33–4.50) for IV-weighted PRS. In the subgroups, the tenth decile PRS exerted an approximately three to fourfold higher risk of CRC incidence compared to the first decile PRS (Additional file 1: Table S6, Fig. 2; Additional file 2: Figures S1-S2).

Fig. 2
figure 2

Hazard ratios and 95% confidence intervals for colorectal cancer risk by deciles of each polygenic risk score. The HRs were estimated using Cox proportional hazard models with adjustment for sex and first-degree family history of colorectal cancer. PRS, polygenic risk score; HR, hazard ratio; CI, confidence interval

Interaction of diets and polygenic risk scores on CRC risk

Both dietary factors and the PRS were independently associated with CRC risk (pinteraction > 0.05) (Table 3). For the WCRF and IV-weighted dietary scores, participants with a high genetic risk had more than double the risk of developing CRC compared to those with a healthy dietary intake and a low genetic risk, regardless of the score category. Notably, the point estimates (HRs) tended to increase across dietary score categories within each PRS category and across PRS categories (Fig. 3). For all food groups and food items, more than double the risk of developing CRC was also observed in participants with unfavorable diets and a high genetic risk compared to those with a healthy dietary intake and a low genetic risk. The estimated joint associations were slightly higher in the men subgroup than in the women subgroup but were equivalent for the colon and rectal cancer subsites (Fig. 3 and Additional file 1: Tables S7-S8).

Table 3 Joint effects of dietary intake and polygenic risk score on colorectal cancer risk in total participants
Fig. 3
figure 3

Overall and subgroup analyses by sex and subsites for joint effects of dietary and polygenic risk scores on colorectal cancer risk IV, inverse variance; PRS, polygenic risk score. Low PRS: 316 to < 454; intermediate PRS: 454 to < 483; high PRS: 483 to ≤ 621

Associations between dietary intake and CRC risk according to PRS tertiles

Of participants with a high genetic risk and an unfavorable diet based on the WCRF dietary score, the cumulative risk of CRC at the age of 80 years was estimated to be 5.08% vs. 2.28% in participants with a low genetic risk and a high adherence to WCRF/AICR dietary recommendations. Of participants with a high genetic risk and an unfavorable diet based on the IV-weighted dietary score, the cumulative risk of CRC at the age of 80 years was estimated to be 5.28% vs. 2.11% in participants with a low genetic risk and a healthy dietary habit (Fig. 4 and Table 4). Subgroup analyses showed a higher cumulative risk in men than in women and for colon cancer and rectal cancer (Fig. 4 and Additional file 1: Tables S9-S10).

Fig. 4
figure 4

Estimates of cumulative risk of developing colorectal cancer at age 80 years according to dietary and polygenic risk score categories CR, cumulative risk; WCRF, World Cancer Research Fund; IV, inverse variance; PRS, polygenic risk score. Low PRS: 316 to < 454; intermediate PRS: 454 to < 483; high PRS: 483 to ≤ 621

Table 4 Estimates of cumulative risk of developing colorectal cancer at age 80 years and associations between dietary intake and colorectal cancer according to polygenic risk score categories in total participants

In further stratification analyses by PRS categories, we found that a low score of the WCRF diet was significantly associated with CRC in individuals with intermediate genetic risk (HR = 1.14, 95% CI = 1.02–1.27), whereas the estimates in low (HR = 1.12, 95% CI = 0.99–1.28) and high (HR = 1.09, 95% CI = 1.00–1.19) genetic risk groups were marginal. In contrast, we found that a high score on the IV-weighted diet was significantly associated with CRC across genetic risk groups, with HRs (95% CIs) of 1.32 (1.12–1.54), 1.34 (1.17–1.53), and 1.21 (1.08–1.35) for individuals with low, intermediate, and high PRS, respectively (Table 4). Similar findings were observed in men, colon cancer, and rectal cancer subgroup analyses for the IV-weighted dietary score but only in the men subgroup for the WCRF dietary score (Additional file 1: Tables S9-S10). Less adherence to the WCRF/AICR dietary recommendation exerted an increased risk of CRC in men with high genetic risk (Additional file 1: Table S9).

Figure 4 presents the cumulative risk of CRC at the age of 80 years for individuals defined jointly by dietary scores and PRS categories. The excess risks were 0.28% (2.56% vs. 2.28%) due to the WCRF diet alone and 1.37% (4.65% vs. 2.28%) due to the PRS. The excess risks were 0.66% (2.77% vs. 2.11%) due to the IV-weighted diet alone and 2.26% (4.37% vs. 2.11%) due to the PRS. Compared to dietary factors, the risk of CRC was more likely attributable to genetic factors in the analysis of the overall study population and subgroup analyses by sex and CRC subsites (Additional file 1: Table S11).

Discussion

In this large prospective study to investigate how dietary intake and genetic risk contribute to the risk of CRC, we found that both an unhealthy diet and increased genetic risk were independently associated with the incidence of CRC without evidence of significant interactions. Regarding dietary components, we found that UK Biobank participants with frequent consumption of red meat, processed meat, or alcohol had an increased risk of CRC, whereas those who commonly drank milk and tea had a decreased risk of CRC. In individuals with high genetic risks, adherence to a healthy dietary habit of 11 foods overall but not the WCRF dietary recommendations alone may be beneficial in CRC risk reductions. Overall, CRC risk was more attributed to increased genetic risk than unhealthy dietary habit.

Evidence from systematic reviews for the association between dietary intake and CRC risk was available in the WCRF Continuous Update Project [34]. Our estimates were similar to those reported from the WCRF Continuous Update Project, with a 12% increased risk of CRC for red meat (relative risk per 100 g/day = 1.12, 95% CI = 1.00–1.25) and an 18% increased risk of CRC for processed meat (relative risk per 100 g/day = 1.18, 95% CI = 1.10–1.28) intake [34]. An umbrella review for updated evidence consistently found positive associations between red meat and processed meat intake and CRC risk [35]. It has been hypothesized that DNA damage is caused by carcinogenic substances, such as heme iron, N-nitroso-compounds, heterocyclic aromatic amines, and polycyclic aromatic hydrocarbons, especially when processing meat or cooking at high temperature [36, 37]. The International Agency for Research on Cancer further emphasized the roles of some heavy metals and other persistent organic contaminants in raw or unprocessed meat, which also indicates the carcinogenicity of red and processed meat consumption [38].

CRC risks can differ between men and women due to endocrine differences, which modulate gene expression after dimerization and translocation to the nucleus [39, 40]. Estrogen exhibits its anti-CRC effects through estrogen receptor (ER) superfamilies, including ER-α and ER-β [39, 40]. While the expression of ER-α is low in both colon cancer and normal colonic cells, there is a significant decrease in ER-β in colonic neoplasms compared to normal colonic mucosa, which results in hyperproliferation, reduced differentiation, and anti-apoptosis [40]. This ER-β reduction was also significantly lower in women than in men [41]. In the present study, the associations of red and/or processed meat with CRC risks appeared to be slightly stronger in men than in women, which might be explained by a lower risk in women than in men due to less meat consumption (mean in men and women: 2.3 and 2.0 times/week for red meat and 2.2 and 1.6 times/week for processed meat, respectively; p < 0.001). In addition, the carcinogenic effect of heme irons in meat-rich diets could be weakened in women because of blood loss during menstruation [42, 43]. Overall, we observed a higher attributable fraction of CRC due to dietary intake in men than in women, which suggested that men may obtain more CRC risk reduction benefits than women by adhering to healthy dietary habits.

The preventive effect of tea consumption was attributed to its active ingredients, such as polyphenols, pigments, polysaccharides, alkaloids, free amino acids, and saponins [44]. Among them, the main active ingredient, epigallocatechin-3-gallate, is mainly absorbed in the intestine and metabolized by the gut microbiome and exerts antioxidation, growth inhibition, and apoptosis induction effects [45]. Evidence in human colon cancer cells supports the bioactivity of epigenetic modifications against colon cancer [45]. Furthermore, the results from a recent meta-analysis of 20 prospective cohort studies showed a similar direction of the association between tea intake and CRC (pooled relative risk 0.97, 95% CI = 0.94–1.01) [46]. However, our study is still limited in assessing several factors that may confer the estimates, such as tea temperature and concentration, age at initiation of drinking, and drinking duration.

In general, the impact of dietary factors on CRC incidence observed in our study is in line with previous reports by Bradbury et al. [23]. However, the present study has extended the investigations of the previous study in several ways. First, we derived both the WCRF and IV-weighted dietary scores to obtain robust findings. A previous study created an a priori dietary score from seven dietary factors and examined its interaction with the genetic risk of upper gastrointestinal cancer [47]. However, dietary components were selected based on recommendations on cardiometabolic health [47]. Another study developed a healthy lifestyle score based on the WCRF/AICR and American Cancer Society guidelines on cancer prevention [24]. Adopting a similar approach of limiting to dietary factors (red and processed meat, fruit and vegetables, and alcohol) only, we did not detect any interactions with the PRS on CRC risk in the present study. We further extended the dietary score calculation by weighting all 11 dietary factors (red meat, processed meat, poultry, fish, milk, cheese, fruit, vegetables, coffee, tea, and alcohol), which were reported by the WCRF/AICR for their effects on CRC risk; however, we still confirmed the independence between dietary and genetic factors on CRC incidence.

Second, we derived variants of CRC susceptibility from largely up-to-date meta-analyses rather than a single GWAS. Previous studies systematically reviewed and used UK Biobank data to externally validate the predictive performance of the PRS in which CRC susceptibility loci were identified from 14 GWASs [48, 49]. The number of SNPs was 120 in Huyghen’s study and from 6 to 63 in other studies, and the AUC of the genetics-plus-age-plus-family history model ranged between 0.65 and 0.68 in men and between 0.61 and 0.65 in women [48, 49]. However, the highest AUC was obtained from the model, including SNPs identified in the Huyghen’ study, which overlapped with UK Biobank data [48]. Our present study determined an overall AUC of 0.60–0.61 according to different weighted approaches, and the estimates from independent datasets that were not taken from UK Biobank can minimize any substantial inflations due to overlap between the base and target data and provide an unbiased evaluation for the developed model. By upweighting the effect of risk alleles, the PRS estimated from the IV-weighted approach showed a greater dose–response association with the CRC risk than that from the unweighted and standard-weighted approach.

Third, by approaching the gene-environment interaction framework, we included the PRS to better understand the role of dietary intake in CRC risk. The precise pathways for the joint effect of dietary intake and the PRS in CRC risk remain unclear. This might be proposed by the natural pleiotropic effect of each factor in overlapping with pathways of CRC development. To explore possible biological mechanisms, we performed a gene-set enrichment analysis of 98 SNPs included in the PRS to identify functional pathways of these variants using a web-based FUMA tool [50]. Accordingly, CRC susceptibility loci included in the PRS were mainly involved in the metabolism of PUFAs (Additional file 2: Figure S3). This result supported findings from a previous study on the link between genetically predicted PUFAs and CRC risk [17] and thus suggested a possible interaction between dietary intake and the PRS for CRC via PUFA metabolism pathways. A large-scale study from the Genetics and Epidemiology of Colorectal Cancer Consortium and Colon Cancer Family Registry examined the effect modification of candidate CRC susceptibility on selected dietary factors, such as red meat, processed meat, fruit, vegetables, and fiber [51]. However, none of the significant interactions was detected after accounting for multiple comparisons [51]. Our findings of a greater number of SNPs in the PRS also suggested no evidence of gene-diet interactions in the risk of CRC.

There are also some limitations that need to be addressed. First, the FFQ was used to quantify information on an individual’s intake during the preceding year. There could be difficulty in recalling some food items accurately. Second, our findings require validation in populations of ethnic backgrounds other than White British.

In summary, the current study provided evidence that unhealthy dietary intake and genetic risk were independently associated with the risk of CRC. Adherence to a healthy dietary habit regarding the above 11 foods may attenuate CRC incidence in individuals at high genetic risk.