Introduction

Colorectal cancer (CRC) (or bowel cancer) is the second and fourth most common cancer diagnosis and cause of cancer death in Europe and the USA, respectively1,2,3. CRC mortality is 25–30% lower in locations where national bowel cancer screening programmes exist such as in the UK, Austria and Switzerland1. However, only a small proportion of CRCs are diagnosed through population-based screening programmes. For example, in the UK, only 10% of CRCs are diagnosed using population-based screening, with the vast majority diagnosed through symptomatic services (67%) or as an emergency presentation (23%)4.

Approximately 15 million colonoscopies were undertaken in the USA in 2012 and 900,000 were performed in the UK in 2016 (refs5,6). The number of colonoscopies performed seems likely to continue to increase owing to the expansion of population screening programmes, a drive for earlier CRC diagnosis and an increasing surveillance burden from individuals identified as at risk of future colorectal neoplasia (on the basis of a genetic predisposition, for example, Lynch syndrome, or previous diagnosis of colorectal adenomas), which will further stretch currently overburdened endoscopy services7,8. For example, in a UK national audit in 2011, more than 65% of colonoscopies were undertaken for diagnostic purposes, roughly 20% were undertaken for surveillance (ongoing assessment of a known high-risk group) and ~10% were undertaken during screening of average-risk members of the general population9. Overall, CRC was diagnosed in only 4% of procedures, one or more colorectal polyps were detected in 27% and 42% of examinations were normal9. There are several reasons for this low neoplasia yield: eligibility for ‘one size fits all’ population-based screening programmes is based solely on age; the evidence supporting the best surveillance interval for known high-risk groups (such as Lynch syndrome) is poor and is guided solely by colorectal adenoma characteristics10,11,12; and lower gastrointestinal symptoms correlate poorly with the presence of disease13,14. Guidelines such as those of the UK National Institute for Health and Care Excellence (NICE) on the investigation of suspected CRC are based on age, presence of specific symptoms and examination findings only, all of which are only weakly predictive of the presence of CRC15. By comparison, Australian guidelines suggest triage of patients for colonoscopy based on age and symptoms but also consider the duration of symptoms and any faecal occult blood test result16.

The increasing demand in colonoscopy, driven by several factors, including the expansion of population-based screening, an ageing population and an extension of the indications for colonoscopy, coupled with restrictions to endoscopic services imposed by factors such as cost and the lengthy training required to achieve competence in high-quality colonoscopy, all of which have been exacerbated by the effects of coronavirus disease 2019 (COVID-19)17, mean that the need to avoid unnecessary investigations and treatment has never been more important. Thus, health-care services must further refine how patients are selected for colonoscopy beyond using existing (basic) criteria. A personalized medicine approach to disease treatment and prevention has been adopted in many areas of medicine; for example, the treatment of certain cancers based on their molecular tumour profile or the prevention of vascular events (myocardial infarction and stroke) based on lipid and blood pressure levels18,19. However, this approach has yet to be adopted for the prevention and diagnosis of CRC.

Systematic reviews of the several CRC risk prediction models developed to date for use in asymptomatic populations are available20,21. Thus, the main focus of this Perspective will be the latest developments towards a risk-stratified approach to CRC prevention and diagnosis, with particular emphasis on the emerging role of laboratory biomarkers for risk stratification. In general, we use the UK health-care system to exemplify current limitations in the prevention and diagnosis of CRC, but the concepts and developments that we discuss are broadly relevant and applicable to all advanced health-care systems worldwide.

Benefits of risk stratification

The potential benefits of effective risk stratification for CRC prevention and early diagnosis are threefold: to individuals who are considered for investigation (henceforth, termed ‘patients’ for the purposes of this article), to health-care professionals and to service providers (Fig. 1).

Fig. 1: Benefits of risk stratification for CRC prevention and diagnosis.
figure 1

Overlapping benefits of improved risk stratification approaches for colorectal cancer (CRC) to patients, health-care professionals and service providers. For example, improved decision-making about investigation by patients and health-care professionals leads to better utilization of finite resources by institutions. COVID-19, coronavirus disease 2019.

Personal stratification for CRC risk has the potential to enable patients to better understand their individual risk of colorectal neoplasia and therefore to appreciate the need for colonoscopy, thereby improving patient compliance and attendance for colonoscopy. In a systematic review in 2015 of 11 randomized trials of 7 cancer risk tools (including decision aids for patients, self-completed risk assessment tools or a personalized risk prevention message) encompassing data on 7,677 patients and including 4 trials of CRC risk assessment in primary care, it was concluded that risk assessment tools were associated with improved patient knowledge, understanding of the importance of screening and the perception of CRC risk22. Additionally, in one of the studies, the intention to participate in CRC screening and screening uptake was increased in the intervention group compared with the control group (43% versus 5%)23. Advanced colorectal neoplasia is usually defined as a composite of CRC and ‘advanced’ colorectal adenomatous polyps (≥10 mm in size, or highly dysplastic lesions, or those with some villous histological architecture, considered clinically significant for detection based on the risk of malignant progression)24. The identification of those at highest risk of colorectal neoplasia can translate into earlier diagnosis of advanced colorectal neoplasia, thereby improving patient outcomes based on the well-established association between earlier stage of CRC at diagnosis and better long-term outcomes, in addition to the importance of identifying and removing precancerous lesions24. Avoidance of colonoscopy in those at low risk of colorectal neoplasia will reduce exposure to an unpleasant procedure, which can be associated with substantial anxiety, pain or discomfort as well as carrying a risk (albeit small) of haemorrhage, perforation, other adverse events and even death25,26,27. With the advent of COVID-19, endoscopy avoidance also reduces infection risk in patients, many of whom are in high-risk groups for serious COVID-19 outcomes28,29.

For health-care professionals, risk stratification could help improve the identification of patients at greatest risk of harbouring or developing advanced colorectal neoplasia, thereby improving the diagnostic yield. This aspect has the further important benefit of identifying patients who might benefit from chemoprevention or other targeted prevention strategies such as lifestyle modification (for example, weight loss, dietary change or smoking cessation)20. Risk stratification can enable tailoring of the post-polypectomy surveillance interval to the individual patient, which has been identified as an important area for further research12.

Lastly, for service providers, effective risk stratification should translate into improved efficiency and utilization of services, leading to a reduction in ‘wasted’ resources and thus allowing increased capacity for ‘appropriate’ procedures and decreased waiting times and health-care cost savings. Targeted colonoscopy should also help mitigate the substantial environmental footprint caused by endoscopy30.

Colonoscopy clearly has an important role in the diagnosis of many other conditions and in the management of non-cancer-related patient symptoms such as inflammatory bowel disease and investigation of diarrhoea31; however, these areas are beyond the scope of this article, which confines itself to the investigation of neoplasia.

Current risk stratification

A systematic review by Usher-Smith et al.20 identified 52 CRC screening risk models encompassing 87 different variables. Peng et al.21 focused their systematic review on 17 CRC risk models derived only from asymptomatic individuals undergoing colonoscopy. However, less than half of these (21 of 52 models) have been externally validated. Moreover, the discriminatory ability of models to identify individuals with colorectal neoplasia is variable20,21,32,33,34.

Risk prediction models in the context of CRC screening have used a median of five risk factors, with the most commonly used variables being age, sex, family history and BMI21. A few models have used genetic markers, including multiple SNPs identified by previous case–control studies20,21. In risk models in symptomatic patient cohorts, specific symptoms (rectal bleeding or altered bowel habit) or examination findings (abdominal mass) have also been included32. Although there is a known association between metabolic syndrome and increased CRC risk (relative risk ~1.4)35, there has been relatively little investigation of the potential value of clinical biomarkers of excess body weight and metabolic health (such as waist circumference, waist–hip ratio, diagnosis of diabetes and routinely available blood investigations such as levels of glycosylated haemoglobin and C-reactive protein) as predictors of risk. There has also been little investigation of whether or how ethnicity should be included in risk models, likely owing to the dearth of data from large cohorts with sufficient numbers of different ethnicities included to enable interpretation20,21.

Laboratory biomarkers

A range of laboratory biomarkers have been or are currently being evaluated for use in CRC risk prediction models (Table 1). Faecal haemoglobin detection by the faecal immunochemical test (FIT) is the biomarker with most promise. It is already widely used in screening programmes internationally as a primary screening test or to triage individuals (subject to a positivity threshold) using symptomatic investigation algorithms, in which it can have a high specificity and negative predictive value for identifying CRC, depending on the threshold faecal haemoglobin value that is set36,37,38. For example, a 2019 systematic review of FIT use in asymptomatic individuals reported a sensitivity of 80% for CRC detection and a specificity of 91% at a positivity threshold of <10 μg/g (ref.38). In symptomatic patients, a primary care study in the UK reported a sensitivity and specificity of 91% for CRC for a FIT result ≥10 μg/g (ref.39). In this respect, FIT can be considered as the first quantitative laboratory biomarker in use for CRC risk stratification. The evaluation and use of FIT for risk stratification and triage for endoscopic investigation will undoubtedly accelerate as endoscopy utilization is reduced by COVID-19 control measures. The potential quantitative use of the absolute faecal haemoglobin level to predict CRC risk and all-cause mortality has been reported40,41,42. Other faecal biomarkers, such as faecal calprotectin levels, have been previously studied and have not been shown to be predictive of colorectal neoplasia in prospective studies43.

Table 1 Examples of potential laboratory biomarkers of CRC risk

The multi-target faecal DNA test (based on the detection of mutant KRAS, methylated BMP3 and methylation of the NDRG4 promoter in faecal DNA) has excellent sensitivity for CRCs (>90%) but has limited performance in the detection of advanced colorectal adenomas (~40%) and is relatively expensive compared with faecal occult blood testing44. A clinical and economic comparison of the effectiveness of faecal DNA testing compared with FIT at a threshold haemoglobin value that provides similar diagnostic performance measures has yet to be performed45.

Intuitively, given the hereditary component of CRC risk, the incorporation of genetic markers should be associated with increased discriminatory power in risk prediction models. Genome-wide association studies have identified up to 70 independent genetic loci associated with CRC risk46. A Genetic Risk Score (GRS) summarizes the effects of multiple SNPs associated with CRC and each GRS might vary depending on the number of SNPs included, the statistical analysis used and the model used to develop the risk score47. The addition of a GRS to CRC risk prediction models has been shown to improve the discriminatory power in some models addressing population risk48,49,50. For example, Iwasaki et al.49 compared three models (amongst 349 men and 326 women with CRC compared with an equivalent number of age-matched controls without a CRC diagnosis): a non-genetic model (incorporating age, BMI, alcohol intake and smoking), a second model with age and GRS, and a third model incorporating the non-genetic model and GRS. The investigators found that the addition of GRS improved the predictive performance for CRC risk (c-statistic: 0.60, 0.63, 0.66, respectively) in men49. However, another study used data from the UK Biobank to construct a weighted GRS (which included 41 SNPs) and found that the inclusion of GRS to two previously developed risk models did not improve performance (c-statistic: Model 1, from 0.68 to 0.69; Model 2, from 0.67 to 0.67)51. A study of 95 common CRC genetic risk variants52 concluded that the polygenic risk score predicted early-onset CRC, opening up the possibility of risk stratification of young (<50 years) individuals for prevention measures.

The availability of technologies to assess the gut microbiota has led to the exploration of the predictive value of a faecal microbial signature to determine colorectal neoplastic risk. However, a review highlighted that studies to date are all small case–control or discovery cohort studies, with only a few studies including a validation cohort53. Despite these limitations, findings are promising: the diagnostic performance of tests based on both 16S ribosomal RNA and shotgun metagenomic sequencing as well as specific bacterial PCR panels is consistently good, with an area under the receiver-operator curve (AUROC) greater than 0.7 (ref.53). No study has yet to combine a specific microbiome profile or specific microbial signature with other risk factors (such as aspects of diet) in a multivariate risk prediction model, although a few studies have extended analyses to combine data with faecal occult blood testing53. The characterization of ‘dysbiosis’ associated with colorectal neoplasia and its link to dietary patterns is required before the incorporation of microbiome data into multivariate prediction models. The predictive value of the presence and abundance of certain pathobiont species (for example, Fusobacterium nucleatum) also remains to be determined in a clinical setting54,55.

Volatile organic compounds (VOCs), which are gaseous substances released from biological samples such as urine, faeces and breath, are of interest for noninvasive cancer diagnosis, in which a profile of multiple VOCs could be associated with a particular cancer or pre-malignant state. A pilot study of 137 individuals found that a faecal VOC profile could be a potential biomarker for CRC56. Urinary VOCs have also been studied as an alternative noninvasive biomarker for CRC, although the evidence is limited at present57,58. One study has suggested that urinary VOCs could be used in combination with FIT, on the basis that it improves the diagnostic performance for CRC detection (sensitivity 0.97, specificity 0.77, negative predictive value (NPV) 1.0) compared with FIT alone (sensitivity 0.80, specificity 0.93, positive predictive value 0.44, NPV 0.9)56.

The complexity and size of datasets generated by ‘–omic’ analyses will be challenging for the incorporation of these laboratory biomarker outputs into clinically useful risk models. However, the rapid pace of progress in the fields of artificial intelligence and machine learning holds great promise for the future use of complex data alongside clinical factors in risk models59. Such tools could potentially enable multiple complex factors such as ‘–omics’ to be more rapidly assessed and clinically applied.

For population CRC screening

CRC screening programmes vary widely in different health-care systems60. For example, in the USA, Germany and Poland, screening is colonoscopy based but ‘opportunistic’ rather than population centred60. Guidelines by the European Society of Gastrointestinal Endoscopy recommend FIT-based screening61. Screening programmes are consistently focused on age60. However, rather than relying on age alone as the only risk factor for population screening for CRC, a more intelligent screening process could include risk factors such as sex, family history or BMI as well as other clinical biomarkers. For example, men have an increased incidence of CRC3 and sex has consistently been included in symptomatic patient risk prediction models20,21. However, no factors are currently adopted into national screening programmes to modulate, for example, the frequency of screening or the FIT threshold value.

In the systematic review by Usher-Smith et al., the AUROC of studies using data from self-completed questionnaires, routine clinical data and/or genetic biomarkers were compared20. No clear trend was observed between an increasing number of variables in a screening risk model and diagnostic performance20. The incorporation of laboratory blood results, such as haemoglobin levels, plasma glucose and lipid levels, or SNPs, is not consistently associated with an improved AUROC20. Encouragingly, two studies have found that that the predictive ability of a positive threshold FIT in combination with other factors is better than that of a positive threshold FIT alone (Stegeman et al., AUROC increased from 0.69 to 0.76, variables were age, sex and history of previous screening; Cooper et al., AUROC increased from 0.63 to 0.66, variables were age, calcium intake, number of family members with CRC and smoking)62,63. With the increasing use of FIT in national CRC screening programmes, FIT-based risk stratification for population screening is an important area for immediate research, in particular exploring the utility of the absolute faecal haemoglobin concentration obtained from the FIT as a factor in multivariate risk stratification models64.

For symptomatic patients

Separate risk models have been developed for CRC risk prediction in symptomatic patients to aid diagnosis. These models have incorporated a limited number of patient-related factors (such as age, sex, family history or BMI) and usually include these factors in combination with symptoms (such as rectal bleeding or change in bowel habit), abnormal examination findings (such as rectal or abdominal mass) and routinely available blood test results65,66,67,68,69,70. In a systematic review, risk models in symptomatic patients had a good discriminatory ability for the presence of CRC (AUROC 0.83–0.97)32.

Symptoms alone are poorly discriminatory for CRC, primarily because lower gastrointestinal symptoms are so ubiquitous. However, rectal bleeding and weight loss have a stronger association with CRC than other non-specific symptoms such as change in bowel habit or abdominal pain71. The addition of the presence or absence of these symptoms into a risk model containing age, sex and medical history has been associated with an increased diagnostic performance (AUROC increased from 0.79 to 0.85)70.

The use and evaluation of FIT in the symptomatic population is increasing worldwide, with intense interest driven by the reduced endoscopic provision forced by infection control measures due to COVID-19. Current evidence suggests that a low threshold FIT (<12 μg/g) has a high NPV for CRC (99%)43,69 and adenomas (94%)43. However, it has lower sensitivity and specificity for adenomas (69% and 56%, respectively) than for CRC (84% and 93%, respectively), suggesting that FIT is best placed to exclude clinically significant (advanced) colorectal neoplasia in symptomatic patients to avoid unnecessary investigations43.

Risk prediction models, such as the Bristol–Birmingham equation, the COLONPREDICT model or the FAST model, are superior at detecting CRC than the NICE referral guidelines: AUROC for the Bristol–Birmingham equation was 0.83 (0.82–0.84) versus 0.64 for NICE guideline CG27 (ref.68); AUROC for COLONPREDICT was 0.92 (0.91–0.94) versus 0.53 for NICE guideline NG12 (ref.65); and AUROC for FAST was 0.87 versus 0.53 for NICE guideline NG12 (ref.65). Interestingly, FIT level alone was also superior to NICE guideline NG12 (AUROC 0.86 and 0.53, respectively)65.

For CRC surveillance procedures

Surveillance colonoscopy places considerable demands upon endoscopic services. Indeed, the need to rationalize surveillance colonoscopy for CRC risk is recognized internationally and is deemed a key research priority12,72,73,74. However, even under the new UK guidelines, surveillance will still be determined entirely by polyp number and characteristics, which have been shown to have limited predictive value for subsequent CRC risk12,75. Improved risk stratification using patient factors has the potential to enable clinicians to identify a genuine high-risk population for careful surveillance and to avoid procedures in lower-risk patients who might require less frequent surveillance or a non-colonoscopic approach to surveillance76.

High-risk patients could particularly benefit from healthy lifestyle advice and support to modify behaviour known to prevent the occurrence or recurrence of neoplasia77. Research into the effects of modifiable lifestyle risk factor interventions linked to screening and surveillance on colorectal neoplastic risk is needed. Additionally, chemoprevention could be targeted at this group of patients (discussed later).

The role of FIT to risk-stratify for surveillance interval has been evaluated in a small number of studies. Terhaar sive Droste et al. evaluated FIT at a threshold of 50 ng/ml haemoglobin levels in patients attending surveillance colonoscopy because of a previous history of adenomas or CRC and a family history of CRC and found that the sensitivity was 80% for CRC but that the sensitivity for advanced adenomas was poor, at 28%78. Additionally, a second FIT sample prior to colonoscopy only slightly improved the sensitivity for advanced adenoma (33%)78. However, the NPVs for CRC and colorectal adenomas were high (99.9% and 92%, respectively)78. Similarly, the ‘FIT for follow-up’ study by Cross et al. predicted that, if annual FIT replaced three-yearly colonoscopic surveillance in patients with previous multiple colorectal adenomas, a substantial proportion of advanced neoplasia would be missed (28%)76.

To date, no risk model to direct personalized surveillance has been described, making this a priority area for further research.

Risk-based chemoprevention

Chemoprevention (the use of drugs or natural substances to prevent cancer) is a recognized prevention strategy for CRC79. Despite compelling evidence for the anti-CRC activity of aspirin, it is currently not used for primary or secondary prevention of CRC due to the rare but clinically significant risk of bleeding, both gastrointestinal and cerebral80. Furthermore, clinicians are currently unable to balance individual risk and benefit from aspirin due to the lack of a predictive model for aspirin chemoprevention efficacy. Additionally, there is a lack of understanding for how best to utilize chemoprevention in an effective way in combination with endoscopic screening and surveillance81. To date, the US Preventive Services Task Force has recommended the use of aspirin for CRC prevention in individuals with an existing increased risk of cardiovascular events and NICE recommended aspirin in individuals with Lynch syndrome in the UK82,83. The demonstration that aspirin and the omega-3 fatty acid eicosapentaenoic acid have differential preventive activity against conventional colorectal adenomas compared with serrated polyps in a high-risk patient group84, along with data suggesting that aspirin dose should be tailored to body weight85, should stimulate a precision medicine approach to chemoprevention. This approach will require the development of a risk tool that will need to balance potential harm against cancer and the overall mortality risk reduction.

Other CRC chemoprevention agents, including the non-steroidal anti-inflammatory drugs sulindac and celecoxib, have been used in patients with the rare CRC predisposition syndrome familial adenomatous polyposis based on evidence from polyp prevention trials86. However, the effectiveness of these and other chemoprevention agents for the delay or avoidance of colectomy as well as for the reduction in CRC risk itself has not been evaluated to date.

An ideal CRC risk model

Development and validation

An ideal model would constitute only factors that are strongly associated (either positively or negatively) with CRC risk (and are independent of, or at least not highly correlated with, one another) so that the model can discriminate between those with and without disease. In reality, things are not this simple. The value of a predictor depends not only on the strength of its association with the outcome but also on its distribution and/or frequency in the population (which is one reason why models developed in different populations vary). In addition, many clinical CRC risk factors are related; for example, excess body weight and diabetes87.

During the assessment of risk model performance, internal and external validation should be undertaken. Internal validation assesses the model utilizing the data upon which it was derived through methods such as bootstrapping or cross-validation. External validation assesses the model using a new independent dataset from a different cohort of patients88. However, many CRC models lack external validation. Given the existence of different CRC risk prediction models, a direct comparison of the performance of different models in the same validation cohort is also needed89. Lastly, model development should be reported in a clear and transparent way, adhering to the TRIPOD statement and checklist for the reporting of multivariable prediction model studies90.

Implementation and barriers

Risk stratification is likely to complement and inform clinical practice rather than replace clinical decision-making. An ideal model for use in practice would have face validity, be quick and easy to use, fit seamlessly within standard consultations, and generate easily understood results, enabling health-care professionals to categorize patients into different risk groups. In addition, obtaining the data required to estimate a patient’s risk should not involve additional health-care episodes, expensive tests or tests that are poorly tolerated by patients.

Thus far, although some published risk models perform at a level that is considered ‘clinically acceptable’ (AUROC >0.7), these models have yet to be adopted in clinical practice. One probable explanation for the lack of clinical implementation is the absence of evidence that the use of a CRC risk prediction model in clinical practice will enhance clinical decision-making and improve health outcomes for patients. Kappen et al. argued that prospective comparative studies (ideally cluster randomized trials) are needed to quantify the impact on clinical decision-making of risk prediction models and suggested various considerations in the conduct of these so-called impact studies91.

Further reasons for the non-adoption of CRC risk prediction models probably include the lack of knowledge about: acceptability to patients, health professionals in primary and secondary care, and service providers; feasibility, especially in terms of obtaining the information on the risk factors that comprise the models, and use within routine consultations or screening programmes; and cost-effectiveness compared with standard of care. One study that tested the QCancer Risk tool in a primary care setting in Australia identified a range of practical barriers around implementation in daily practice, including difficulties in introducing the tool into the workflow of the consultation92. In addition, some users found that the output of the risk model was ‘confrontational’, confusing and alarming to patients and others (often the more experienced practitioners) distrusted the model output when it conflicted with their clinical judgment92. Similar findings emerged from another study by the same group that explored issues around the implementation of the CRISP CRC risk prediction tool93. This finding suggests that considerable work is still needed to address the barriers to implementation before these models can be widely adopted.

A concern for any risk prediction model is the scenario in which a patient is categorized as low-risk but harbours clinically significant, advanced colorectal neoplasia or, conversely, is categorized as high-risk but does not have neoplasia. Clearly ‘low risk’ does not equate to ‘no risk’ and ‘high risk’ does not equate to cancer. All tests or scores (even the most well accepted and commonly used biochemical markers) can result in a false-positive or false-negative result. For example, in the study by Stegeman et al. of 101 patients with advanced neoplasia, 7% of FIT-negative individuals were ‘risk positive’ according to the model (FIT plus risk questionnaire), whereas 2% were FIT positive but ‘risk negative’63. These issues need to be appreciated by those using the models in practice and conveyed to patients in a way that is understandable and minimizes both the possibility of false reassurance or cancer worry.

A further potential concern is that providing patients with personalized cancer risk information could cause psychological harm. The limited work on this topic (most of which is not focused on CRC) suggests that using self-completed risk assessment tools or personalized risk prevention messages might not result in increased patient anxiety prior to the investigation22,94; however, further research specific to CRC diagnosis and prevention is needed.

Finally, a challenge remains with regards to dealing with conflict between the results of current widely accepted testing (for example, FIT screening results) and results from a risk-prediction model. For example, strategies for managing those with high-risk scores but negative or low FIT results and those with low-risk scores and positive or high FIT results are needed as both scenarios conflict with current routine practice and could have the potential to lead to harms for the patient (such as unnecessary procedures or missed cancer).

Challenges and research priorities

Despite the multiple potential benefits of risk stratification in CRC, as yet, there is insufficient evidence demonstrating the clinical effectiveness and cost-effectiveness of risk stratification in different settings, including screening, surveillance and assessment of symptomatic patients, compared with current practice. It is unlikely that a ‘one size fits all’ model will be developed that is applicable in multiple settings and health-care systems.

We have identified several priorities for research that should hasten the use of a risk-stratified approach to the prevention of CRC (Box 1). The development of acceptable and feasible risk-stratified approaches to CRC prevention, screening and early diagnosis need to be prioritized. Large, prospective cohort data or, ideally, anonymized data from national databases of sufficient quality (since these would be population based and not subject to selection biases) are required to provide sufficient power to evaluate individual factors in multivariate models, to undertake external validation of developed models and to compare the performance of different models. There could be a role for artificial intelligence in analysis as demonstrated by Nartowt et al., who showed the potential of this approach for CRC based on information available on medical databases (such as age, sex, family history of CRC or presence of ischaemic heart disease)95.

The research community needs to embrace the testing effectiveness of ‘real-world’ screening, surveillance and early diagnosis risk stratification algorithms in terms of their influence on clinical decision-making and patient outcomes. The effects on health-care services and patient care as well as on patient experience of care need careful evaluation. However, we also need to better understand how patients and clinicians perceive the utilization of individual risk stratification (with the attendant risk of missed lesions or psychological harms of being labelled high-risk) compared with the traditional approach of considering population risk for health-care provision. It is important that research into improved CRC risk stratification is not limited to endoscopic prevention measures but that it also addresses other prevention interventions, including better, more targeted use of chemoprevention agents and lifestyle modification. Ultimately, risk stratification has the potential to modify the patient journey by the individualization of patient care. Further work remains to identify clinical markers and new biomarkers that would provide the best yield of neoplasia for patients attending screening, surveillance and diagnostic colonoscopies.

Conclusions

Improved risk stratification for the prevention and early diagnosis of CRC is essential to best utilize stretched clinical services and improve outcomes. This approach will enable resources to be targeted optimally and reduce unnecessary investigations in low-risk individuals. The potential of combining knowledge of clinical risk factors for CRC with laboratory biomarkers, such as FIT and VOC profiles, is only just now being realized. Careful reporting and validation of multivariable risk prediction models in individual clinical settings will be essential in order to derive clinically relevant tools for the precision diagnosis and prevention of this common malignancy.