Main

In recent years there has been a significant shift in the way clinicians access practice-related resources. Medical oncology, like many other specialties, has embraced web-based technology. Research from the United States, Canada and Australia demonstrates that the Internet is being used increasingly as a major information source to support the treatment decisions of cancer clinicians, and to guide the delivery of care (Wong and Veness, 2005; Ousley et al, 2010). Several organisations have developed oncology websites, typically featuring clinical practice guidelines and/or treatment protocols, both of which synthesise the rapidly changing medical evidence-base. Given the increased awareness and use of web-based resources, there is a growing need to evaluate the quality of the information used in day-to-day oncology practice (Wong and Veness, 2005; Masters, 2008).

The literature abounds with evaluations of the ‘quality’ of web-based health information however, most research to date has focused primarily on content targeting patients (Kim et al, 1999; Eysenbach et al, 2002; Deshpande and Jadad, 2009). Further, there is no consensus amongst researchers about what constitutes a quality evaluation tool (Kim et al, 1999; Eysenbach et al, 2002; Gagliardi and Jadad, 2002; Carden et al, 2007). There are currently more than 50 instruments, comprising hundreds of criteria, being used to evaluate web-based health information. The paucity of research evaluating clinician-targeted websites is somewhat surprising, given their prolific use in clinical practice (Kuperman et al, 2001; Westbrook et al, 2004; Smith et al, 2006; Pearson et al, 2009). Therefore, the aim of this paper is to evaluate clinician-focused oncology websites. In the absence of a ‘gold standard’ instrument to compare website quality, we conducted a targeted assessment of specific oncology website content and features commonly used in clinical practice, namely guidelines and chemotherapy protocols.

The Appraisal of Guidelines for Research and Evaluation (AGREE) instrument (AGREE Next Steps Consortium, 2009) was designed and validated to evaluate the quality of clinical guidelines according to six independent domains. In the past decade, AGREE has been cited in over 150 peer-reviewed publications, and has been used to evaluate an extensive range of clinical guidelines across various areas of healthcare, including cancer (Harpole et al, 2003; Burgers et al, 2004). One previous evaluation, focusing predominantly on paper-based breast cancer guidelines, concluded oncology guidelines were of superior quality compared to guidelines developed for other medical specialties (Burgers et al, 2004). The AGREE has not been used previously to evaluate protocol quality, however, use in protocol development has been proposed as one of the several applications of the instrument (besides guideline assessment; Brouwers et al, 2010a). Indeed, the instrument's domains are equally applicable to protocol development and content. Therefore, in this paper we report on the quality of web-based guidelines and protocols for breast cancer and sarcoma. These cancers were chosen because of their contrasting prevalence rates and the differences in the evidence-base for the treatment of these diseases. We also report on a number of key features of the websites, and conduct a basic navigability assessment. This research provides clinicians with a quick reference guide on the strengths and limitations of oncology resources across several major websites.

Materials and methods

Website selection

To provide a meaningful commentary on the quality of web-based oncology resources, we consulted collaborators involved in cancer care and guideline/protocol development (RW–also a medical oncologist; and two external collaborators) to identify international websites disseminating resources that potentially satisfied the following criteria:

Website features

The website disseminated clinical guidelines and/or protocols. We defined guidelines as statements developed systematically to assist practitioner's decision-making in cancer care (Woolf et al, 1999), and defined protocols as evidence-based statements assisting practitioners in implementing treatment choices (specific information on implementing a particular chemotherapy protocol, e.g., dosing, scheduling, and monitoring). We did not evaluate sites containing only ‘information summaries’, which are evidence-based statements providing a general overview across various oncology topics (e.g., diagnosis, treatment).

Target audience

The website targeted health professionals. Sites could also disseminate patient-focused information; however, we excluded sites focused solely on patient information.

Scope

The website contained detailed information on more than one cancer type and more than one mode of cancer treatment.

Fourteen potential websites were identified by our collaborators and after reviewing them we found that eight websites satisfied the aforementioned criteria (Table 1). Websites were excluded for being patient-focused (n=3); not containing guidelines or protocols (n=1); and containing guidelines/protocols on only one aspect of cancer care (e.g., supportive care or screening only, n=2).

Table 1 Characteristics of the eight oncology websites included in our evaluation

Cancer selection

We evaluated guidelines and protocols for two cancers with different prevalence rates; breast cancer is the most prevalent cancer globally and sarcoma has prevalence rates of less than 1% (Parkin et al, 2005). Notably, these cancers also differ in the extent of the available treatment evidence-base; MEDLINE searches (15 February 2011) using mapped terms returned 124 548 articles for ‘breast cancer’ and 77 488 for ‘sarcoma’. In addition, our search of the US-based clinical trials registry (http://clinicaltrials.gov) revealed 2717 phase I, II and III breast cancer clinical trials, and 680 for sarcoma.

Guideline/protocol selection

We selected one guideline and protocol for each cancer from the eight websites (if available). We aimed to evaluate resources that were as similar as possible; however, we did encounter some variability in the scope of resources across sites. We selected adjuvant breast cancer guidelines and chose a commonly administered adjuvant protocol common to all websites. For sarcoma, we chose guidelines and protocols covering soft tissue sarcoma. In the absence of this option, we selected an alternative guideline/protocol that was most similar to the other websites (this was the case for the National Institute of Clinical Excellence (NICE) sarcoma guideline and OncologySTAT sarcoma protocol).

Appraisers

Four independent appraisers (JL, AD, LM, and SP; all healthcare professionals and/or researchers) pilot tested the AGREE-II instrument using six guidelines and protocols so as to become familiar with the tool and assess its appropriateness for rating protocols. Following pilot testing, three raters (AD, JL, and LM) evaluated all the resources presented in this paper. Rating procedures were conducted in accordance with the AGREE-II manual and were as follows:

  • Appraisers independently rated up to five resources each week for 6 weeks (10 January 2011–18 February 2011).

  • Scores were entered into a database weekly and discrepancies of >2 between scores of at least two raters were identified and re-rated, blind to the scores of other raters.

  • A fourth reviewer (SP) rated items with outstanding disagreements.

Instrument

AGREE was developed collaboratively by international researchers for use in guideline development and evaluation, and AGREE-II was launched in 2010 with improved validity and reliability, as well as a comprehensive users’ manual (AGREE Next Steps Consortium, 2009).

AGREE-II has 23 items grouped into six independent domains: Scope and Purpose, Stakeholder Involvement, Rigour of Development, Clarity of Presentation, Applicability, Editorial Independence, and a general question rating overall quality and whether appraisers recommend the guideline for use. Items are scored on a 7-point Likert scale (7=strongly agree; 1=strongly disagree). The AGREE-II has acceptable internal consistency (Cronbach's α=0.64–0.89), face, construct, and criterion validity (Brouwers et al, 2010b, 2010c).

Reporting and analysis

Website characteristics

We documented website characteristics such as country of origin, site charges, advertising, and other key features. We also conducted a basic navigability assessment, defined as the number of pages (clicks) from the website homepage to reach a given guideline/protocol without using an internal search engine. Two independent raters (JL, AD) tallied the number of clicks to navigate to each of the protocols and guidelines evaluated. Scores were crosschecked; however, there were no discrepancies.

Guideline and protocol evaluation

We report the pattern of scores for guidelines and protocols across eight oncology websites. It was not our intent to hypothesis test or make direct comparisons between guideline and protocol quality. As detailed in the AGREE-II manual, we calculated a percentage score for each domain using the following formula; ((score obtained by raters−minimum score possible)/(maximum score possible−minimum score possible) × 100), such that the maximum achievable score is 100% and the minimum achievable score is 0%. On the basis of our pilot testing, we excluded question 18 from the protocol evaluation, as it was deemed not applicable. This item addresses the facilitators and barriers to applying recommendations; protocols by definition are about implementation after these factors have been considered and/or treatment decisions made. The AGREE-II manual recognises that some questions may not apply to all circumstances.

The AGREE-II does not provide guidance on what constitutes high vs low scores. Instead it advises that scores should be interpreted within the context of the project.

Inter-rater reliability

We assessed inter-rater reliability of the three assessors (JL, AD, and LM) using intra-class correlation coefficients (Streiner and Norman, 2003) according to the resource type (guidelines/protocols) and cancer type (breast/sarcoma). We also recalculated these statistics, based on the final scores, after disagreements were arbitrated by the fourth independent rater (SP).

Results

Website characteristics

We evaluated eight websites developed in North America, Europe, the United Kingdom, and Australia. All sites provided guideline access free of charge, with three websites charging additional fees for membership and/or access to other resources such as journal subscriptions. With the exception of National Comprehensive Cancer Network (NCCN), all websites provided access to protocols, free of charge. Only one website contained commercial advertising (OncologySTAT); however, most publicised relevant journals, clinical trials, and upcoming oncology conferences. The number of clicks required to navigate from the homepage to a given guideline/protocol ranged from two to five, suggesting resources were generally easy to find (Table 1).

Guidelines

Domain scores

The combined mean domain scores for breast guidelines compared with combined sarcoma guidelines varied by a maximum of 10%. Therefore, we report the overall mean domain scores for breast and sarcoma guidelines combined (Table 2).

Table 2 Breast and sarcoma guideline domain scores, and mean combined scores from oncology websites

Scores across the six domains and general rating item were highly variable (range 29–73%). Average scores of greater than 50% were obtained in four of six domains plus the general rating item. Relative to other domains, Rigour of Development and Editorial Independence had the greatest range in scores (ranges of 5–85% and 2–92%, respectively). The Scope and Purpose domain scores were consistently high across all the websites (72.6±11.2%), with the Applicability domain scoring the lowest (29.0±17.3%). Overall, NICE breast and Cancer Care Ontario (CCO) sarcoma guidelines achieved the highest scores across most domains, with the British Columbia Cancer Agency (BCCA) guidelines scoring lowest across most domains. Guidelines scoring well in one domain generally scored well across all the domains.

Additionally, domain and overall rating scores for breast and sarcoma guidelines disseminated by the same website were generally consistent. Only 7 of 35 scores varied by more than 10%, and 4 of these were from NICE, in which the sarcoma guideline scored lower than the breast guideline.

Item scores

We attained mean scores of 5 or more (maximum score 7) on seven items: all questions in the Scope and Purpose domain (Items 1–3), clearly defining target users (Item 6) and all items in the Clarity of Presentation domain (Items 15–17). Scores of less than 3 were obtained for items referring to the process of seeking views and preferences of the target population, and providing descriptions of the strengths and limitations of the evidence informing the guidelines (Items 5, 9). Items relating to guideline implementation such as consideration of facilitators and barriers to application and resource implications of applying the recommendations also achieved scores of less than 3 (Items 18, 20) (Table 4).

Protocols

Domain scores

The difference in mean domain scores for breast protocols combined and sarcoma protocols combined ranged from 1 to 6%. Therefore, we report the overall mean domain scores for breast and sarcoma protocols combined (Table 3).

Table 3 Breast and sarcoma protocol domain scores, and mean combined scores from oncology websites

There was a high degree of variability in the overall scores across the six domains and general rating item (31–71%). Average scores of greater than 50% were obtained for two of six domains, plus the general rating item. Relative to other domains, Rigour of Development and Editorial Independence had the greatest range in scores (10–68% and 0–89%, respectively). Clarity of Presentation was the highest scoring domain across protocols (70.6±17.6%); with the Rigour of Development score being the lowest (30.8±20.0%). Overall, CCO and eviQ protocols achieved the highest scores across most domains, with OncologySTAT protocols scoring the lowest across most domains for both breast and sarcoma. Protocols scoring well in one domain generally scored well across all domains.

We found little variation between domain scores for breast and sarcoma protocols disseminated by the same website. Only 3 of 35 scores yielded differences more than 10% with no difference of greater than 20%.

Item scores

Scores greater than 5 (maximum score 7) were obtained for two items in the Clarity of Presentation domain (providing recommendations that are specific and unambiguous; easily identifiable; Items 15, 17), and advice on putting recommendations into practice (Item 19). Scores of less than 3 were obtained for items addressing the process of seeking the views and preferences of the target population (Item 5), and five of the eight items in the Rigour of Development domain (Items 7–10, 13). Scores of less than 3 were also obtained for items such as consideration of the resource implications of applying the recommendations and recording/addressing the competing interests of development group members (Items 20, 23) (Table 4).

Table 4 Mean item scores for oncology guidelines and protocols assessed with the AGREE-II instrument (maximum seven, minimum one)

Inter-rater reliability

The overall intra-class correlation coefficient was 0.80 before arbitration of discrepant scores by the fourth independent reviewer and 0.85 thereafter, suggesting the overall variance between raters was minimal. This was also the case when considering guidelines/protocols and breast/sarcoma cancers separately (range: 0.75–0.83). As expected, we found a noticeable improvement in the intra-class correlations after arbitration of discrepant scores (range: 0.84–0.85).

Discussion

Our study is the first comprehensive evaluation of the quality of clinical guidelines and treatment protocols disseminated on websites designed for cancer clinicians. Using a psychometrically robust evaluation tool we identified areas of strength and deficiency, and distinguished between the quality of different oncology guidelines and protocols. Overall, the quality of oncology resources was modest.

In general, we found little variation between breast and sarcoma resources with respect to combined domain scores and for resources disseminated by the same website. This finding is not surprising, and suggests the processes underpinning the development of oncology resources do not vary by cancer type. We found the greatest variability in domains addressing the processes used to locate and summarise evidence, and to formulate and update recommendations. Further, items addressing the way in which bias and competing interests are minimised during development were highly variable. Our findings provide clear guidance to resource developers as to the way in which they can improve their processes and/or the way in which they are communicated to end-users. Clearly, collaboration across international boundaries would be prudent so as to minimise duplication of effort and costs of developing these tools (Brouwers et al, 2011). However, organisations must retain the capacity to tailor resources to the needs of the clinicians working in their own healthcare jurisdictions.

From the perspective of oncology practice, it is difficult to ascertain whether the differences demonstrated in resource quality will have meaningful effects on the uptake and use of oncology guidelines and protocols, or whether the use of higher quality resources equates to better quality care and improved outcomes for patients. The premise of oncology guidelines and protocols is to reduce treatment variation, adverse events, and patient mortality (Gandhi et al, 2005; Norton and Baker, 2007; Pavlidis et al, 2007; Manchon Walsh and Borras, 2009; Schwappach and Wernli, 2010). Research suggests the use of guidelines in oncology can reduce practice variation (Smith and Hillner, 2001; Ray-Coquard et al, 2005), yet there are very few examinations of the impact of oncology resources and computerised support systems on patient outcomes (Voeffray et al, 2006). Clearly, future research efforts should address this evidence void.

Systematic reviews in the area of prescriber behaviour change and computerised clinical support highlight repeatedly that the provision of clinical resources does not guarantee uptake (Kawamoto et al, 2005; Mollon et al, 2009; Moxey et al, 2010). Key factors critical to the success of computerised support systems include relevant content, having the ability to save limited clinical time, end-user involvement in development, and the ability of resources to be integrated into the clinical workflow. In the process of our evaluation, we noted that oncology websites typically fare well across these factors. All of the websites under review were geared to assist clinicians maintain an up-to-date and comprehensive knowledge of their field of practice (Grol and Grimshaw, 2003) by providing links to the oncology evidence-base, and updating their guidelines and/or protocols routinely (or at the very least providing dates of last update). Further, content development groups typically include end-users and resources were presented in a clear and accessible manner. Finally, given clinicians report their lack of time as a factor affecting the uptake of web-based tools (Bates et al, 2003; Hains et al, 2009, 2010), features such as clarity of presentation, quick reference summaries alongside guidelines, advice on implementing treatment (protocols), and easy-to-locate resources observed across the eight websites, suggests they are ideal for point-of-care use.

Our study is not without limitations. We acknowledge that our evaluation was not systematic, rather websites were selected by a small group of content experts and we evaluated a relatively small sample of guidelines and protocols. As such, our evaluation may not be representative of all oncology guidelines/protocols disseminated online. Further, our evaluation group is based in one healthcare setting and does not have members representing all of the websites under evaluation. Finally, we note that the AGREE-II was not developed to evaluate protocol quality. However, our initial psychometric evaluation demonstrated that the tool's properties were robust for this type of assessment, which may serve as preliminary evidence that the AGREE-II can be used for protocol assessment research.

Web-based information systems have revolutionised healthcare delivery and have the distinct advantage over paper-based systems because of their capacity to accommodate changes in best-practice evidence in real time. Given the increasing popularity of web-based tools in clinical practice, it is paramount we monitor and evaluate the quality of clinician-targeted resources disseminated online. It is unreasonable to expect individual practitioners to replicate the efforts detailed in this paper. As such we encourage continued evaluations of this kind, which provide a quick reference tool for oncology clinicians to scrutinise the resources they are using in their day-to-day practice. Further our research pinpoints where resource developers should direct their efforts to obtain high-quality guidelines and protocols.