Background

All randomized clinical trials (RCTs) require a protocol describing the rationale, methods, proposed analysis plan and organizational/administrative details from trial inception to reporting of results. Throughout a trial, diverse groups use the trial protocol, including investigators, participants and personnel, funding agencies, research ethics committees/institutional review boards (REC/IRB), journal editors and systematic reviewers. Transparent and clearly written protocols are important to guide trial conduct. They enable thorough assessment of adherence to scientific and ethical standards prior to trial inception [16] and monitoring of changes made throughout a trial that could bias interim or final trial results [5]. Some journals now require the submission of protocols with trial manuscripts, which are then included in the peer review process [710].

Unfortunately, a high proportion of trial protocols do not adequately describe important methodological details, decreasing their utility for trial implementation and critical appraisal of trials. For example, protocols often fail to designate primary outcomes [11] or detail allocation concealment [12], sample size calculations [13] and sponsor and investigator roles in trial conduct [14], all of which have been associated with biased trial results and conclusions. Additionally, comparisons of trial protocols with corresponding journal publications have consistently shown important, unacknowledged discrepancies, including discrepancies in primary outcomes [5] and statistical methods [13, 15]. With recent calls for greater access to trial protocols [16, 17] and trial registration [18], the content of trial protocols is receiving increased attention.

Reporting guidelines have been developed to improve the transparency of other research documents such as reports of research findings for journal publication [1926]. Indeed, the implementation and endorsement of some of these guidelines, including the CONSORT Statement (CON solidated S tandards O f R eporting T rials) [27] for reports of RCT findings, have been empirically shown to improve report quality [2830]. However, development methods of reporting guidelines vary, potentially impacting their utility to various stakeholders [20]. Some groups advocate that reporting guidelines should be developed using rigorous, systematic and transparent methodology and recommendations for reporting guideline development have recently been proposed [31].

Guidelines for clinical trial protocol content are available from varied sources, such as textbooks, funding applications and institutional guidelines. However, to our knowledge, their characteristics and methods of development have not been reviewed. In this paper we report a systematic review with the following objectives: 1) to identify reporting guidelines relevant to RCT protocols; 2) to examine their characteristics and development methods; and 3) to review their content.

Methods

The systematic review protocol was developed with input from trial and systematic review methodologists (See Appendix A in Additional file 1). This report describes the results of an updated review (original completed 2007; JT, AWC, ACT, DM; unpublished).

Eligibility criteria

Documents were eligible if they included a guideline that explicitly itemized content that should be included in protocols for human RCTs and that recommended methodological details beyond a generic heading for ‘Methods’; the term ‘proposal’ was added to the original eligibility criteria as it is sometimes used to refer to the protocol. Documents describing only common or typical protocol content (without recommending content) were excluded; when intent was unclear, the report was included. Tools were excluded if they recommended content specific to a narrow health care research area (for example, disease stage based on a specific classification system) as we intended to focus on guidelines that could be generalized to other research topics; to guide specific protocol aspects such as quality of life assessment; or to assess clinical trial quality. For practical reasons, guidelines were limited to those available in English or French. Both published and unpublished guidelines were eligible.

Information sources

Relevant guidelines were identified via two methods: 1) systematic review of indexed literature and 2) targeted search of major RCT funding agencies.

Systematic review of the literature

Searches were conducted in MEDLINE including in-process and other non-indexed citations (1948 to September Week 4 2010, Ovid interface); EMBASE including EMBASE Classic (1947 to 2010 Week 40, Ovid interface); and the Cochrane Methodology Register (The Cochrane Library 2010, Issue 4, Wiley interface). An information specialist (MS) developed the search strategies. The MEDLINE search, presented in Appendix B in Additional file 1, was amended for the other electronic databases (available upon request). Overlap in journal coverage between MEDLINE and EMBASE was removed using a Batch Citation Matcher technique [32], and remaining duplicates were removed in Reference Manager (Thomson Reuters Corporation, New York, USA). We also searched reference lists of included studies, the PubMed ‘related articles’ link, SCOPUS to identify reports citing the included studies, major clinical trials registries (clinicaltrials.gov, controlled-trials.com), and the EQUATOR Network (E nhancing the QUA lity and T ransparency O f health R esearch) database of reporting guidelines [19] for additional relevant guidelines. A sample of books was also reviewed and were identified based on book title through reference lists and via searches on Amazon.com [33], WorldCat.org [34], and local library portals using the search terms ‘protocols’ or ‘clinical trials’.

Targeted searching of RCT funding agencies

We reviewed a sample of guidelines from major clinical trial funding agencies as we expected some key guidelines would not be readily identifiable through electronic database searching. A short questionnaire was circulated by email to a convenience sample of six key informants who were nominated by the research team, each representing one of six countries previously identified as the top ‘health-related publication producers’ [35]: United States, United Kingdom, Japan, Germany, France and Canada. Informants provided up to two nominations from each of the following sectors within their country: major public (for example, governmental), non-governmental, non-profit (for example, charitable), and for-profit (for example, pharmaceutical industry) clinical trial funding organizations. We reviewed the funding agencies’ websites and contacted the agencies/companies by email to identify relevant guidelines.

Study selection and data extraction

Two reviewers independently screened titles and abstracts followed by potentially relevant full text articles using pre-defined eligibility criteria and pilot-tested forms. Disagreements were resolved by consensus or by the involvement of a third reviewer.

The following data were extracted from the included studies: report characteristics (publication year/version, source); guideline characteristics (format, intended scope, funding, and endorsement); contributors (number, country of residence and expertise of contributors); guideline development process (including use of consensus methods or evidence to inform content, pilot and validity testing, time-frame and dissemination methods); and guideline content (number and content of recommended items).

Two reviewers independently extracted data from a 25% random sample of studies using a standardized pilot-tested form (JT and JK). The remaining characteristics were then extracted by one reviewer, with key data (methods, search for evidence and number of guideline items) extracted in duplicate independently by a second reviewer. Where multiple reports of a guide or associated website were identified (that is, companion reports), all known associated sources were consulted and the most updated or complete version was treated as the primary source.

Guideline content

We itemized guideline content to compare recommendations across guidelines. This analysis was limited to evidence-informed guidelines, those with explicitly described methodology and those with either explicit or probable endorsement of the guideline by a recognized institution or organization. This subset was chosen to select guidelines that were potentially more rigorously developed (that is, those with methodology beyond the consensus of a few authors’ opinions) or more widely acknowledged. To aid in this comparison, we referred to the 2005 version (PR spreadsheet of elements V2.0; David Gemzik, personal communication) of Clinical Data Interchange Standards Consortium’s (CDISC) Protocol Representation Model [36], which aims to comprehensively list potential protocol concepts (to support the interchange of protocol information). Guideline content was mapped, where possible, to one of 264 concepts included in this model. Where no suitable concept existed or where the concept had a different level of granularity than the CDISC concepts, a new category was created. Content mapping was conducted by one reviewer (JT) and verified in full by a second reviewer (JK).

Synthesis of results

Data were summarized using descriptive measures. The following pre-specified sub-groups were compared descriptively: guidelines limited to RCTs versus those with a broader scope; guidelines with versus those without explicit development methods or cited evidence; and guidelines with versus those without described funding sources. A sensitivity analysis compared guidelines explicitly intended for ‘protocols’ versus those for ‘proposals’. Due to the nature of the review, no formal reporting bias assessments were conducted.

Results

Electronic searches yielded a total of 5,147 records (Figure 1), and 76 records were identified from other methods. Ten guidelines were identified from nominated funding agencies or their websites, eight of which were relevant; nine agencies confirmed no relevant guideline and no response was received (and no guideline located) for seven agencies (see Appendix C in Additional file 1 for list of agencies). After screening titles and abstracts, 384 full-text documents were reviewed and 46 were included in this review. Six guidelines were presented in two separate reports [3748], leaving 40 unique guidelines for data extraction [38, 40, 42, 44, 46, 4882] (Table 1).

Figure 1
figure 1

Flow of information through the systematic review.

Table 1 Guidelines included in this review (primary source)

Guideline characteristics and methods of development

We present the general characteristics of the guidelines in Table 2. The majority of guidelines were published as journal articles (n = 22; 55%); most were completed/published from 1991 to present (n = 24; 60%); and most were presented as checklists, tables or bullet lists (n = 27; 68%), some with additional explanatory text. Seven (18%) were specific to RCTs [46, 50, 51, 56, 65, 71, 75]. Fifteen guidelines were disseminated via websites (either specific to the tool or independent sites with links to view the guideline), and five reports described conference or lecture presentations [49, 51, 59, 62, 73]. Six (15%) guidelines reported a funding source, all of which were non-profit. Most guidelines (n = 25; 62%) had no clear institutional endorsement. Guidelines contained a median of 23 recommended items (IQR = 14, 34; range = 7 to 109).

Table 2 General characteristics of included guidelines (N = 40)

In Table 3, we report aggregate results of guideline development methodology; details for each guideline are included in Appendix D in Additional file 1.

Table 3 Overview of guideline development methods (N = 40)

Eleven reports (28%) did not identify the contributors to their development. Of those that did, the majority (n = 24/29, 83%) listed at most 3 contributors (median (IQR) = 2 (1,3)) and most included contributors from a single country (n = 24/29; 83%); exceptions are noted in Appendix D in Additional file 1. Contributors’ areas of expertise were clearly reported for only 8 (20%) guidelines while other reports (for example, [43]) listed names and affiliations of all contributors without explicitly stating their areas of expertise. Contributors’ stated areas of expertise included clinical researchers/clinicians (five guidelines), methodologists/statisticians (two guidelines) and bioethicists, trial managers and information technology personnel (one guideline each).

Only eight reports (20%) described any development methods (Table 3 and Appendix D in Additional file 1): six journal articles [46, 5456, 59, 63] (two of which were specific to RCT protocols [46, 56]), and two international reports [42, 70]. Four of these guidelines describe more detailed and comprehensive methodology: the Council for International Organizations of Medical Sciences (CIOMS) International Ethical Guidelines for Biomedical Research Involving Human Subjects [42], the International Conference on Harmonization Tripartite Guideline for Good Clinical Practice E6 (ICH E6) [70], the PRACTIHC tool (Pragmatic Randomized Controlled Trials in HealthCare) [46] and templates developed for the United States National Institute of Allergy and Infectious Diseases [59] (See Appendix D in Additional file 1).

Stated development methods included informal consensus procedures (including consensus meetings) (n = 4; 10%) [42, 46, 59, 70]; pilot testing (n = 2; 5%) [46, 59]; soliciting input from a broader stakeholder group (for example, public/experts) prior to dissemination (n = 5; 13%) [42, 46, 54, 59, 70] and formal tool validation (n = 1; 3%) [56]. No report stated all of these methods. Two reports included a prospective request for public feedback after dissemination [55, 56] and one stated future plans for formal tool validation [59]. No report indicated the use of a formal consensus process (for example, Delphi consensus, Nominal Group Technique) for guideline development.

Additionally, no report described a systematic search for existing guidelines or empirical evidence to inform guideline content. One reported searching personal files [59] for previous guidelines and another reported a non-systematic search of the Internet and reference lists, and contacting experts to identify previous guidelines and evidence [46]. Three reports cited empirical evidence for some of the items [53, 58, 71] without describing methods for identifying this evidence.

Guideline content

We extracted content from a subset of 23 guidelines. The recommended content varied substantially between the guidelines (Table 4). Over 380 concepts were recommended (median (IQR) = 31 (24,80) concepts per guideline; range = 16 to 150), over half of which were each recommended in only one guideline (including both distinct concepts such as conflicts of interest and sub-concepts of existing headings such as rationale for choosing specific outcomes). We present the most commonly recommended concepts in Table 4.

Table 4 Common concepts requested in the 23 protocol guidelines with explicit methodology or institutional adoption (numbers in parentheses represent number of guidelines)*

Subgroup comparisons

Few differences were noted between pre-specified subgroups by scope, development methods, and funding source. The number of guidelines in each subgroup was small, thus limiting the ability to make definitive conclusions. In Table 5 we present the most notable differences between the subgroups. No differences were found between guidelines intended for ‘protocols’ versus those for ‘proposals’.

Table 5 Major differences between pre-defined subgroups*

Discussion

Our review identified numerous guidelines aiming to inform the content of clinical trial protocols. However, recommended concepts varied substantially across guidelines and the vast majority of guidelines did not describe their methods of development. When described, most included informal methods with limited stakeholder involvement and limited use of evidence to inform their recommendations. Similar findings have been reported elsewhere [20,][83].

Very few concepts were recommended consistently across guidelines, including several whose importance is supported by empirical evidence. For example, only half of the more recent guidelines [59, 70, 71, 74, 75, 80, 81] included an item recommending that primary outcomes be stated, despite preceding research showing biased modifications of primary outcomes throughout trials [3, 5, 11]. Similarly, only three [6, 66, 71] explicitly requested information regarding allocation concealment, the absence of which has been associated with inflated trial effect sizes [8486], although many requested general allocation methods. Conflicts of interest and roles of the sponsor in the trial were explicitly recommended in only one guideline [81], despite being required by the Declaration of Helsinki [87] and despite research showing that trials with financial competing interests report positive results more often than other trials [6, 88, 89]. Only three [71, 75, 81] guidelines explicitly recommended including or citing a systematic review as part of the trial rationale despite the problems associated with non-systematic literature searches [90, 91]. Finally, only 4 [46, 71, 72, 81] of 15 guidelines published after the introduction of trial registration requirements in 2005 [18] specifically requested registration information. No guideline recommended all of these important concepts.

The reasons for the variation and omissions are unclear. Few of the guideline reports in our sample described their development methods, preventing assessment of the validity of the recommendations. If not properly developed, guidelines could potentially ultimately be of limited use and may not improve the reporting of elements that are important to key users of protocols. Of the eight guidelines that did detail methodology, four seem relatively comprehensive [42, 46, 59, 70]. Although these four shared many common elements, considerable variation in recommended content was also present.

For a guideline to be widely acceptable, we believe it should be developed using robust methodology that engages key stakeholders during development and is guided by empirical evidence, where possible. In addition, the methodology should be clearly reported and accessible to enable understanding of the process and assessment of its validity. Recommendations for reporting guideline development have recently been proposed [31] and include a series of steps akin to those recommended for clinical practice guideline development [92]: involvement of multidisciplinary expert panel for a formal consensus process (for example, Delphi consensus) and consensus meeting(s), literature reviews to identify key evidence, pilot testing, active dissemination and impact evaluation. Recent research conducted by the EQUATOR group on the development of health care reporting guidelines [20, 83] suggests that such extensive methods are rarely employed. This is congruent with our current findings.

This review has some limitations. Although comprehensive in searching indexed periodicals, our review was not exhaustive in the search for institutional guidelines or books. However, our main findings would not likely substantively change with the inclusion of guidelines from these sources, as most guidelines available outside of journal articles did not describe development methods. Our results are also based on the methodology stated in included reports; we did not contact authors for additional information. Finally, the process of mapping and comparing concepts across guidelines was challenging due to the varied terminology used and the many sub-concepts of general headings that were recommended. To decrease bias we employed a systematic method and a second reviewer verified the process.

Our systematic review highlights some potential limitations of existing clinical trial protocol content guidelines. Given the evidence of protocol deficiencies [1114] and the importance of trial protocols to diverse stakeholders we believe there is a need for standard guidance that is developed using rigorous methods, broad consultation with key stakeholders and is based on empirical evidence, where possible. Development of reporting guidelines requires substantial resources and time [31], and the conduct of this review is an important inaugural step to justify undertaking such an initiative. Since the initial version of this review, an international collaboration known as the SPIRIT Initiative (S tandard P rotocol I tems: R ecommendations for I nterventional T rials) has convened to produce such guidance by systematically developing recommendations for minimum content of clinical trial protocols [93]. The primary aim of SPIRIT is to improve the content and utility of clinical trial protocols.

Conclusions

This review identified many guidelines for clinical trial protocols; the recommendations provided by these guidelines varied substantially and potentially important concepts were often not recommended. Most guidelines did not describe their methods of development and none of the reports described replicable methods of development including formal consensus of key stakeholders or a thorough search for relevant empirical evidence. Given the importance of trial protocols to diverse stakeholders and evidence of protocol deficiencies, we believe that development of a guideline meeting such standards is needed.

Financial disclosure

No direct funding was received for this study. Some authors were personally salaried by their institutions during the period of writing though no specific salary was set aside or given for the writing of this paper. Dr. Moher is supported, in part, by a University (of Ottawa) Research Chair. No funding bodies had any role in the study design, data collection, analysis, decision to publish or preparation of the manuscript.