Background

A discrete choice experiment (DCE) is a method used to explore the relative importance of different attributes within a decision-making process. Based on random utility theory [1], DCEs operate on the tenet that there are multiple attributes influencing decision-making and that all decisions involve trade-offs between the range of elements that constitute the influential attributes. DCEs offer a means through which the nuances of decision making can be understood, by providing insights into the often implicit trade-offs made, which are not easily accessed through other research methods. DCEs have been increasingly adopted to examine attributes influencing decision-making in areas of healthcare, including stroke rehabilitation [2, 3].

Traditionally, the broad study design of a DCE is informed by literature review, expert opinion, theoretical arguments and/or qualitative work [4]. However, there is no standard development process for a DCE for identifying a comprehensive list of attributes that may influence the decision of interest and the subsequent optimal selection of attributes for inclusion in the final design, necessary due to the frequently large numbers and variable nature of type of attributes that may be identified. Furthermore, there is a lack of guidance on the optimal presentation of choice sets or decision alternatives in DCEs. The processes involved in the design of DCEs are often poorly described, with little or no detail reported regarding the procedures for selection of attributes/levels in the final design. In particular, there is a dearth of guidance on procedures that can be undertaken to select the most salient attributes and their associated levels from the full range of attributes that may influence the decision of interest and, when reported, this information has been characterised as “excessively brief” [4].

The International Society for Pharmacoeconomics and Outcomes Research (ISPOR) has published two reports on recommended best research practice in the development, administration, analysis and reporting of DCEs [5, 6]. However, they do not offer a clear approach to certain important aspects of DCE development; specifically how to select the most salient attributes/levels for inclusion in choice scenarios from the full range of attributes that may have an influence on the decision of interest; and administration in terms of the optimal method of presenting choice scenarios to respondents to maximise process/face validity of the choice task.

With some notable exceptions [7, 8], there is a dearth of examples of a clear, transparent DCE study design processes in the published literature. Although researchers have advocated the use of qualitative methods such as one-on-one interviews, focus groups, analysis of policy documents, and expert opinion to inform the design of DCEs [4, 9], recent evidence suggests that the uptake and reporting of qualitative methods to inform DCE design are lacking. Vass et al. [10] found that of 254 healthcare DCEs examined, 44% did not report using any qualitative methods and only 11% reported relying on qualitative methods extensively. This underuse, or lack of reporting of, qualitative methods to inform DCE development inhibits a clear understanding of whether a DCE design process has been rigorous and systematic and thus inhibits the ability to effectively evaluate the quality of the research. Quantitative approaches, including ranking exercises and nominal group techniques to support the process of selecting factors for inclusion in DCEs have been outlined [7, 8] but to our knowledge, there are no examples where both qualitative and quantitative methods have been employed to inform DCE design.

The validity of the results obtained from a DCE is contingent on the quality of the study design process [11]. When DCEs are conducted in circumstances where there is little or no empirical evidence to guide study design, documenting the design process is especially important. Current reporting on DCE design often does not provide sufficient information to allow for an informed judgement to be made regarding the quality of a study [4, 9]. Increased transparency at this stage of the process would offer valued guidance to researchers, allowing them to draw on approaches and methods that best suit the objectives at hand. It would also support more effective critical appraisal of the findings of the reported studies utilising DCEs. Robust methods should underpin DCE design, yet it is often not possible to judge these processes, or whether there was any structured development process involved, from the limited information presented in many published papers.

There have been calls for a minimum level of reporting on the DCE development process, including the processes used to collate an initial list of attributes, the analyses conducted during this design stage (including sample details and information on type of analysis conducted), processes undertaken in reducing attributes to a manageable number, and a brief description of the results of these processes [9, 10]. As each decision made during the DCE development process can have implications for both the final design and the validity of results obtained, the adherence of researchers to these basic reporting conventions would improve the transparency of DCE research and enable readers to better judge the quality of such studies. By employing a DCE to explore decision-making about thrombolysis in acute stroke care as an exemplar, we describe a systematic, transparent and iterative multi-method process to design an online DCE.

Decision context: Thrombolytic treatment in acute stroke care

Intravenous thrombolysis using recombinant tissue plasminogen activator is an effective medical treatment for acute ischaemic stroke. Though there is clear evidence of its efficacy and benefit in certain patient groups and it is promoted and supported by national guidelines [12,13,14,15], it remains underused, as approximately one in five patients eligible for thrombolysis do not receive it [16]. Decision making about whether or not to treat a stroke patient with thrombolysis is complex due to the time limited window for treatment, potential difficulties in obtaining consent, and the many clinical attributes that might influence the balance between risk and benefit for individual patients [17]. Early treatment of acute ischaemic stroke with thrombolysis is associated with more favourable outcomes, yet there is a small but significant risk of adverse outcomes as result of treatment [18]. Symptomatic intracranial haemorrhage occurs as a complication in approximately 3–4% of patients, which increases the probability of long-term disability and death [16]. A DCE offers on optimal means of better understanding the complexities of clinician decision-making about intravenous thrombolysis in a way that reflects decision making in practice.

The aim of this paper was to build on existing DCE guidelines by describing a systematic and rigorous process for DCE construction and choice set presentation, using the exemplar case of clinical decision-making for thrombolytic treatment during acute stroke care.

Methods

Overview of DCE 5-stage design process

We designed hypothetical patient vignettes to mimic the decision of interest as closely as possible, using a 5-stage iterative process described by Fig. 1. This was process was adopted because qualitative research was required to inform the design, but also in order to consider the nature and relative importance of factors influencing choice (clinical and non-clinical), and it was important to check the material being constructed and validity of the choices with clinicians. The decision of interest was in the form of a binary response (decision to offer thrombolysis or not) as this reflects routine clinical practice whereby clinicians are faced with one patient at a time and must decide whether or not to offer thrombolysis. Clinicians were likely to be familiar and comfortable with vignettes as they are regularly used in training and continuing professional development.

Fig. 1
figure 1

Summary of key stages of development process

Ethical approval was obtained from Newcastle University Research Ethics Committee (reference: 00720/2013). Consent to participate was obtained for all those involved in the study.

STAGE 1. Exploratory work

We considered all attributes that might be influential on thrombolysis decision-making, as omitting important attributes in the DCE may bias findings [2]. We drew on previously gathered qualitative data collected for a previous research programme examining decision-making in acute stroke [19] and on work to inform the development of a computerised decision support tool for thrombolysis in acute stroke care; COMPASS (COMPuterised decision Aid for Stroke thrombolysiS) [20]. This included qualitative interview data on the views and experiences of stroke clinicians about thrombolysis decision-making, ethnographic data collected through non-participant observation of thrombolysis decision-making discussions, and data from the usability and feasibility testing of COMPASS. Relevant literature was also reviewed to identify further attributes that have been documented to affect decision-making about thrombolysis and attributes influencing medical decision-making more broadly [21,22,23,24,25,26].

In addition, nine telephone-based semi-structured interviews were conducted with stroke clinicians and experts in stroke research. Purposive sampling was used to capture the views of experienced clinicians identified through the literature review (who had not contributed to the development of the COMPASS tool). We also targeted clinicians from hospitals in both the upper and lower quartiles of thrombolysis rates, based on figures from national audits [27]. These interviews aimed to identify any additional patient-related attributes or levels that represented the borderline and/or difficult cases; in particular the nature of trade-offs made between influential attributes and levels when making thrombolysis decisions. Telephone interviews were conducted due to the geographic spread of participants. Interviews were audio-recorded and transcribed verbatim.

Anonymised transcripts were imported into QSR International’s NVivo 9 to facilitate the coding and analysis of data using a framework approach, where the initial framework was guided by the interview guide (deductive approach) and additional codes were generated where required (inductive approach) [28]. Codes were themed where relevant and coding was discussed among the research team to reach consensus. Analysis of interview data and review of secondary data sources from the previous research programme on stroke [19], alongside relevant literature, facilitated the generation of an initial long list of potentially influential patient-related attributes on clinical decision-making regarding thrombolysis.

STAGE 2. Expert panel discussions – Procedure for screening attributes and levels

Exploratory work commonly identifies too many attributes/levels to include in DCEs and, due to sample size constraints and the potential for respondent cognitive burden, it is often not practical or feasible to include all possible combinations that may influence decision-making. By including all potentially influential attributes and levels of interest, the statistical power of a DCE to detect effects may be reduced, due to the large number of possible combinations of levels and the inability of a small sample size to adequately assess all these various combinations. Methods are needed to select the most important attributes and levels. Although there are publications that provide generic advice on DCE design [4, 9], there is a dearth of explicit guidance on this process. A well-designed DCE has been described as “one that has sufficiently rich set of attributes and choice contexts, together with enough variation in the factor levels necessary to produce meaningful behavioural responses.” [29] Abiiro et al. [7] advise employing quantitative methods, such as ranking exercises, to support the process of selecting attributes for inclusion to a manageable number. Prior to the ranking exercise at the next phase, the expert multidisciplinary panel in the current study (consisting of three stroke clinical researchers, one trainee stroke physician, two patient representatives (from the UK Stroke Association), two chartered psychologists, two health economists, and an expert in shared decision-making) screened the long list of attributes (generated during Stage 1) in terms of whether they would be feasible or meaningful to include in a DCE, which would then be further scrutinised using a structured prioritisation/ranking exercise. Following this ranking exercise and subsequent discussions, a set of inclusion and exclusion criteria were agreed with reference to the study objectives.

STAGE 3 - phase 1. Design of online structured prioritisation exercise (SPE)

In order to identify the relative importance of various patient-related attributes for clinical decision-making and areas of uncertainty on specific attributes, an online structured prioritisation exercise (SPE) was developed. This was a ranking exercise performed by clinicians on a single occasion to prioritise attributes related to clinical decision-making for inclusion in the cases, to ensure that the DCE was relevant without being excessively long. It offered a means to elucidate a broader range of views to inform the selection of the attributes and in particular, questions were phrased to identify where uncertainty and ‘cut-offs’ existed on individual attributes regarding the suitability of the patient for thrombolysis. This step was important for the current study as it was widely recognised that individual clinicians had different thresholds or cut-offs for decision-making. The findings would inform expert panel discussions in the next phase of the design process. Free text boxes were also provided after each question. The SPE was hosted on Qualtrics (www.qualtrics.com) and 30 stroke clinicians in the north east of England were invited to participate. The full SPE is available in Additional file 1.

STAGE 3 - phase 2. Using the SPE to inform selection of attributes/levels

The same expert panel assembled in Stage 2 then independently ranked each attribute in order of priority for inclusion in the DCE, indicated how they would operationalise them, and provided suggestions on possible levels for each patient-related attribute. Results were then aggregated and fed back to the panel to inform discussion on selection of the final list of attributes and levels for inclusion. There was a need to compromise on the perceived importance of attributes to ensure concerns regarding sample size/design feasibility, ecological validity (i.e., ensuring information presented is as consistent as possible with information that would be available in a real life situation) and the potential impact of unconscious attributes (e.g., patient ethnicity) were considered in the final attribute selection process. Issues considered during these panel discussions were:

Ensuring clinical face validity

Realism of the patient information presented in vignettes and that attributes/levels could plausibly occur together was vital. Certain attribute levels could not reasonably appear together meaningfully and therefore some initial constraints were imposed on the design. For instance, a modified Rankin score of 0 or 1 (indicating patient is able to carry out all usual duties and activities) is implausible for a patient described as having severe dementia.

Ensuring sufficient information was present in vignettes

It was crucial that information on certain attributes was provided in the vignettes (as either variable or fixed attributes) in order for clinicians to reach a decision about the offer of thrombolysis and for the decision-making process to mimic real-life decisions as closely as possible. For instance, stroke severity score and the results of the CT scan were considered as vital information.

Anticipated sample size and resulting design considerations

Whilst more levels can define the true relationship between attributes and their influence on decision-making, there is an inherent trade-off between this and the number of attributes and levels that can feasibly be included in the DCE design [1]. Furthermore, there are important considerations regarding the maximum numbers of attributes and levels a participant can process simultaneously. Due to the limited size of the pool of potential study participants in the current research, restricting the number of variable attributes and levels per factor was necessary to control the number of potential combinations in order to ensure the DCE would have sufficient statistical power to detect important effects. Based on previous studies and estimates of the total number of clinicians treating acute stroke patients in the United Kingdom, a sample size of 150–200 was estimated as achievable in the current study. This estimate represents approximately 25–45% of the 422 clinicians registered as ‘Full Members’ of the British Association of Stroke Physicians [30]. Moreover, to address cognitive burden, we judged that 12–16 hypothetical patient vignettes would be the upper limit of decision vignettes that should be administered to each participant (and this would be tested/clarified in the piloting of the DCE in Stage 4).

Attributes that were considered important to provide basic clinical information (but not included as variables in the DCE) were included as fixed attributes in the design to remove subjectivity around interpretation of additional issues. This provides common interpretation of such attributes across respondents, whilst retaining face validity. In the current study, blood glucose level was included as a fixed attribute in each vignette (6.0 mmol/L). An identical text description of a patient’s CT scan result was included in all vignettes to remove subjectivity around imaging interpretation. Definitions were also included to standardise the interpretation of attributes and levels. For instance, the standard definitions for modified Rankin scores [31] used to assess pre-stroke dependency were presented in each vignette. Definitions were also provided for pre-stroke cognitive functioning and the calculation for each level of stroke severity using the National Institute of Health Stroke Scale (NIHSS) [32] to ensure consistent interpretation.

Consideration of explicit versus implicit influences on decision-making

There is strong evidence that implicit attributes can be influential in medical decision-making [23]. Therefore the expert panel also considered attributes that did not emerge from the exploratory work and SPE. For instance, gender may have a population-level effect on thrombolysis administration, with evidence of under-utilisation in women [33], although this may reflect a different age and presentation profile in women [34]. This is despite a pooled analysis of randomised controlled trials demonstrating that women are more likely than men to benefit from thrombolysis [35, 36]. Gender was not considered to be a conscious consideration at the individual, bed-side level by the panel. Other attributes that may have an implicit effect on decision-making were ethnicity and social class. Panel discussions were focused on striking a balance between the inclusion of both implicit and explicit attributes, which may influence thrombolysis decision-making.

STAGE 4. Pilot testing

A pilot DCE was developed based on the decision made by the expert panel. The generation of a fully balanced design with all possible combinations of attributes and levels was not possible. Therefore, the software programme, NGene (v1.1.1; http://www.choice-metrics.com/), was used to generate a fractional factorial (d-efficient) design to generate a design that was as balanced as possible, given the imposed constraints. A fractional factorial design was employed due to the relatively large number of variable attributes (nine) included in the final DCE, with number of levels varying from two to six on different attributes. This allowed for the testing of a subset of possible combinations. Use of a blocked design allowed the number of vignettes presented to any one participant to be reduced to a number considered manageable in pilot testing to avoid overburdening participants. In each block, 13 hypothetical patient vignettes were presented to each participant.

The pilot testing of the initial DCE employed a ‘think aloud’ approach that was guided by best practice guidelines [37]. This method asks participants to verbalise what they are thinking when responding to the survey, which can help to reveal vague or confusing questions, or other issues in need of clarification. The aim of this pilot stage was to facilitate further testing of the credibility of the vignettes among the population of interest, examine participants’ understanding and interpretation of the task and questions, check the usability of the survey on different systems and browsers, and gauge how long the survey might take to complete. Changes were made to the design and wording of vignettes and phrasing of questions during the piloting process to ensure any amendments could be tested. Furthermore, post-pilot testing was undertaken by the expert panel prior to the formal data collection phase. The pilot testing protocol is available in Additional file 2.

STAGE 5. Final expert panel discussions

At this stage, the same expert panel from Stage 2 were engaged to finalise the DCE. Refinements were made to the online survey based on outcomes of pilot testing and final amendments to questions and the instructions to the DCE were agreed.

Results

Given the underreporting/lack of development work in DCE design in the field, we have developed a 5-stage process which offers a replicable and transparent strategy, with methods to identify a comprehensive list of attributes that may potentially influence a decision of interest. Underreporting of the design process highlights that procedures used to select the most salient attributes for inclusion in the final design need to be based on explicit criteria to permit a judgement about bias and external validity of the findings to be made. Furthermore, presentation of choice sets in DCEs using traditional methods may lack sensitivity to decision-making contexts, which indicated a need to identify the optimal mode of presenting choice sets to respondents to maximise face/clinical validity of the choice task that is important for augmenting external validity of findings from DCEs.

Stage 1. Full list of attributes identified from initial exploratory work

Table 1 displays all the patient-related attributes identified as potentially important in thrombolysis decision-making based on: (i) the analysis of new and existing qualitative data sets on the views and experiences of stroke clinicians involved in thrombolysis decision-making; and (ii) a review of relevant literature.

Table 1 Patient-related attributes that could influence decision-making about thrombolysis

Stage 2. Expert panel discussions

The expert panel reviewed the list generated by Stage 1 and concluded it was comprehensive and relevant. There was a need at this stage to begin the process of reducing the list to the most important attributes. A set of criteria were agreed to focus the scope of the study:

Removal of attributes that could be considered as related to uncertainty regarding the diagnosis of acute ischaemic stroke

This was considered prudent based on the rationale that: (i) incorporating diagnostic uncertainty has methodological implications for the design of the patient vignettes and interpretation of results from the DCE. Given that a diagnosis of acute ischaemic stroke represents a gradient of certainty (and includes consideration of differential diagnoses such as transient ischaemic attack and stroke mimics) this does not lend itself well to the DCE framework. Furthermore, diagnostic uncertainty was beyond the scope of the aims stated in the study protocol [38] to explore attributes influencing decisions to offer thrombolysis, as opposed to what attributes influence diagnosis.

Removal of redundant/uninformative questions from the SPE

Questions were excluded if the panel agreed that the area of uncertainty on any attribute had been clearly identified by the literature and/or qualitative data. For instance, a question regarding patient’s pre-stroke status using the modified Rankin Scale (mRS) was removed as the panel were confident that most clinical variation in decision-making exists between mRS 2 and 3.

Furthermore, attributes were omitted that have been shown to have a population level effect in research studies but were not considered by the expert panel (or were not identified in exploratory interviews) to be important for decision-making about thrombolysis at the individual patient level (e.g., current use of aspirin).

Stage 3 (phase 1). Results of on-line survey (SPE)

The online SPE was live for six weeks and 11 clinicians completed the survey (37% response rate). Results are summarised in Additional file 3.

Stage 3 (phase 2). Results of second round of expert panel discussions

The ISPOR report on good research practices advises that discussion with experts is an approach that can be employed to help reduce the list of attributes [6]. The expert panel completed a ranking exercise to generate discussion and to help work towards a consensus regarding the importance of attributes for inclusion, and which attribute levels should be varied and which should be fixed. The results of this exercise are presented in Table 2. In consideration of the potential influence of implicit attributes on decision-making about thrombolysis, the panel elected to include ethnicity and gender as variable attributes in the DCE. Socio-economic status was also considered for inclusion in the DCE. However, due to the challenging nature of representing this information in vignettes and the potential influence of type of employment (e.g. manual work) in judgements of risk/benefit, it was decided to omit this attribute. Based on these discussions, it was agreed to include seven fixed attributes and nine variable attributes (see Table 3).

Table 2 Stage 3 (Phase 2) Expert panel ranking exercise results (n = 6)
Table 3 Fixed attributes used in DCE and rationale for inclusion

The expert panel also considered the attribute levels for inclusion and discussions were informed and guided by the results of the exploratory work and by the SPE, which specially explored the ‘grey’ areas of decision making, or attributes where respondents felt there was inadequate evidence to inform decision-making (e.g., cognitive impairment). Based on the staged process, the expert panel reviewed results and selected attribute levels to explore these ‘grey’ areas of decision-making.

Stage 4. Results of pilot testing

Sixteen clinicians were invited to take part in pilot testing. Six agreed to take part and fully completed the pilot. The ‘think aloud’ approach helped to refine language used in the survey, ensured questions were interpreted by participants as intended, and confirmed sufficient information was presented in the vignettes to allow clinicians to reach a decision. The testing revealed that vignettes were credible, although a number of implausible combinations were identified and added as constraints to the design. These included low NIHSS scores with aphasia combined with moderate/severe dementia. These combinations were therefore excluded from the design.

Unexpected issues also emerged. For example, when two clinicians read that symptom onset began 4 h and 15 min ago, they factored in time to prepare the thrombolysis bolus into their decision-making (as the thrombolysis time window for treatment is up to 4.5 h post symptom onset). Therefore, thrombolysis bolus preparation time was included as an additional fixed attribute, stating that the treatment dose could be prepared within 5 min to eliminate this as an influence on clinicians’ decision-making (Table 3).

Stage 5. Final attributes and levels for inclusion in DCE

Final amendments were made by the expert panel based on feedback received during the pilot stage. As pilot participants felt that the patient vignettes contained sufficient information to allow them to reach an informed decision about the offer of thrombolysis, the nine attributes (and associated levels) used in the pilot were retained for inclusion in the final DCE (Table 4). Definitions for variable attribute levels in the vignettes were also finalised by the expert panel based on the feedback from pilot participants (see Additional file 4).

Table 4 Final list of variable attributes and levels in the DCE

Key considerations and final design of DCE

Key considerations of the development of this online DCE included the size of the population of interest, the number of combinations of attribute levels, respondent burden and likely response rate. It was crucial for the expert panel to consider these issues alongside the perceived importance of various attributes and levels. Based on this five-stage design process, we were confident that we had provided sufficient information to allow clinicians to reach a decision by using both fixed and variable attributes. A sample vignette is displayed in Fig. 2 to demonstrate how the hypothetical patient attributes and levels were presented to clinicians.

Fig. 2
figure 2

Sample patient vignette

Discussion

The aim of this paper was to build on existing DCE design guidelines by describing a systematic and rigorous development process using an exemplar case describing the development of a clinically-valid DCE to investigate clinical decision-making about thrombolytic treatment in acute stroke care.

Previous literature has often pointed to the underreporting of aspects of DCE design. The under-reporting (or lack of DCE development work) on crucial aspects of this design process makes a strong case for adding additional details to the minimum reporting guidelines in order for readers to judge quality and to lend confidence in DCE results. Despite existing guidance on DCE design, development and evaluation [2, 4] and calls to improve reporting on this process [4], underreporting is still prevalent [10]. However, as noted by others [11], this may be due to tensions between journal editors and authors regarding space constraints and word limits. Nonetheless, journals are increasingly offering authors the opportunity to include supplementary material as online appendices and we urge authors to avail themselves of this, where possible, to ensure adequate and comprehensive reporting on DCE research.

The value and strength of the results obtained from a DCE are contingent on the quality of the study design process. When DCEs are conducted in circumstances where there is little or no empirical evidence to guide study design, documenting the design process is especially important. Increased transparency at this stage would offer guidance to researchers, allowing them to draw on approaches and methods that best suit the objectives at hand, or to develop, extend or replicate existing DCEs. Robust methods should underlie DCE design, yet it is often not possible to judge these processes, or whether there was in fact any development process involved, from the limited information presented in many published papers.

Previous work by Coast et al. [4] indicates that underreporting of processes to design DCEs is a problem in the wider DCE published literature. As each decision made during the DCE development process can have implications for the final design and the results obtained, we strongly urge that, at a minimum, researchers should be required to report details of the development processes of the DCE design including the rationale for the approach adopted, details of analyses conducted, processes to narrow down attributes/levels, results of analyses, and details of the sample. Furthermore, piloting of the DCE is a crucial step and researchers should document what the aim of piloting was, the results of piloting, and provide details on the sample used and any changes made post-piloting. Finally, the content and face validity of the DCE must be considered and maximised during the design process in order for the study and its results to be considered meaningful. The adherence of researchers to these basic reporting conventions would improve the transparency of DCE research.

To exemplify this, we have established and described an iterative process that is transparent, rigorous and systematic, providing rationales for decisions made at each stage of the DCE design. This five-stage development process was used to design an online DCE utilising choice sets in the form of patient vignettes to examine attributes that influence the clinical decision to offer thrombolysis to acute ischaemic stroke patients. A review of relevant literature and secondary analysis of qualitative data in Stage 1 enabled the compilation of a comprehensive list of attributes that might influence clinical decision-making about thrombolysis. Through expert panel discussions, and applying explicit inclusion/exclusion criteria, Stage 2 enabled selection of attributes/levels that could be meaningfully included in the choice sets of this DCE. Discussions in Stage 2 informed the design of a structured prioritisation exercise (SPE) in Stage 3 that consisted of a ranking exercise to prioritise attributes and levels for inclusion in the patient vignettes. Findings of ranking exercises guided subsequent expert panel discussions to reach consensus on the DCE choice sets and on-line survey design. This was followed by a pilot phase incorporating a think aloud approach (Stage 4), and in turn finalising the design of the DCE choice sets using fractional factorial methods (Stage 5).

This paper contributes to the literature on DCE development by clearly documenting a systematic and rigorous, iterative process to derive the final number and type of attributes, attribute levels and associated operational definitions for inclusion in an online DCE. It builds on ISPOR guidelines [5, 6] by providing explicit guidance on a five-stage process, which may be particularly beneficial when there is little or no previous research to guide DCE design. This transparency in the development process of a DCE (process validity) to understand the decision of interest (decision to offer thrombolysis or not), as well as information used by clinicians in their decision-making, confers confidence in the rigour and reliability of findings and applicability to the real-world (ecological validity) [39]. Although we have outlined a linear approach to achieve this, this development process is flexible and may adapted as appropriate.

This development work highlights the crucial importance of an iterative design process for augmenting the clinical face validity of the choice task in terms of (i) mode of administration (paper-based or electronic); (ii) identification and optimal selection of variable and fixed attributes and associated levels to include in choice sets; (iii) form of choice set presentation (textual versus traditional tabular approach, and order of information presentation;) and (iv) the necessity of piloting to augment clinical face validity, and in turn external validity, and test acceptability of choice sets by the target population, prior to the data collection phase.

Identification and optimal selection of variable and fixed attributes, and associated levels, to include in choice sets is critical. Given the limited size of the population of interest in the current DCE, the size of the research sample (based on explicit inclusion criteria) and likely response rate were important considerations to establish the feasible DCE design options that were likely to be achievable and testable within an acceptable level of statistical power. Additional issues to consider alongside selection processes are (i) whether levels of attributes cohere together meaningfully within choice sets to identify any potential design constraints, (ii) whether the information within choice sets provide sufficient information to clinicians to make a decision (i.e., are specific levels of fixed attributes warranted for the purposes of clinical face validity), and (iii) if the levels of variable attributes elicit variability in the respondents in terms of the clinical decision-making about thrombolysis.

The presentation of attributes and levels in the form of hypothetical patient vignettes in accordance with preferences of clinicians differs from the traditional form of presentation of DCE choice sets and constituted a novel aspect of the study. The intended respondents are familiar with this form of information presentation during clinical training; this was therefore the optimal method of information presentation to enhance clinical and external validity. Furthermore, attribute order/placement in the vignette was decided based upon the information that would be available to the clinician immediately and the typical order in which additional information may be obtained. The order of presentation of attributes was also informed by work to develop the COMPASS decision support tool [20].

Whilst this DCE was designed for the specific purpose of elucidating attributes influencing the decision to offer thrombolytic therapy for patients with acute stroke, we contend that a key strength of this study is the development of a five-step approach that can be transferable to other research contexts. We believe the steps undertaken can assist researchers in determining the attributes of interest and ensuring broad input into the DCE design process. The number of interviews to conduct with expert informants or the number of participants required to adequately inform an attribute prioritisation exercise will likely differ considering the study subject and the availability of previous relevant research.

It is often difficult to predict or know in advance the number of development stages required for any particular study [9]. We propose that this design framework offers structure to researchers to apply to inform design decisions. The exploratory phase and involvement of subject experts beyond the research team is crucial to avoid narrowing the scope too early. After this stage, a structured exercise to agree priorities can be useful to understand those attributes that are strong influences or demonstrate high variability on a decision of interest. Depending of the aims of the research, one or both of these issues may prove more important to consider in DCE design. The use of the ‘think aloud’ approach is a very useful technique for pilot testing prior to data collection. It can alert researchers to ambiguous questions or terms, explore usability of the instrument and can highlight if participants feel the choice set is missing important information required to make a clear decision. The expert panel, comprising individuals with various perspectives and expertise on the decision of interest, was a useful resource during the research and we would advocate the use of such a panel to support decision-making in DCE design. It also enabled consideration of attributes that could potentially exert an unconscious influence on decision-making (as suggested by the evidence base) that may not have emerged during the qualitative research.

Among the challenges inherent in our approach is the valuation of data and how best to use the multiple data sources that were generated during the development process. Stages 1–4 shaped the study in terms of mode of administration and facilitated selection of variables identified as likely to be contributing most to variation in the decision of interest. Practical issues such as adherence to good practice guidance on design of survey questionnaires to maximise interpretability and response rates [40, 41] are also key considerations, including establishing predicted sample size informed by robust estimates of population size and response rates to similar studies. Sample size estimates are typically determined by the number of attributes and levels and what constitutes a meaningful effect size in the context of the study. These aspects are often not known a priori and therefore it can be challenging to incorporate this in the design planning process. We believe the steps described in this paper can be adopted and adapted to suit many other areas of research to inform appropriate DCE design. We encourage others to replicate the approach adopted here to explore the appropriateness of the design method to other domains.

Conclusions

The sub-optimal reporting on crucial aspects of DCE design makes a strong case for minimum reporting guidelines. The onus is on both researchers and journal editors to ensure a minimum reporting standard is adhered to in order to indicate quality and bestow confidence in DCE results. Our five-stage design process has demonstrated that adopting a rigorous staged approach to the development of DCEs yields enhanced face validity of the choice task and strengthens the external validity of findings. The iterative, transparent development process described here could be applied to other settings and index decisions and would be a valuable adjunct to existing guidance on development of DCEs.