FormalPara Key Points for Decision Makers

We consider understanding in the context of discrete choice experiments (DCEs) as an overarching concept that can be defined as participants choosing rationally based on comprehending both the choice context and choice task, and being willing to participate.

We identify a broad range of definitions, measurement approaches and proposed implications of participant understanding in the DCE literature.

While some common approaches emerge, there is little consensus around how to maximise participant understanding of DCEs in the health context.

There remains significant scope for further work on measuring and improving DCE participant understanding in health, which may draw on the use of DCEs in other fields, behavioural and experimental economics, and psychology.

1 Introduction

Discrete choice experiments (DCEs) are a quantitative method used to measure preferences for goods and services. The application of DCEs to measure preferences for health and health care is increasing in geographic scope, areas of application, and sophistication of the design and analysis [1].

The health context can present unique challenges that make DCEs complex to complete. They often include attributes related to risks or probability, and these concepts can be difficult for people to understand and accurately interpret [2,3,4]. Many health DCEs include unfamiliar medical terminology and some people, particularly those with lower health literacy, may have difficulty understanding the context of the choice they are asked to make [5]. The emotional intensity associated with making decisions about health and healthcare may result in heuristics that change people’s choice behaviour [6] and many health scenarios may be difficult to imagine, such as being in a wheelchair or having to pay for health care in a universal healthcare system [7, 8]. Features of the DCE design such as the number of attributes, levels, alternatives and choice tasks [9,10,11], and the use of choices with utility balance or in the ‘magic-p’ range [12, 13], can also lead to increased choice difficulty.

If people have difficulty understanding a DCE this may impact the validity of their responses. Difficult questions in surveys can reduce response rates [14] and increase the use of simplifying strategies [15], opt out selection [16], non-transitivity of choices [17], non-trading [18] and missing responses [19], and result in added variance and measurement error [20]. Each has the potential to reduce the quality or validity of the resulting preference data and therefore compromise results and interpretation of a DCE. An expert stakeholder panel of people doing and using DCEs suggested ensuring participant understanding was both the most desirable and most actionable characteristic of high-scientific-quality preference studies [21].

The concept of understanding relates to participants responding (choosing) rationally within a DCE based on comprehending both the choice context and choice task and being willing to participate. We conceptualise understanding as an overarching concept that is impacted by the task design and instructions, the attributes and levels and the choice context, as well as the personal characteristics of the respondents and the familiarity or relevance of the topic. There is no consistent or recommended approach to assess participant understanding of DCEs. Options include pretesting and piloting of the attributes and levels during the design phases [22, 23], including rationality or validity checks such as duplicate or dominated choice sets within the design [1, 24,25,26,27], using econometric approaches to assess rationality [28], or collecting self-reported participant understanding of the DCE [25]. However, most of these assessments are designed to assess a single component or indicator of understanding (such as consistent choices) rather than overall understanding in terms of choice context, choice task and/or task engagement.

There is limited previous research about whether and how people understand health DCEs and, more broadly, how ‘understanding’ is conceptualised. Ryan and Gerard [29] found “there did appear to be a significant minority who felt they had some difficulty, but given the nature of the question it was not clear whether this arose from the choosing per se or because the instructions were unclear”, recommending further research around how comprehension is impacted by the number of attributes, number of levels, and presentation format. Others describe respondent understanding in terms of participant comprehension and minimising choice task ambiguity, with rigorous development processes and effective communication as strategies to achieve understanding [21]. A recent review considered pre-choice processes of decision making rather than broader understanding [30], describing methods such as eye-tracking, brain imaging, time to complete and think-aloud interviews to investigate the processes individuals follow in completing a choice task. The aim of this review was therefore to identify how participant understanding is conceptualised in the health DCE literature, including how understanding is defined, measured, and used. This can inform future research to improve ease of understanding of DCE choice tasks, approaches to maximise validity of DCE findings, and how to appropriately interpret DCE results for use in changing practice and policy.

2 Methods

A scoping review was conducted following the methods of the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) guidelines [31, 32] to capture the scope of the literature and conceptualisation of understanding in health DCEs [33]. As a scoping review, registration in PROSPERO was not applicable.

2.1 Review Questions

How is participant ‘understanding’ (1) defined, (2) measured, and (3) used within the health DCE literature?

2.2 Eligibility Criteria

Eligible papers included peer-reviewed journal articles in the field of health, related to a DCE or BWS3 (best–worst scaling case 3 studies, an extension of DCE where respondents are presented with three or more profiles and instead of choosing the single profile they most prefer, they choose both the best and the worst profiles [34]), and including some consideration or assessment of participant understanding. The concept of understanding was deliberately kept broad using words such as comprehension and misunderstanding to allow the full scope of how understanding is conceptualised in the health DCE literature to be captured.

Conference abstracts, review articles or non-English papers were excluded. Additionally, papers describing how preferences or choice responses differed by subgroups such as sex or age were excluded unless they assessed differences by level of understanding.

2.3 Information Sources, Search Strategy and Selection of Sources of Evidence

Searches were conducted in June 2019 in the MEDLINE, EMBASE, PsychInfo and Econlit databases, consistent with other recent systematic reviews of DCEs [35, 36]. Search terms were based on previous systematic reviews of DCEs [1, 35,36,37]. Keywords for the concept of understanding were used in a proximity-based search with keywords for patient, participant and respondent. The full search strategy for each database is presented in ‘Supplementary Material A: Database-Specific Search Terms’. Hand-searching reference lists of identified papers, a recent systematic review of DCEs in healthcare [1], and personal collections of authors was also conducted.

A single reviewer reviewed the title and abstract for each paper identified. All authors participated in the full-text review of any studies that appeared eligible, or which could not be identified as ineligible from the title and abstract, with each paper reviewed by two reviewers and uncertainty discussed with an additional reviewer to achieve consensus.

2.4 Data Extraction and Synthesis

A descriptive analytical approach was used to summarise the relevant data from each eligible paper (and online appendices). Data extraction was initially developed and trialled on six studies and circulated to all authors to ensure all relevant information was captured. Final data extraction included publication year, country, clinical area, subject group (e.g. general public, patients, or health professionals), sample size, study design (e.g. online survey, qualitative interviews), DCE or BWS3, number of attributes, the range of number of levels, number of choice sets, definition of understanding, how understanding was tested, results of the understanding tests, and how the understanding information was used. Each study was also categorised based on how understanding was conceptualised and used within the study. Consistent with the best practice guidelines for scoping reviews [32], methodological quality and risk of bias assessment were not undertaken for this study.

3 Results

The search strategy identified 306 unique articles for potential inclusion. Review of titles and abstracts excluded 31 papers and full-text review excluded a further 200 papers, resulting in 75 eligible studies (Fig. 1).

Fig. 1
figure 1

PRISMA diagram for the scoping review, detailing the database searches and other methods of identifying articles, the number of abstracts screened, and full-text publications reviewed. DCE discrete choice experiment, PRISMA Preferred Reporting Items for Systematic Reviews and Meta-Analyses

Tables 1, 2 and 3 show the characteristics of the eligible studies, while Tables 4, 5 and 6 show how understanding was defined, measured, and used within each eligible study. The majority of studies were DCEs (or studies about DCEs), with two BWS3 studies [38, 39] and three studies including both a DCE and a BWS3 design [40,41,42]. Most were conducted in Europe (n = 24) and the UK (n = 17), followed by the US (n = 15) and Australia (n = 10). Over half of the studies (n = 42) had a population of patients or carers, and 31 studies included the general population or target population (e.g. for prevention studies, the general population is the target population). Two studies examined clinician preferences [39, 43]. The mean sample size was 421 (range 18–4287, median 231). In the DCE studies, the number of attributes ranged from 3 to 12 (mean = 6) and the number of levels ranged from 2 to 6. On average, participants faced 11 choice sets (range 5–27).

Table 1 Characteristics of publications included in the scoping review, by category: applied DCEs
Table 2 Characteristics of publications included in the scoping review, by category: pretesting studies
Table 3 Characteristics of publications included in the scoping review, by category: studies of understanding in DCEs
Table 4 Conceptualisation of understanding in the publications included in the scoping review, by category: applied DCEs
Table 5 Conceptualisation of understanding in the publications included in the scoping review, by category: pretesting studies
Table 6 Conceptualisation of understanding in the publications included in the scoping review, by category: studies of understanding in DCEs

Studies were categorised as (1) applied DCEs (n = 52), which included an assessment of understanding in the pretesting, analysis, or discussion; (2) pretesting studies (n = 7), which were stand-alone studies conducted prior to a DCE, designed to identify relevant attributes and levels of pilot DCE instruments; and (3) studies of understanding (n = 16), which were a mixture of qualitative studies and survey methods papers. Often these studies of understanding were a substudy to an applied DCE and aimed to assess the concept of understanding within the context of a choice experiment. Many focused on the factors influencing understanding or completion of a DCE.

3.1 Applied Discrete Choice Experiments (DCEs)

Fifty-two applied DCEs were identified. Most were conducted with patients (n = 35) and sample sizes ranged from 30 [39] to 1791 [44]. The most common clinical areas were diabetes and cancer. Twenty-four of the DCEs were conducted online, 18 were paper-based, and 8 were face-to-face interviews. An additional two DCEs were protocol papers describing DCEs yet to be conducted [45, 46].

Understanding was assessed in three ways: using pretesting to test or confirm participant understanding (n = 18); including tests of understanding within the DCE survey, such as a repeated or dominated choice set (n = 22); and referring to the concept of understanding in the discussion (n = 4). There were eight additional studies that addressed understanding in both the pretesting phase and with tests in the DCE survey.

3.1.1 Applied DCEs Addressing Understanding in the Pretesting Phase

The 18 applied DCEs addressing understanding in the pretesting phase all used the concept of understanding as a way to modify or revise the DCE instrument before final roll out [44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61]. Most used pretesting to test or confirm understanding in relation to the terminology of attributes and medical concepts. For example, “… comprehension and relevance of the attributes … as well as the survey instructions” [52] and “wording used in the questionnaire was correct and understood by the target population” [58]. However, many also took a broader view of understanding and used the pretesting process to assess understanding of the choice task; for example, whether participants “understood the attribute definitions, accepted the hypothetical context of the survey, and were able to complete the choice questions as instructed” [50]. Similarly, Bridges et al. [53] completed six pretesting interviews to assess whether participants “understood the survey and were willing to trade off among the attributes and levels”. Despite each of the studies in some way describing the use of pretesting to assess or confirm understanding, nearly one-third (n = 7) did not report the results of their assessment. Others gave minimal information, such as “minor changes were made to the wording to improve respondent comprehension” [50].

All except one study implied that understanding was established to be good enough to continue with the DCE. The exception was a protocol paper that described focus groups and 1:1 interviews with cancer survivors and oncology health professionals to develop a DCE about preferences for cancer care among people with cancer [46]. They found “Half of the participants felt that the questionnaire was confusing and difficult to interpret, which suggested that the questionnaire might be particularly burdensome and cognitively demanding to our group of patients”, and subsequently made changes to the text and layout to improve understanding [46].

3.1.2 Applied DCEs Including a Test of Understanding Within the DCE Task

The 22 applied DCEs that included a test of understanding within the DCE task were spread across year, country, clinical area, sample size and survey administration, and the definition of understanding was similarly varied [62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83]. While many studies referred to relatively generic concepts of understanding, such as the task or attributes and levels [e.g. 82], other studies took a more targeted approach and related the concept of understanding back to the underlying assumptions of rational decision making, referring to concepts of consistency [62,63,64, 77], dominance or non-satiation [67] and rationality [77, 78]. Some studies referred to engagement with the task as an indicator of understanding [70, 71, 79], while others referred to participants understanding the concept of making a choice [68, 72].

Despite these different definitions of understanding, methods to assess understanding were relatively consistent. Over half used a dominated choice set [63, 64, 68,69,70,71, 73,74,75,76,77,78, 82], with other approaches including a repeated choice set [64, 77, 79], analysis of dominant preferences [63, 67], inclusion of debriefing questions [65, 76, 80, 83], inclusion of quiz questions [65, 66], inclusion of a numeracy scale [66, 75] and use of the root likelihood approach in Sawtooth [62]. Fourteen studies used the understanding results to exclude participants from the analysis [62,63,64, 67, 68, 70, 71, 73, 74, 77,78,79,80, 82], two used them to describe the sample [65, 69], one included them only in the discussion section [72], and the remaining did not describe how they used the results of their understanding assessment [66, 75, 76, 81, 83].

Where reported, rates of passing the various tests of understanding varied and were spread across the studies that excluded participants based on the results of these tests and those that did not. Many studies reported rates of < 5% failing their tests [64, 68,69,70,71, 74, 76, 79, 80, 82]; however, some were much higher, for example, 12.5% provided inconsistent responses [62], up to 7.6% failed a dominated choice set, and up to 17% displayed dominant preferences [67], while 18% failed a rationality test [78] and 33% failed a repeated choice task [77]. Liu et al. [67] stated “These proportions of irrational choices are well within the acceptable standard in a DCE” and cited Johnson et al. [84], although it is not clear where in this work a threshold for irrational choice prevalence is given. Other studies simply stated when testing the impact of including or excluding those with inconsistent/irrational responses, there were no significant differences in the results [63, 64]. Some studies gave only descriptive summaries of the results, such as “Responses to the knowledge assessment questions also indicated good understanding” [65] and “Many found it difficult to understand the concept of choosing between hypothetical situations and hence took up to an hour to complete” [72].

Two studies compared irrational responses by sociodemographic characteristics [78, 82]. Adam et al. [82] found that the five participants (of 263) who gave irrational responses were more likely to be male, older, and less educated than the average respondent in their DCE [82]. Hol et al. [78] identified that participants who had previous experience of the health test under study (colorectal cancer screening) were significantly more likely to pass the rationality test than previously unscreened participants.

In studies where debriefing questions were used, the rates of people reporting the survey difficult to complete were higher than the proportion who failed the rationality test. For example, one study reported 3% failed, while 9% rated the survey as difficult [80], and, in another study, 7% failed and 24% reported the survey to be difficult [81]. In a study that only included debriefing questions, 12% of respondents reported they found the scenarios difficult to complete [83]. When numeracy was tested rather than rationality, it was reported that “The relatively high score for subjective numeracy score indicates that our sample probably did understand the risks and percentages they had to compare in the DCE task” [75].

Eight additional studies assessed understanding in both pretesting and within the DCE survey [43, 85,86,87,88,89,90,91]. These studies were similar to the other reported studies, with pretesting results used to refine the wording and formatting of the surveys, and the rationality tests using dominant or repeated choice tests and debriefing questions. Insufficient details were available to assess whether the rates of people failing the rationality/consistency tests were lower when pretesting for understanding had also been conducted.

3.1.3 Applied DCEs that Referred to Understanding in the Discussion

Of the 52 applied DCEs identified, four simply referred to the concept of understanding within the discussion [38, 39, 92, 93]. In these studies, understanding was referred to as an indicator of validity of the results. For example, “That all surveyed haematologists completed the self-administered discrete choice exercise without any missing data suggests a clear understanding of the issues involved in this experiment” [39], and “For a small minority, there was a suggestion of counter-intuitive preferences … this may be because of a lack of participant understanding or engagement or the result of an artefact of the survey design” [93].

3.2 Pretesting Studies

Seven studies described the process to develop a DCE—either identifying attributes and levels and/or pretesting the draft DCE instrument with an assessment of understanding [40, 94,95,96,97,98,99]. Most were published after 2015 and half were conducted in the US. DCE topics covered a range of clinical areas, such as diabetes [40], cancer [99], dermatology [98], and workforce retention [96]. All used qualitative interviews [40, 94,95,96, 98, 99] and/or focus groups [96, 97].

Most studies focused their assessment of understanding on participants’ understanding of the terminology in the attributes, choice sets, or survey. For example, in their focus group discussions, Helter and Boehler [97] defined understanding as “The final step in the process of attribute development aims to ensure that the desired meaning is evoked and that the terminology is understandable for the respondents”, and Coast and Horrocks [98] report testing participants’ ‘understanding of specific terms’. These types of studies usually presented results descriptively, such as “Names and descriptions of the attributes were adapted throughout the pretesting process to maximise participants’ understanding” [40].

However, some studies also included an assessment of understanding of the choice task itself. In a think-aloud study, Katz et al. [95] found “Most subjects … completed the DCE questionnaire without difficulty; however, not all subjects were fully engaged, and in some cases the interviewer … documented signs of cognitive overload or exhaustion. In a few cases, subjects struggled with the idea of choosing between two hypothetical medications … Most subjects showed good comprehension of the choice tasks and clearly considered two or more attributes in choosing between medications”. Similarly, in their cognitive interviews and focus group discussions, Abdel-All et al. [96] examined multiple aspects of understanding, which they describe as “comprehension (ability to understand the questions as intended), retrieval of information (thinking about the question and drawing conclusions), judgement and selection of response to the question”. They found “The DCE was well received … and … they did not find it difficult to understand the choice sets presented to them” [96].

One study did not specifically define understanding, but detailed the issues identified during their patient interviews [94]. These included respondents’ understanding of terminology, i.e. “two respondents read the word ‘ability’ as ‘aim’ because these two words are close in Danish. This fact proved to make the attributes difficult to understand and pointed to the need for reframing them”, and understanding the choice task, i.e. “… it became very clear that the included six attributes were too many ... and two respondents ended up answering lexicographically or using heuristics with emphasis on the last or the last two attributes, despite not ranking these the highest” [94].

3.3 Studies of DCE Understanding

Sixteen studies were specifically about the concept of understanding in the context of DCEs [9, 10, 18, 27, 41, 42, 100,101,102,103,104,105,106,107,108,109]. Published between 2002 and 2019, they were primarily from the UK [18, 27, 101, 105, 108, 109] and Europe [10, 42, 102, 104], and were conducted in a range of clinical areas, including vaccination [100, 102, 104], screening [27, 102, 105] and primary care [18, 101, 103, 106]. All were conducted in the general (or target) population, with the exception of the study by Veldwijk et al. [102], who included both the general population and patient groups, and Ryan and Bate [26] and San Miguel et al., [101] who used only a patient sample. The sample sizes ranged from 19 [105] to 4287 [18]. All included a DCE, with two also including a BWS3 task to compare the two approaches [41, 42].

3.3.1 Qualitative Studies of DCE Understanding

There were six qualitative studies in this group [27, 42, 100, 102, 103, 105]. Two were retrospective investigations of understanding in participants who had previously completed a DCE, and aimed to verify the appropriateness of DCEs as a method in their clinical area [100, 102]. They considered understanding to relate to both the information presented, such as definitions of attributes and levels, and the ‘complex decision strategies’ used when the task is understood. Both studies found that participants appeared to understand the technical information within the choice task, and the choice task itself. For example, “The majority of the participants seemed to have understood the provided information about the choice tasks, the attributes, and the levels. They used complex decision strategies (continuity axiom) and are therefore capable to adequately complete a DCE. However, based on the participants’ age, educational level and health literacy additional, actions should be undertaken to ensure that participants understand the choice tasks and complete the DCE as presumed” [102].

Three papers were prospective think-aloud studies, to assess decision-making processes [103] within a DCE and confirm participants were trading off [105] and choosing rationally [27]. All reported that most, but not all, participants were able to understand the concept of making a choice and make rational trade-offs in line with the assumptions of decision making. For example, one study found some participants reinterpreted the attributes, reacted in different ways to the cost attribute (from dismissing it to considering it a dominant attribute), showed evidence of non-trading, and brought their own experiences to the choice task [103].

The last qualitative study [42] compared the feasibility of DCE and BWS in their clinical context (genetic testing) by interviewing participants about their understanding of the choice format, as well as including a dominant choice set and self-reported difficulty questions. They found respondents completing the BWS task were more likely to report difficulties in both understanding and answering the choice task than those completing the DCE task [42].

3.3.2 Quantitative Studies of DCE Understanding

Among the 16 studies of DCE understanding, 9 were online or paper-based surveys [10, 18, 41, 101, 104, 106,107,108,109]. Four of these were designed to compare two different presentation versions of a DCE instrument, such as words or graphics [104], level overlap [107], general presentation format [106], and the impact of number of choice sets on choice difficulty [10]. The other studies examined various assumptions of decision making, such as rationality [101, 108] or dominant preferences [18], or compared a DCE choice task with another choice task—BWS3 [41] or Constant Sum Paired Comparison [109]. These studies measured understanding through consistency of preferences [101, 104, 107], dominated preferences [106, 107], duplicate choice set [107], participant perception of difficulty or format preferences [10, 18, 41, 101, 104, 106, 107, 109], time to complete the survey instrument [18], and survey fatigue [107].

The results were mixed for how task presentation impacts consistency and dominance. While some studies found alternative task presentations could improve consistency [104, 106], others did not [107]. It was noted that participants who reported the task as being difficult were less likely to satisfy tests of rationality such as dominance or consistency [18, 101]. Neither of the studies comparing the ease of DCE completion with other study designs found a significant difference in self-assessed difficulty of the different tasks [41, 109].

Finally, there was one laboratory-based study among the 16 studies of DCE understanding, in which 32 participants undertook a DCE within an eye-tracking system [9]. The study explored how the number of attributes in the choice scenario impacted attribute non-attendance. Understanding was conceptualised as ‘processing additional information’, and the study found “participants reported being engaged with the survey and although many stated that the choice sets with more information took longer to process, the information itself was not difficult to understand” [9]. However, this paper does not directly link the processing of information or attribute non-attendance with the concept of ‘understanding’ within a DCE.

4 Discussion

This scoping review aimed to identify how the concept of participant understanding is conceptualised in the DCE literature in the health context. We identified 75 DCE and BWS3 studies that included some consideration or assessment of participant understanding. They included 52 applied DCEs, 7 pretesting studies and 16 studies specifically to examine some aspect of DCE understanding.

4.1 How is Participant Understanding Defined Within the Health DCE Literature?

As with many key terms in the DCE literature, there is inconsistency in how participant understanding is defined. It appears that there are two primary ways ‘understanding’ is conceptualised within the health DCE literature. The first is the understanding of the general choice scenario, such as the medical terminology or levels of risk used in the attributes and levels. There has been consistent evidence in the general survey and DCE literature that specialised medical terminology [110] and calculations or presentations of risk and probability [2, 3] are difficult for many people to accurately interpret. This misinterpretation can negatively impact survey responses, including reduced response rates [14] and increased use of simplifying strategies [15], opt out selection [16], non-transitivity of choices [17], and non-trading [18] and missing responses [19]. Our review finds this type of understanding is typically referred to in applied DCE studies and tested using qualitative methods in the pretesting phase of DCE development. However, a growing body of work is examining the influence of alternative presentation formats for concepts and choices, to aid participant understanding and increase the validity of participant responses. Additionally, although not identified in this review, there is also interest in the use of novel presentation formats, such as video [111], to improve the instructions and definitions given to DCE participants. It is possible that having personal experience of the choice scenario may assist with understanding of attributes and complex concepts such as risk, as found in a study of preferences for colorectal cancer screening where previously screened subjects were more likely to pass the rationality test compared with screening-naive subjects [78]; however, to date there is insufficient literature to confirm this.

The second way ‘understanding’ is conceptualised within the DCE literature is the need to understand the concept of making a choice. Our review suggests this is seen in the literature as a key component of the rational decision-making process. This is reliant on participants feeling like the choice is realistic and one that they are ‘qualified’ to make [112], and although it can also be linked to the clarity of instructions and formatting within the survey instrument, it is more often related to the face validity of the decision-making process. Our review found that while this aspect of understanding is addressed in applied DCEs using tests such as duplicate or dominant choice sets, or studied specifically through both qualitative and quantitative methods, there is little consensus of the best way to measure or maximise this component of understanding. Assumptions that participants with good numeracy or health literacy will understand a DCE task also require further investigation.

Rarely seen in the literature in this review was the concept of engagement as a key component to understanding. Both understanding the choice scenario and understanding the concept of making a choice are reliant on participants being willing to engage with the task and process the information. It is clear from general survey literature that people may be more or less willing to engage with survey tasks due to feelings about the relevance or importance of the task [113, 114], personal characteristics (e.g. intellectual capability, attention span) [115], or familiarity with the concepts (such as having done DCE surveys before) [78]. While DCE response rates are often reported and have been associated with DCE complexity and relevance [116], there has been relatively little exploration of a broader conceptualisation of participant engagement as a component of understanding in the context of DCEs, such as dropout rates, time taken, topic relevance, and use of incentives.

4.2 How is Participant Understanding Measured and Used Within the Health DCE Literature?

There appear to be three primary methods to measure or assess participant understanding in the health DCE literature: qualitative methods in pretesting, rationality tests within the DCE, and self-reporting within the DCE. We did not find any studies including econometric approaches, such as the analysis of attribute non-attendance [28] or controlling for inconsistency and fatigue by allowing for unobserved preferences or scale heterogeneity [17, 117]. Regardless of the method used to assess understanding, the reporting of these methods, their results, and the implications for analysis and interpretation are not consistent.

While the use of qualitative methods, particularly in the pretesting phase, are important in developing high-quality DCEs [22], their use does not guarantee understanding during choice task completion. Data collection modes or participant characteristics may differ from pretesting to final rollout [118]. For example, pretesting of attributes and levels is often recommended or conducted using face-to-face interviews, even for paper-based or online DCEs [23]. Similarly, pilot testing may use a sample recruited from a different pool of potential respondents.

There were insufficient details in the included studies to assess whether studies including pretesting for understanding had lower rates of participants failing rationality tests in the final DCE.

Within the DCE instrument itself, quantitative approaches such as duplicate or dominant choice sets are often used to assess rational decision making. Rationality tests or validity checks are frequently cited to assess particular axioms of human decision making [1, 24,25,26], and exclude irrational, inconsistent or incomplete responses [24, 26, 27]. However, in this review, the meaning and implications of these assessments is described in various ways by different authors, suggesting a lack of clarity around the true purpose and interpretation of these tests. Liu et al. [67] suggested that their proportion of ‘irrational choices’ was “well within the acceptable standard in a DCE”, however these standards are not widely known or used. It also appears that while many studies state they will use the results for exclusion of ‘irrational’ respondents, this is rarely done in the primary analysis. Using these checks as a sensitivity analysis is consistent with both qualitative [27, 119] and quantitative [24] work, suggesting that ‘failing’ tests such as these does not necessarily reflect irrational or uninformative responses, or a lack of understanding of the task [27, 100, 101, 119].

Finally, self-reported ‘debriefing’ questions are often used to assess understanding within the DCE instrument. Ryan and Gerard [29] found 10 of 34 studies reported the results of debriefing questions around ease of completion; 40% of these studies reported that participants expressed difficulty, ranging from 20 to 35% of participants. More recently, a survey of health DCE authors found about half included debriefing questions, but the wording varied widely and the results were often not analysed or reported [25]. There appears to be limited literature to establish whether self-reported difficulty with a DCE is related to response validity or reliability.

4.3 Implications and Next Steps

There is inconsistency in how participant understanding is defined and characterised. This review provides a standardised conceptualisation of participant understanding and how it fits with other key components of DCE reliability and validity (summarised in ‘Supplementary Material B: Definitions of Key Terms in Relation to Participant Understanding’).

It would appear the concept of ‘understanding’ is commonly discussed as an important component of establishing the validity of DCE instruments, and numerous DCEs use pretesting, rationality checks and debriefing questions to measure or improve understanding. Despite this, there are relatively few studies specifically examining the concept of understanding within DCEs and how it can be measured or improved. The studies that do assess the impact of understanding on some aspect of DCE completion are primarily around alternative presentation formats, or design aspects such as the number of alternatives or choice sets. The existing assessments are primarily designed to assess a single aspect of understanding, such as debriefing questions asking about understanding the concept of making a choice, or an indicator for a lack of understanding, such as a dominated choice set. While assessments of rationality in completing a choice task are important, many may capture some aspects of understanding as well as some aspects of task completion that are unrelated to understanding itself, and more work is required to tease apart these concepts. It is likely that to truly capture overall understanding in terms of the choice context, choice task and task engagement, a suite of tests or assessments will be required.

For those conducting DCEs, our results suggest a need to carefully consider which aspects of understanding are critical to the choice context and ensure the methods to improve and assess understanding can be justified. Ensuring DCEs can be understood by participants will allow individuals and groups who may not participate in DCEs if the tool is too complex to have their preferences recorded. Similarly, improved understanding of a DCE instrument will increase the robustness of the results and ideally lead to improved health or healthcare experience for the community under study. For those using the results of DCEs to inform policy or practice, understanding and identifying the importance, potential risks and implications of participant understanding in DCEs may be an important component of interpreting the results, and improved participant understanding may increase the impact of these studies through increased confidence in their results.

There remain several avenues of future research to improve how we conceptualise, measure, and improve participant understanding in DCEs. Figure 2 illustrates how the three components of DCE understanding identified in this review, i.e. understanding the concept of making a choice, understanding the scenario, and being willing to engage, relate to different aspects of the DCE survey and participant. Future research to establish how each of these components relates to the processes of rational decision making by DCE participants and how they can best be measured is required and may allow identification of methods or strategies to improve understanding by targeting specific components [30, 120].

Fig. 2
figure 2

Model of participant understanding in the context of DCEs. DCE discrete choice experiment

The methods to measure or assess understanding identified in this review, i.e. qualitative pretesting, rationality tests, and self-reported debriefing questions, appear to be focused primarily on understanding the scenario or understanding the concept of choosing, with considerable intersection between the two. These (and other) strategies may be mapped to one or more of the components of understanding in Fig. 2, allowing a greater appreciation for the implications of failing these assessments. Ongoing research into the impact of various design features, such as task or risk presentation formats on the different components of understanding, and how they are best assessed is also required [111, 121], along with the value of including psychological instruments [122] and econometric techniques [24]. There remains little literature around the use of debriefing questions within DCEs [25], and an opportunity to develop and test a standardised set of questions that can address the different components of understanding may be useful.

4.4 Strengths and Limitations

This is a broad scoping review of how understanding is conceptualised within the health DCE literature. While this allows a general overview of the topic and is particularly useful as a first examination of the topic, this review has several limitations.

The broad conceptualisation of understanding taken in searching and extracting the literature made it difficult to balance sensitivity and specificity within the search strategy. Given the frequency of ‘understanding patient preferences for …’ statements in the stated preference literature, the search strategy needed to be relatively generic, leading to the need for high levels of manual review and hence potential bias and errors in the study selection and data extraction. There may also be pockets of literature that address the concept of understanding but which use different terminology to describe the concept, and thus were missed in our search.

We also likely did not capture all the applied DCEs that include some aspect of understanding within our search strategy, representing a possible selection bias. This is consistent with the scoping review methodology, where comprehensiveness is not necessarily the goal, and still demonstrates that understanding is often considered and included in applied choice experiments despite having relatively little research around how it is conceptualised or what it means. In addition, we know that many people who include tests of understanding within their DCE do not always analyse or report their results, therefore this is probably an underestimate of the scope of understanding within applied DCEs. We recommend authors to report their approach to and results of assessing understanding, at least in an appendix to their manuscript.

In addition, we did not look at DCE research outside the field of health. There could certainly be methodological papers that have addressed the concept of understanding within DCEs in other topic areas. If this is the case, it is interesting that there has been relatively little carry through to the health context, perhaps reflecting the unique characteristics of choices made about health and healthcare compared with other, less personal or emotional topics.

Finally, using the scoping review methodology means we did not include a formal evaluation of the quality of the evidence, and, by the nature of the study, we are unable to provide a synthesised result or answer to a specific question [123]. Rather, we provided an overview of the available literature and how it conceptualises understanding [123].

5 Conclusion

While the concept of understanding appears to be an important component of DCE studies in health, relatively few studies specifically examine the concept within their DCEs, or consider how understanding can be measured or improved. Within the health DCE literature, understanding is typically defined in two ways: understanding of the general choice scenario, such as the medical terms or levels of risk, or understanding the concept of making a choice. However, the impact on understanding of participant willingness to engage with the choice task is a gap in the literature.

The most common methods to measure or assess participant understanding in the health DCE literature are qualitative pretesting, rationality tests within the DCE, and self-reported debriefing questions. Regardless of the method used to measure understanding, the reporting of these methods, their results, and the implications for analysis and interpretation are not consistent, and there remains significant scope for further research into each of these methods and how they relate to and address participant understanding.

For those conducting or using the results of health DCEs, we suggest careful consideration of which aspects of participant understanding are most important in the specific choice context, a critical selection and justification of methods to assess understanding, and transparent reporting of these methods and results.