Abstract
Obtaining quantitative survey responses that are both accurate and informative is crucial to a wide range of fields. Traditional and ubiquitous response formats such as Likert and visual analogue scales require condensation of responses into discrete or point values—but sometimes a range of options may better represent the correct answer. In this paper, we propose an efficient interval-valued response mode, whereby responses are made by marking an ellipse along a continuous scale. We discuss its potential to capture and quantify valuable information that would be lost using conventional approaches, while preserving a high degree of response efficiency. The information captured by the response interval may represent a possible response range—i.e., a conjunctive set, such as the real numbers between 3 and 6. Alternatively, it may reflect uncertainty in respect to a distinct response—i.e., a disjunctive set, such as a confidence interval. We then report a validation study, utilizing our recently introduced open-source software (DECSYS), to explore how interval-valued survey responses reflect experimental manipulations of several factors hypothesised to influence interval width, across multiple contexts. Results consistently indicate that respondents used interval widths effectively, and subjective participant feedback was also positive. We present this as initial empirical evidence for the efficacy and value of interval-valued response capture. Interestingly, our results also provide insight into respondents’ reasoning about the different aforementioned types of intervals—we replicate a tendency towards overconfidence for those representing epistemic uncertainty (i.e., disjunctive sets), but find intervals representing inherent range (i.e., conjunctive sets) to be well-calibrated.
Similar content being viewed by others
Introduction
The collection, analysis, utilization (and monetization) of data, obtained largely from people, has increased exponentially over the last decades. This growth reflects rapid progress in information and communications technologies, which have been successfully exploited to improve access to these data. Many areas of research are already seeing great practical benefits from technological advances—consider the growing prevalence of remotely administrable digital questionnaires (Behrend et al., 2011; Gnambs & Kaspar, 2015; Krantz & Reips, 2017; Schmidt, 1997). Moreover, development and evaluation of methods and best practices for collecting and interpreting survey response data have long been important topics, subject to both extensive discussion and empirical research (cf. Converse & Presser, 1986; Fowler, 1995; Groves et al., 2011; Krosnick & Fabrigar, 1997; Meade & Craig, 2012; Payne, 1951/2014; Saris & Gallhofer, 2014; Thurstone & Chave, 1929). However, in stark contrast to the rapid development in methods of reaching people to acquire data, the predominant modes of capturing, encoding, and quantifying responses to social, behavioural, and psychological surveys remain fundamentally those developed early in the twentieth century—i.e., ordinal, ‘Likert-type’ and continuous ‘visual analogue’ scales (Freyd, 1923; Likert, 1932). We propose that it is time to pause, take stock, and ask whether technological developments permit advances in the type of data that can be efficiently collected through surveys—i.e., moving beyond point responses—as well as the means to access and analyse them.
Specifically, this paper focuses on an issue that we believe has received relatively little attention given its potential significance—the efficient handling of uncertainty or range that may be inherent in individual responses. These may arise from a variety of factors, including lack of knowledge or information available to the respondent, inherent randomness, variability, or vagueness in the answer, and ambiguity or ill-definition in the question. This paper begins by discussing the potential value of capturing such information through interval-valued responses, which may represent either ‘conjunctive’ or ‘disjunctive’ sets (cf. Couso & Dubois, 2014; Dubois & Prade, 2012), before contextualising this in comparison with the capabilities of conventional methods. It then puts forward and empirically evaluates an ellipse-based response mode—which is designed to efficiently capture and quantify the uncertainty and range associated with each response in the form of an interval, increasing both fidelity and information content. This method thus articulates the response level uncertainty that is often overlooked by point response modes.
Changing the actual format of the data (from traditional point estimates, to interval-valued) is a very substantial shift requiring an equally substantial and sustained research effort across all facets of the process of questionnaire-based research—from capture to analysis. Recent work by the present authors has included development of open-source software for efficient digital administration of interval-valued surveys (DECSYS—cf. Ellerby et al., 2019b), which is available to download (via https://www.lucidresearch.org/decsys.html) from GitHub. This paper builds upon this—reporting a study designed to provide an initial validation of the underlying approach. In this study, participants completed an electronic questionnaire with questions designed to vary systematically along three primary dimensions expected to influence response interval width: inherent variability or range in the appropriate response, adequacy and availability of supporting information, and clarity of question phrasing. We examine whether interval-valued responses reliably reflect induced variability along each of these dimensions—with experimental hypotheses that observed variance in interval widths will be significantly associated with each factor. Results find consistently in the affirmative, providing an initial demonstration of the efficacy of interval-valued responses to systematically capture additional information across a variety of circumstances, relative to discrete response modes.
The added value of intervals
Sometimes the correct response to a question is clear and discrete. How many sides does a square have? Here, the answer four is both correct and complete. However, this is not true for all questions. In many cases the appropriate response is uncertain, due to either ambiguity in the question or lack of knowledge of the respondent (cf. Converse & Presser, 1986; Coombs & Coombs, 1976; Bruine de Bruin et al., 2011; Fowler, 1995; Payne, 1951/2014), or because the correct response intrinsically comprises a range of values (cf. Budescu et al., 2009; Harris et al., 2017; Liu & Mendel, 2008; Wagner et al., 2013; Wu et al., 2012). These three features are illustrated in the following examples.
First, consider the uncertainty induced by the question ‘What number will I roll?’ The context of the question is not specified—what is to be rolled? If dice, then how many of them? Of how many sides? And on how many occasions? Viewed alone, the ambiguity in the question makes it impossible to confidently answer. Second, consider the question ‘I will roll two fair six-sided dice one time; what sum total will I roll?’ Here, the context and the meaning of the question are quite clear. However, the correct answer is inherently unpredictable. Given the stochastic nature of rolling dice, although a best guess can be made, a specific number remains impossible to answer with certainty. The only thing we can say for sure is that the correct answer falls within the interval [2,12]Footnote 1. Third, consider the range of values associated with the question ‘I have one standard six-sided die; what numbers are shown on its faces?’ The question is worded clearly, and all relevant information is known. However, the answer comprises a set of distinct values and, to completely express a correct response, the full range of these values must be represented. The minimum complexity response format with the capacity to do so is an interval, with the answer [1,6].
In each of the cases described, any discrete response is insufficient to convey all information necessary for a correct answer. When presented with a discrete response mode, each situation will manifest as response uncertainty—in terms of which single response option represents the most appropriate answer. The act of forcing respondents to condense their answer into a discrete approximation can be viewed as adding noise to the data, as extraneous factors may influence which discrete response option is chosen at the time. Limiting the impact of collapsing a potentially complex response, by enabling interval-valued responses—which intrinsically comprise a range of values (cf. Cloud et al., 2009)—should permit the capture of higher-fidelity response data. We argue that there are many real-world cases where it would be both more valuable and more efficient to offer respondents the opportunity to provide interval-valued responses, rather than currently prevalent point response modes.
One important consideration concerning interval-valued responses is that they may capture different types of information. Specifically, intervals reflecting uncertainty generally comprise a disjunctive set, in the sense that they describe the (lack of) knowledge of the respondent about a specific quantity—as described in the first two cases above—rather than an actual range (cf. Couso & Dubois, 2014; Dubois & Prade, 2012). Confidence intervals are a common example of such disjunctive sets. At the same time, intervals also frequently represent conjunctive sets—i.e., true ranges of values existing in the real world—as described in the third case above. In this paper, we do not focus on defining the nature of the information captured by the interval; we aim rather to demonstrate that capturing intervals can fundamentally afford the systematic and efficient capture of information not available using discrete response modes.
Practical applications
One well-documented application for interval-valued judgements is in direct elicitation of uncertainty in respect to individual estimates (i.e., confidence intervals)—whether from experts or otherwise (cf. Cooke, 1991; Hemming et al., 2018; Kahneman et al., 1982; Klayman et al., 1999; Morgan, 2014; Soll & Klayman, 2004; Speirs-Bridge et al., 2010). As alluded to above, however, it is important to note that the information which can be captured by intervals is not limited to confidence intervals.
In practice, interval-valued responses may be particularly useful in the context of decision research, where participants are routinely asked to make unfamiliar choices and estimate probabilities of outcomes under uncertainty (cf. Cubitt et al., 2015; Ellerby & Tunney, 2017, 2019; Tversky & Kahneman, 1974). However, aside from eliciting intervals with an explicit degree of confidence, interval size can also be used more generally to identify questions for which participants are more or less uncertain or confident about the most appropriate response (cf. Yaniv & Foster, 1995, 1997). Consider the extreme case of ‘Don’t Know’ responses, which may be expressed by selecting the midpoint of an ordinal scale; interval-valued responses allow for the clear distinction between respondents who are uncertain or ambivalent and those who are genuinely neutral, as shown in Fig. 1. Difficulty in doing this is a fundamental limitation of point response formats (Coombs & Coombs, 1976; Klopfer & Madden, 1980; Krosnick & Fabrigar, 1997) and has led to the frequent introduction of an additional ‘Don’t know’ response option separate from the scale, increasing complexity for respondents and analysis (discussed further in the next subsection).
In questionnaire pretesting, interval-valued responses could be used to identify questions that are causing confusion for participants. Here, unusually broad intervals could be interpreted as indicating questions that are ambiguous, imprecise, or otherwise difficult to understand. Complementing existing tools (e.g., QUAID—Graesser et al., 2000), this would allow for easy identification of areas that require improvements in phrasing or presentation, and could also be particularly useful for informing the nature of any observed differences between native and non-native speakers.
In attitudinal surveys, intervals could also be a valuable tool for informing the degree of conviction with which a respondent holds a given attitude or belief, which may be orthogonal to the strength of the attitude or belief itself. For one case—consumer preference research—understanding the degree of conviction or flexibility associated with consumer preferences or attitudes could valuably inform the likelihood of these motivating future behaviour, as well as openness versus resistance to persuasion and potential for establishing consensus (Ellerby et al., 2020; Petty et al., 2002; Rucker et al., 2014).
For another example, consider political science, where facilitating the capture of each respondent’s level of commitment regarding their voting intentions could be valuable in informing more accurate models of voting behaviour (cf. Burden, 1997). Specifically, in polling, we could reduce the noise arising from forcing the respondent to collapse their response model (a complex belief) into a single value at the point of completing the survey. As the same participant may collapse the same belief to a different discrete response when actually voting (e.g., based on immediate external influence, such as a salient news event on the day of polling), it is preferable to capture the actual range of the belief as completely as possible—thus providing not just voter intent at the polling stage, but information regarding the level of voter uncertainty. In fact, a study conducted by Aldrich et al. (1982) positively evaluated the possibility of allowing respondents to select multiple response points along an ordinal response scale. They concluded that their results suggested that discrete measures ‘force a false precision of response… expressed ambiguity is a real phenomenon. If so, the current practice of forcing respondents to select a single point not only introduces another source of measurement error into already imprecise data, but also loses theoretically important information.’ (p. 411). Nevertheless, discrete response scales predominate to this day.
Finally, intervals will of course be valuable in any case where the most correct and complete answer to a question is itself an interval, such as interpretation of probability phrases (cf. Budescu & Wallsten, 1985, 1995; Harris et al., 2017; Karelitz & Budescu, 2004) or other linguistic terms (Navarro et al., 2016; Wu et al., 2012). Here, they will inherently improve the fidelity of the response.
Limitations of conventional response modes
The putative benefits of any new survey response mode must be considered in relation to existing paradigms. This section focuses on two of the most prevalent, Likert-type and visual analogue scales. Since their development 85 years ago, Likert scales, and in particular the corresponding response format (Likert, 1932), have become established as a ubiquitous method of data collection and basis for analysis over a broad range of research areas. This type of response was designed to capture attitudes; it is therefore commonly used in market research, to obtain customer feedback, and in psychometric measures such as personality inventories. To be precise, although commonly used to describe the range of allowed responses to a single questionnaire item, the term Likert scale refers only to the combined responses relating to multiple items (Carifio & Perla, 2007).
Standard Likert questions require a discrete, ordinal response. These are often selected from a range of 1 to 5, or along a corresponding axis of Strongly Disagree to Strongly Agree. This response format is generally agreed to constitute an ordinal scale, though it is often assumed to represent interval dataFootnote 2, in the sense that each point is assumed to reflect a value that is equidistant from each adjacent point (Blaikie, 2003). This conjecture has been widely criticised (cf. Bishop & Bishop & Herron, 2015; Jamieson, 2004; Knapp, 1990; Kuzon et al., 1996), with an important consideration being whether the data in question represent a true scale (i.e., comprising multiple collective responses) or only individual Likert-type questions (Boone & Boone, 2012; Carifio & Perla, 2007, 2008). Nevertheless, counterarguments have quite equivocally proposed that parametric statistics are generally robust for the purposes of analysing Likert-type data, and that their use in these cases does not substantially increase the risk of ‘coming to the wrong conclusion’ (Norman, 2010, p. 7), by comparison with non-parametric alternatives.
The standard Likert response format offers relatively low granularity in its most common 5-point form, though this can of course be remediated by increasing the number of response options. Research suggests that doing so has some effect on key data characteristics such as internal consistency, but that this is not transformational (cf. Dawes, 2008; Maydeu-Olivares et al., 2009). Alternatively, visual analogue scales (VAS) permit continuously variable responses, by allowing the respondent to select any single point along the represented continuous dimension (cf. Ahearn, 1997; Aitken, 1969; Couper et al., 2006; Freyd, 1923). In theory, therefore, VAS allow for maximum potential response precision. They have also been shown to approximate an interval-2 rather than ordinal-scale level of measurement (Reips & Funke, 2008). In practice, however, VAS have been found to provide responses comparable to Likert-type scales, while the latter have sometimes been considered preferable in terms of ease of use for the survey respondent (Guyatt et al., 1987; Kuhlmann et al., 2017; Van Laerhoven et al., 2004).
However, as alluded to previously, conventional application of these response formats does not consider—or attempt to measure—each individual respondent’s uncertainty, or range (e.g., vagueness) inherent in the response. That is, both Likert-type and VAS questions force the respondent to condense their answer to a single-valued data point, effectively hiding and resulting in the loss of this associated information. At the data analysis level, statistics concerning between-subject (i.e., inter-source) variance can be calculated from discrete responses post hoc. However, as a method for estimating uncertainty, this is relatively inefficient, in that it requires multiple subjects. More importantly, both within-subject (i.e., intra-source) uncertainty and range inherent in each response are fundamentally distinct from this—reflecting heterogeneity within individual responses. Capture of this information is the focus of this paper.
It is possible to obtain information on individual response uncertainty using conventional methods. Consider the problem of ‘Don’t know’ responses (cf. Coombs & Coombs, 1976; Klopfer & Madden, 1980; Krosnick & Fabrigar, 1997). The inclusion of a separate ‘Don’t know’ option offers one workaround, but with its own fundamental limitations. Uncertainty is not all or nothing, but of varying degrees—when a ‘Don’t know’ option is present, participants are forced to choose between giving a best guess, to communicate their limited information regarding the subject in question, or instead to opt out, to communicate information only about the presence of some unspecified (though presumably substantial) degree of uncertainty. However, as they cannot express both, in any case where the respondent is neither entirely certain nor entirely ignorant, the response will be incomplete.
Alternatively, dedicated additional questions may be asked about the uncertainty associated with each judgement—which may also be framed in terms of confidence (cf. Cheesman & Merikle, 1984; Dienes et al., 1995; Peirce & Jastrow, 1884; Sporer et al., 1995; Tunney & Shanks, 2003). However, this approach is inefficient, effectively doubling the number of responses that participants are required to provide. Moreover, although this question is not directly examined in the current study, the lack of cohesion between initial responses and subsequently elicited uncertainty responses may also limit response fidelity—for instance, confidence may differ depending upon response time, and if confidence ratings are provided after initial responses, then these ratings may have continued to evolve in the intervening period (cf. Baranski & Petrusic, 1998; Pleskac & Busemeyer, 2010). Finally, although they often correspond, we note that ‘confidence’ is a complex notion inequivalent to (un)certainty—and reported confidence may have distinct determinants from uncertainty reported through interval width (Teigen & Jorgensen, 2005). For example, a respondent may be perfectly confident of an inherently interval-valued response (e.g., the conjunctive set of numbers shown on a standard die, as mentioned earlier).
Enabling efficient interval capture using ellipses
An emerging ellipse-based response mode has been developed to allow efficient, coherent, and intuitive capture and quantification of uncertainties and ranges associated with individual survey responses. This method is designed to leverage the growing prevalence and public familiarity with modern information technologies—such as accurate and portable touchscreen devices—and builds upon widespread familiarity with ‘circling’ areas of interest to minimise, and potentially eliminate, any added effort investment required to obtain this richer response information. It is designed as an efficient compromise between traditional and currently ubiquitous methods, such as discrete ordinal scales, and more complex or esoteric approaches, such as qualitative interviews, various methods of eliciting (specifically) probability distributions (cf. Bruine de Bruin et al., 2011; Morris et al., 2014; Speirs-Bridge et al., 2010), including SHELF (Gosling, 2018; O’Hagan, 2019), and the ‘fuzzy graphic rating scale’ (FRS), (cf. Hesketh et al., 1988; Hesketh et al., 1992; Lubiano et al., 2016; Quirós et al., 2016).
In contrast to currently predominant point response formats (e.g., Likert-type and VAS), the ellipse response mode elicits intervals. While a variety of approaches could enable provision of interval-valued responses, in this case respondents do so by drawing an ellipse along a continuous scale, delineating the interval that they believe best represents the answer. Crucially, participants are encouraged to use the width of this ellipse (i.e., the size of the interval) to indicate the degree of range (e.g., uncertainty, variability, vagueness, or ambiguity) in their response. This is illustrated in Fig. 2, by comparison with more traditional Likert-type and VAS response formats.
Therefore, this interval-valued response mode extends beyond the benefits of existing continuous scales, such as VAS. It not only has the potential to capture responses that are precise, but also makes possible the capture of fundamentally distinct information, such as individual response uncertainty or vagueness. This information is usually lost when using point-valued response formats. Moreover, ellipses achieve this in a single, integrated, and cohesive response. These attributes should make the approach intuitive to use and quickly and easily administrable, facilitating broader adoption while reducing training requirements necessary to obtain high-fidelity responses. The ellipse approach is designed primarily to streamline the process of interval-valued data collection so as to provide an alternative to point response modes (e.g., Likert-type or VAS), where it offers a substantial informational advantage at a minimally increased, or potentially even reduced, workload—e.g., by counteracting choice paralysis when selecting between multiple potentially appropriate discrete alternatives or requiring fewer questions to be asked.
Previous research on the elicitation of confidence intervals (an example of disjunctive sets) has consistently indicated that interval-valued estimates tend to be overconfident (Juslin et al., 1999; Klayman et al., 1999; Tversky & Kahneman, 1974; Yaniv & Foster, 1995, 1997). Further evidence suggests that singular interval estimates, such as those elicited by the ellipse response mode, are prone to greater overconfidence than those split into multiple steps (Soll & Klayman, 2004; Speirs-Bridge et al., 2010). At the same time, the coherent, singular nature of ellipse responses is a key asset in avoiding the added complexity and effort induced by soliciting multiple responses for each survey item.
While this goes beyond the scope of the present paper, there is evidence to suggest that for uncertainty capture (i.e., the disjunctive case), ellipses may be better placed to capture respondents’ subjectively ‘reasonable’ bounds—with the potential for follow-up questions to determine subjective confidence in the accuracy of the interval provided (Teigen & Jorgensen, 2005; Winman et al., 2004), and potentially calibrate or convert these into standardised credible intervals or probability distributions (cf. Hemming et al., 2018; Speirs-Bridge et al., 2010). However, as discussed earlier, confidence intervals and, more generally, disjunctive sets are only one example of the information which can be captured by intervals. In practice, the most appropriate method to use will depend upon both the questions to be answered and the time and effort available to invest. Here we focus on the ellipse response mode as an alternative to day-to-day surveys that use discrete responses.
In different disciplines, related research has already been conducted into practical foundations for using interval-valued responses. Much of this has focused on the exploration and development of appropriate mathematical methods to effectively extract, process, and interpret the resulting data (cf. Anderson et al., 2014; Ferson et al., 2007; Havens et al., 2017; Kabir et al., 2021; Liu & Mendel, 2008; McCulloch et al., 2019, 2020; Miller et al., 2012; Wagner et al., 2013; Wagner et al., 2015; Wu et al., 2012)—leveraging existing theory from the fields of interval arithmetic (cf. Cloud et al., 2009; Moore, 1966; Nguyen et al., 2012) and fuzzy set theory (Zadeh, 1965). The present authors have also directed substantial recent work into developing a practical software tool (DECSYS—Ellerby et al., 2019b, 2020), which now makes efficient digital collection of interval-valued survey data scalable and widely accessible.
Nonetheless, despite the promise of and growing interest in this response mode, it remains to be conclusively demonstrated whether interval-valued responses provided through ellipses reliably capture set-valued information (either disjunctive—e.g., arising from uncertainty, or conjunctive—e.g., arising from vagueness); or whether, due to either participant satisficing or careless responding (cf. Krosnick, 1991; Meade & Craig, 2012; Vannette & Krosnick, 2014), they might instead be arbitrary or dependent upon other extraneous factors. The present study was designed to address this gap in the evidence base and determine whether this ‘quick and easy’ approach can indeed provide robust and valuable additional information associated with individual responses—establishing a foundation for further research and development of methods for efficient interval capture.
Experimental summary
The aim of the experiment was to assess the efficacy of interval-valued survey responses to capture meaningful additional information, across a variety of circumstances and relating to three potentially distinct sources, which may each be present when responding to conventional survey questions. Over a three-section digital questionnaire, we induced varying degrees of each—inherent variability in the appropriate response, adequacy and availability of supporting information, and clarity of question phrasing—to determine whether the variance in size of elicited intervals is systematic, with reference to established ground truth in each of these cases.
In Section 1, participants were asked to simply identify and reproduce data from a chart, which itself comprised a range of specific values (i.e., a conjunctive set). The objective here was simply to establish whether respondents, given minimal instruction, are able to understand the basic concept of an interval-valued response, and to use the ellipse format to communicate ranges that they are aware of—i.e., which were in this case explicitly provided.
In Section 2, participants made their own judgements concerning stimuli with experimentally controlled degrees of both uncertainty and variability. The aim here was to examine how respondents apply the interval-valued response format to represent their own experienced degrees of stimulus-related uncertainty and variability (i.e., both disjunctive and conjunctive aspects), as well as the interplay between these two factors.
In Section 3, participants responded to a broader variety of items, which were designed to differ systematically in levels of ambiguity, specificity, and comprehensibility. In this section we explore how respondents use interval-valued responses to communicate their own subjectively perceived ambiguities or degrees of uncertainty, when induced by lack of clarity in, or comprehension of, the question or associated stimuli.
Eliciting interval-valued responses from participants in each of these cases, following only brief instruction regarding how to use this response format at the outset of the study, should provide insight into the potential for practical utility and real-world adoption of this response mode across a range of contexts and situations.
Method
Participants
A total of 40 participants completed this experiment, recruited in an opportunity sample. As the study was advertised primarily across University of Nottingham UK campuses, via physical posters and email list invitations, members of the university community were disproportionately represented. Insofar as is feasible within this population, the sample was varied—with participants originating from a variety of schools, UK campuses, and job roles (a substantial proportion of staff participants were non-academic). Each participant was paid a fixed inconvenience allowance of £5 (GBP), to compensate for their time spent taking part. Nineteen of these self-identified as male and nineteen female, two declined to report their gender. Ages ranged from 18 to 50 (M = 25.65, SD = 7.75). Twenty-four reported as native English speakers and fifteen as non-native speakers, while one declined to report this.
A range of different statistical analyses were conducted; statistical power will differ substantially between these. For the simplest cases, i.e., one-sample and paired-samples parametric t tests, this final sample size was determined to offer a statistical power of > .99, .87, and .23, to detect a large, medium, or small effect, respectively (two-tailed, α = .05, d = .8, .5, and .2—cf. Cohen, 1988). For ANOVAs, power depends heavily upon specifics—including number of measurements, as well as factors that are not known a priori, such as correlations among repeated measures and sphericity of the data to be obtained. Nonetheless, calculations using default values for these parameters indicated, across all planned ANOVAs, a minimum power of > .99 to detect a large effect, .94 to detect a medium effect, and .26 to detect a small effect (α = .05, corr. among rep. measures = .5, sphericity assumed, F = .4, .25, .1, respectively—cf. Cohen, 1988). When applying worst-case scenarios for sphericity, power across all analyses remained at minimum .95, .62, and .15, for F = .4, .25, and .1. All power calculations were made using G*Power (Faul et al., 2007).
Questionnaire items
All questionnaire items were developed for the purposes of this study. These were classified into three distinct sections. The purpose and content of all three sections was briefly explained to each participant before the beginning of the study, and this was reiterated for each individual section before it was begun.
In Section 1, participants observed a chart showing claimed natural life expectancies for several different household pets, if given proper care and attention (see Fig. 3). This was created using data obtained from injaf.org, under a Creative Commons agreement for non-commercial use, and with written permission from the copyright holder. The data provided were already in the form of intervals (e.g., ranging between 15 and 20 years for a domestic cat)—the task of respondents was therefore simply to reproduce these data as best they could, over a series of eight specific animals, using the interval-valued scale.
In Section 2, participants were presented with a series of six sets of 35 marbles—each comprising some proportion of blue versus yellow marbles. Each set was organised into five rows, of seven marbles each. Respondents were tasked with providing their best estimate for how many blue marbles were present overall in each row of the given set—with a response scale ranging between zero and seven. Initially, for each set, six out of seven marbles in each row were hidden, entailing a substantial degree of uncertainty around this estimate. In the following questions, three more, and then all marbles in each row of the set were revealed. This process was repeated for each of the six sets, for a total of 18 questions. Importantly, for some sets the number of marbles in each row was identical, but for others there was a discrepancy between rows. This between-row variability entailed a degree of range in the correct interval-valued response. Example stimuli are shown in Fig. 4.
In the third and final section, respondents answered a collection of more varied items. These questions were designed to hold varying degrees of vagueness, ambiguity, or comprehensibility in their wording or meaning. Crucially, although questions were deliberately worded to induce qualitatively more or less uncertainty, ambiguity, or vagueness in the response, the precise quantity of these depended upon the subjective interpretation of the respondent. First, questions were asked concerning the temperature in England, either during a specific month (December, July) or without qualification, the aim being to establish whether interval widths could be used to distinguish between more and less specifically phrased questions. Second, questions were asked concerning a scenario in which a percentage value either increased by 50% or doubled—with the former being possible to interpret as either a 50% or 50-percentage-point increase on the original amount. The aim here was to determine whether interval widths could discriminate questions with clear versus potentially ambiguous interpretations. Third, respondents were asked to report how well they were described by a series of 12 personality-related words—four of these were commonly used (talkative, aggressive, lazy, quiet), four were low-frequency (garrulous, bellicose, indolent, taciturn), and four were not real words, but created solely for use in the study (brendacious, apoccular, lombardistic, revenotile). The objective here was to establish whether intervals could identify questions within which a word was poorly, or not, understood. Fourth, respondents were asked a series of double-barrelled questions, followed by their single-barrelled counterparts (e.g., How well does this statement describe you?—I like reading books and watching television, vs I like reading books, vs I like watching television. Also: watching and playing sports, drinking tea and coffee, cooking and eating). The objective here was to further establish whether intervals could discriminate more and less specifically phrased questions.
The final two question types in this section moved the focus from the question wording alone, to the combination of question and associated stimulus. Fifth, respondents were asked to judge how many cars, and then how many vehicles, were present within each of two images (see Fig. 5)—each of these contained eight vehicles (the more inclusive term) in total, of which a potentially disputable number could be considered cars. Sixth, and finally, respondents were asked to judge the number of blue marbles contained in three rows of eight marbles each (see Fig. 6). For the first of these, the correct response was clear, with marbles each being either blue (4), green (2), or yellow (2)Footnote 3,Footnote 4. For the next, a gradient was introduced from blue to green to make the number of blue marbles blurred, or vague. For the final set, marbles each contained only blue or yellow3, but five marbles were patterned, containing varying proportions of each colour. As the question did not clearly specify whether the term blue was intended to refer only to completely blue, or also to mainly, or even partially blue marbles, the appropriate response could be considered ambiguous. The aim in each of these cases was to further examine whether the width of interval-valued responses would enable discrimination between question–stimulus combinations that were clear, and those that may be interpreted as vague, unclear, or ambiguous.
In the subjective feedback section of the questionnaire, participants were asked four questions. These were whether they found the response format ‘easy to use’, ‘unnecessarily complex’, and whether it allowed them to ‘effectively communicate [their] desired response’, as well as whether ‘Overall, [they] liked the response format’. These were administered using a conventional 5-point ordinal scale, ranging from ‘Strongly Disagree’ to ‘Strongly Agree’.
Experimental design
The study used a repeated-measures design, in which each participant completed the entire 63-item questionnaire. All responses were made through the ellipse-based interval-valued response mode, using a touchscreen Microsoft Surface Pro computer, together with a stylus. All surveys were administered through a local area network, using the DECSYS software’s Workshop Mode (Ellerby et al., 2019b)Footnote 5. In this paper, response intervals were deconstructed into two separate dependent variables, namely their position (captured by the interval mean) and width (the distance between interval endpoints)Footnote 6,Footnote 7. Participants also completed a short feedback questionnaire—adapted from the Systems Usability Scale (Brooke, 1996)—for which the dependent variable was degree of agreement, represented along a 5-point ordinal response scale.
Question order was not randomised for each respondent, due to design factors combined with limitations to the randomisation capabilities of the early version of DECSYS used for the study. In Section 2, for instance, it was necessary to present the three questions relating to each set of marbles together and in sequence, with six, then three, then no hidden marbles. Question order must therefore be considered as a potential factor when interpreting results; for example, participants had more experience providing interval-valued responses by the time they were completing the latter stages of the survey.
Procedure
Prior to the study, ethical approval was obtained from the University of Nottingham, School of Computer Science—where the lead authors are based. Before beginning the experiment, each participant was presented with a project information sheet, and informed consent was obtained. All participants were free to withdraw from the study at any time and without giving a reason. Following this, participants were individually seated before a Microsoft Surface Pro touchscreen computer and began the survey process. Initially, participants read a general instructions page, in which the appropriate use of the interval-valued questionnaire response format was briefly explained to them (see Appendix 1). This instructed participants to use an ellipse to mark each answer, and that a wider ellipse should be used to indicate greater uncertainty, range, or vagueness in the desired response. Illustrated examples were provided here of both a more and a less certain response (see also Fig. 2c, d). After this, participants read through three slides of task instructions, which provided basic information regarding each section of the survey (see Appendix 2). Participants were then asked to indicate whether they had read and fully understood the task information, and to raise any outstanding questions with the experimenter. Once they had done so, they proceeded to complete all three sections of the survey—questions are detailed in the Questionnaire items section. Total task duration was generally 30–40 minutes. Following completion of the main survey, participants were asked to complete a few more short questions, this time on paper, designed to elicit their subjective feedback about the interval-valued response format. Once these were completed, participants were informed that they had finished the study and were provided with their inconvenience allowance.
Results
A series of statistical analyses were conducted, relating to response data obtained within the different sections of the survey. These are detailed in the following subsections.
It is important to bear in mind that for the purposes of the analyses performed here, we extract only certain characteristics from the intervals (i.e., interval position—as the mean of the interval endpoints, and interval width—as the difference between the left and right endpoints)6,7.
Section 1: Reproducing presented intervals
The purpose of this section was to assess the capability of respondents to reproduce intervals explicitly provided to them, using the ellipse response mode. Visualisations of the interval-valued responses provided in relation to two of the items in this section are shown in Fig. 7; one shows the raw intervals provided by each respondent, and the second represents an aggregation of these responses, based on the Interval Agreement Approach (IAA—cf. Wagner et al., 2015). The IAA represents the degree of agreement across the group—i.e., overlap between all intervals—as a two-dimensional distribution. This may be interpreted as a fuzzy set (cf. Zadeh, 1965), with the agreement between intervals at each x-value determining the corresponding degree of membership.
For comparative analysis, we determined both Pearson correlation coefficients (r values) and mean square error values (MSE), which were based on a standardised response scale over [0,100] (to permit comparison with Section 2 results). This was done for both the midpoints and widths of the intervals drawn by each of the 40 respondents, over all eight questions, by comparison with those originally provided on the animal lifespan chart (Fig. 3). The resulting groups of 40 r values (representing each respondent) were then compared against zero using one-sample t tests (two-tailed), to examine the group-level relationship between stimulus and response intervals. As these data were consistently found to significantly deviate from normality, we used bootstrapping to ensure robustness (10,000 samples). We also report MSE values with bootstrapped 95% confidence intervals.
These tests revealed a significant and strong positive correlation between original and drawn interval midpoints, p = .003 (M = .97, 95% CIs .93 to 1.00); also, between original and drawn interval widths, p = .002 (M = .89, 95% CIs .78 to .97). MSE values—with scale range standardised to [0,100]—were as follows: for interval midpoints M = 55.29 (95% CIs 6.70 to 122.27), for interval widths M = 33.72 (95% CIs 8.99 to 65.65). These results indicate that the interval-valued responses provided in this section closely approximated the original (ground-truth) intervals upon which they were based—in terms of both position and width.
Section 2: Reporting disjunctive and conjunctive intervals
The purpose of this section was to examine whether participants were able to use the ellipse response mode to generate intervals representing observed degrees of both uncertainty and variability—and then to assess just how the intervals provided reflect these two factors. Two independent forms of analysis were applied to achieve these objectives. First, the objectively definable maximum and minimum bounds on the possible number of blue marbles in each row were determined for each question, as a benchmark (ground truth) for comparison. Then, as in the previous section, midpoints and widths of these intervals were compared against the corresponding responses provided by each of the 40 participants, to obtain both r and MSE values (based on a standardised response scale from [0,100]) for each respondent across all questions. As in Section 1, these data were found to significantly deviate from normality, so r values were compared against zero using bootstrapped one-sample t tests (two-tailed, 10,000 samples), and MSE values are reported with bootstrapped 95% confidence intervals Fig. 8.
These tests revealed a significant and strong positive correlation between benchmark and drawn interval midpoints, p < .001 (M = .90, 95% CIs .86 to .94), as well as between benchmark and drawn interval widths, p < .001 (M = .75, 95% CIs .68 to .82). In this case, MSE values were: M = 65.72 (95% CIs 37.60 to 99.52) for interval midpoints, and M = 954.01 (95% CIs 663.76 to 1267.01) for interval widths. These results indicate that the interval-valued responses provided in this section were again strongly associated with expected values—in this instance representing minimum and maximum bounds on the possible range of blue marbles, across all rows in any given set. However, in contrast with the previous section, where the task was simply to reproduce intervals, deviation from benchmark interval width values was substantially greater.
Mixed-effects modelling
This second stage of analysis applied linear mixed-effects modelling to examine how midpoints and widths of respondent-provided intervals were influenced by three salient factors that were visible to each respondent—two being putative sources of interval width, representing disjunctive and conjunctive aspects, respectively. These factors were the proportion of visible marbles that were blue, the proportion of marbles that were hidden (i.e., epistemic uncertainty—disjunctive), and the discrepancy in the visible number of blue marbles between rows (i.e., ontic range—conjunctive)—this was proportional to the maximum possible discrepancy, of seven. This analysis was conducted separately for the dependent variables of midpoint and width—each also represented as a proportion of the size of the entire response scale.
These three factors were entered into each model, alongside three two-way interaction terms. Two random intercepts were also incorporated, one for participants and another for questionnaire items. These allow the models to account for differing baseline positions and widths between subjects and between items. Models were estimated using the restricted maximum likelihood method, by means of the fitlme MATLAB function. The model formula to explain interval midpoint (\( {\gamma}_{i,j}^m \)) and interval width (\( {\gamma}_{i,j}^w \)) captured by \( {\gamma}_{i,j}^z \) was therefore as follows:
where z represents the outcome variable, which may be either midpoint (m) or width (w) for participant i on item j. b0 denotes the fixed intercept, while μi and μj denote respective random intercepts for participant and questionnaire item. The remaining b terms denote the coefficients of the six respective fixed effects: proportion of visible marbles that were blue xB, proportion of hidden marbles xH, the discrepancy in blue marbles visible between rows xD, plus two-way interaction terms. xi, j represents the observation for each participant and item, and εi, j the error term. Results are shown in Tables 1 and 2.
Regarding the midpoint of participant estimates, results indicate that this increased directly in line with the proportion of blue marbles (of those visible), and at a 1:2 ratio with the number of hidden marbles—consistent with participants assuming a 50% likelihood of each hidden marble being blue. The two-way interaction term between these two variables closely compensates for overestimation in cases where there is both a high proportion of blue visible marbles and a high proportion of hidden marbles (i.e., this effect regulates the effect of proportion of visible marbles that were blue according to the proportion of total marbles that were visible, such that the midpoint increases in line with the proportion of total marbles that were blue). Interestingly, a greater discrepancy between rows also led to higher average estimates, although this effect was ameliorated in cases with a high proportion of either hidden or blue visible marbles.
Regarding the widths of participant estimates, results indicate that these increased in direct proportion to the observed discrepancy between rows, but at only an approximately 4:5 ratio with the number of hidden marbles. Note that the upper 95% confidence interval of the latter effect is below 1—indicating that this tendency to underestimate uncertainty concerning hidden marbles was statistically significant. The two-way interaction term between these two variables indicates that the combined effect of hidden marbles and visible discrepancy is also less than the sum of its parts. This is consistent with two potential explanations: first, that respondents find it difficult to integrate conjunctive and disjunctive sources of interval width—i.e., the presence of between-row variability detracts from accurate uncertainty reporting, exacerbating the general tendency to underestimate this; second, that when more marbles are hidden from view, it is reasonable to give less weight to observed discrepancies, as these may yet ‘balance out’ when the hidden marbles are revealed—but as more marbles become visible it becomes increasingly likely that any visible discrepancy will be reflected in the final set. All effects relating to the proportion of blue visible marbles were found not to have any significant effect on response widths.
Section 3: Generating subjective intervals
The purpose of this section was to examine how respondents used intervals to describe and communicate their own subjectively perceived uncertainty about, or range inherent in, the appropriate response, arising from the phrasing or subject of a given question. For this section, a variety of statistical analyses were applied, as appropriate, to examine differences in responses between items within each of the different subsections. Note that ANOVA model residuals were consistently found to violate the normality assumption. Therefore, we also performed non-parametric robustness checks in each case, based upon Monte Carlo resampling (10,000 samples). Results were found to match the parametric test outcome in every instance, and are detailed in Appendix 3.
First, we assessed questions deliberately worded in a more or less specific manner—relating to subjective estimates of temperature in England—to establish whether intervals reflected (and could be used to identify) this level of specificity. One-way repeated-measures ANOVAs revealed significant effects of specifying time of year upon both interval midpoints, F(2.079,81.093) = 75.528, p < .001, η2p = .659, and interval widths, F(2.260,88.141) = 18.015, p < .001, η2p = .316. Specifically, the results showed significantly higher absolute temperature estimates for July, and lower for December, as well as significantly lower interval widths for both of these months, by contrast with two questions (asked before and after the month-specific questions) that did not specify a time of year.
Second, we examined whether intervals could discriminate a question with ambiguously interpretable phrasing from an unambiguous question, in this case relating to hypothetical increases in percentage recycling rates. Bootstrapped paired-samples t tests—used due to violations of normality—revealed significantly higher average interval midpoints when respondents were asked to ‘increase [10%] by 50%’ than when asked to ‘double’ this same initial rate, p = .016. However, despite mean interval width being approximately three times the size for the former question (.33 vs .11), this difference was not found to be statistically significant, p = .161. These results are consistent with a bimodal distribution for the percentage question, comprising groups of respondents who interpreted ‘increase by 50%’ differently—i.e., some interpreted it as a 50-percentage-point increase, leading to a higher average response midpoint than the ‘double’ question. The results do not however provide evidence that individual respondents commonly identified and reported the potential ambiguity in the question. These interpretations are supported by descriptive results showing the actual interval response distributions (see Fig. 9).
Third, we examined whether intervals could identify questions within which a word was not understood by the respondent—likely to be of particular concern when designing surveys aimed at non-native speakers—in this case items related to personality characteristics. Two separate 4 × 3 repeated-measures factorial ANOVAs were conducted, with dependent variables of interval midpoint and width, respectively. Independent variables were characteristic (four levels: talkative, aggressive, lazy, quiet) and word frequency (three levels: high, low, and pseudo-word—see Questionnaire items section for more detail)—note that the four pseudo-words were randomly allocated between characteristic conditions. For interval midpoints, significant main effects were found for both characteristic, F(2.386,93.056) = 5.250, p = .004, η2p = .119, and word frequency, F(1.684,65.686) = 7.352, p = .002, η2p = .159, as well as a significant two-way interaction term, F(3.466,135.180) = 8.900, p < .001, η2p = .186. For interval widths, there was a significant main effect of word frequency, F(1.556,60.680) = 59.922, p < .001, η2p = .606, but neither a significant main effect of characteristic, F < 1.0, nor a significant two-way interaction, F(4.332,168.941) = 1.166, p = .328. These findings reflect that respondents tended to identify with certain characteristics more than others. Also, on average, they were more likely to disagree with real words, but this was not the case for all characteristics—e.g., ‘How well does the [high-frequency] word talkative describe you?’ received more agreement than both its low-frequency counterpart ‘garrulous’ and control pseudo-word ‘brendacious’. Regarding interval widths, these were largest for pseudo-words, becoming progressively smaller for real but low-frequency and then high-frequency words—in a pattern highly consistent with expected levels of word comprehension (Fig. 10).
Fourth, we used double- versus single-barrelled questions to examine in more depth whether intervals could discriminate between questions phrased precisely or more ambiguously. The same approach was taken as previously: two separate 4 × 3 repeated-measures factorial ANOVAs were conducted, with dependent variables of interval midpoints and widths, respectively. Independent variables were the general topic of the question (four levels: reading books and watching television vs watching and playing sport vs drinking tea and coffee vs cooking and eating) and the specific subject of the question (three levels: e.g., cooking and eating vs cooking vs eating). For interval midpoints, this revealed a significant main effect of topic, F(2.166,84.474) = 15.483, p < .001, η2p = .284, but no significant main effect of subject, F < 1.0, and no significant two-way interaction, F(3.254,126.887) = 1.011, p = .395. By contrast, for interval widths, this revealed a significant main effect of subject, F(1.428,55.687) = 14.375, p < .001, η2p = .269, but no significant main effect of topic, F(2.411,94.022) = 2.257, p = .100, or a significant two-way interaction, F( < 1.0). These results reflect significant differences in overall liking between topics, but not between the specific subjects within each topic. Nonetheless, significantly broader intervals were provided for more ambiguous double-barrelled items than for their single-barrelled counterparts (Fig. 11).
Fifth, we examined judgements regarding the definition of stimuli falling within more or less ambiguous categories—in this case images of sets of eight vehicles, of which a potentially disputable number could be classified as ‘cars’ (see Fig. 5, both panels). Two further 2 × 2 repeated-measures ANOVAs were conducted, with respective dependent variables of interval midpoints and widths. In this case, independent variables were the specific image stimulus used (i.e., Fig. 5, panel a vs b), and the clarity of the category over which a judgement was made (i.e., cars vs vehicles). For interval midpoints, significant within-subject main effects were found for both image stimulus, F(1,39) = 99.194, p < .001, η2p = .718, and clarity of category, F(1,39) = 85.403, p < .001, η2p = .687, as well as a significant two-way interaction, F(1,39) = 99.545, p < .001, η2p = .719. These indicate that a greater number of stimuli were judged to be vehicles than cars in each image, and also that a greater number of stimuli were judged to be cars in the second image (Fig. 5, panel b), but an equal number were judged to be vehicles in each image. For interval widths, a significant main effect of category clarity was found, F(1,39) = 7.374, p = .010, η2p = .159, but neither a significant main effect of image stimulus, F(1,39) = 1.746, p = .194, nor a significant interaction term, F( < 1.0). These reflect that broader intervals were provided for the relatively ambiguous category ‘cars’, across both image conditions, than for the more clearly inclusive category ‘vehicles’.
Sixth, and finally, we examined perceptual judgements over more or less ambiguous image stimuli in another context—coloured marbles (see Fig. 6, all panels). Two one-way ANOVAs were conducted, with dependent variables of interval midpoint and width, respectively. For interval midpoints, a significant effect of stimulus was found, F(1.498,58.418) = 12.889, p < .001, η2p = .248. The highest interval midpoints were provided for the colour-gradient set of marbles (panel b: M = 54.74), followed by the solid colour set (panel a: M = 50.21), and finally the patterned set (panel c: M = 42.05). Post hoc paired-samples t tests were conducted to examine paired differences; these were bootstrapped (10,000 samples) to ensure robustness to violations of normality. These revealed both increments to be statistically significant: p = .011 and p = .005. For interval widths, a significant effect of stimulus was also found, F(1.490,58.126) = 14.196, p < .001, η2p = .267. The widest intervals were drawn for the patterned marble set (panel c: M = 20.56), followed by the colour-gradient set (panel b: M = 10.29), and the smallest intervals for the unambiguous, solid colours set (panel a: M = 1.90). Again, bootstrapped post hoc paired-samples t tests (10,000 samples) revealed each increment to be statistically significant: p = .014 and p = .010. This finding indicates that the width of participants’ response intervals reflected the degree of uncertainty induced by each stimulus, as hypothesised. Results, illustrated using the IAA method, are shown in Fig. 12.
Subjective feedback
After finishing the main questionnaire, participants completed four subjective feedback questions, designed to assess their perceptions of the interval-valued response format. These were obtained using a traditional 5-point ordinal response scale (ranging from 1—Strongly Disagree, to 5—Strongly Agree). Descriptive results are shown in Table 3. By comparison with the midpoint of the response scale (3.00), these indicate significant agreement with the ellipse response mode being easy to use, allowing effective communication of desired responses, and being liked overall, but significant disagreement with it being unnecessarily complex.
Discussion
The present study applied empirical methods to test the efficacy of a novel interval-valued survey response mode, in terms of capturing uncertainty and range in individual responses, arising from multiple sources. Capturing these is unfeasible using conventional discrete or point-valued response formats, such as Likert-type and visual analogue scales, without asking specific follow-up questions, at the cost of substantially increased questionnaire duration and complexity. By contrast, we hypothesised that the proposed interval-valued response mode—which can now be administered digitally with a high degree of efficiency—can achieve this effectively, within a single coherent response, by providing participants with an additional dimension to their answers. Specifically, this response format enables respondents to vary the width of an interval, in the form of an ellipse drawn along a continuous scale, according to how certain and specific they wish to indicate each answer to be.
The core aim of this study was to determine whether the proposed interval-valued response mode is capable of systematically capturing vagueness and uncertainties in individual responses. This was done by examining interval-valued responses to questions of a specially designed survey, within which respondent knowledge (e.g., information availability), question clarity, and inherent variability in the correct response were each experimentally manipulated. In operational terms, it was hypothesised that variance in respondents’ interval widths would reliably reflect induced between-item variance in each of these factors.
Results consistently indicated in the affirmative, across a variety of contexts. In the first section of the survey, respondents used interval widths to effectively replicate observed differences in ranges of expected lifespan between animal species—accurately representing both facets of the data presented to them in interval format. This demonstrates the practical utility of this response mode as a quick and coherent method enabling participants to provide answers that inherently comprise a range (i.e., conjunctive sets). It also suggests that the response mode enables individuals to respond reliably in this context, even though this was not explicitly tested (e.g., through a test–retest design), and will require future study.
In the second section of the survey, respondents’ intervals were found to consistently reflect their degree of uncertainty about a set of stimuli, in terms of information hidden from them, as well as inherent variability observed between cases within a given set. We note that the former here would result in a disjunctive set, representing the (lack of) knowledge of a true value, while the latter results in a conjunctive set describing an actual range of values (as no single discrete answer is correct). In turn, it is interesting to note that while each of these factors significantly contributed to response widths, results indicated that participants tended to substantially underestimate the former in their response intervals, mirroring a large body of evidence for overconfidence in probabilistic interval estimates (cf. Juslin et al., 1999; Klayman et al., 1999; Soll & Klayman, 2004; Speirs-Bridge et al., 2010; Tversky & Kahneman, 1974; Yaniv & Foster, 1995, 1997). By contrast, we found no such evidence for underestimation of conjunctive response ranges—which appeared well-calibrated. That is, while intervals provided did not tend to encompass all possible outcomes when a proportion of the set was hidden, they did closely represent the full range of observed outcomes within a set—at least when a high proportion of marbles were visible. This effect was evident in certain cases as an otherwise incongruous increase in mean interval widths corresponding to an increase in available information, as greater disparity was revealed within a set than may have been expected (see Fig. 8, sets 4 and 6). Future research might be able to leverage a conjunctive set framing as a pathway to reducing overconfidence.
It was also found that when there was a high amount of both hidden information and stimulus variability, this led to further underestimation of total response width; this may reflect participants’ difficulty in dealing with these two sources of interval width together (i.e., forcing them to effectively integrate disjunctive and conjunctive sets), exacerbating the existing tendency to underestimate epistemic uncertainty. Alternatively, it may reflect a deliberate interaction between the two factors, whereby lower information about the set of stimuli leads to lower confidence in the reliability of any variability that is currently observed—that is, participants may have assumed that observed discrepancies were more likely to even out once more hidden marbles were revealed than they were to increase.
On the whole, these findings demonstrate the capacity of the proposed response format to efficiently capture and quantify degrees of both disjunctive and conjunctive range in responses—although it is not designed to discriminate the two. This could prove valuable across a wide range of circumstances where assessments of real-world variables, solicited from experts or the broader population, are crucial to making predictions and informing subsequent policy, strategy, or investment decisions—e.g., political polling, supplier selection, marketing, environmental planning, cyber-security (i.e., vulnerability), and other risk or impact assessments. Likewise, in the context of social, behavioural, and psychological research, intervals may facilitate richer capture and understanding of participants’ preferences, attitudes, or choices, considering the relevant information available to them, and in combination with more subjective factors such as beliefs and their associated confidence or conviction. Intervals also provide the necessary information to guide potential follow-on qualitative research stages, such as to explore the underlying origins and drivers of uncertainties or ambiguities once they have been identified.
In the third section of the survey, respondents were found to systematically vary their interval sizes across a broader range of situations—each relating to their degree of subjective uncertainty, induced by the wording of the question or its associated stimulus. Participants drew significantly smaller intervals to represent temperature estimates for specific months than for non-time-specific questions. But while respondents were inconsistent in their interpretations of a deliberately ambiguous question (‘increase [10%] by 50%’), only a minority explicitly recognised and communicated the potential ambiguity in their response by providing a range from 15% to 60%. When presented with personality judgements concerning adjectives of varying degrees of obscurity, interval widths reflected word frequency in the manner expected. For high-frequency words, interval widths fluctuated between respondents and questions, but these remained relatively small compared with the other conditions, with no respondents circling the entire scale. For low-frequency words, mean widths were substantially greater, indicating uncertainty, with approximately half of responses covering the entire response scale, indicating that the appropriate response was unknown. For fabricated pseudo-words, this method of indicating complete uncertainty became the majority response—although it is worth noting that some respondents did select only the centre point in these cases. It is likely that these participants reverted to a habitualised response, as would be given on a discrete response scale. This may be a greater problem over longer questionnaires, for which response fatigue may increase this effect—but it can be expected that, as the interval-valued response mode becomes more widely used, familiarity with this format will overcome any inertia working against its appropriate use.
In addition to this, interval widths for double-barrelled questions were also found to be significantly broader than those for their more specific single-barrelled counterparts. Furthermore, questions requiring judgement over stimulus category membership found significantly broader intervals for a contextually ambiguous category term (car) than for a more clearly inclusive term (vehicle). Likewise, significantly larger intervals were provided for membership of the same category term (blue) when the stimuli over which the judgement was made were less clearly classifiable. These cases highlight the potential for interval-valued responses to identify questions which respondents struggle to understand—whether due to unclear stimuli or imprecise question wording. This capability could be of great value, for example in questionnaire pretesting, particularly in cases where a questionnaire will be taken by non-native speakers.
Participants’ subjective feedback was also collected and assessed. This was found to be generally positive on all four questions asked—consistent with findings in other contexts (Ellerby et al., 2020; Ellerby & Wagner, 2021). On average, respondents agreed both that the response format was easy to use and that it allowed them to effectively communicate their desired response, while they disagreed that it was unnecessarily complex. Overall liking received greater agreement than any of the more specific questions, which may be due to some respondents having been more uncertain in their responses to the subcomponents, as a result of their greater complexity. Ironically, capturing interval-valued feedback rather than using a traditional discrete response scale may have shed more light on this effect.
Summary and future work
To summarise, in the present paper we put forward and empirically examine a new ellipse-based questionnaire response format. This is designed to capture interval-valued data with a focus on efficiency and user experience, including through increasingly ubiquitous information technologies, minimising the time and effort requirements for both survey respondents and administrators. First, we present a case for the potential added value of this response mode, by contrast with conventional alternatives. Here, we identify three primary situations in which interval-valued responses might capture new and valuable information: where responses inherently comprise a range of values, where they are uncertain due to a lack of knowledge about the correct answer, and where they are ambiguous, or otherwise uncertain, due to imprecision or ill-definition in either the question or associated stimulus.
Following this, we document a validation study within which interval-valued responses are collected across a range of questionnaire items—designed to exemplify each of the three cases described above. Interval widths were found to systematically reflect experimentally induced differences between items, in relation to each of these factors. These findings provide a basis of empirical evidence for the efficacy and practical value of this new interval-valued response mode, but further research is needed to determine the extent to which these generalise—assessing the impact of individual, or sample, differences on the realisation of this value in practice; as well as exploring the effects of variations in interval-valued scale design upon outcomes including data fidelity, reliability, and participant workload.
Future work is also required to more thoroughly explore the broad range of potential benefits offered by this information-rich type of response—as well as to establish how to best retain and interpret this additional information, through development of novel methods of statistical analysis, which may build upon a broad body of existing work across academic disciplines including symbolic data analysis, interval arithmetic, interval ranking, fuzzy sets, and interval aggregation through fuzzy integrals (cf. Anderson et al., 2014; Billard, 2006; Billard & Diday, 2004; Ferson et al., 2007; Liu & Mendel, 2008; McCulloch et al., 2019; Miller et al., 2012; Nguyen et al., 2012; Wagner et al., 2013; Wagner et al., 2015; Wu et al., 2012). Another practical concern remains in demonstrating the efficacy of this approach across a broader variety of real-world contexts—expanding upon a growing body of research (cf. Ellerby et al., 2019a, 2020; Navarro et al., 2016; Wallace et al., 2016). Finally, it is important that future empirical research is conducted to directly compare this interval-valued response mode against existing methods of both lower (e.g., Likert, VAS) and higher complexity (e.g., multi-step interval elicitation, SHELF, FRS)—to establish and quantify the trade-offs between information capture and important practical concerns regarding response efficiency (i.e., workload, survey duration, perceived complexity, training requirements), which have a large influence on likelihood of broader adoption.
Notes
Uncertainty can be broken down into epistemic (reducible, subjective) and aleatoric (irreducible, objective), (cf. Brugnach et al., 2007; Couso & Dubois, 2014; Dewulf et al., 2005; Dubois & Prade, 2012; Fox & Ülkümen, 2011; Kwakkel et al., 2010; Walker et al., 2003; Zandvoort et al., 2018). The second dice example is of aleatoric uncertainty, as the correct response is yet to be determined and is therefore currently unknowable. For an example of epistemic uncertainty, relating instead to a lack of knowledge regarding a knowable answer, consider the question ‘I have just rolled two fair six-sided dice one time; what sum total did I roll?’ when the dice remain hidden from the respondent’s view. Here the answer is knowable but remains uncertain from the perspective of the respondent, until more information can be obtained (e.g., viewing the dice in question). For the purposes of the present study, we focus on the breadth of underpinning ‘drivers’ of interval width, which may be encountered when completing a conventional questionnaire, rather than arguing for a particular typology of uncertainties. It is thus convenient to note that both of these cases fall within the single category of disjunctive sets—i.e., the response interval reflects the inadequacy of knowledge or information necessary for the respondent to make a precise and certain judgement.
Here the term interval refers to the definition designated by Stevens (1946). This is semantically different from how this term is more commonly used throughout the rest of this paper, and indeed most of the sciences. Outside of this section an ‘interval’ will refer to the mathematical notion of a closed set of values on a continuous real line, which is defined by two endpoints and possesses the property of width, which is zero only in the case that both endpoints are identical.
Note that as these stimuli were shown on an LED computer screen, a perceiver with high enough image resolution could argue that none of the marbles truly contained any yellow—as this percept is created through the combination at a specific ratio of green and red light output.
This software is available as an open-source package which enables interested parties to run their own instance of the platform to conduct surveys. Beyond this, DECSYS is designed to enable hosting for example on standard cloud infrastructure such as Microsoft Azure, providing a similar user experience to commercial tools such as SurveyMonkey® or Qualtrics®. Here, the software provides the functionality of multiple survey administrator accounts—allowing, for example, individual researchers to develop and administer their own surveys and access the resulting data. All data here are accessible exclusively by the specific survey administrator. In a geographical sense, the physical location of the data stored is determined by whether DECSYS is being used as a service hosted on a personal server (e.g., within a university) or on cloud infrastructure. In the case of the present study, all data were stored locally.
This enables analysis at a reasonable level of complexity, using established methods. However, in general, it is complete intervals which capture individual responses. With the application of a rapidly advancing body of work around advanced techniques for handling interval-valued data, we expect that there is significant potential in the development of new interval-based quantitative analysis methods, offering an efficient approach to rapidly collate, compare, and assess rich quantitative information from survey respondents.
The interval midpoint is used to represent interval location in our analysis, and we do not assume that the midpoint of each interval is necessarily a ‘best estimate’. In cases where the response interval represents a disjunctive set, a respondent’s ‘best estimate’ may or may not be located at the centre of the interval provided.
References
Ahearn, E.P. (1997). The use of visual analog scales in mood disorders: A critical review. Journal of Psychiatric Research, 5, 569-579.
Aitken R.C. (1969). Measurement of feelings using visual analogue scales. Proceedings of the Royal Society of Medicine, 62:989–993.
Aldrich, J., Niemi, R., Rabinowitz, G., & Rohde, D. (1982). The measurement of public opinion about public policy: A report on some new issue question formats. American Journal of Political Science, 26(2), 391–414. https://doi.org/10.2307/2111047.
Anderson, D. T., Havens, T. C., Wagner, C., Keller, J. M., Anderson, M. F., & Wescott, D. J. (2014). Extension of the fuzzy integral for general fuzzy set-valued information. IEEE Transactions on Fuzzy Systems, 22(6), 1625-1639.
Baranski, J. V., & Petrusic, W. M. (1998). Probing the locus of confidence judgments: Experiments on the time to determine confidence. Journal of Experimental Psychology: Human Perception and Performance, 24, 929 –945
Behrend, T. S., Sharek, D. J., Meade, A. W., & Wiebe, E. N. (2011). The viability of crowdsourcing for survey research. Behavior research methods, 43(3), 800.
Billard, L. (2006). Symbolic data analysis: what is it?. In Rizzi, A., Vichi, M. (eds.), Compstat 2006 - Proceedings in Computational Statistics. Physica-Verlag HD. https://doi.org/10.1007/978-3-7908-1709-6_20.
Billard, L., & Diday, E. (2004). Symbolic data analysis: Definitions and examples. https://basepub.dauphine.fr/handle/123456789/6895.
Birch, J. (2012). Worldwide prevalence of red-green color deficiency. JOSA A, 29(3), 313-320.
Bishop, P. A., Herron R.L. (2015). Use and misuse of the Likert item responses and other ordinal measures. International Journal of Exercise Science, 8:297–302.
Blaikie N. (2003). Analysing Quantitative Data. Sage Publications.
Boone, H. N., & Boone, D. A. (2012). Analyzing Likert data. Journal of extension, 50(2), 1–5.
Brooke, J. (1996). SUS-A quick and dirty usability scale. Usability evaluation in industry, 189(194), 4-7.
Brugnach, M., Dewulf, A., Pahl-Wostl, C., & Taillieu, T. (2007). Towards a relational concept of uncertainty: incorporating the human dimension. 1-25. In: International Conference in Adaptive and Integrated Water Management, Coping with complexity and uncertainty. http://www.newater.uos.de/caiwa/data/papers%20session/B4/2007-10-22%20Brugnach%20et%20al.2007-Caiwa.pdf.
Bruine de Bruin, W., Manski, C. F., Topa, G., & Van Der Klaauw, W. (2011). Measuring consumer uncertainty about future inflation. Journal of Applied Econometrics, 26(3), 454–478.
Budescu, D. V., & Wallsten, T. S. (1985). Consistency in interpretation of probabilistic phrases. Organizational Behavior and Human Decision Processes, 36(3), 391-405.
Budescu, D. V., & Wallsten, T. S. (1995). Processing linguistic probabilities: General principles and empirical evidence. In: Psychology of Learning and Motivation (Vol. 32, pp. 275-318). Academic Press.
Budescu, D. V., Broomell, S., & Por, H. H. (2009). Improving communication of uncertainty in the reports of the Intergovernmental Panel on Climate Change. Psychological science, 20(3), 299-308.
Burden, B. C. (1997). Deterministic and probabilistic voting models. American Journal of Political Science, 41, 1150-1169.
Carifio J, Perla R.J. (2007). Ten Common Misunderstandings, Misconceptions, Persistent Myths and urban legends about Likert scales and Likert response Formats and their Antidotes. Journal of Social Sciences, 3(3):106.
Carifio J, Perla R.J. (2008). Resolving the 50-year Debate Around using and Misusing Likert Scales. Medical Education, 42(12):1150–1152.
Cheesman, J., & Merikle, P. M. (1984). Priming with and without awareness. Perception & Psychophysics, 36, 387-395.
Cloud, M. J, Moore, R. E, Kearfott, R.B. (2009). Introduction to Interval Analysis. Philadelphia: Society for Industrial and Applied Mathematics (SIAM). ISBN 0-89871-669-1.
Cohen, J. (1988). Statistical Power Analysis for the Behavioral Sciences (2nd ed.). Routledge.
Converse, J. M., & Presser, S. (1986). Survey questions: Handcrafting the standardized questionnaire (Vol. 63). Sage.
Cooke, R. (1991). Experts in uncertainty: opinion and subjective probability in science. Oxford University Press on Demand.
Coombs, C. H., & Coombs, L. C. (1976). “Don't Know” Item Ambiguity or Respondent Uncertainty? Public Opinion Quarterly, 40(4), 497-514.
Couper, M., Tourangeau, R., Conrad, F., & Singer, E. (2006). Evaluating the effectiveness of visual analog scales: A web experiment. Social Science Computer Review, 24, 227-245.
Couso, I., & Dubois, D. (2014). Statistical reasoning with set-valued information: Ontic vs. epistemic views. International Journal of Approximate Reasoning, 55(7), 1502-1518.
Cubitt, R. P., Navarro-Martinez, D., & Starmer, C. (2015). On preference imprecision. Journal of Risk and Uncertainty, 50(1), 1-34.
Dawes J. (2008). Do Data Characteristics Change According to the Number of Scale Points Used? An experiment Using 5 point, 7 point and 10 Point Scales. International Journal of Market Research, 50, 61-104.
Dewulf, A., Craps, M., Bouwen, R., Taillieu, T., & Pahl-Wostl, C. (2005). Integrated management of natural resources: dealing with ambiguous issues, multiple actors and diverging frames. Water science and technology, 52(6), 115-124.
Dienes, Z., Altmann, G. T. M., Kwan, L., & Goode, A. (1995). Un-conscious knowledge of artificial grammars is applied strategically. Journal of Experimental Psychology: Learning, Memory, & Cognition, 21, 1322-1338.
Dubois, D., & Prade, H. (2012). Gradualness, uncertainty and bipolarity: making sense of fuzzy sets. Fuzzy sets and Systems, 192, 3-24.
Ellerby, Z. W., & Tunney, R. J. (2017). The Effects of Heuristics and Apophenia on Probabilistic Choice. Advances in Cognitive Psychology, 13(4), 280.
Ellerby, Z., & Tunney, R. (2019). Probability Matching on a Simple Simulated Foraging Task: The Effects of Reward Persistence and Accumulation on Choice Behaviour. Advances in Cognitive Psychology, 15(2), pp. 37-52.
Ellerby, Z., & Wagner, C. (2021). Do People Prefer to Give Interval-Valued or Point Estimates and Why? In 2021 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE). IEEE.
Ellerby, Z., McCulloch, J., Wilson, M., & Wagner, C. (2019a). Exploring how component factors and their uncertainty affect judgements of risk in cyber-security. In S. Nadjm-Tehrani, (Ed.), Critical Information Infrastructures Security. CRITIS 2019. Lecture Notes in Computer Science (Vol. 11777). Cham: Springer. https://doi.org/10.1007/978-3-030-37670-3_3.
Ellerby, Z., McCulloch, J., Young, J., & Wagner, C. (2019b). DECSYS–Discrete and Ellipse-based response Capture SYStem. In 2019 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE). IEEE.
Ellerby, Z., Miles, O., McCulloch, J., & Wagner, C. (2020). Insights from interval-valued ratings of consumer products—a DECSYS appraisal. In 2020 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE). IEEE.
Faul, F., Erdfelder, E., Lang, A.-G., & Buchner, A. (2007). G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods, 39, 175-191.
Ferson, S., Kreinovich, V., Hajagos, J., Oberkampf, W., & Ginzburg, L. (2007). Experimental uncertainty estimation and statistics for data having interval uncertainty. Sandia National Laboratories, Report SAND2007-0939. https://doi.org/10.2172/910198.
Fowler, F. J. (1995). Improving survey questions: Design and evaluation (Vol. 38). Sage.
Fox, C. R., & Ülkümen, G. (2011). Distinguishing two dimensions of uncertainty. In Brun, W., Keren, G., Kirkebøen, G., & Montogomery, H. (eds) Perspectives on thinking, judging, and decision making (pp. 21–35). Oslo: Universitetsforlaget.
Freyd, M. (1923). The graphic rating scale. Journal of Educational Psychology, 14, 83-102.
Gnambs, T., & Kaspar, K. (2015). Disclosure of sensitive behaviors across self-administered survey modes: a meta-analysis. Behavior research methods, 47(4), 1237-1259.
Gosling, J. P. (2018). SHELF: The Sheffield Elicitation Framework. In Dias, L., Morton, A., & Quigley, J. (eds), Elicitation. International Series in Operations Research & Management Science (Vol. 261). Cham: Springer. https://doi.org/10.1007/978-3-319-65052-4_4.
Graesser, A. C., Wiemer-Hastings, K., Kreuz, R., Wiemer-Hastings, P., & Marquis, K. (2000). QUAID: A questionnaire evaluation aid for survey methodologists. Behavior Research Methods, Instruments, & Computers, 32(2), 254-262.
Groves, R. M., Fowler Jr, F. J., Couper, M. P., Lepkowski, J. M., Singer, E., & Tourangeau, R. (2011). Survey methodology (Vol. 561). John Wiley & Sons.
Guyatt, G. H., Townsend, M., Berman, L. B., & Keller, J. L. (1987). A comparison of Likert and visual analogue scales for measuring change in function. Journal of chronic diseases, 40(12), 1129-1133.
Harris, A. J., Por, H. H., & Broomell, S. B. (2017). Anchoring climate change communications. Climatic change, 140(3-4), 387-398.
Havens, T.C., Wagner, C., & Anderson, D.T. (2017). Efficient modeling and representation of agreement in interval-valued data. In 2017 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE). IEEE.
Hemming, V., Burgman, M. A., Hanea, A. M., McBride, M. F., & Wintle, B. C. (2018). A practical guide to structured expert elicitation using the IDEA protocol. Methods in Ecology and Evolution, 9(1), 169-180.
Hesketh, B., McLachlan, K., & Gardner, D. (1992). Work adjustment theory: An empirical test using a fuzzy rating scale. Journal of Vocational Behavior, 40(3), 318-337.
Hesketh, T., Pryor, R., & Hesketh, B. (1988). An application of a computerized fuzzy graphic rating scale to the psychological measurement of individual differences. International Journal of Man-Machine Studies, 29(1), 21-35.
Jamieson S. (2004). Likert scales: how to (ab)use them. Medical Education, 38: 1212–1218.
Juslin, P., Wennerholm, P., & Olsson, H. (1999). Format dependence in subjective probability calibration. Journal of Experimental Psychology: Learning, Memory, and Cognition, 25(4), 1038.
Kabir, S., Wagner, C., & Ellerby, Z. (2021). Towards Handling Uncertainty-at-Source in AI--A Review and Next Steps for Interval Regression. arXiv (preprint). https://arxiv.org/abs/2104.07245.
Kahneman, D., Slovic, P., & Tversky, A. (eds.). (1982). Judgment under uncertainty: Heuristics and biases. Cambridge University Press.
Karelitz, T. M., & Budescu, D. V. (2004). You say" probable" and I say" likely": improving interpersonal communication with verbal probability phrases. Journal of Experimental Psychology: Applied, 10(1), 25.
Klayman, J., Soll, J. B., Gonzalez-Vallejo, C., & Barlas, S. (1999). Overconfidence: It depends on how, what, and whom you ask. Organizational behavior and human decision processes, 79(3), 216-247.
Klopfer, F. J., & Madden, T. M. (1980). The middlemost choice on attitude items: Ambivalence, neutrality, or uncertainty?. Personality and Social Psychology Bulletin, 6(1), 97-101.
Knapp T.R. (1990). Treating ordinal scales as interval scales: an attempt to resolve the controversy. Nursing Research, 39(2):121–123.
Krantz, J. H., & Reips, U. D. (2017). The state of web-based research: A survey and call for inclusion in curricula. Behavior research methods, 49(5), 1621-1629.
Krosnick, J. A. (1991). Response strategies for coping with the cognitive demands of attitude measures in surveys. Applied cognitive psychology, 5(3), 213–236.
Krosnick, J. A., & Fabrigar, L. R. (1997). Designing Rating Scales for Effective Measurement in Surveys. In Lyberg, L., Biemer, P., Collins, M., De Leeuw, E., Dippo, C., Schwarz, N., & Trewin, D. (eds), Survey Measurement and Process Quality (pp. 141–164). https://doi.org/10.1002/9781118490013.ch6.
Kuhlmann, T., Dantlgraber, M., & Reips, U. D. (2017). Investigating measurement equivalence of visual analogue scales and Likert-type scales in Internet-based personality questionnaires. Behavior research methods, 49(6), 2173-2181.
Kuzon W.M., Urbanchek M.G., McCabe S. (1996). The Seven Deadly Sins of Statistical Analysis. Annals of Plastic Surgery. 37(3):265–272.
Kwakkel, J. H., Walker, W. E., & Marchau, V. A. (2010). Classifying and communicating uncertainties in model-based policy analysis. International journal of technology, policy and management, 10(4), 299-315.
Likert R. (1932). A Technique for the Measurement of Attitudes. Archives of Psychology. 140: 1–55.
Liu, F., & Mendel, J. M. (2008). Encoding words into interval type-2 fuzzy sets using an interval approach. IEEE Transactions on Fuzzy Systems, 16(6), 1503-1521.
Lubiano, M. A., de Sáa, S. D. L. R., Montenegro, M., Sinova, B., & Gil, M. Á. (2016). Descriptive analysis of responses to items in questionnaires. Why not using a fuzzy rating scale?. Information Sciences, 360, 131-148.
Maydeu-Olivares, A., Kramp, U., García-Forero, C., Gallardo-Pujol, D., & Coffman, D. (2009). The effect of varying the number of response alternatives in rating scales: Experimental evidence from intra-individual effects. Behavior Research Methods, 41(2), 295-308.
McCulloch, J., Ellerby, Z., & Wagner, C. (2019). On Comparing and Selecting Approaches to Model Interval-Valued Data as Fuzzy Sets. In 2019 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE). IEEE.
McCulloch, J., Ellerby, Z., & Wagner, C. (2020). Choosing Sample Sizes for Statistical Measures on Interval-Valued Data. In 2020 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE). IEEE.
Meade, A. W., & Craig, S. B. (2012). Identifying careless responses in survey data. Psychological methods, 17(3), 437.
Miller, S., Wagner, C., Garibaldi, J. M., & Appleby, S. (2012). Constructing general type-2 fuzzy sets from interval-valued data. In 2012 IEEE International Conference on Fuzzy Systems. IEEE.
Moore, R.E. (1966). Interval Analysis. Prentice-Hall. ISBN 0-13-476853-1.
Morgan, M. G. (2014). Use (and abuse) of expert elicitation in support of decision making for public policy. Proceedings of the National academy of Sciences, 111(20), 7176-7184.
Morris, D. E., Oakley, J. E., & Crowe, J. A. (2014). A web-based tool for eliciting probability distributions from experts. Environmental Modelling & Software, 52, 1-4.
Navarro, J., Wagner, C., Aickelin, U., Green, L., & Ashford, R. (2016, July). Exploring differences in interpretation of words essential in medical expert-patient communication. In: 2016 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE) (pp. 2157-2164). IEEE.
Nguyen, H. T., Kreinovich, V., Wu, B., & Xiang, G. (2012). Computing statistics under interval and fuzzy uncertainty. Springer Verlag.
Norman, G. (2010). Likert scales, levels of measurement and the “laws” of statistics. Advances in Health Science Education. https://doi.org/10.1007/s10459-010-9222-y
O’Hagan, A. (2019). Expert knowledge elicitation: subjective but scientific. The American Statistician, 73(sup1), 69-81.
Payne, S. L. B. (2014). The Art of Asking Questions: Studies in Public Opinion, 3 (Vol. 451). Princeton University Press. (Original work published 1951).
Peirce, C. S., & Jastrow, J. (1884). On small differences of sensation. Memoirs of the National Academy of Sciences, 3, 75-83.
Petty, R. E., Briñol, P., & Tormala, Z. L. (2002). Thought confidence as a determinant of persuasion: the self-validation hypothesis. Journal of personality and social psychology, 82(5), 722.
Pleskac, T. J., & Busemeyer, J. R. (2010). Two-stage dynamic signal detection: a theory of choice, decision time, and confidence. Psychological review, 117(3), 864.
Quirós, P., Alonso, J. M., & Pancho, D. P. (2016). Descriptive and comparative analysis of human perceptions expressed through fuzzy rating scale-based questionnaires. International Journal of Computational Intelligence Systems, 9(3), 450-467.
Reips, U.-D., Funke, F. (2008). Interval level measurement with visual analogue scales in Internet-based research: VAS Generator. Behavior Research Methods, 40, 699–704.
Rucker, D. D., Tormala, Z. L., Petty, R. E., & Briñol, P. (2014). Consumer conviction and commitment: An appraisal-based framework for attitude certainty. Journal of Consumer Psychology, 24(1), 119-136.
Saris, W. E., & Gallhofer, I. N. (2014). Design, evaluation, and analysis of questionnaires for survey research. John Wiley & Sons.
Schmidt, W. C. (1997). World-Wide Web survey research: Benefits, potential problems, and solutions. Behavior research methods, instruments, & computers, 29(2), 274-279.
Soll, J. B., & Klayman, J. (2004). Overconfidence in interval estimates. Journal of Experimental Psychology: Learning, Memory, and Cognition, 30(2), 299.
Speirs-Bridge, A., Fidler, F., McBride, M., Flander, L., Cumming, G., & Burgman, M. (2010). Reducing overconfidence in the interval judgments of experts. Risk Analysis: An International Journal, 30(3), 512-523.
Sporer, S. L., Penrod, S., Read, D., & Cutler, B. (1995). Choosing, confidence, and accuracy: A meta-analysis of the confidence-accuracy relation in eyewitness identification studies. Psychological Bulletin, 118(3), 315.
Stevens S.S. (1946). On the Theory of Scales of Measurement. Science, 103 (2684): 677–680.
Teigen, K. H., & Jorgensen, M. (2005). When 90% confidence intervals are 50% certain: On the credibility of credible intervals. Applied Cognitive Psychology: The Official Journal of the Society for Applied Research in Memory and Cognition, 19(4), 455-475.
Thurstone, L. L., & Chave, E. J. (1929). The measurement of attitude: A psychological method and some experiments with a scale for measuring attitude toward the church. University of Chicago Press.
Tunney, R. J., & Shanks, D. R. (2003). Subjective measures of awareness and implicit cognition. Memory & Cognition, 31, 1060-1071
Tversky, A., & Kahneman, D. (1974). Judgment under uncertainty: Heuristics and biases. Science, 185(4157), 1124-1131.
Van Laerhoven, H., Van Der Zaag-Loonen, H. J., & Derkx, B. H. (2004). A comparison of Likert scale and visual analogue scales as response options in children's questionnaires. Acta Paediatrica, 93(6), 830-835.
Vannette, D. L., & Krosnick, J. A. (2014). A comparison of survey satisficing and mindlessness. The Wiley Blackwell handbook of mindfulness, 1, 312.
Wagner, C., Miller, S., & Garibaldi, J. M. (2013). Similarity based applications for data-driven concept and word models based on type-1 and type-2 fuzzy sets. In 2013 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE). IEEE.
Wagner, C., Miller, S., Garibaldi, J. M., Anderson, D. T., & Havens, T. C. (2015). From interval-valued data to general type-2 fuzzy sets. IEEE Transactions on Fuzzy Systems, 23(2), 248-269.
Walker, W. E., Harremoës, P., Rotmans, J., Van Der Sluijs, J. P., Van Asselt, M. B., Janssen, P., & Krayer von Krauss, M. P. (2003). Defining uncertainty: a conceptual basis for uncertainty management in model-based decision support. Integrated assessment, 4(1), 5-17.
Wallace, K. J., Wagner, C., & Smith, M. J. (2016). Eliciting human values for conservation planning and decisions: a global issue. Journal of environmental management, 170, 160-168.
Winman, A., Hansson, P., & Juslin, P. (2004). Subjective probability intervals: how to reduce overconfidence by interval evaluation. Journal of Experimental Psychology: Learning, Memory, and Cognition, 30(6), 1167.
Wright, W. D. (1952). The characteristics of tritanopia. JOSA, 42(8), 509-521.
Wu, D., Mendel, J. M., & Coupland, S. (2012). Enhanced interval approach for encoding words into interval type-2 fuzzy sets and its convergence analysis. IEEE Transactions on Fuzzy Systems, 20(3), 499-513.
Yaniv, I., & Foster, D. P. (1995). Graininess of judgment under uncertainty: An accuracy-informativeness trade-off. Journal of Experimental Psychology: General, 124(4), 424.
Yaniv, I., & Foster, D. P. (1997). Precision and accuracy of judgmental estimation. Journal of behavioral decision making, 10(1), 21-32.
Zadeh L.A. (1965). Fuzzy sets. Information and Control, 8, 338-353.
Zandvoort, M., Van der Vlist, M. J., Klijn, F., & Van den Brink, A. (2018). Navigating amid uncertainty in spatial planning. Planning Theory, 17(1), 96-116.
Acknowledgement
This research was partially supported by UK Research and Innovation, Engineering and Physical Sciences Research Council [research grants EP/P011918/1 & EP/M02315X/1].
Open Practices Statement
The DECSYS software used to implement the survey documented in this paper is accessible via .https://www.lucidresearch.org/decsys.html. The software is hosted on GitHub to invite community engagement and contributions under the open-source GNU Affero licence. Full code and documentation are stored at the following repository https://github.com/decsys/decsys.
Interval-valued data and commented code for extraction and creation of IAA charts and interval plots—as shown in Figs. 7, 9 & 12—are stored at https://osf.io/pb4ne/ to provide a working example of this process. Further data and study materials are also available via the same link, including original survey stimuli. The experiment was not preregistered.
Author information
Authors and Affiliations
Corresponding author
Additional information
Open Practices Statement
The DECSYS software used to implement the survey documented in this paper is accessible via https://www.lucidresearch.org/decsys.html The software is hosted on GitHub to invite community engagement and contributions under the open-source GNU Affero licence. Full code and documentation are stored at the following repository https://github.com/decsys/decsys
Interval-valued data and commented code for extraction and creation of IAA charts and interval plots—as shown in Figs. 7, 9, and 12—are stored at https://osf.io/pb4ne/ to provide a working example of this process. Further data and study materials are also available via the same link, including original survey stimuli. The experiment was not preregistered.
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix 1: Instructions for the ellipse response mode
Thank you for agreeing to take part in this experiment.
This task is designed to assess how well people are able to respond using a newly developed survey response-format. Here, responses are made by drawing ellipses on a continuous scale, instead of picking a single response option. How to use this type of response-format is explained below, with examples. Please read this information carefully and ask the experimenter if you have any questions.
Appendix 2: Survey instructions (text content)
There will be three sections to this survey, each with slightly different tasks. Examples of how you can respond are provided for each.
Section 1: Here you will be provided with a chart, showing the life expectancies of different animals. In a series of questions, you will be asked to correctly identify the life expectancy of an animal from this chart, and to communicate this as accurately as you can, through drawing an ellipse on the scale provided.
Section 2: Here, you will be shown a series of sets of marbles. Each set will be made up of 5 rows of 7 marbles each, but some may be hidden. All marbles will be either blue or yellow, and you will be asked to estimate the number of blue marbles that are in each row. You will only make one estimate per set of marbles (5 rows). If any marbles are hidden, then you should make your estimate including these hidden marbles. You can illustrate your degree of uncertainty using the width of the ellipse.
Section 3: Here, you will be asked a series of more varied questions. Respond as you feel appropriate, applying the same approach as in the previous sections.
Please indicate when you have read and fully understood this information sheet, and you are ready to proceed. Ask the experimenter if you have any further questions.
Appendix 3: Monte Carlo ANOVA results (robustness checks)
Time of year, one-way (4-level) ANOVAs
Midpoints p < .001
Widths p < .001
Personality words, 4×3 ANOVAs
Midpoints: Characteristic p = .002, Word Freq. p < .001, Int. p < .001
Widths: Characteristic p = .638, Word Freq. p < .001, Int. p = .318
Double-barrelled questions, 4×3 ANOVAs
Midpoints: Topic p < .001, Subject p = .398, Int. p = .414
Widths: Topic p = .082, Subject p < .001, Int. p = .632
Vehicle questions, 2×2 ANOVAs
Midpoints: Stimulus p < .001, Category Clarity p < .001, Int. p < .001
Widths: Stimulus p = .190, Category Clarity p = .004, Int. p = .359
Marble questions, one-way (3-level) ANOVAs
Midpoints p < .001
Widths p < .001
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ellerby, Z., Wagner, C. & Broomell, S.B. Capturing richer information: On establishing the validity of an interval-valued survey response mode. Behav Res 54, 1240–1262 (2022). https://doi.org/10.3758/s13428-021-01635-0
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.3758/s13428-021-01635-0