Service user involvement in instrument development is increasingly recognised as important, but is often not done and seldom reported. This has adverse implications for the content validity of a measure. The aim of this paper is to identify the types of items that service users felt were important to be included or excluded from a new Recovering Quality of Life measure for people with mental health difficulties.
Potential items were presented to service users in face-to-face structured individual interviews and focus groups. The items were primarily taken or adapted from current measures and covered themes identified from earlier qualitative work as being important to quality of life. Content and thematic analysis was undertaken to identify the types of items which were either important or unacceptable to service users.
We identified five key themes of the types of items that service users found acceptable or unacceptable; the items should be relevant and meaningful, unambiguous, easy to answer particularly when distressed, do not cause further upset, and be non-judgemental. Importantly, this was from the perspective of the service user.
This research has underlined the importance of service users’ views on the acceptability and validity of items for use in developing a new measure. Whether or not service users favoured an item was associated with their ability or intention to respond accurately and honestly to the item which will impact on the validity and sensitivity of the measure.
There has been an increasing commitment to Patient Reported Outcome Measures (PROMs) designed to measure day-to-day health improvements that matter most to service users [1, 2]. The methodological quality of PROMs has been determined by the extent to which they meet the recognised properties of measurement of reliability and validity [3,4,5]. A key property of validity is the extent to which a measure captures what it is intended to measure. In the absence of a gold standard measure, researchers have often used indirect methods for testing validity such as factor analysis, Rasch, or Item Response Theory (IRT). These are important in instrument development, but are insufficient on their own to fully establish the validity of an outcome measure . It is also important to measure content validity; the extent to which the set of items comprehensively covers the different components of health to be measured  and face validity; whether the items of each domain are sensible, appropriate, and relevant to the people who use the measure on a day-to-day basis .
The content and face validity of many outcome measures currently in use are based on the judgements of researchers and health care professionals, with limited input from service users [8, 9]. However, what may be regarded as a good outcome by a clinician or researcher may differ from what is regarded as important to service users  and only service users can determine whether the measure captures these outcomes favourably . Active input from service users in all the development stages of a measure can improve the acceptability, relevance, and the quality of the measure and related research [9, 11, 12].
In recent years, there has been a change in mental health policy from an emphasis on services that focus only on symptom reduction, towards those that take a more holistic and positive approach of service user-defined recovery and quality of life [13,14,15]. This growing movement towards positive mental health has given rise internationally to recovery-oriented mental health services [16,17,18,19,20]. It is of particular importance that PROMs used in these services reflect the key areas that service users consider relevant to recovery, rather than the traditional focus on symptom reduction . Aspects of life that are considered important to ‘recovery’ have been shown to be consistent with those important to ‘quality of life’ [22, 23]. The themes service users consider important to quality of life: well-being, relationships and a sense of belonging, activity, self-perception, autonomy, hope, and physical health [24,25,26] are similar to those regarded as important to recovery in mental health: connectedness, hope, identity, meaning, and empowerment (CHIME) .
The new measure Recovering Quality of Life (ReQoL) in mental health was developed in four stages: (1) generation and subsequent shortlisting of candidate items; (2) testing face and content validity of shortlisted items; (3) psychometric evaluation; and (4) selection if the final items which involved combining the qualitative and quantitative evidence from stages 2 and 3. Importantly, service user opinion and input was utilised at all stages with a panel of six expert service users being involved in the selection of the shortlisted items from the pool of candidate items through to the decisions surrounding the inclusion of the final items. Psychometric testing involved over 4000 service users completing either a 60 or 40 item version of the measure before the selection of the final items. For details on all the stages involved in the development of the measure see Keetharuth et al. .
In this paper, we specifically report on the second of four stages which involved seeking the opinions of service users on a pool of items to help inform the selection of items which would go forward for psychometric testing. This builds on a systematic review of qualitative research and primary qualitative research involving service user interviews which identified the seven themes service users considered important to quality of life referred to above [24,25,26]. A pool of potential items (n = 1597) which best represented these seven domains was generated from current quality of life and recovery instruments and from service user interviews. The item set contained both positively (e.g. I felt happy) and negatively (e.g. I felt sad) worded items. These were subjected to initial sifting using a set of criteria adapted from those originally proposed by Streiner and Norman . After consideration by clinicians, researchers, and service users who were members of the research study’s Scientific, Advisory, Stakeholder and Expert User Groups, the item set was reduced to 88 potential items for the new ‘Recovering Quality of Life’ measure . The aim of this paper is to identify the key themes of face validity which should be considered when developing a quality of life instrument, using the ReQoL as an example. We also report on service users’ views on acceptability and validity of an item set.
A qualitative study using face-to-face structured interviews and focus groups with service users was undertaken.
There was a requirement that ReQoL be suitable for mental health service users over the age of 16. Therefore, adults (aged 19–79) and young adults (aged 16–18 years) from National Health Service (NHS) mental health services and a local charity were invited to participate. In order to obtain views from across the spectrum of mental health service users, broad inclusion criteria were applied; the only exclusion concerned people experiencing acute episodes of their mental health condition, those not well enough to take part, and those who could not speak English or give informed consent. This allowed for maximum variation of mental health problems, severity of problems, and current service contact.
Adult participants were recruited from four UK NHS Trusts providing mental health services and a UK charity in the north of the country. Two trusts were located in the south of England and two in the north. Recruitment of young adults (aged 16–18) took place from two further NHS trusts based in the Midlands and the North of England. Recruitment was undertaken on our behalf by healthcare staff and clinical studies officers within the individual trusts. Recruitment procedures followed what was usual and practical for each individual Trust. This variously involved healthcare staff approaching service users on inpatient wards, when attending therapy sessions, those who had previously agreed to be approached for research purposes, attendees at a Recovery College and at a Rehabilitation and Recovery Centre, and members of established PPI (Public and Patient Involvement) groups. At the time of the interview, demographic, care service, and diagnostic information was taken. In order to obtain a representative and diverse sample, this information was used to identify recruitment gaps and recruitment personnel were subsequently informed and asked to target these groups.
A total of 59 adult service users took part: 40 participated in individual interviews; 11 attended two focus groups of seven and four participants, respectively; and four interviews took place with two participants together. A total of 17 young adults participated: 15 participated in individual interviews and two participants chose to be interviewed together. Interviews lasted between 20 min and 1 h 40 min. For information on the sample demographics please refer to Table 1.
Interviews and focus groups were conducted between September 2014 and June 2015. Interviews and focus groups were undertaken at NHS sites familiar to the participants apart from one which was on university premises. Participants were provided with an information sheet prior to the interviews. Written consent was obtained at the time of the interview prior to any data collection. Participants were asked to complete a short demographics form indicating their gender, age, ethnicity, employment, education level, mental health diagnosis, and their own perception of their mental health problem (which may or may not be the same as the diagnosis). Interviews and focus groups with adult service users were conducted by three experienced qualitative researchers, one of whom is a service user (JCo, JCa, AG). Interviews with young adults were undertaken by an experienced qualitative researcher and a clinician who specialises in child and adolescent mental health (ETB). All participants were given a £20 shopping voucher in recognition of their time.
The interview process was the same for both adult and young adult service users. The items were grouped by domain, ordered, and presented one domain at a time. To reduce fatigue effect, the items were presented in a different order by starting with a different domain for each subsequent interview. To help establish the validity of each potential item, participants were asked whether it was meaningful and relevant to their quality of life; whether it was clear, understandable and easy to answer; and the reason they either liked or disliked an item, or preferred it to another. They were also asked for their preferred item within a group or pair of items thought by them to relate to a similar concept (e.g. between ‘I felt relaxed’ and ‘I felt calm’). Alternative wordings to items were elicited if participants thought an item was important to their quality of life but unclear or difficult to answer.
An iterative approach was undertaken with the interviews. Adult participants were initially presented with a set of 88 potential items. At approximately the halfway stage after interviews had been undertaken at two of the trusts, 12 items (which were primarily a reformulation of existing questions) were added as a result of the feedback provided by those participants who had been interviewed. This increased the item set to 100 potential items for the remaining interviews. As a result of the findings from the adult participants, a meeting was held with scientific and user group members and decisions were made to remove some items from the set and to change others. This reduced the number of items to 61. Young adults were presented with this reduced item set. Due to the large size of the item set, not all participants provided their views on all items. The average number of items commented on by the adult group was 68; the lowest number was 5 by a person who was unwell at the time and the interview was terminated early, and also 19 items by a focus group of 7 people. The maximum number of responses was all 100 items. The average number of items commented on by the young adults was 34; the lowest number was 21 and the highest was all 61 items. The comfort and enthusiasm of the interviewee to continue with the interview was considered at all times. The majority of the interviews were recorded and these recordings were transcribed. Notes were taken for the three adult interviewees who preferred their interview not be recorded. All identifying information was removed from the transcripts and notes prior to analysis.
A pragmatic approach was taken in the analysis of the interview data. From the transcripts, the comments made by each participant for each item were charted into a spreadsheet framework with items across the horizontal axis and participants on the vertical axis. A traffic light system was used to highlight negative (red), positive (green), and neutral or ambiguous comments (orange). Those items with relatively high levels of acceptability and unacceptability were identified. A thematic analysis of the comments was undertaken to establish the underlying reasons for the popularity, or lack of popularity, of the items. This information was used as a starting point for discussion with the scientific, advisory, and expert user groups to establish whether or not an item should remain as a potential item in the ReQoL measure. In the final stages of the development of the measure, this information was used in conjunction with other evidences such as the psychometrics and feedback from clinicians  to make a decision on the final items.
Analysis required that the items included in the measure should be relevant and meaningful, be unambiguous and easy to answer when feeling distressed, did not cause further distress, and were non-judgemental. Importantly, this was from the perspective of the service user. Interviewee quotes relating to these issues and the respective items can be found in Table 2.
There were some items which were universally liked and had few objections. What these items had in common was that the service users could relate to the item as being something they experienced regularly. As a result, responses to these questions did not require much thought and were considered easy to answer by the vast majority of participants. Items which fell into this category were ‘I had difficulty getting to sleep or staying asleep’; ‘My health limited day to day activities’; ‘I felt able to trust others’ (Quote 1); ‘I felt anxious’ and ‘I felt confident in myself’ (Quote 2). These items were considered to be particularly relevant because of the further impact these experiences or feelings had on other aspects of their quality of life, for example, if you were unable to trust people, then you would never be happy, and how self-confidence impacted on self-worth.
Some items were described by the service users as being irrelevant or meaningless, either to their own mental health problems or to quality of life. For example, when considering the item ‘I felt accepted as who I am’, service users felt that is was more important to quality of life that they accepted themselves rather than be accepted by others. It was felt that it was not necessary to ‘feel loved’ (Quote 3) for a good quality of life, and was perhaps a bit of a luxury, but feeling ‘cared for’ was important and less specific to a certain type of relationship. It was also felt impractical and unachievable to have ‘everything under control’ or be able to ‘do all the things I wanted to do’ (Quote 4).
There were objections to some items because they were thought to be too general and not specific to mental health difficulties. Examples included ‘I avoided things I needed to do’—it was felt that there may be very good reasons to avoid doing things you did not want to do and doing so could enhance your mental health; ‘I felt irritated’ which was felt to be a normal reaction and not necessarily linked to mental health; and ‘I felt tired and worn out’ (Quote 5) which could be due to physical health as well as mental health. A few of the young adults thought that being ‘confused about who I am’ (Quote 6) was a natural thing for individuals in their age group.
Ease of response
Some items were considered difficult to answer by some participants because they were too abstract, thus requiring too much thought. Whilst under normal circumstances this would not be a problem, it was something they felt could be particularly difficult when in a distressed state upon first accessing mental health services. This particularly applied to items where they thought they were required to consider the thoughts of others, e.g. ‘I thought people did not understand me’; ‘I felt discriminated against’ (Quote 7); ‘I felt accepted as who I am’; and ‘I thought people did not want to know me’ (Quote 8). Rather than considering whether they thought people understood them (the intention of the item), some would try to think about eg whether people did understand them.
There were some items that the service users thought may be difficult to provide an honest answer to because of their own low level of self-awareness whilst ill. It was sometimes only in retrospect that they may realise they had ‘neglected themselves’ (Quote 9) or were not ‘thinking clearly’ (Quote 10).
Some items could be interpreted in more than one way, for example, whether an item related to physical or emotional health. ‘I was in pain’ was considered by some to be about emotional rather than physical pain, and there was some ambiguity whether ‘I had problems with self-care, washing or dressing’ was related to emotional or physical problems, or both. There were also items that were intended to be negative (i.e. indicative of a poor quality of life) but in some circumstances could be regarded as positive. For the item ‘I felt guilty’ (Quote 11), a couple of the interviewees indicated that this could be a positive change when they had done something regrettable whilst ill. Similarly, items that were intended to be positive could be interpreted negatively. For example, ‘thinking clearly’ could be due to a total lack of emotion and doing something ‘bad’ could be ‘enjoyable or rewarding’. For young adults, the item ‘I felt everything was under control’ (Quote 12) was not necessarily indicative of good quality of life as it meant being in charge, and the item ‘I was able to cope with everyday life’ was preferred as it indicated they were dealing with it without having to take over control. Young people also thought the item ‘I felt hopeless’ (Quote 13) had two interpretations of ‘lack of hope’ and ‘feeling useless’.
Distressing or sensitive items
One of the most common reasons given for objecting to an item, and having the view that it should not be included in a quality of life measure, was that it would cause upset. Some items were considered to be too negative. These were often related to suicidal thought and intent. It was the extreme negative (and direct) wording within items that participants found distressing, such as ‘I had thoughts of killing myself’ (Quote 14) and ‘I thought I would be better off dead’. The wording was described as being ‘upsetting’, ‘harsh’, and ‘too strong’ and could actually provoke suicidal thoughts. However, the majority of participants thought that suicidal intent was an important indicator of quality of life but expressed a preference for items with a more indirect, sensitive approach. The items ‘I did not care about my own life’ and ‘I thought my life was not worth living’ were preferred. Other items considered to be too extreme by some participants, and thus described as upsetting, related to feelings about the self. The items ‘I felt humiliated or shamed by other people’; ‘I felt useless’; ‘I felt shame’; ‘I felt stupid’; and ‘I detested myself’ (Quote 15) were described as ‘too personal’, ‘embarrassing’, ‘not nice’, and ‘traumatising’. One participant stated that such a negative question about the self would make ‘the voices’ more prominent. For items relating to the self, the positive items (e.g. ‘I felt confident in myself’; ‘I felt ok about myself’) were preferential. Of the negative items, again a gentler approach was favoured, e.g. ‘I disliked myself’ rather than ‘I detested myself’.
Due to its sensitive nature, a number of people responded that they would not like to admit to certain things and therefore would either not respond to the item or would answer dishonestly. The reasons given were that they would find it embarrassing [‘I felt humiliated or shamed by other people’; ‘I had problems with self-care, washing or dressing’ (Quote 16)] and had concerns surrounding the consequences of disclosure [‘I had thoughts about killing myself’; ‘I have threatened or intimidated another person’ (Quote 17)].
Some items were felt to be insensitive because they were too positive, in particular the items ‘I felt full of life’ (Quote 18), ‘I felt I could bounce back from my problems’, and, to a lesser extent, ‘I felt happy’. Participants felt these items to be unrealistic, as they thought they were never likely to feel this way even when they were well. Such items were described as ‘patronising’ and ‘daft’, particularly if asked when they were accessing a mental health service for the first time when they would be feeling particularly distressed.
Some positively worded items were thought to be too judgemental and reflected an opposing value system. This particularly applied to those items that were related to doing ‘good’ things ‘I was able to do things that helped others’ (Quote 19); ‘I felt I made a contribution’ (Quote 19); ‘I did things that I found worthwhile’; and ‘I felt useful’. It was felt by some participants that helping others was not necessary for quality of life and that people with mental health problems were not necessarily in a position to be able to help others or to do things that were worthwhile, and doing such activities could make them feel worse. Furthermore, participants noted that if the items were answered truthfully (that they did not do things that helped others) this may result in feelings of guilt. The concept of ‘independence’ was also thought to be judgemental because of the assumption that independence was something to be valued (Quote 20).
Recent guidance advocates clear and transparent reporting of measure development and assessment of instrument properties . Reliability and psychometric properties of instruments are frequently reported, but a key stage of the development of any measure is that of content and face validity. The assessment of content and face validity and the acceptability of items by those people who will ultimately use the measure can only be achieved through qualitative work with service users . This paper demonstrates the importance of considering service users’ views on potential items. Service users provided their views on potential items to be considered in the ReQoL measure. In summary, they expressed concerns about items that were sensitive and could potentially cause distress; judgemental of what was a good outcome; not relevant to their mental health and quality of life; ambiguous in their interpretation; and difficult to answer either because they were too vague or abstract. The service users indicated that the potential consequences of including such items would be that they would not respond to the item accurately, truthfully, or at all, which in turn would have a detrimental effect on the validity of the measure. This is especially concerning as most of the items tested were from, or adapted from, measures currently in use.
Our findings are consistent with those of Crawford et al.  who retrospectively sought the views of service users about the appropriateness of commonly used measures in mental health. One of the primary concerns expressed in their study was where an item was judgemental in its criteria of ‘good’ outcome: for example, the view that people who got on well with family members had better social functioning or quality of life. The items susceptible to this criticism in our set of potential items related to the concepts of ‘independence’, ‘helping others’, and ‘contributing to society’ by doing things that were useful or worthwhile.
Crawford et al.  also recommended that a measure should not comprise a long list of questions about difficulties associated with mental ill health but rather it should consist of positive as well as negative items. We did find that some items could be considered too extreme in their negative or positive nature. Those considered too negative by our participants predominantly related to suicidal thoughts and those considered to be too positive related to higher levels of well-being (happiness, joy, and fun). This presented some concerns for the ReQoL team as it was considered important that a valid quality of life measure should cover the full spectrum of mental health from the worst to the best achievable . Additionally the polar aspects of the stages of recovery should also be represented; from a profound sense of loss and hopelessness to living a full and meaningful life . So, whilst service user concerns were acknowledged, a decision was made to retain to the next stage of development those items which were considered the least extreme and least objectionable to service users but still reflected a sense of hopelessness, and positive well-being. A similar dilemma occurred with items relating to a negative sense of self. The theory of stages of recovery starts with a notion of ‘negative identity’ through to a ‘positive sense of self’ . There were no objections related to items having an overly positive sense of self but there were some that caused upset because they were unduly negative (e.g. I detested myself, I felt stupid). Again, in order to cover a broad spectrum of ill health, those that were negative but considered less insensitive were retained.
Some participants objected to items they thought could easily be applied to any member of the general public rather than being specific to those with mental health problems (feeling irritated, avoiding things, tiredness). After discussion, it was decided that this did not necessarily warrant omission of these items as the measure is designed for people at different stages of their recovery.
To maximise the acceptability, validity, and reliability, the findings from this research informed the development of the ReQoL measure at every stage. As a result of the feedback from service users, some items were omitted, whilst others were reworded. After psychometric assessment of item performance, the findings were again used to inform the final item selection for the ReQoL measure alongside clinician input .
For an instrument to measure the concepts most significant and relevant to a service user’s difficulties, it is important that those providing feedback are representative of the target population . This was considered as far as possible and people with a wide variety of diagnoses, with different severity levels and from inpatient, outpatient, and recovery services were recruited and interviewed. However, it should be noted that a small proportion of those with the greater severity of problems at the time of the interview were less able, or motivated, to provide a comprehensive response. They were happy to indicate whether or not they liked an item but did not always articulate why. A greater depth of response was given by those who had milder conditions or those who were relatively well at the time of the interview, though they were able to reflect back on when they were less well. Due to time constraints, we were unable to interview people from primary care services.
Whilst care was taken at the initial selection of items using an established list of criteria for choosing the best items, this stage made it clear that the responses of the service users were not always those expected by the researchers. This further reinforces the fact that a key component of measure development is that of content validity to the target population. Here we have shown that certain questions can be considered irrelevant, judgemental, distressing, ambiguous, or difficult to answer and care should be taken when developing a measure to be cognisant of this. However, the items that one population of service users find objectionable is likely to differ from another. This paper demonstrates that the involvement of service users within this important methodological stage is paramount to the development of a reliable and valid measure that will be acceptable to those who complete it. Further independent research on service users’ acceptance of the measure and items compared with other quality of life measures would be welcomed.
Availability of data and material
The data that support the findings of this study are available from the corresponding author. However, restrictions apply to the availability of these transcripts as the participants only consented to their data being accessible to the researchers involved in the project.
Department of Health (2010). The NHS outcomes framework 2011–2012, London: Department of Health.
US Department of Health and Human Services FDA Center for Drug Evaluation and Research, US Department of Health and Human Services FDA Center for Biologics Evaluation and Research, US Department of Health and Human Services FDA Center for Devices and Radiological Health. (2006). Guidance for industry: Patient-reported outcome measures: Use in medical product development to support labelling claims: Draft guidance. Health and Quality of Life Outcomes, 4, 1–20.
Fitzpatrick, R., Davey, C., Buxton, M., & Jones, D. (1998). Evaluating patient based outcome measures for use in clinical trials. Health Technology Assessment, 2, 1–74.
Streiner, D. L., Norman, G. R., & Cairney, J. (2014). Health measurement scales: A practical guide to their development and use, (5th edn.). Oxford: Oxford University Press.
Mokkink, L. B., Terwee, C. B., Patrick, D. L., Alonso, J., Stratford, P. W., Knol, D. L., et al. (2010). The COSMIN checklist for assessing the methodological quality of studies on measurement properties of health status measurement instruments: An international Delphi study. Quality of Life Research, 19, 539–549.
Patrick, D. L., Burke, L. B., Gwaltney, C. J., Leidy, N. K., Martin, M. L., Molsen, E., & Ring, L. (2011). Content validity—establishing and reporting the evidence in newly developed patient-reported outcomes (PRO) instruments for medical product evaluation: ISPOR PRO Good Research Practices Task Force report: Part 2—assessing respondent understanding. Value in Health, 14(8), 978–988.
Holden, R. R. (2010) Face validity. Corsini Encyclopedia of Psychology. https://doi.org/10.1002/9780470479216.corpsy0341
Staniszewska, S., Haywood, K. L., Brett, J., & Tutton, L. (2012). Patient and public involvement in patient-reported outcome measures. The Patient-Patient-Centered Outcomes Research, 5(2), 79–87.
Wiering, B., Boer, D., & Delnoij, D. (2017). Patient involvement in the development of patient-reported outcome measures: A scoping review. Health Expectations, 20(1), 11–23.
Perkins, R. (2001). What constitutes success? British Journal of Psychiatry, 179(1), 9–10.
Rose, D., Evans, J., Sweeney, A., & Wykes, T. (2011). A model for developing outcome measures from the perspectives of mental health service users. International Review of Psychiatry, 23(1), 41–46.
Groene, O. (2012). Patient and public involvement in developing patient-reported outcome measures. Patient, 5(2), 75–77.
Ennis, L., & Wykes, T. (2013). Impact of patient involvement in mental health research: Longitudinal study. British Journal of Psychiatry, 203(5), 381–386.
Jacobson, N., & Curtis, L. (2000). Recovery as policy in mental health services: Strategies emerging from the states. Psychiatric Rehabilitation Journal, 23(4), 333.
Bracken, P., Thomas, P., Timimi, S., Asen, E., Behr, G., Beuster, C., et al. (2012). Psychiatry beyond the current paradigm. British Journal of Psychiatry, 201(6), 430–434.
Bohanske, R. T., & Franczak, M. (2010). Transforming public behavioral health care: A case example of consumer-directed services, recovery, and the common factors. In B. L. Duncan, S. D. Miller, B. E. Wampold & M. A. Hubble (Eds.), The heart and soul of change: Delivering what works in therapy. American Psychological Association, Worcester, pp. 299–322.
Australian Health Ministers Advisory Council (2013). A national framework for recovery-oriented mental health services: Guide for practitioners and providers. Commonwealth of Australia, Canberra. http://www.health.gov.au/internet/main/publishing.nsf/content/67D17065514CF8E8CA257C1D00017A90/$File/recovgde.pdf. Accessed 22 Nov 2016.
Piat, M., Sabetti, J., & Bloom, D. (2010). The transformation of mental health services to a recovery-orientated system of care: Canadian decision maker perspectives. International Journal of Social Psychiatry, 56(2), 168–177.
Boardman, J., & Shepherd, G. (2009). Implementing recovery: A new framework for organisational change. London: Sainsbury Centre for Mental Health. https://socialwelfare.bl.uk/subject-areas/services-client-groups/adults-mental-health/centreformentalhealth/128566implementing_recovery_paper.pdf. Accessed 15 Nov 2016.
Mental Health Commission (2007). The journey of recovery for the New Zealand mental health sector. Wellington: Mental Health Commission https://www.mentalhealth.org.nz/assets/ResourceFinder/te-haererenga-1996-2006.pdf. Accessed 14 Nov 2016.
Hogan, M. F. (2003). The President’s new freedom commission: Recommendations to transform mental healthcare in America. Psychiatric Services, 54(11), 1467–1474.
Boardman, J., & Shepherd, G. (2013) Assessing recovery: Seeking agreement about the key domains. Report for the Department of Health. Unpublished (available from the authors).
Shepherd, G., Boardman, J., Rinaldi, M., & Roberts, G. (2014). Supporting recovery in mental health services: Quality and outcomes. Centre for Mental Health and Mental Health Network, NHS Confederation, 34. https://www.centreformentalhealth.org.uk/recovery-quality-and-outcomes. Accessed 18 Nov 2016.
Connell, J., Brazier, J., O’Cathain, A., Lloyd-Jones, M., & Paisley, S. (2012). Quality of Life of People with Mental Health Problems: A synthesis of qualitative research. Health and Quality of Life Outcomes, 10(1), 138.
Brazier, J., Connell, J., Papaioannou, D., Mukuria, C., Mulhern, B., Peasgood, T., et al. (2014). A systematic review, psychometric analysis and qualitative assessment of generic preference-based measures of health in mental health populations and the estimation of mapping functions from widely used specific measures. Health Technology Assessment, 18, 34.
Connell, J., O’Cathain, A., & Brazier, J. (2014). Measuring quality of life in mental health: Are we asking the right questions? Social Science and Medicine, 120, 12–20.
Leamy, M., Bird, V., Le Boutillier, C., Williams, J., & Slade, M. (2011). Conceptual framework for personal recovery in mental health: Systematic review and narrative synthesis. British Journal of Psychiatry, 199, 445–452.
Keetharuth, A. D., Brazier, J. E., Connell, J., Bjorner, J. B., Carlton, J., Buck, T., E., et al (2018). Recovering Quality of Life (ReQoL): A new generic self-reported outcome measure for use with people experiencing mental health difficulties. British Journal of Psychiatry, 212, 42–49.
Crawford, M. J., Robotham, D., Thana, L., Patterson, S., Weaver, T., Barber, R., et al. (2011). Selecting outcome measures in mental health: The views of service users. Journal of Mental Health, 20, 336–346.
Huppert, F. A., & Whittington, J. E. (2003). Evidence for the independence of positive and negative well-being: Implications for quality of life assessment. British Journal of Health Psychology, 8(1), 107–122.
Andresen, R., Oades, L., & Caputi, P. (2003). The experience of recovery from schizophrenia: Towards an empirically validated stage model. Australia & New Zealand Journal of Psychiatry, 37(5), 586–594.
The authors would like to thank all the participants in this research and the staff who have been involved in the recruitment of participants. Special thanks to Rob Hanlon, Jo Hemmingfield, and John Kay for providing us with service users’ perspectives.
The study was undertaken by the Policy Research Unit in Economic Evaluation of Health and Care interventions (EEPRU) funded by the Department of Health Policy Research Programme. The research was also part-funded by the National Institute for Health Research Collaboration for Leadership in Applied Health Research and Care Yorkshire and Humber (NIHR CLAHRC YH). http://www.clahrc-yh.nihr.ac.uk. The views and opinions expressed are those of the authors, and not necessarily those of the NHS, the NIHR, or the Department of Health.
Conflict of interest
All authors declare that they have no conflict of interest.
Ethical approval was obtained from the Edgbaston National Research Ethics Service Committee, West Midlands (14/WM/1062). Governance permission was obtained from each of the participating NHS Trusts.
Informed consent was obtained from all participants in the study. Participants have consented to the possibility of using quotations in the research reports or other publications, after they have been anonymised.
About this article
Cite this article
Connell, J., Carlton, J., Grundy, A. et al. The importance of content and face validity in instrument development: lessons learnt from service users when developing the Recovering Quality of Life measure (ReQoL). Qual Life Res 27, 1893–1902 (2018). https://doi.org/10.1007/s11136-018-1847-y