Publication Cover
Journal of Human Development and Capabilities
A Multi-Disciplinary Journal for People-Centered Development
Volume 23, 2022 - Issue 3
1,126
Views
0
CrossRef citations to date
0
Altmetric
Articles

A Deliberative Approach to Valuing Capabilities: Assessing and Valuing Changes in the Well-Being of those Close to Patients Receiving Supportive End of Life Care

ORCID Icon, ORCID Icon, ORCID Icon, ORCID Icon & ORCID Icon

ABSTRACT

Aim: Explore the use of deliberative valuation to elicit relative weights for a set of capabilities identified as being important and relevant to those close to patients receiving supportive care at the end of life. Methods: Focus groups, involving the general UK population (n =38) and policy-makers (n =29) with experience of, and influence on, priorities for end of life care. Public participants completed two valuation tasks (budget pie and visual analogue scale (VAS)) individually, discussed their responses, and then recorded a final (individual) response. Policy-makers completed the VAS tasks in a separate series of focus groups. Quantitative and qualitative analyses of participants’ responses are reported. Results: Individual values were aggregated to form relative weights for the capabilities. Capabilities given greatest weighting were “good communication between care providers and close persons” and “practical support for close persons”. The quantitative impact of deliberation on weights overall was negligible, but qualitative findings indicated that disclosure of personal experiences did appear to prompt others to consider issues from new perspectives. Discussion: Deliberative valuation was found to be a potentially feasible method for generating weights. However, further consideration needs to be given as to how to optimise recruitment whilst ensuring that participants actively engage with the task.

Introduction

Scarce healthcare resources necessitate difficult decisions regarding prioritisation. Economic evaluation can provide evidence to inform resource allocation decisions. Economic evaluation can be defined in broad terms as the systematic comparison of alternative courses of action in terms of their associated costs and outcomes (Drummond et al. Citation2005). The specific form of economic evaluation required in health and care contexts by decision-making organisations internationally, including the National Institute for Health and Care Excellence (NICE) in the UK, is cost-utility analysis (CUA) (NICE Citation2013). CUA rests on the premise that the objective of the healthcare system is to maximise health across the population (Coast Citation2009; Culyer Citation1989).

Increasing interest in the application of Sen’s capability approach within health and care contexts has been noted previously (Kinghorn Citation2015). Two motivations for operationalising the capability approach in health and care contexts were identified from that previous debate (Kinghorn Citation2015, 600): “(i) as an alternative to utilitarian health maximization … and (ii) facilitating agreement on a core concept of health … with which to drive policy reform”. With respect to the latter of these two motivations, Ruger seeks to specify an objective global standard for assessing health by setting out her Health Capability Paradigm, which focuses on the capability to avoid premature mortality and the capability to avoid premature morbidity (Ruger Citation2009). With respect to the former, the maximisation of health outcomes has since been argued to be particularly inappropriate within the context of supportive end of life care (Kinghorn and Coast Citation2019). Normand (Citation2009) highlights the paradox of evidence, suggesting that people wish to give priority to palliative care interventions which would not be deemed cost-effective when evaluated using a QALY framework. The capability to experience a good death (Coast Citation2014) has been proposed as a more suitable conceptual approach.

One set of outcome measures noted as aligning to the first motivation (which can be thought of as expansionist because of the desire to broaden the evaluative space beyond health) are the ICECAP measures (Al-Janabi, Flynn, and Coast Citation2012; Grewal et al. Citation2006), which capture the impact of health conditions and public services on the self-reported well-being of those directly impacted. The ICECAP-Supportive Care Measure (ICECAP-SCM) is a measure of care-related well-being, developed for use in economic evaluations of supportive end of life care (Sutton and Coast Citation2014), which can be interpreted as assessing elements of the capability for a good death.

The negative impact of bereavement on close persons has been well evidenced (Stroebe, Schut, and Stroebe Citation2007), and was famously acknowledged by Dame Cicely Saunders (a pioneer of current models of hospice care): “How people die remains in the memory of those who live on” (Saunders Citation1989). In terms of national guidance, NICE lists sixteen “Quality Statements” relating to adult end of life care in the UK, which includes the following explicit reference to families and carers (referred to in this paper as “close persons”) (NICE Citation2017):

Families and carers of people approaching the end of life are offered comprehensive holistic assessments in response to their changing needs and preferences, and holistic support appropriate to their current needs and preferences.

A measure has been developed (the ICECAP-CPM) to complement the use of ICECAP-SCM by capturing outcomes of importance to close persons (see Canaway et al. Citation2016 for further details). The measure has six attributes:

  • Communication with those providing care services

  • Privacy and space (for example, the close person being able to spend time with the patient in a private, peaceful location)

  • Practical support (for example, support to enable the close person to care for the patient)

  • Emotional support

  • Preparing and coping

  • Emotional distress (to the close person, related to the condition of the decedent).

Each attribute has five levels, ranging from “fully able/[able to achieve the outcome] all of the time” to “completely unable/none of the time”. The “best” level is coded as 5 (for the purposes of data entry), and the “worst” level is coded as 1. 15,625 (56) unique states are defined by the measure. The measure is available to view at www.bham.ac.uk/icecap.

It is important that the relative weighting of attributes within an economic measure reflect the underlying value judgements and priorities of society. Whilst any progress (in terms of enabling the capabilities on ICECAP-CPM) will be associated with positive benefit, in light of resource scarcity, innovation and/or investment enabling those capabilities most important to close persons should, ceteris paribus, be prioritised over those which enable less important capabilities. Some way of gauging and quantifying public priorities and values is therefore necessary, as this will be reflected in the relative weighting of attributes. A scoring system will combine information on the weighting of attributes and attribute levels, so as to be able to summarise any state of the world defined by the measure as a single numerical score, often ranging between a defined lower anchor and a defined upper anchor. Such a scoring system exists for ICECAP-SCM (Huynh et al. Citation2017), and was developed through the use of a Discrete Choice Experiment (DCE) and Best-Worst Scaling, administered through an online survey. This type of methodology readily facilitates a large sample size, from a representative sample of the population, and enables sophisticated analysis, in the case of ICECAP-SCM, for example, capturing interaction effects between the attributes.

The purpose of this study was to explore deliberative valuation methods as a means of eliciting relative weights for the states defined by ICECAP-CPM. Deliberation was deemed appropriate because of its conceptual alignment with the guidance of Amartya Sen, who advocates tackling some of the more practical steps in operationalising the capability approach (such as establishing the value of particular capabilities) via reasoned consensus, involving public discussion and democracy (Sen Citation1999). Public deliberation has been defined as “an approach designed to capture in-depth and informed public perspectives on complex topics” (Carman et al. Citation2015). Deliberation allows respondents time to reflect upon and discuss their values and to share the experiences which have influenced those values, as well as being exposed to the experiences of others. The deliberative process is intended to prompt respondents to articulate, reflect upon and possibly defend their values.

In contrast to traditional economic methods for the elicitation of individual preferences (which either prompt respondents to or assume that they will naturally adopt the perspective of individual welfare maximising consumers), deliberative methods often seek to elicit citizen perspectives or group verdicts (Jacobs Citation1997). For economists, deliberation facilitates the formation of numerical expressions of the relative importance of different options or attributes in contexts where respondents are unlikely to have clear pre-existing values (Jacobs Citation1997).

This work represents a significant methodological and normative shift away from the elicitation of individual values from a large online sample, towards the elicitation of considered, societal values from a smaller sample of participants actively engaged through a facilitated and interactive process of data collection.

Here we report the use of focus groups to facilitate the sharing of views and experiences relating to end of life care, to inform the reporting of numerical values by the public and professionals, indicating the relative importance of a set of pre-determined capabilities (Canaway et al. Citation2016). Qualitative analysis was used to identify: participants’ willingness to share experiences relating to a sensitive topic; acknowledgement of the views of others and any explicit acknowledgment of the influence of this on participants’ own views; reaction to and reflections on the valuation tasks. We are able to compare and contrast numerical values and qualitative accounts from both members of the public and a separate sample of professionals with career experience relevant to end of life care.

Methods

One series of focus groups was held with participants from the general population and a separate series of focus groups was held involving professionals/policy-makers with experience and influence relevant to supportive end of life care.

Ethical approval for the study was obtained from the Science, Technology, Engineering and Mathematics Ethical Review Committee at the University of Birmingham (ERN_11-1296).

Public participants were recruited over Summer 2014 via the edited (open) electoral register, from purposefully selected electoral wards from across the West Midlands region (UK). Six electoral wards were chosen to achieve diversity with respect to: rural versus urban; high, mid and low deprivation; and provision of hospice services.

For public participants, names were selected at random from the electoral registers and those selected were sent an invitation letter and information leaflet by post, with a target of recruiting approximately six to 10 participants per focus group. Initially, 300 invitations were sent per electoral ward, but numbers were increased where response levels were low, meaning that, in total, 2050 invitations were sent out across the 6 area. Those recipients who were interested in participating were asked to return an expression of interest form; they could also contact the research team by email or telephone if they had questions or concerns. Individuals who had been bereaved within a six-month period were excluded for ethical reasons. The focus groups were held at accessible locations within the selected electoral wards. Participants were refunded for travel costs, but no incentive was offered for participation. Groups lasted for between 75 and 120 minutes. The Information Leaflet is presented as Appendix D.

Policy-makers and service managers (referred to in general terms as policy-makers from hereon in) were recruited to participate in one of four focus groups. Policy-maker focus groups either “piggybacked” on planned meetings of existing local/regional networks (for example, Clinical Commissioning Groups and forums for healthcare leaders) or brought together individual national and regional policy-makers with experience and influence relating to end of life care identified and recruited directly by the research team.

Focus groups began with introductions from the research team and from participants. An open discussion was then invited and facilitated relating to end of life care generally. For both public and policy-makers, a budget pie task was used to explore the relative importance/value of the six attributes (see Appendix A). In addition, public participants completed visual analogue scales (VAS) to explore the relative value of the attribute levels. Policy-makers were not asked to complete the more time consuming VAS tasks as the focus groups were typically shorter in duration and policy-makers discussed some additional topics relevant to the wider research project, but falling outside of the scope of this paper.

The VAS tasks were completed first in the public focus groups. All respondents were presented with a VAS for each of the six attributes, printed within a workbook (see appendix A). The “best” level (portraying the highest capability) was fixed at 100 on the VAS and represented the upper anchor; the “worst” level (portraying the lowest level of capability) was fixed at zero, and represented the lower anchor. Respondents were asked where on the scale they felt that the remaining three attribute levels would fit, according to how good or bad they perceived them as being, and relative to the two anchor points. Respondents completed the tasks individually and then shared their responses with the group as a stimulus for discussion. Following this discussion, respondents were asked to record a final response; there was no pressure for respondents to change their responses, but they had the ability to do so at this point.

For the budget pie method, respondents were presented with a circle, divided into six (approximately equal) segments (within the workbook). Each segment was labelled with one of the six attributes (Appendix A). Respondents were asked to divide 100 tokens between the 6 attributes (segments). Respondents were advised to allocate more tokens to attributes they considered to be of greater importance, to allocate all 100 tokens, and that, should they wish, it was acceptable to allocate zero tokens to one or more of the attributes. Budget pie was considered to be a relatively simple task, which could be completed with “paper and pen”, but which encompassed an element of scarcity and necessitated respondents to make trade-offs. Respondents completed the task individually but then shared their responses with the group as a prompt for comparison and discussion. Following this discussion, respondents were asked to record a final response.

All focus groups were digitally audio recorded; recordings were transcribed verbatim and pseudonymised. Qualitative analysis of the transcripts was conducted following a constant comparative approach (Strauss and Corbin Citation1990; Glaser and Strauss Citation1968) to provide a deeper understanding on how individuals weigh attributes, their openness to deliberation, and the impact of deliberation on their values. Following initial familiarisation with the transcripts, transcripts were divided into two batches: transcripts for focus groups where changes were recorded, following discussion, and focus groups where no changes were made. A coding framework was developed for transcripts within the same batch, before the development of analytical accounts. Finally, there was comparison of analytical accounts between batches (Coast and Jackson Citation2017). Where quotations are used to illustrate qualitative themes, “G” is used to identify members of the general public, and “P” to identify policy-makers. Letters are also used after the initial prefix of “G” or “P” to indicate which focus group the participant contributed to (A to E for the public, and A to D for policy-makers). A number denotes the specific participant within that group, e.g. “GA1” for participant one, within public group A. The symbol “>” is used to indicate the continuation of a conversation or exchange (a quotation directly follows that above in conversation).

Descriptive statistics are presented for quantitative responses. VAS and budget pie responses were both rescaled from a zero to 100 scale onto a zero to one scale. A paired t-test was used to explore differences in attribute weightings between sample subgroups, defined by age and gender, and between the public and policy-maker samples.

Scores were calculated by multiplying weights for attribute levels by the relative weighting of the attribute. The scoring system (denoted in Equation (1)) is therefore additive: the overall score (V) is calculated by adding the weights reflecting the relative goodness of the level of capability (L), adjusted by the relative attribute weighting (A), across the six attributes. L is the value from the VAS task (it will be 1 for the best attribute level and 0 for the worst attribute level); A is the value from the budget pie task (it will also be a value between 0 and 1). Hence, a score of 1 reflects the highest possible level of capability (the best level on all six attributes, full capability) and 0 represents the worst possible state (the bottom level on all six attributes, no capability). (1) V=(LCommunicationACommunication)+(LPrivacyAPrivacy)+(LPracticalAPractical)+(LEmotionalAEmotional)+(LPreparingAPreparing)+(LDistressADistress)(1)

The scoring system presented is based on public values only, because policy-makers did not complete the VAS task.

Results

This section begins by providing details of those who participated in the research (public and policy-makers). Quantitative results from the public sample are presented, and then compared with quantitative results from policy-makers. We explore differences between participants’ initial responses and their responses following discussion within the group, before presenting themes from the qualitative analysis (views relating to the attributes and justifications given for relative weightings, and the nature of the deliberative process).

Recruitment and Participants

In response to 2050 invitation letters, there were 72 initial expressions of interest (an initial response rate of 3.5%); initial response rates varied (by electoral ward) between 0.5% and 5.4%. Seven potential respondents were excluded at initial screening due to recent bereavement. Because of the extremely low response rate of 0.5% from one electoral ward, this focus group was cancelled. Across the remaining five electoral wards, 65 individuals were sent confirmation letters inviting them to attend; 38 did attend. Where response rates were high from a particular electoral ward, two focus groups were conducted.

One focus group was held in each of wards A (urban, mid-deprivation; n =3), B (urban, high deprivation; n =1) and C (urban, mid-deprivation; n =6). Two focus groups were held in wards D (n =6 for D1 and n =6 for D2) and E (n =9 for E1 and n =7 for E2). Wards D and E were rural, low deprivation areas. Because only one participant attended the ward B “focus group”, values from this participant are excluded from analysis.

Older people, those reporting their ethnicity as white British and females were over-represented in the sample (). Approximately a third of participants had experienced bereavement between six months and two years prior to receiving the invitation letter.

Table 1. Respondent characteristics for public workshops.

29 policy-makers participated across the four focus groups targeted at this sample group. Groups A (n =9) and B (n =7) involved existing teams; group C (n =3) an existing network. Groups A, B and C were local teams/networks of professionals with responsibility for the commissioning and provision of health and end of life care; Group D (n =10) brought together experts from across the UK in the field more generally, including those with expertise in terms of both local and national policy making.

Quantitative Results from the Public Workshops

reports the weighting of the attribute levels (values are rescaled from zero to one). reports the relative weighting of the attributes, from the budget pie task (again, with values rescaled from zero to one). The attribute which received the greatest weight in the budget pie task was “good communication” between close persons and those providing care services. The combined weighting for “communication” and “practical support” (0.549) is greater than the combined score for the remaining four attributes. “Emotional support” (for the close person) received the lowest weighting.

Table 2. Weighting of the levels across attributes.

Table 3. Attribute weighting.

When results are examined for subgroups, defined according to age (below 65 years (n =16) and above 65 years (n =22)), “communication” remains as the attribute with the greatest weight (0.291 and 0.284 respectively). The biggest difference between the two sub-samples defined according to age, is observed for the “emotional distress” attribute, given a weighting of 0.179 by the younger sub-sample, but just 0.088 by the older sub-sample. The “emotional distress” attribute was the only attribute for which there was a statistically significant difference (at the 5% level) in the weighting between the two age sub-samples. There were no statistically significant differences (at the 5% level) between the weighting of attributes between sub-samples defined according to either gender (male/female) or bereavement status (bereaved or not bereaved within the last two years).

A scoring system for the ICECAP-CPM is presented in ; as an example, the state 4,4,4,4,4,4 would have a score of:

Table 4. ICECAP-CPM scoring system.

0.2479+0.2245+0.0990+0.0788+0.0905+0.0965=0.8372

Public Versus Policy-Maker Weighting

“Privacy and space” was the only attribute for which there was no statistically significant difference (at the 5% level) between the mean scores of the public and of policy-makers (). Members of the public gave greater weight to “communication” and “practical support” than did policy-makers. There was greater variation/distinction in the weighting given across the attributes by the public (in terms of the difference between the attribute receiving the greatest weighting and the attribute receiving the lowest weighting) than by policy-makers, although the attributes at the two extremes are the same for both groups. Three of the policy-makers gave equal weight to all six attributes, but no members of the public gave equal weight in this way. Other policy-makers gave very similar weight across the six attributes.

Variation across participants was narrower in the case of responses from policy-makers, compared to the public, for every attribute. The range of values across public participants for the same attribute was equal to or exceeded 0.25 across all attributes (and was 0.4 or above for four of the six). The range across policymaker values for the same attribute was below 0.4 for all of the attributes, with a range of values across policy-maker participants of below 0.25 for three attributes.

Impact of Deliberation Upon Quantitative Responses

For the attribute weighting task (budget pie), there were five public participants (13.5%) who made at least one change to their initial response following discussion (Appendix B). To change the weighting of one attribute and still have scores totalling 100, it was clearly necessary to also change the weighting of at least one other attribute. Three of the five participants who chose to change their response on the weighting task were in public group C. Two of the five participants made four changes each. Just over half of the 14 changes (n =8) brought the participant’s weighting closer to the initial group mean value. In changing the weighting of one attribute so that it was closer to the initial mean/modal value for the group, some participants adjusted the weighting for another attribute such that it diverged further from the initial group mean/mode.

To give an indication of the limited scale of the impact of specific discussion of values on the final scoring system (as reported in , using public values only), had participants’ initial responses been used to calculate the scoring system (rather than the final responses), level 2 of the attribute “emotional support” would have been 0.0745 rather than 0.0788; this was the largest difference at −0.0043.

Only one policy-maker changed their responses following discussion, reducing their weighting of privacy and space and allocating those “tokens” instead to “practical support”.

Discussion Relating to the Attributes

There was generally more discussion relating to attributes that were given lower weights than there was for attributes given greater weight. Participants often gave little more justification for giving high weight to communication and practical support than to state that they were important, and the general consensus within the group meant that this view was unchallenged. Practical support was seen as important for protecting the physical and mental well-being of the close person, and for ensuring the safety of the patient.

… if you don’t have the support you need to help look after them then that’s going to affect them quite adversely. (GE68, aged 18–29)

… if you haven’t got the practical help, how are you going to function? (GE7, Aged 30–44)

Good communication was seen as important at a point in time that was referred to as being “scary” for both the patient and close persons. Some participants discussed close persons taking on an advocacy role on behalf of the patient and helping the patient to absorb and process information at a challenging time for them. Good communication was also referred to as enabling close persons to have a sense of control and to act with more confidence in supporting the patient practically.

you, as a carer, have to give trust to somebody else, but actually you want the communication as well, you want to know what’s happening because that’s the only bit of control that you might have. (PA3)

If you don’t know you can’t act for yourself, let alone anyone else. (GE4, Aged 65+)

There was a common sense that if communication and practical support are done well then achievement on the remaining attributes will follow:

I thought they all needed a score of some sort but I happened around good communications and practical support as the highest because lots of the others all depend on those two. (GE61, Aged 65+)

I think that communication with family and friends relieves most of the other things … so if healthcare professionals … do that effectively and competently and honestly, then that reduces the distress that they’re feeling. (PD5)

A variety of views were expressed in relation to emotional distress. One view was that emotional distress in response to the death of a loved one is a natural (even desirable) human response, a response which cannot be avoided. A contested view was that emotional distress would diminish over time; other participants countered this viewpoint based either upon their own personal experiences or those of their friends and family.

I don’t think you’re going to avoid emotional distress. (GD2, Aged 65+)

for them, losing that person can’t be resolved; it’s just something you’ve got to deal with. (PA1)

In my experience it doesn’t go away. (GA3, Aged 65+)

There was a feeling by some participants that family members witnessing loss of dignity (which was referred to in the description of the emotional distress attribute) is totally unacceptable. The emotional distress attribute was seen as relating more directly to the care of the patient and hence those wishing to prioritise the patient tended to give it a relatively high weighting.

Many participants regarded emotional support as something that would be provided (exclusively) through family, social and religious networks, and questioned the need to fund external support, although a minority of participants stressed the value of external/professional sources of support, such as counselling.

I-mean it’s important to get emotional support from family and friends but I don’t see how that can be arranged from an outside agency? (GA3, Aged 65+)

There was also a view that priority should be given to avoiding emotional distress, and that as a result, the need for emotional support would be alleviated. In response to the emotional support attribute, a minority of participants suggested that people should “get over it” (GE2, Aged 65+); whereas one participant noted that their generation do not have the “stiff upper lip” mentality of older generations (GE7, Aged 30–44).

For me it’s, this is nice to have, at the risk of sounding a cold fish. (GE63, Aged 45–64)

With respect to being prepared, some participants questioned whether you can ever be prepared for the death of a loved one, and some questioned whether it is possible to prepare for something that is uncertain.

… nobody can [be prepared for the death of a loved one], it doesn’t matter who you are; either that or you’ve got a block of stone inside. (GC5, Aged 65+)

Others considered the need for close persons to find some “head space” to process and come to terms with events and hence begin to prepare, particularly if they were heavily burdened with practical care responsibilities. C1, a carer for a terminally ill son, expressed this view:

say [if] you’re a full time carer and you-know you were very busy and you couldn’t prepare for it, so, to be given the opportunity to be able to prepare for it would be lovely, as in even being told what’s going to happen. (GC1, Aged 18–29)

Whilst some participants highlighted the possible cost of providing privacy and space, as well as questioning the practical feasibility of doing so, one participant felt that if the close persons could be offered privacy and space then this would help to alleviate emotional distress.

Some policy-makers found it difficult to generalise across the diverse cases they had observed, in terms of identifying one capability that would be of significantly greater importance, and hence inclined towards giving equal (or near equal) weight amongst the capabilities.

I’d look at one and think, “Oh yes, that is really important; maybe more so than the others” and then you’d think about different families … and you think, “Well actually, no because that was more important to them” or “that was more important to them” and it’s, it’s such a personal thing. (PB2)

The Nature of the Deliberative Contribution

Public participants were observed to contribute in several ways. The first was through facilitating debate by demonstrating active listening, and/or by supporting and challenging the contributions of others. In the second, a small minority of participants revealed an “agenda”, in terms of either their a priori expectations and interpretations of the research topic and/or their particular enthusiasm for discussing a specific topic (potentially steering discussion back to this topic at separate time-points over the course of the discussion). The third type of contribution involved those who drew primarily upon their own personal or professional experience(s) in relation to the tasks and when engaging in the discussion. It would be false to claim that all (or even most) participants could be categorised as conforming exclusively to “one type or another”; instead, they often contributed at different times in different ways.

  • (i) “Active listening”

Active listening is defined here as explicitly acknowledging the views of others, often acknowledging differences of opinion, and in some cases either confirming agreement with or challenging the views of others. This “supporting and challenging” extended to the nature of the task and instruction and background information given by the research team.

Can I touch on that, [be]cause that’s quite relevant actually with what [C5] was saying … . (GC2)

Acknowledgement of other people’s views did not necessarily indicate that a participant was prepared to change their own:

I think I’ll keep my original, but it’s you-know it’s very fluid isn’t it? (GA3)

I think it’s important but doubt whether it’s attainable. I get your point … . (GD4)

“Challenge” was observed (and the term challenge is not used to imply hostility) to a limited degree in group D (in which individual policy-makers were brought together), but not in those groups involving existing teams. An example of challenge from within group D can be drawn from discussion of the emotional distress attribute (see below).

I assessed it as the impact on the carer, these guys have assessed it as the patient … . (PD9)

> No, I’ve assessed the fact that the patient is dying in pain or distress as having an impact on the carer. (PD8)

Across all policy-maker groups, participants acknowledged where there was similarity in the distribution of their weights.

  • (ii) Revealing an agenda

Three explicit examples of participants seemingly revealing an agenda were identified. GA1 drew upon his experience of having pancreatic cancer, and appeared to use his experience both as an illustration of (as a means of signalling) his own strength (both in challenging the initial prognosis and seeking alternative medical support, and in the way he has “stayed strong” for his family) and in terms of promoting his view about the duty of family to provide care, a theme he returns to throughout the focus group. Participant E4 anticipated that the focus group would consider voluntary euthanasia and sought confirmation of this early in the opening discussion; GC2 also made reference to assisted suicide.

… a person should be cared for preferably by the family. And the family should be encouraged to look after the person unless there’s a professional reason not to do. (GA1)

I’m a really strong man but my family isn’t strong you-know, my children are not strong. (GA1)

But I was going to ask the question whether this ties in with this new debate that’s going on about you-know assisted suicide and stuff. (GC2)

No text from the policy-maker focus groups was coded as relating to this theme, although it should be acknowledged that discussion in these groups was typically briefer and that they were attending the focus group in a professional capacity.

  • (iii) Drawing heavily on personal experiences

One participant (a younger adult) who brought a particularly unique perspective was the parent of a terminally ill child, another spoke about her experience relating to the sudden death of a sibling; a third (another younger adult) disclosed that her mother is terminally ill (GE6). Those sharing personal or professional experiences which introduced new and distinct perspectives to the discussion (as in the cases illustrated above) appeared to influence the group by doing so.

It’s a good point. The point about the young man – we’ve all been thinking, I think, about older people dying or in the process of dying. We didn’t think about the ones that have that sort of problem. (GD4)

… I was going on seventeen and my sister died at twelve, she had a car accident and she died a week later, the vent was taken off her and I wasn’t told, my parents just came home distraught … and I knew nothing about it, and even though -yes it’s about forty five years ago- it still hurts. (GC3)

Well I was just going to say about hospices, for instance my son is terminally ill and he’s nine at the moment … . (GC1)

> I think what [C1] said has set something off in me that I hadn’t really considered, is that you assume the communication is coming one way and you’re getting your questions answered, but like [C1] said for her son, she needs to communicate to them to enable them to care for her son because she knows him best. (GC2)

GC1 was the last participant to speak during the opening discussion in Group C. E6 said relatively little throughout focus group E. However, what both participants did say had an emotional impact. The fact that these participants disclosed emotionally impactful personal experiences without appearing to dominate or steer discussion distinguished them from those identified as revealing a personal agenda, where the agenda was typically acknowledged by others in the group, but did not appear to influence discussion significantly. Participants in group C prompted GC1 at several points during the discussion, enquiring about her experiences. Participants in Group C made more changes to their initial responses than did participants in any other group. E8 only revealed during the final/closing discussion that her father had committed suicide when she was aged 26, perhaps after developing a level of trust with other group participants; tasks, by this point had been completed.

Some participants in the public focus groups drew upon professional experiences, using medical/specialist terminology and acronyms (albeit rarely); they also challenged comments from other participants which they regarded as inaccurate, for example, GC3 challenged GC4 when GC4 said that staff coming in and taking a pulse would be reassuring: “Why would they want to take your pulse?”.

Where groups/teams of policy-makers were interdisciplinary, some participants made reference to their specific specialism shaping their perspective, and policy-makers drew on their personal as well as their professional experiences. Participants in the policy groups also drew upon their professional experiences, to varying degrees.

Discussion

This is the first paper to report values for the ICECAP-Close Person Measure (CPM). Values were elicited following deliberation amongst focus group participants; a methodology specifically selected because of the suggestion by Sen to tackle practical steps in operationalising the capability approach via public discussion and democracy (Sen Citation1999).

Enabling close persons to have good communication with those professionals caring for the dying patient and enabling close persons to access adequate practical support as they themselves provide care were attributes that were singled out by public participants as being of particularly high value or importance; they also received the greatest weighting from policy-makers, although with less differentiation in the weights across the attributes from this latter group. The only generational difference of statistical significance within the public groups related to weighting of the emotional distress attribute. Conversely, the more explicit generational differences identified from the qualitative analysis related to the emotional support attribute, with some older participants suggesting either that this should exclusively be provided by family networks, or that those experiencing bereavement should “get over it”, but with younger participants seemingly having more awareness and acceptance of external sources of support, and rejecting a “stiff-upper-lip” mentality. Because participants appeared to link emotional distress and emotional support, with similar attitudes being expressed by participants in response to both attributes, the generational differences more explicitly observed in discussion coded as relating to emotional support are likely to have applied equally to emotional distress.

Scores can be used within a cost-effectiveness analysis, whereby the results of an economic evaluation would be the cost per improvement in close person outcomes. There is also potential to combine outcomes for close persons with outcomes for patients, as per the suggestion of Coast et al. (Citation2021).

Policy-makers came to the focus groups having already considered issues relating to end of life care and some focus groups “piggybacked” on existing networks, where participants may already have had experience of working together to reach joint decisions. It could either be that policy-makers were more homogenous in terms of their views, or they could have been less inclined to disagree because of a reluctance to harm ongoing team dynamics and working relationships. This may explain why there was less variation in the values reported by policy-makers, why challenge was rarely identified in qualitative analysis of the policy-maker groups, and why only one policy-maker changed their values following discussion. Policy-makers however, did – to some extent – draw upon personal as well as professional experiences and did not necessarily find the budget pie task any easier than did members of the public. Policy-makers were more inclined to give equal or similar weight across the attributes than were members of the public, possibly reflecting the difficulty of distilling the diversity of cases they have observed into a “one size fits all” value.

Quantitatively, there was negligible difference in the values reported by the public following discussion, as compared to their initial values. There are relatively few cases of deliberation being used to elicit numerical values that we are aware of, but limited impact of deliberation on numerical values was also identified in previous research by Karimi et al., who elicited time trade-off values before and after deliberation within focus groups (Karimi, Brazier, and Paisley Citation2019). There were significant differences between this research and that reported by Karimi et al., in the sense that a relatively simple task was used here, with greater transparency within the focus groups with respect to the values, and in the sense that whereas changes were observed in the responses of 75% of participants in the study by Karimi et al. (which essentially “averaged out” at an aggregate level, hence the overall impact being negligible), here only 13.5% of the public changed their responses on the budget pie task.

Hence, whilst it has been reported that participants are willing to publicly acquiesce or compromise to reach a group decision (Delli Carpini, Cook, and Jacobs Citation2004), few of them changed their private (numerical) values in this study, where there was no pressure to reach a group decision. As well as the lack of pressure to publicly acquiesce, the negligible quantitative impact of the specific discussion of differing values may reflect the fact that exposure to the views and experiences of others during the opening group discussion may already have begun to influence the views of participants before they were presented with the valuation tasks. It is even possible that the views and values of respondents were influenced through discussions with friends, family members and colleagues prior to attending the group, given that the topic of the focus group was known to respondents in advance of attending. It should therefore be acknowledged that, whilst it is possible to quantify the difference between initial responses (respondents’ first attempt at the task) and final responses (following discussion), this is an imperfect approximation of the impact of deliberation. Alternatively, it is possible that the minimal time available within the focus groups is inadequate for impactful deliberation.

Karimi et al. partly attribute the lack of systematic change in aggregated values elicited during their focus groups to the fact that discussion in their focus groups did not result in new information being generated. Some participants in our study made significant personal disclosures, which did appear to resonate with other participants; it prompted them to consider the topics from alternative perspectives (most noticeably in Group C), and perhaps to some extent made the hypothetical tasks seem more real. These significant personal disclosures suggest a level of comfort and trust amongst participants.

Although participants shared their own values with the group, they were not presented with summary statistics, such as a mean (this aggregate level analysis was undertaken retrospectively), and this was a limitation in terms of participants having a clear metric with which to judge how their own values compared to those of others. One potential limitation of the simple way in which the tasks were presented was that attributes were presented in the same order for all participants and all groups, and that the budget pie task presented participants with approximately equal sized segments for each attribute. This may have introduced bias as either participants may have spent more time considering attributes presented first (perhaps even implying importance from their ordering on the task), or been more inclined to give equal weight across attributes when presented with equal segments in the budget pie. Indeed, both public and policy-maker participants gave most weight to communication and practical support attributes, which were considered first, although qualitative analysis reveals considered reasons for this. The weighting assigned to different attributes in the budget pie task differed considerably for the public groups.

We do not recommend combining weights from the public with those from policy-makers as the latter sample only completed the budget pie task to weight the attributes, and because there were statistically significant differences between the means of the two samples.

Whilst it was possible to identify themes through the qualitative analysis, the themes only reflect views that participants chose to verbalise within the group and where there was similarity in the responses given by participants they often felt little need to justify their responses.

Recruitment proved to be challenging and the limited sample size and the over-representation of older and female participants, as well as those living in rural, low deprivation areas is a clear limitation of this research. Difficulty with recruitment may reflect the sensitive nature of the topic of end of life care, that recruitment followed a single route, involving only the mailing out of invitation letters, and that no financial incentive was offered to participants. A multi-pronged recruitment strategy involving online and media forums and/or an element of personal interaction could be explored in future. There is a trade-off between enabling sufficient time for meaningful discussion and requiring participants to give up significant amounts of their time. Generous financial or other incentives are likely to be necessary to achieve acceptable sample sizes that are representative of the wider population.

It was seen in this study that participants participated in a variety of ways. Achieving sample sizes that are sufficiently large will be important so as to include participants with diverse views and experiences but may also be important in terms of including diverse character traits. Some participants contributed to creating a “safe space” by acknowledging the views of others, as well as differences of opinion, by seeking clarity in relation to the tasks and in prompting others to contribute to the discussion.

Other ICECAP measures have been valued using Best-Worst Scaling or DCE. These methods are typically complex, trading off multiple attributes simultaneously. These are challenging methods to facilitate in terms of “paper and pen” completion, and usually rely on individuals making choices that would impact upon their own welfare. The simpler VAS and budget pie tasks that we used here elicited individual values and separated the estimation of the weights given to the attributes from the weights given to the levels. There was, however, perhaps more scope for respondents’ responses to reflect societal considerations. Group discussion will have exposed respondents (to some extent) to the views and experiences of others, as well as potentially prompting them to consider attributes and the meaning of attributes more carefully than they would have if completing the tasks individually (for example, online). It is unclear whether there would be value in facilitating group discussion prior to respondents then completing a DCE or BWS task individually; given that it might be difficult for respondents to reflect societal considerations within individual pairwise choices, such face to face discussion may be an expensive way of prompting respondents to consider the topic of focus, as well as significantly restricting sample sizes. This type of deliberative valuation may also be most effective when valuation methods are used which do allow respondents to express a more societal viewpoint.

Conclusion

A provisional scoring system has been developed for ICECAP-CPM, a measure designed to capture the ability of individuals to provide care for and to spend time with a dying loved one, in a safe and appropriate environment, with appropriate emotional and practical support and without being subjected to experiences which they find distressing. Mean weights from both members of the public and policy-makers indicate that two of the six attributes, relating to communication and practical support are of particular importance. A range of personal and professional experiences were shared within the focus groups, and may have influenced participants’ responses, although there is limited scope to capture this, with quantitative differences between pre and post-discussion responses on the valuation tasks being negligible.

For future work, researchers will need to consider the methods used for definitive value elicitation (and the extent to which these reflect societal considerations or individual welfare maximising decisions), the value of deliberation in relation to the chosen elicitation methodology and the impact on sample sizes of using deliberative valuation.

Supplemental material

Supplementary_Materials.docx

Download MS Word (66.7 KB)

Acknowledgements

We would like to thank all those who generously gave their time to participate in the focus groups. Pauline Mutimer provided valuable administrative support to the project.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data availability statement

Anonymised quantitative data can be requested from [email protected].

Additional information

Funding

This work was supported by the European Research Council [grant number 261098 EconEndLife].

Notes on contributors

Philip Kinghorn

Philip Kinghorn, Dr, is a Senior Lecturer at the University of Birmingham. His primary research interests relate to the conceptualisation, assessment and valuation of outcomes for inclusion within economic analysis. Philip has experience of conducting research in the contexts of palliative care, chronic pain and social care. Philip has particular interest in the application of the Capability Approach, and the use of qualitative methods within health economics.

Alastair Canaway

Alastair Canaway, Dr, is a health economist at the Centre for Health Economics at Warwick. He is experienced with working with both quantitative and qualitative data, and is interested in research in methodologically challenging areas including: end of life care, mental health, and child outcome measurements.

Cara Bailey

Cara Bailey, Dr, is a Senior Lecturer in the School of Nursing at the University of Birmingham. Her research focuses on the end of life care with a particular interest in the quality of care and patient outcomes. She has experience in mixed methods research with expertise in qualitative research methods with people who are at the end of life, seriously ill, and bereaved.

Hareth Al-Janabi

Hareth Al-Janabi is a Professor of Health Economics at the University of Birmingham. His research interests include the measurement of wellbeing and preferences, family (informal) care, the economics of mental health, and research methodology in health economics. Hareth uses a wide variety of research methods, including qualitative approaches (such as in-depth interviews, meta-ethnography, and cognitive interviews) and quantitative approaches (such as preference elicitation and econometrics).

Joanna Coast

Joanna Coast is Professor in the Economics of Health & Care at the University of Bristol. Jo’s research interests lie in the theory underlying economic evaluation, developing capability measures of outcome for use in economic evaluation, priority setting, end-of-life care and the economics of antimicrobial resistance. She also has a methodological interest in the use of qualitative methods in health economics.

References

  • Al-Janabi, H., T. Flynn, and J. Coast. 2012. “Development of a Self-Report Measure of Capability Wellbeing for Adults: The ICECAP-A.” Quality of Life Research 21 (1): 167–176.
  • Canaway, A. G., H. Al-Janabi, P. Kinghorn, C. Bailey, and J. Coast. 2016. “Development of a Measure (ICECAP-Close Person Measure) Through Qualitative Methods to Capture the Benefits of End-of-Life Care to Those Close to the Dying for Use in Economic Evaluation.” Palliative Medicine. doi:10.1177/0269216316650616.
  • Carman, K., C. Mallery, M. Maurer, G. Wang, S. Garfinkel, M. Yang, D. Gilmore, and A. Windham. 2015. “Effectiveness of Public Deliberation Methods for Gathering Input on Issues in Health: Results from a Randomized Trial.” Social Science & Medicine 133: 11–20.
  • Coast, J. 2009. “Maximisation in Extra-Welfarism: A Critique of the Current Position in Health Economics.” Social Science & Medicine 69: 786–792.
  • Coast, J. 2014. “Strategies for the Economic Evaluation of End of Life Care: Making a Case for the Capability Approach.” Expert Review of Pharmacoeconomics & Outcomes Research 4: 473–482.
  • Coast, J., C. Bailey, A. Canaway, and P. Kinghorn. 2021. ““It Is not a Scientific Number It Is Just a Feeling”: Populating a Multi-Dimensional End-of-Life Decision Framework Using Deliberative Methods.” Health Economics 30 (5): 1033–1049.
  • Coast, J., and L. Jackson. 2017. “Understanding Primary Data Analysis.” In Qualitative Methods for Health Economics, edited by J. Coast, 92–118. London: Rowman & Littlefield.
  • Culyer, A. J. 1989. “The Normative Economics of Health Care Finance and Provision.” Oxford Review of Economic Policy 5 (1): 34–58.
  • Delli Carpini, M., F. L. Cook, and L. R. Jacobs. 2004. “Public Deliberation, Discursive Participation, and Citizen Engagement: A Review of the Empirical Literature.” Annual Review of Political Science 7: 315–344.
  • Drummond, M., M. Sculpher, G. W. Torrance, B. O’Brien, and G. Stoddart. 2005. Methods for the Economic Evaluation of Health Care Programmes. 3rd ed. New York: Oxford University Press.
  • Glaser, B. G., and A. L. Strauss. 1968. The Discovery of Grounded Theory: Strategies for Qualitative Research. London: Weidenfeld and Nicolson.
  • Grewal, I., J. Lewis, T. Flynn, J. Brown, J. Bond, and Joanna Coast. 2006. “Developing Attributes for a Generic Quality of Life Measure for Older People: Preferences or Capabilities?” Social Science & Medicine 62: 1891–1901.
  • Huynh, E., J. Coast, J. Rose, P. Kinghorn, and T. Flynn. 2017. “Values for the ICECAP-Supportive Care Measure (ICECAP-SCM) for use in Economic Evaluation at end of Life.” Social Science & Medicine (1982) 189: 114–128.
  • Jacobs, M. 1997. “Environmental Valuation, Deliberative Democracy And Public Decision-Making Institutions.” In Valuing Nature: Economics, Ethics and Environment, edited by J. Foster, 211–231. London: Routledge.
  • Karimi, M., J. Brazier, and S. Paisley. 2019. “Effect of Reflection and Deliberation on Health State Values: A Mixed-Methods Study.” Value in Health 22 (11): 1311–1317.
  • Kinghorn, P. 2015. “Exploring Different Interpretations of the Capability Approach in a Health Care Context: Where Next?” Journal of Human Development and Capabilities 16 (4): 600–616.
  • Kinghorn, P., and J. Coast. 2019. “Appropriate Frameworks for Economic Evaluation of End of Life Care: A Qualitative Investigation with Stakeholders.” Palliative Medicine. doi:10.1177/0269216319839635.
  • NICE. 2013. “Guide to the Methods of Technology Appraisal 2013.” Process and Methods Guides, April 4. http://publications.nice.org.uk/pmg9.
  • NICE. 2017. “End of Life Care for Adults.” Accessed October 2. https://www.nice.org.uk/guidance/qs13.
  • Normand, C. E. 2009. “Measuring Outcomes in Palliative Care: Limitations of QALYs and the Road to PalYs.” Journal of Pain and Symptom Management 38 (1): 27–31.
  • Ruger, J. 2009. Health and Social Justice, p81. New York: Oxford University Press.
  • Saunders, C. 1989. “Pain and Impending Death.” In Textbook of Pain, edited by P. Wall and R. Melzak, 472–478. Churchill Livingstone.
  • Sen, A. 1999. Commodities and Capabilities. New Delhi: Oxford University Press.
  • Strauss, S., and J. Corbin. 1990. Basics of Qualitative Research. Grounded Theory Procedures and Techniques. London: Sage.
  • Stroebe, M., H. Schut, and W. Stroebe. 2007. “Health Outcomes of Bereavement.” Lancet 370 (9603): 1960–1973.
  • Sutton, E., and J. Coast. 2014. “Development of a Supportive Care Measure for Economic Evaluations of End-of-Life Care Using Qualitative Methods.” Palliative Medicine 28 (2): 151–157.