ArticlePDF Available

Development of a brief, generic, modular resource-use measure (ModRUM): piloting with patients

Authors:

Abstract and Figures

Background Bespoke self-report resource-use measures (RUMs) are commonly developed or adapted for each new randomised controlled trial. Consequently, RUMs lack standardisation and validation is rarely conducted. A new generic RUM, ModRUM, has been developed using a rigorous process, including consultation with health economists and patients. ModRUM includes a concise core healthcare module, designed to be included in all trials, and depth-adding questions, which can replace or be added to core questions as needed. Modules covering other sectors are under development. The aim of this study was to test the acceptability, feasibility, and criterion and construct validity of the healthcare module of ModRUM. Methods Patients who had a recent appointment at their GP practice were invited to complete ModRUM (core module or core module with depth questions), a characteristics form and the EQ-5D-5L. Acceptability was assessed via response rates and questionnaire completion time. Feasibility was assessed by reviewing issues observed in participants’ responses and question completion rates. Construct validity was tested via hypothesis testing and known-group analyses, using Wilcoxon rank-sum and Kruskal–Wallis tests, and a generalised linear model. Criterion validity was tested by comparing ModRUM results with primary care medical records. Sensitivity, specificity, and agreement using Lin’s concordance correlation coefficient (p c ) were estimated. Results One hundred patients participated from five GP practices in the South-West of England. Acceptability was higher for the core module (20% versus 10% response rate). Question completion rates were high across both versions (> 90%). Some support was observed for construct validity, with results suggesting that healthcare costs differ dependent on the number of long-term conditions ( p < 0.05) and are negatively associated with health-related quality of life ( p < 0.01). Sensitivity was high for all questions (> 0.83), while specificity varied (0.33–0.88). There was a good level of agreement for GP contacts and costs, and prescribed medication costs (p c > 0.6). Conclusion This study provided preliminary evidence of the acceptability, feasibility, and criterion and construct validity of ModRUM. Further testing is required within trials and with groups that were less well represented in this study.
Content may be subject to copyright.
Gareldetal. BMC Health Services Research (2023) 23:994
https://doi.org/10.1186/s12913-023-10011-x
RESEARCH
Development ofabrief, generic, modular
resource-use measure (ModRUM): piloting
withpatients
Kirsty Garfield1,2*, Joanna C. Thorn1,2, Sian Noble1,2, Samantha Husbands1,2 and Will Hollingworth1,2
Abstract
Background Bespoke self-report resource-use measures (RUMs) are commonly developed or adapted for each
new randomised controlled trial. Consequently, RUMs lack standardisation and validation is rarely conducted. A new
generic RUM, ModRUM, has been developed using a rigorous process, including consultation with health economists
and patients. ModRUM includes a concise core healthcare module, designed to be included in all trials, and depth-
adding questions, which can replace or be added to core questions as needed. Modules covering other sectors are
under development. The aim of this study was to test the acceptability, feasibility, and criterion and construct validity
of the healthcare module of ModRUM.
Methods Patients who had a recent appointment at their GP practice were invited to complete ModRUM (core
module or core module with depth questions), a characteristics form and the EQ-5D-5L. Acceptability was assessed
via response rates and questionnaire completion time. Feasibility was assessed by reviewing issues observed in par-
ticipants’ responses and question completion rates. Construct validity was tested via hypothesis testing and known-
group analyses, using Wilcoxon rank-sum and Kruskal–Wallis tests, and a generalised linear model. Criterion validity
was tested by comparing ModRUM results with primary care medical records. Sensitivity, specificity, and agreement
using Lin’s concordance correlation coefficient (pc) were estimated.
Results One hundred patients participated from five GP practices in the South-West of England. Accept-
ability was higher for the core module (20% versus 10% response rate). Question completion rates were high
across both versions (> 90%). Some support was observed for construct validity, with results suggesting that health-
care costs differ dependent on the number of long-term conditions (p < 0.05) and are negatively associated
with health-related quality of life (p < 0.01). Sensitivity was high for all questions (> 0.83), while specificity varied
(0.33–0.88). There was a good level of agreement for GP contacts and costs, and prescribed medication costs (pc > 0.6).
Conclusion This study provided preliminary evidence of the acceptability, feasibility, and criterion and construct
validity of ModRUM. Further testing is required within trials and with groups that were less well represented in this
study.
Keywords Resource-use measurement, Self-report, Questionnaire development, Questionnaire validation
Open Access
© The Author(s) 2023. Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which
permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the
original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or
other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line
to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory
regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this
licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecom-
mons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
BMC Health Services Research
*Correspondence:
Kirsty Garfield
kirsty.garfield@bristol.ac.uk
Full list of author information is available at the end of the article
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 2 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
Introduction
Within trial-based economic evaluations, participants
are often required to self-report their use of resources
in resource-use measures (RUMs). Self-report is a prag-
matic approach that allows a wide range of resource-use
data to be collected relatively quickly and cheaply [1, 2].
Despite self-report RUMs being a popular approach,
there is currently no standardised generic RUM that
is relevant and well-utilised in a wide range of trials [2,
3]. Instead, for each new trial, researchers tend to adapt
existing or design bespoke RUMs [3]. is leads to a lack
of standardisation, which inhibits the comparability of
results across trials, which is important when a primary
objective of economic evaluation is to inform resource
allocation decisions [4, 5]. In addition, within the scope
of a trial, it is unlikely that the measurement properties
of a bespoke or adapted RUM, including validity and
acceptability, will be assessed prior to administration
with trial participants.
Recently, a new modular RUM (ModRUM) has been
developed, which is designed for use in a wide range of
trials. ModRUM includes a core healthcare module, to be
collected in all trials, with optional depth questions that
can replace or be added to core questions when more
detailed information is required for increased precision
in cost estimates or when broader healthcare items are
relevant (e.g. paramedic care). e items included in the
healthcare module were informed by a Delphi consen-
sus study with health economists, where they identified
ten core items that should be collected in all trial-based
economic evaluations [4]. e face and content validity
of ModRUM were then assessed in qualitative interviews
with health economists. Once the content of ModRUM
was deemed valid by health economists, the acceptability
and content validity of ModRUM were assessed in quali-
tative ‘think-aloud’ interviews with patients recruited
from primary care [6]. ModRUM was revised based on
findings to improve comprehensibility [6]. ModRUM can
be adapted so that it is relevant to a range of trials of dif-
ferent health conditions (e.g. examples can be changed,
and/or items pertinent to the trial population can be
added). e feasibility of adapting ModRUM was tested
by health economists who adapted ModRUM to hypo-
thetically use it for a recently funded trial.
Once the content and face validity of an instrument
have been assessed, the remaining measurement proper-
ties, including feasibility, acceptability, construct validity
and criterion validity, can be tested in a larger quantita-
tive study [7]. e feasibility of a new instrument requires
the instrument to be viable for patients to complete
and for researchers to administer and analyse [8], while
the acceptability assesses whether the instrument is
well-received by patients [9]. Construct validity can be
established through hypothesis testing to assess whether
the association between scores from the instrument cor-
relate as expected to another instrument measuring the
same or a related construct, or to a patient characteris-
tic which is hypothesised to be associated with the con-
struct of interest [8]. To test the criterion validity of a
new instrument, the scores from the new instrument are
compared with the scores of another measure, which is
ideally the ‘gold-standard’ measure [8].
is paper reports on a study where a paper-based ver-
sion of ModRUM was piloted with patients recruited
from primary care. e aims of this study were to assess
the feasibility, acceptability, and construct and criterion
validity of ModRUM.
Methods
Data collection
GP practices based in the Bristol, North Somerset or
South Gloucestershire regions of England were invited
to participate in this study. GP practices were selected
to represent a range of deprivation scores and areas. GP
practice staff identified, screened and sent postal invita-
tions to eligible patients. Patients were eligible to take
part if they were aged 18 or over, capable of understand-
ing and completing a questionnaire in English, capable of
giving informed consent, and they had had an appoint-
ment (face-to-face or remote) with a member of the clini-
cal team (e.g. GP, nurse) within the previous four weeks.
GP practices were assigned to either send ModRUM
core module (labelled ModRUM-C, hereinafter) (Fig.1),
or ModRUM core with depth questions (labelled Mod-
RUM-CD, hereinafter) (Additional file 1: Figure S1).
All questions referred to a three-month recall period,
which represents a commonly used recall period in tri-
als [10]. e target was 800 invitations. To account for
potentially lower response rates, more invitations were
sent from practices sending ModRUM-CD (n = 450) and
from practices that were rated as more deprived (n = 520)
[11]. Patients were also asked to complete the EQ-5D-5L
and a patient characteristics form, and to self-report
how long it took them to complete ModRUM [12]. e
EQ-5D-5L was collected to assess construct validity and
increase external validity of the study, as the EQ-5D-5L
is commonly completed alongside RUMs in RCTs due to
it being the preferred health-related quality of life meas-
ure in adults by the National Institute of Health and
Care Excellence [13]. Patients who wished to participate
were asked to complete and return the documents, and
a consent form, in a pre-paid return envelope. At least
eight weeks following completion of ModRUM, data on
primary care consultations and prescribed medications
were extracted from primary care medical records by GP
practice staff.
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 3 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
Data analysis
All analyses were conducted in Stata 17 [14]. Self-
reported and medical record resource-use data were
cleaned and costed using appropriate national unit costs
(Additional file1: TableS1) [15, 16]. All unit costs were
for the year 2019. Where unit costs were not available
for 2019, past costs were inflated to 2019 prices using the
NHS cost inflation index [15]. Utility values were esti-
mated from EQ-5D-5L scores using a validated mapping
function from the EQ-5D-3L [17, 18].
Participant acceptability was assessed using question-
naire response rates and participant-reported completion
time. e impact of GP practice deprivation level and
ModRUM version on the response rate was also consid-
ered using logistic regression. Participant feasibility was
assessed using question completion rates and by review-
ing issues participants experienced in answering Mod-
RUM questions.
Construct validity was assessed via hypothesis testing
including known-group analyses [8, 19]. e following
hypotheses were tested: older participants have higher
total healthcare costs than younger participants [20],
participants with more long-term conditions have higher
total healthcare costs than participants with no or one
long-term condition [20, 21] and participants with lower
self-reported quality-of-life have higher total healthcare
costs than those with higher self-reported quality-of-life
[22]. Potential associations were also explored for sex,
age on leaving full time education and GP practice dep-
rivation level. Known-group validity was assessed using
Wilcoxon rank-sum and Kruskal–Wallis H tests [23]. As
ModRUM version was not controlled for within these
tests, total cost estimates were based on the core ques-
tions, which were asked of all respondents. A general-
ised linear model (GLM), with identity link function and
gamma distribution to account for the positively skewed
distribution of costs, was employed to assess the relation-
ship between quality-of-life scores and healthcare costs.
In the model, a clustered sandwich estimator was used to
obtain robust variance estimates that adjust for potential
similarity of participants within GP practices. Explana-
tory variables included ModRUM version, sex, age and
GP practice deprivation score. Multiple model specifica-
tions were considered and compared using linktest, histo-
grams, percentile plots of deviance residuals and Akaike’s
information criterion. To assess the correlation between
explanatory variables, the variance inflation factor (VIF)
was estimated for explanatory variables and a correlation
matrix was formed.
Criterion validity was assessed via sensitivity, specific-
ity, Lin’s concordance correlation coefficient (CCC) and
Bland–Altman plots [2426]. For resource use, sensitiv-
ity is the proportion of participants that report use of a
Fig. 1 ModRUM core module for piloting with patients
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 4 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
resource in their medical records, that are correctly iden-
tified as using the resource in ModRUM [23]. Specificity
is the proportion of participants who have no recorded
use of a resource in their medical records, that are cor-
rectly identified as not using the resource in ModRUM
[23]. Lin’s CCC can be used to compare continuous, non-
normally distributed data [24]. It incorporates measures
of precision (Pearson’s correlation) and accuracy and
is scaled between -1 (perfect reversed agreement) and
1 (perfect agreement) [24]. Following previous studies
assessing agreement between self-report and medical
record data, Lin’s CCC (pc) was interpreted according to
the following categories: poor (less than 0.40), fair (0.40
to 0.59), good (0.60 to 0.74) and excellent (0.75 to 1.00)
[2729].
Results
Five GP practices took part in this study. Participant-
reported data were collected between November 2020
and March 2021, and GP medical record data were
obtained between May and June 2021. 717 patients were
invited to participate, including 449 invites to complete
ModRUM-CD, and 438 invites sent to patients registered
at practices in the five deciles of deprivation considered
most deprived.
Acceptability
e response rate was higher for patients invited to
complete ModRUM-C (53 participants, 20% response
rate) than ModRUM-CD (47 participants, 10% response
rate). After controlling for practice deprivation score, for
patients who received ModRUM-C, the odds of taking
part were 1.74 times as large as for patients who received
ModRUM-CD (95% CI: 1.12 to 2.72, p = 0.014). After
controlling for ModRUM version, a one-unit improve-
ment in the deprivation level of the GP practice, meant
that the odds of patients participating increased by a fac-
tor of 1.11 (95% CI: 1.04 to 1.19, p = 0.003).
Mean and median participant-reported ModRUM
completion times were similar for both versions (Addi-
tional file1: TableS2). e maximum reported comple-
tion time of 25min was reported for ModRUM-C, which
possibly indicates that some participants included time
spent completing all documents in the mail pack. All
other times reported were 12min or less. Once the out-
lier was omitted, the mean completion time for Mod-
RUM-C reduced to 4.9min, compared with 5.7min for
ModRUM-CD; however, this did not alter the median
time, which was 5min for both versions of ModRUM.
Participant characteristics, quality oflife andresource use
Participant characteristics are presented in Additional
file1: TableS3. Most participants were of white ethnicity
(95%), 63% had at least one long term condition, 61%
were female, and 55% were aged 66 or over. e mean
EQ-5D-5L utility score for all participants was 0.750 (SD:
0.249) (Additional file1: TableS4). On average, the utility
score was slightly higher for participants who completed
ModRUM-C, than for participants who completed Mod-
RUM-CD (0.772 [SD: 0.212] versus 0.726 [SD: 0.285]).
Mean healthcare utilisation and costs are presented in
Additional file1: TableS5. In both versions of ModRUM,
remote consultations with a GP were the most commonly
used resource (ModRUM-C: 1.9 contacts, ModRUM-CD:
1.8 contacts). For most resources, the mean number of
contacts was similar across ModRUM versions, with the
exception of GP surgery contacts which was higher for
ModRUM-CD (1.18 versus 0.60). Other healthcare pro-
fessional contacts were higher for ModRUM-C; however,
once other healthcare professional and nurse contacts
were added for ModRUM-CD, the number of contacts
was similar. e mean total cost was higher for Mod-
RUM-CD (£537 (SD: £1045) versus £462 (SD: £802)). e
large standard deviation for both versions reflects that a
minority of patients had costly inpatient stays.
Feasibility
e feasibility of answering questions as intended was
demonstrated, as minimal data cleaning was required
for ModRUM-C. One participant reported only positive
answers and the unanswered questions were assumed
to be zero. For ModRUM-CD, cleaning involved mov-
ing answers that were in the incorrect position to the
relevant question (e.g. when GP contacts were reported
under other healthcare professional, but the GP question
was unanswered).
Prior to cleaning the data, question completion rates
ranged from 96 to 100% for ModRUM-C and 91 to 100%
for ModRUM-CD. One participant missed an entire page
of questions when completing ModRUM-C, while seven
participants who completed ModRUM-CD missed at
least one page. Of these seven, two participants reported
resource use which should have been reported under
missed questions (GP/nurse contacts) under the other
healthcare professional question, suggesting that they
may not have seen the questions, as opposed to missing
them intentionally. For ModRUM-CD, five participants
did not complete the tick box question, but the answer
could often be inferred from answers in the tables. Two
participants recorded remote outpatient appointments
under the face-to-face outpatient appointment ques-
tion which preceded it. Five participants either missed
the number of times a medication was prescribed, or
reported an answer in a different metric to what was
asked for.
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 5 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
Several participants who completed ModRUM-CD
reported issues with the pre-paid return envelope. e
size of the envelope provided was the only option pro-
vided by the mailing service; however, given the high-
quality paper and additional pages of ModRUM-CD, the
study documents only just fitted in the provided pre-paid
envelope. Several participants returned ModRUM-CD in
their own envelope.
e design of ModRUM means that questions that
appear in ModRUM-C are embedded in ModRUM-CD,
where for most items, the core question is the top-level
question in ModRUM-CD, with a table below to record
further details (e.g. clinic type, procedures, length of
stay). For participants who completed ModRUM-CD,
costs could be estimated using top-level questions only
(e.g. number of outpatient appointments) or using more
detailed depth questions (e.g. clinic type, tests/proce-
dures performed and reason for outpatient appointment).
Estimated costs were higher across most resources when
resources were costed using more detailed information
(Additional file 1: Table S6). e largest contributors
to this difference were hospital inpatient and day case
admissions, for which this sample included three par-
ticipants who had inpatient admissions and three partici-
pants who had day case admissions.
Construct validity
e hypothesis that older patients would have higher
total healthcare costs was not supported, as there was
no evidence of a difference in total healthcare costs by
age group (Table 1). However, in the regression analy-
sis the opposite was observed with under 66-year-olds
having higher healthcare costs than over 65 year olds
(p = 0.002) ( Table 2). ere was good evidence against
the null hypothesis that median total healthcare costs are
the same irrespective of number of long-term conditions,
which suggests total costs differ dependent on number of
long-term conditions (p < 0.05). Total healthcare costs as
estimated using ModRUM, were also negatively associ-
ated with health-related quality of life (p < 0.001); in other
words, participants with higher self-reported healthcare
costs, reported lower EQ-5D-5L scores. Total healthcare
costs were positively associated with GP practice depri-
vation score (p < 0.001), with increased healthcare costs
observed for participants registered at less deprived GP
practices.
Criterion validity
High sensitivity across all resources (> 0.83), indicated
that participants were likely to report healthcare use
when it was recorded in the medical records that they
had used the resource (Table3). e low specificity score
for GP contacts and wide confidence intervals (0.33, (95%
CI: 0.10 to 0.65)) were likely impacted by the low propor-
tion of participants who had not had a GP contact. When
compared with healthcare professional contacts, specific-
ity for prescribed medications was relatively high at 0.88
(95% CI: 0.47 to 1.00).
Mean resource use and costs were higher in ModRUM
than the medical records for GP and other healthcare
professional contacts (Table4). For GPs, the mean dif-
ference in contacts and costs were 0.4 (95% CI: 0.1 to
0.7) and £16 (95% CI: £5 to £27), respectively. For other
healthcare professionals, the mean difference in con-
tacts and costs were 0.6 (95% CI: 0.1 to 1.1) and £24 (95%
CI: £8 to £40), respectively. e estimated mean cost of
prescribed medications was 42% higher for GP medical
record data than ModRUM data. Based on Lin’s CCC,
a good level of agreement was observed between Mod-
RUM and medical records for GP contacts and costs, and
prescribed medication costs. For other healthcare profes-
sional contacts and costs, there was a poor level of agree-
ment. Based on the 95% limits of agreement, a smaller
range of differences between data sources for each indi-
vidual was observed for GP than other health care pro-
fessional contacts, while for costs, the largest range was
for prescribed medications.
Discussion
Main ndings
ModRUM was piloted with 100 patients. Despite com-
pletion times being similar, based on response rates,
Table 1 Results from rank tests to assess construct validity, by
patient characteristic
Groups n Rank sum p-value
Sex
Female 54 2379.0 0.521
Male 36 1716.0
Age group
18–30 1 29.5 0.538
31–45 14 745.5
46–55 12 596.5
56–65 15 624.5
66–75 26 1251.5
76 or over 22 847.5
Number of long-term conditions
None 36 1299.0 0.013
One 20 821.0
More than one 30 1621.0
Age on leaving full time education
16 or under 37 1597.5 0.618
17 or 18 16 620.0
19 and over 33 1523.5
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 6 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
ModRUM-C is potentially more acceptable to patients
than ModRUM-CD. High question completion rates
indicate that questions were feasible to answer. e
results of this study provide some evidence for the con-
struct and criterion validity of ModRUM for collecting
resource-use data from patients recruited in a primary
care setting. A revised version of ModRUM is available
for use under license [30].
Strengths andweaknesses ofthis study
is research presents initial testing of ModRUM.
While testing of RUMs in trials prior to administration
is not common [3], researchers will be encouraged to
conduct their own testing prior to administration in a
trial, to test ModRUM in their population and with any
adaptations they have made. Due to Covid-19, patient
recruitment could not be done face-to-face as planned.
Table 2 Results from the generalised linear regression analysis to assess construct validity
a Rescaled to increments of 0.1
b On a scale of 1 to 10, where 1 is most deprived and 10 is least deprived
n Adjusted cost (£) Mean dierence (95% CI) p-value
Marginal mean (95% CI)
Version ModRUM-C 47 639 (454 to 823)
ModRUM-CD 35 364 (187 to 541) -275 (-410 to -140) < 0.001
Sex Male 34 580 (421 to 739)
Female 48 480 (293 to 667) -100 (-180 to 31) 0.049
Age group 65 and under 40 632 (467 to 797)
Over 65 42 416 (220 to 612) -216 (-350 to -82) 0.002
Ethnic group Non-white 3 479 (244 to 715)
White 79 523 (354 to 692) 44 (-133 to 220) 0.629
Number of long-term conditions None 36 287 (243 to 331)
One 19 348 (227 to 468) 61 (-53 to 175) 0.298
More than one 27 956 (409 to 1503) 670 (87 to 1,252) 0.024
Age leaving full time education 16 or under 33 477 (337 to 617)
17 or 18 16 415 (212 to 619) -62 (-135 to 12) 0.102
19 or over 33 617 (424 to 811) 140 (40 to 241) 0.006
EQ-5D-5L scorea82 -47 (-57 to -36) < 0.001
GP practice deprivation scoreb82 22 (15 to 29) < 0.001
Table 3 Estimated sensitivity and specificity of ModRUM compared with medical record data
Sensitivity (95% CI) Specicity (95% CI)
General practitioner contacts
ModRUM
Yes No
Medical record Yes 80 2 0.98 0.33
No 8 4 (0.92 to 1.00) (0.10 to 0.65)
Other healthcare professional contacts
ModRUM
Yes No
Medical record Yes 80 2 0.84 0.55
No 8 4 (0.70 to 0.93) (0.39 to 0.70)
Prescribed medications
ModRUM
Yes No
Medical record Yes 80 2 0.97 0.88
No 8 4 (0.86 to 1.00) (0.47 to 1.00)
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 7 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
e sample was limited to 100 participants. Recruit-
ment was instead conducted via postal invitation and
the number of invitations sent was restricted by budg-
etary constraints. Acceptability was assessed through
response rates and self-reported completion time.
Response rates were higher for ModRUM-C than Mod-
RUM-CD (20% versus 10%). ese rates were consistent
with previous research on response rates to postal sur-
veys [31]. Response rates may have also been impeded
by the Covid-19 pandemic. Invitations were sent during
winter 2020/21, which included periods when England
was in national lockdown, meaning people may have
been less able or more reluctant to participate. In the
context of a trial, for which ModRUM is designed to
be used in, it is anticipated that response rates would
be considerably higher given that trial participants are
more likely to be engaged in the research. Question
completion rates were at least 96% for ModRUM-C and
91% for ModRUM-CD.
Most participants completed the patient character-
istics form and the EQ-5D-5L, which meant that con-
struct validity could be tested. Support for the validity
of ModRUM was obtained for hypotheses made regard-
ing health-related quality of life and number of long-
term conditions. Participants with lower EQ-5D-5L
scores had higher total healthcare costs (p < 0.001).
Participants with long-term conditions had higher total
healthcare costs (p = 0.012). While it was hypothesised
that older patients would have higher healthcare costs,
the opposite was found with under 66-year-olds hav-
ing higher costs (p = 0.002). It is likely that this result
was impacted by the sampling strategy, where to be
recruited to the study, patients were required to have
had a recent appointment at their GP practice. is cri-
terion means that the hypothesis, which was framed on
the general population, may not be valid for this sam-
ple. Further testing of this hypothesis is required.
With the exception of one participant who had left
their GP practice, medical record data was successfully
obtained for all participants. Assessment of criterion
validity was limited to a subset of ModRUM questions,
where corresponding data were available in the primary
care medical records. To assess the criterion validity
of questions not assessed in this study, data would have
needed to have been accessed from additional sources.
Lower values of specificity may indicate that ModRUM
is picking up resource use not captured in the medi-
cal records, or it could be a result of incorrectly report-
ing resource usage when it had not occurred within the
recall period (telescoping). is included other primary
and community-based healthcare professionals; how-
ever, these are unlikely to be comprehensively covered
in the medical records, with services such as NHS 111
captured in ModRUM, but not in medical records. With
good agreement observed for GP contacts and costs
and prescribed medication costs, and results on average
higher for GP and other health care professional contacts
and costs in ModRUM, this study provides support for
using ModRUM as an alternative to medical records for
resource-use data. e increased proportion of contacts
observed in ModRUM indicates that primary care medi-
cal records may not be a ‘gold standard’ for questions on
primary and community care in ModRUM.
It was initially planned that patients would be recruited
from GP practice waiting rooms; however, the Covid-19
pandemic meant that this was not feasible, so invitations
and study documents were sent via the post, which lim-
ited the number of invitations that could be sent. Several
strategies were adopted to maximise recruitment, such as
invitations being addressed from patients’ GP practices.
Although the response rate was comparable to previ-
ous postal surveys [31], other strategies to increase the
response rate, such as reminders, may have increased the
response rate further. However, medical record access
Table 4 Estimated agreement between ModRUM and medical record healthcare contacts and costs, by healthcare item
a Lin’s concordance correlation coecient
n ModRUM Medical records Mean dierence pca95%
limits of
agreement
Mean (SD) Mean (SD) (95% CI)
General practitioner
Contacts 94 2.8 (2.2) 2.4 (2.09) 0.4 (0.1 to 0.7) 0.693 (-2.8 to 3.6)
Cost (£) 94 80 (69) 64 (57) 16 (5 to 27) 0.602 (-92 to 124)
Other healthcare professionals
Contacts 91 1.7 (2.4) 1.1 (1.3) 0.6 (0.1 to 1.1) 0.224 (-4.0 to 5.2)
Cost (£) 91 36 (75) 13 (21) 24 (8 to 40) 0.021 (-126 to 174)
Prescribed medications
Cost (£) 44 69 (106) 97 (166) -29 (-61 to 3) 0.702 (-234 to 177)
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 8 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
restrictions meant that this would have needed to have
been led by the GP practice, which would have added
burden on the practices, and required additional funds.
As anticipated, practice deprivation level was asso-
ciated with whether a patient participated. Increasing
the number of invitations sent from practices in more
deprived practices meant there was good representa-
tion from these practices. e sample included slightly
more participants (55%) from the GP practices in the
least deprived areas and more female participants (61%).
Participants were mostly balanced across categories for
number of long-term conditions and age on leaving full
time education. e sample was predominantly of white
ethnicity (95%), meaning non-white ethnic groups were
not well represented in this pilot. Given the low propor-
tion of people of non-white ethnicity at the participating
GP practices, an alternative sampling approach may have
helped to recruit people from non-white ethnic groups
(e.g. a GP personally inviting patients to take part, and/
or sending more invites to people from non-white ethnic
groups). Equal representation was not achieved across
age groups, with only one participant recruited from
the 18 to 30 age group and 55% of participants aged 66
or over. is was expected based on the identification
process, as patients were required to have had a clinical
appointment at their GP practice in the last four weeks,
and older patients were expected to visit the GP practice
more frequently. Also, having a larger proportion of older
participants may be more representative of the ages of
people participating in many trials.
As recruitment in person was unfeasible, invitations
were sent from GP practices via a hybrid mail service.
GP practices uploaded patient details, and the mail ser-
vice printed and sent the invitation packs. ere were
issues that may have impacted response rates and ques-
tion completion rates for ModRUM-CD. First, due to the
high-quality paper and the number of pages to return,
the pre-paid envelope was almost too small. Several
patients added notes to say the envelope was too small
or they used their own envelope. Others may have com-
pleted the questionnaire, but not returned it. Second, the
documentation was sent as individual sheets. A booklet
may have improved question completion rates. For par-
ticipants who had missed entire pages of ModRUM-CD,
it was likely that the participant had not seen the ques-
tion as opposed to missing it due to being unable to recall
the data.
Comparison toexisting literature
To date, piloting of existing RUMs has been conducted to
identify issues and refine RUMs, with the aim of improv-
ing acceptability to respondents and increasing data qual-
ity [27, 3234]. Construct validity was assessed by Ness
etal. for the Multiple Sclerosis Health Resource Utiliza-
tion Survey [35]. All results were significant, including
health-related quality of life, which was negatively associ-
ated with total costs, which is consistent with the result
found in this study [35].
Many studies have compared self-report data with
medical record data. Noben et al. reviewed studies
reporting comparisons of self-report and administrative
data and concluded that the evidence did not support
one method over the other [36]. Of the six included stud-
ies that were considered to have sufficient quality, they
concluded that patients generally reported lower esti-
mates for resource use when compared with administra-
tive data, which contrasts with the results of this study
[36]. Patel etal. compared patient-report data from an
adapted version of the client-service receipt inventory
(CSRI) with GP medical record data for a random sample
of primary care patients [37, 38]. ey found no signifi-
cant difference between data sources for number of con-
tacts, with agreement, as estimated using Lin’s CCC, high
(pc = 0.756) [37]. More granular information was captured
from participants for costing, which allowed self-report
GP contacts to be costed by duration of consultation [37].
Byford etal. also used the CSRI and observed relatively
high agreement for GP contacts (pc = 0.631) [28]. A low
level of agreement for other healthcare professionals
in this study, was consistent with the findings of Byford
et al., who found low levels of agreement for practice
nurse, community psychologist and community psychiat-
ric nurse contacts (all pc < 0.350) [28]. Despite a difference
in average cost, the good level of agreement of prescribed
medication costs observed in this study (pc = 0.702), was
in contrast to existing research, where poor agreement
has been observed [27].
Implications forresearch practice
In this study, GP and other healthcare professional
contacts were higher in ModRUM than GP medical
records, suggesting that ModRUM may be more com-
prehensive for these resources. As healthcare providers
become more diverse, for economic evaluation, a vali-
dated patient-report measure, which is relatively cheap
to implement and easy to access, may be preferable to
collecting medical record data from a range of sources.
For administrative data to remain a feasible option for
economic evaluations, increased diversity of providers
means that data would ideally be obtained from a higher
level (i.e. at integrated care system level, as opposed to
individual providers).
For prescribed medications, high levels of sensitiv-
ity and specificity indicated that participant report was
generally consistent with medical records for binary
responses on whether participants had used medications.
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 9 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
While the agreement between data sources was consid-
ered good, the total cost of prescribed medications was
42% higher when costed using medical record data. is
could be the result of less accurate recall by participant
report, assuming that it is unlikely that medications
that were not prescribed would be included in medical
records. e estimated cost difference could also have
been impacted by the alternative costing approaches,
with more detailed medical record data, allowing for
increased precision in cost estimates.
Researchers should carefully consider the amount of
detail they ask participants to provide. Both the response
rate and question completion rates were higher for the
shorter version of ModRUM. Where researchers are
uncertain what level of detail is appropriate to collect,
depth questions could be included in the feasibility or
internal pilot phase of a randomised controlled trial.
Costing this information using both top-level core ques-
tions and detailed information from tables, could indi-
cate whether questions can be made more concise. If
the researcher chooses to use core questions in the main
trial, the detail provided during the internal pilot or feasi-
bility study could also be used to inform the most appro-
priate unit costs to use in the final analysis. For example,
if a large proportion of outpatient appointments are
performed in Orthopaedics, the researcher may choose
an orthopaedics unit cost to cost all appointments, as
opposed to a generic outpatient unit cost.
Unanswered questions andfuture research
Further testing of ModRUM is required in a larger sam-
ple, in trials and with groups that were not well repre-
sented in this study (non-white ethnic groups and people
aged 30 and under). Further research is underway to
increase the breadth of ModRUM, with modules cover-
ing social care and informal care. is research reports
the development of a paper-based version of ModRUM,
an electronic version is also being developed.
Conclusion
is study provides preliminary evidence for the feasibility,
acceptability, and construct and criterion validity of Mod-
RUM in a sample of patients recruited from primary care.
Future testing of ModRUM is required within trials, and
with groups that were less well-represented in this study.
Abbreviations
CCC Concordance Correlation Coefficient
CSRI Client Service Receipt Inventory
GP General practitioner
ModRUM Modular resource-use measure
ModRUM-C ModRUM core module
ModRUM-CD ModRUM core module with depth questions
RUM Resource-use measure
Supplementary Information
The online version contains supplementary material available at https:// doi.
org/ 10. 1186/ s12913- 023- 10011-x.
Additional le1: Figure S1. ModRUM core module with depth ques-
tions. TableS1. Unit costs, by healthcare resource. TableS2. Participant-
reported time to complete ModRUM. TableS3. Participant characteristics.
TableS4. EQ-5D-5L scores, by ModRUM version. TableS5. Healthcare
utilisation and costs, by ModRUM version. TableS6. Comparison of costs
for participants who completed ModRUM-CD, using information from
core and depth questions
Acknowledgements
The authors would like to thank all the participants and GP practices involved in
this study. The authors would also like to thank delegates at the HESG Leeds in
January 2022 for providing feedback on an earlier version of this manuscript.
Authors’ contributions
All authors (KG, SH, JCT, SN, WH) were involved in conception and design
of the study. KG managed the study, performed data input, cleaning and
analyses, and drafted the manuscript. All authors contributed to and reviewed
subsequent drafts and the final manuscript. All authors approved the submit-
ted manuscript.
Funding
This work is supported by the MRC Network of Hubs for Trials Methodology
Research (MR/L004933/2- PhD Award R9). JCT, WH and SN acknowledge fund-
ing from the Pecunia project (EU Horizon 2020, grant agreement No 779292).
The funding bodies had no role in the design of the study and collection,
analysis, and interpretation of data and in writing the manuscript.
Availability of data and materials
The datasets used and analysed during the current study are available from
the corresponding author upon reasonable request.
Declarations
Ethics approval and consent to participate
Ethical approval for the research was provided by South Central - Berkshire
B Research Ethics Committee (REC reference 19/SC/0244). Participants were
provided with a participant information sheet. All participants provided writ-
ten informed consent. All methods were performed in accordance with the
relevant guidelines and regulations.
Consent for publication
Not applicable.
Competing interests
The authors declare no competing interests.
Author details
1 Health Economics Bristol, Population Health Sciences, Bristol Medical School,
University of Bristol, 1-5 Whitladies Road, Bristol BS8 1NU, UK. 2 Bristol Trials
Centre, Population Health Sciences, Bristol Medical School, University of Bristol,
1-5 Whiteladies Road, Bristol BS8 1NU, UK.
Received: 31 October 2022 Accepted: 8 September 2023
References
1. van Lier L, Bosmans J, van Hout H, Mokkink L, van den Hout W, de Wit G,
et al. Consensus-based cross-European recommendations for the identi-
fication, measurement and valuation of costs in health economic evalua-
tions: a European Delphi study. Eur J Health Econ. 2018;19:993–1008.
2. Franklin M, Thorn J. Self-reported and routinely collected electronic
healthcare resource-use data for trial-based economic evaluations: the
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 10 of 10
Gareldetal. BMC Health Services Research (2023) 23:994
fast, convenient online submission
thorough peer review by experienced researchers in your field
rapid publication on acceptance
support for research data, including large and complex data types
gold Open Access which fosters wider collaboration and increased citations
maximum visibility for your research: over 100M website views per year
At BMC, research is always in progress.
Learn more biomedcentral.com/submissions
Ready to submit your research
Ready to submit your research
? Choose BMC and benefit from:
? Choose BMC and benefit from:
current state of play in England and considerations for the future. BMC
Med Res Methodol. 2019;19(8):1–13.
3. Ridyard CH, Hughes DA. Methods for the collection of resource use data
within clinical trials: a systematic review of studies funded by the UK
health technology assessment program. Value Health. 2010;13(8):867–72.
4. Thorn JC, Brookes ST, Ridyard C, Riley R, Hughes DA, Wordsworth S, et al.
Core items for a standardized resource use measure (ISRUM): expert
Delphi consensus survey. Value Health. 2018;21(6):640–9.
5. Brazier J, Ratcliffe J, Salomon J, Tsuchiya A. Measuring and valuing health
benefits for economic evaluation. Oxford: Oxford University Press; 2017.
6. Garfield K, Husbands S, Thorn JC, Noble S, Hollingworth W. Development
of a brief, generic, modular resource-use measure: qualitative interviews
with health economists. Value Health. 2020;23(1):S292–3 PNS48.
7. de Vet H, Terwee C, Mokkink L, Knol D. Measurement in medicine: a
practical guide. Cambridge: Cambridge University Press; 2011.
8. Streiner DL, Norman GR, Cairney J. Health measurement scales: a practical
guide to their development and use. 5th ed. Oxford: Oxford University
Press; 2015.
9. Fitzpatrick R, Davey C, Buxton M, Jones D. Evaluating patient-based out-
come measures for use in clinical trials: a review. Health Technol Assess.
1998;2(14):1–74.
10. Thorn J, Ridyard C, Riley R, Brookes S, Hughes D, Wordsworth S, et al.
Identification of items for a standardised resource-use measure: review of
current instruments. Trials. 2015;16(Suppl 2):O26.
11. Public Health England. National General Practice Profiles; 2019. (Available
from: https:// finge rtips. phe. org. uk/ profi le/ gener al- pract ice/ data# page/8/
page- optio ns/ map- ao-4).
12. Herdman M, Gudex C, Lloyd A, Janssen M, Kind P, Parkin D, et al. Develop-
ment and preliminary testing of the new five-level version of EQ-5D
(EQ-5D-5L). Qual Life Res. 2011;20(10):1727–36.
13. National Institute for Health and Care Excellence. NICE health technology
evaluations: the manual 2022; 2022.
14. StataCorp. Stata statistical software: release 17. College Station: StataCorp
LLC; 2021.
15. Curtis LA, Burns A. Unit costs of health & social care 2020. Canterbury:
University of Kent, PSSRU; 2020.
16. NHS England and NHS Improvement. National Schedule of NHS Costs
2018/19; 2020. (Available from: https:// www. engla nd. nhs. uk/ natio nal-
cost- colle ction/).
17. National Institute for Health and Care Excellence. Position statement on
use of the EQ-5D-5L value set for England (updated October 2019); 2019
(Available from: https:// www. nice. org. uk/ about/ what- we- do/ our- progr
ammes/ nice- guida nce/ techn ology- appra isal- guida nce/ eq- 5d- 5l).
18. van Hout B, Janssen MF, Feng Y-S, Kohlmann T, Busschbach J, Golicki D,
et al. Interim scoring for the EQ-5D-5L: mapping the EQ-5D-5L to EQ-
5D-3L value sets. Value in Health. 2012;15(5):708–15.
19. Mokkink LB, Terwee CB, Patrick DL, Alonso J, Stratford PW, Knol DL, et al.
The COSMIN study reached international consensus on taxonomy, ter-
minology, and definitions of measurement properties for health-related
patient-reported outcomes. J Clin Epidemiol. 2010;63(7):737–45.
20. Li J, Green M, Kearns B, Holding E, Smith C, Haywood A, et al. Patterns
of multimorbidity and their association with health outcomes within
Yorkshire, England: baseline results from the Yorkshire Health Study. BMC
Public Health. 2016;16:649.
21. Department of Health. Long term conditions compendium of informa-
tion. 3rd ed. London: Department of Health; 2012.
22. Sullivan P, Slejko J, Sculpher M, Ghushchyan V. Catalogue of EQ-5D scores
for the United Kingdom. Med Decis Making. 2011;31(6):800–4.
23. Kirkwood BR, Sterne JAC. Medical Statistics. 2nd ed. Malden: Blackwell
Science; 2003.
24. Lin L-K. A concordance correlation coefficient to evaluate reproducibility.
Biometrics. 1989;45:255–68.
25. Steichen TJ, Cox NJ. Concordance correlation coefficient. Stata Technical
Bulletin, StataCorp LP. 1999;8(43):35-9. https:// EconP apers. repec. org/
RePEc: tsj: stbull: y: 1999:v: 8:i: 43: sg84.
26. Bland JM, Altman DG. Statistical methods for assessing agree-
ment between two methods of clinical measurement. The Lancet.
1986;327(8476):307–10.
27. Pinto D, Robertson MC, Hansen P, Abbott JH. Good agreement between
questionnaire and administrative databases for health care use and costs
in patients with osteoarthritis. BMC Med Res Methodol. 2011;11:45.
28. Byford S, Leese M, Knapp M, Seivewright H, Cameron S, Jones V, et al.
Comparison of alternative methods of collection of service use data
for the economic evaluation of health care interventions. Health Econ.
2007;16(5):531–6.
29. Cicchetti DV. The precision of reliability, validity estimates revisited:
distinguishing between clinical and statistical significance of sample size
requirements. J Clin Exp Neuropsychol. 2001;23(5):695–700.
30. ModRUM-modular resource-use measure. www. brist ol. ac. uk/ modrum.
Accessed 23 June 2022.
31. Sahlqvist S, Song Y, Bull F, Adams E, Preston J, Ogilvie D. Effect of ques-
tionnaire length, personalisation and reminder type on response rate
to a complex postal survey: randomised controlled trial. BMC Med Res
Methodol. 2011;11:1–8.
32. Guzman J, Pelosi P, Bombardier C. Capturing health care utilization after
occupational low-back pain: development of an interviewer-adminis-
tered questionnaire. J Clin Epidemiol. 1999;52(5):419–27.
33. Beresford B, Mann R, Parker G, Kanaan M, Faria R, Rabiee P, et al. Reable-
ment services for people at risk of needing social care: the MoRe mixed-
methods evaluation. Health Serv Res Deliv. 2019;7(16):1-254.
34. Cooper NJ, Mugford M, Symmons DP, Barrett EM, Scott DG. Development
of resource-use and expenditure questionnaires for use in rheumatology
research. J Rheumatol. 2003;30(11):2485–91.
35 Ness N-H, Haase R, Kern R, Schriefer D, Ettle B, Cornelissen C, et al. The
multiple sclerosis health resource utilization survey (MS-HRS): develop-
ment and validation study. J Med Internet Res. 2020;22(3):e17921.
36. Noben CY, de Rijk A, Nijhuis F, Kottner J, Evers S. The exchangeability of
self-reports and administrative health care resource use measurements:
assessment of the methodological reporting quality. J Clin Epidemiol.
2016;74:93–106.
37. Patel A, Rendu A, Moran P, Leese M, Mann A, Knapp M. A comparison of
two methods of collecting economic data in primary care. Fam Pract.
2005;22(3):323–7.
38. Beecham J, Knapp M. Measuring mental health needs. In: Thornicroft G,
editor. Costing psychiatric interventions. 2nd ed. London: Gaskell; 2001. p.
200–24.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in pub-
lished maps and institutional affiliations.
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
1.
2.
3.
4.
5.
6.
Terms and Conditions
Springer Nature journal content, brought to you courtesy of Springer Nature Customer Service Center GmbH (“Springer Nature”).
Springer Nature supports a reasonable amount of sharing of research papers by authors, subscribers and authorised users (“Users”), for small-
scale personal, non-commercial use provided that all copyright, trade and service marks and other proprietary notices are maintained. By
accessing, sharing, receiving or otherwise using the Springer Nature journal content you agree to these terms of use (“Terms”). For these
purposes, Springer Nature considers academic use (by researchers and students) to be non-commercial.
These Terms are supplementary and will apply in addition to any applicable website terms and conditions, a relevant site licence or a personal
subscription. These Terms will prevail over any conflict or ambiguity with regards to the relevant terms, a site licence or a personal subscription
(to the extent of the conflict or ambiguity only). For Creative Commons-licensed articles, the terms of the Creative Commons license used will
apply.
We collect and use personal data to provide access to the Springer Nature journal content. We may also use these personal data internally within
ResearchGate and Springer Nature and as agreed share it, in an anonymised way, for purposes of tracking, analysis and reporting. We will not
otherwise disclose your personal data outside the ResearchGate or the Springer Nature group of companies unless we have your permission as
detailed in the Privacy Policy.
While Users may use the Springer Nature journal content for small scale, personal non-commercial use, it is important to note that Users may
not:
use such content for the purpose of providing other users with access on a regular or large scale basis or as a means to circumvent access
control;
use such content where to do so would be considered a criminal or statutory offence in any jurisdiction, or gives rise to civil liability, or is
otherwise unlawful;
falsely or misleadingly imply or suggest endorsement, approval , sponsorship, or association unless explicitly agreed to by Springer Nature in
writing;
use bots or other automated methods to access the content or redirect messages
override any security feature or exclusionary protocol; or
share the content in order to create substitute for Springer Nature products or services or a systematic database of Springer Nature journal
content.
In line with the restriction against commercial use, Springer Nature does not permit the creation of a product or service that creates revenue,
royalties, rent or income from our content or its inclusion as part of a paid for service or for other commercial gain. Springer Nature journal
content cannot be used for inter-library loans and librarians may not upload Springer Nature journal content on a large scale into their, or any
other, institutional repository.
These terms of use are reviewed regularly and may be amended at any time. Springer Nature is not obligated to publish any information or
content on this website and may remove it or features or functionality at our sole discretion, at any time with or without notice. Springer Nature
may revoke this licence to you at any time and remove access to any copies of the Springer Nature journal content which have been saved.
To the fullest extent permitted by law, Springer Nature makes no warranties, representations or guarantees to Users, either express or implied
with respect to the Springer nature journal content and all parties disclaim and waive any implied warranties or warranties imposed by law,
including merchantability or fitness for any particular purpose.
Please note that these rights do not automatically extend to content, data or other material published by Springer Nature that may be licensed
from third parties.
If you would like to use or distribute our Springer Nature journal content to a wider audience or on a regular basis or in any other manner not
expressly permitted by these Terms, please contact Springer Nature at
onlineservice@springernature.com
... 5. Resource use. The ModRum questionnaire will be used to evaluate the use of healthcare services, providing basic health economics information of the intervention [43]. ...
Article
Full-text available
Falls among older adults pose a significant public health challenge, as they lead to severe outcomes such as fractures and loss of independence. Research has shown that training cognitive function and balance simultaneously, termed Dual-Task (DT) training, improves mobility and reduces fall risks in older adults. This study aims to evaluate the feasibility and acceptability of a blended supervised and self-directed technology-based DT training programme for older adults who have high risk of falling. This is a single-arm, non-randomised feasibility study employing quantitative and qualitative methods. Fifty healthy adults aged 65 years or above will be recruited from the NHS primary and secondary care pathways and from the community. Participants will undergo supervised cognitive and balance DT training for 12 weeks, followed by self-directed DT training for an additional 12 weeks. The cognitive training will be delivered using a commercial mobile application (app) available from the AppStore or Google Play. The balance training will involve static (Marching on the spot, Tandem Stand, Hip Abduction & Extension, Squats, Tiptoe Stand, and Pendulum/Sideways Sway) and dynamic (Figure of Eight Walk, Walking Forwards and Backwards, Lunges, Functional Reach, Toe Tapping, Upper Limb Strength Exercises, and Side-Steps/Simple Grapevine) exercises focused on improving balance, postural stability and strength. Feasibility outcomes will be recruitment, adherence, usage of the app, and attrition. Outcomes measure data, that will be collected at baseline and at 24 weeks, includes the Timed- Up and Go (TUG) test (likely primary outcome in any future trial), along with self-reported questionnaires assessing cognition, fear of falling, quality of life, healthcare service usage, and the self-reported number of falls. Focus group interviews will be conducted with thirty participants and thirty healthcare professionals for in-depth exploration of the feasibility and acceptability of the DT training programme.
Article
Full-text available
Background: Survey-based studies are frequently used to describe the economic impact of multiple sclerosis (MS). However, there is no validated health resource survey available, preventing comparison of study results and meaningful conclusions regarding the efficiency of long-term treatments. Objective: The aim of this study was to develop and validate a tablet- and paper-based MS health resource utilization survey. Methods: We developed and validated the Multiple Sclerosis Health Resource Utilization Survey (MS-HRS), consisting of 24 cost items for paper and tablet users. Data for validation came from two large German observational studies. Survey practicability was assessed according to the response rate. Reliability was described using test-retest reliability as well as Guttman lambda. Construct validity was assessed as convergent and discriminant validity via correlations with associated patient-reported outcomes and known-group analyses. Results: In total, 2207 out of 2388 (response rate: 92.4%) patients completed the survey and were included to determine psychometric properties. The test-retest reliability had an intraclass correlation coefficient of 0.828 over a course of 3 months. Convergent validity analyses showed that total costs correlated positively with increased disability (r=0.411, P<.001). For discriminant validity, correlations of total costs with the Treatment Satisfaction Questionnaire for Medication ranged from −0.006 (convenience) to −0.216 (effectiveness). The mean annual cost was €28,203 (SD €14,808) (US 39,203;SDUS39,203; SD US 20,583) with disease-modifying therapies. Conclusions: The MS-HRS is a multilingual, reliable, valid, electronically available, and easy-to-administer questionnaire providing a holistic cross-sectional and longitudinal assessment of resource utilization in patients with MS.
Article
Full-text available
Background Reablement is an intensive, time-limited intervention for people at risk of needing social care or an increased intensity of care. Differing from home care, it seeks to restore functioning and self-care skills. In England, it is a core element of intermediate care. The existing evidence base is limited. Objectives To describe reablement services in England and develop a service model typology; to conduct a mixed-methods comparative evaluation of service models investigating outcomes, factors that have an impact on outcomes, costs and cost-effectiveness, and user and practitioner experiences; and to investigate specialist reablement services/practices for people with dementia. Methods Work package (WP) 1, which took place in 2015, surveyed reablement services in England. Data were collected on organisational characteristics, service delivery and practice, and service costs and caseload. WP2 was an observational study of three reablement services, each representing a different service model. Data were collected on health (EuroQol-5 Dimensions, five-level version) and social care related (Adult Social Care Outcomes Toolkit – self-completed) quality of life, practitioner (Barthel Index of Activities of Daily Living) and self-reported (Nottingham Extended Activities of Daily Living scale) functioning, individual and service characteristics, and resource use. They were collected on entry into reablement ( n = 186), at discharge ( n = 128) and, for those reaching the point on the study timeline, at 6 months post discharge ( n = 64). Interviews with staff and service users explored experiences of delivering or receiving reablement and its perceived impacts. In WP3, staff in eight reablement services were interviewed to investigate their experiences of reabling people with dementia. Results A total of 201 services in 139 local authorities took part in the survey. Services varied in their organisational base, their relationship with other intermediate care services, their use of outsourced providers, their skill mix and the scope of their reablement input. These characteristics influenced aspects of service delivery and practice. The average cost per case was £1728. Lower than expected sample sizes meant that a comparison of service models in WP2 was not possible. The findings are preliminary. At discharge (T1), significant improvements in mean score on outcome measures, except self-reported functioning, were observed. Further improvements were observed at 6 months post discharge (T2), but these were significant for self-reported functioning only. There was some evidence that individual (e.g. engagement, mental health) and service (e.g. service structure) characteristics were associated with outcomes and resource use at T1. Staff’s views on factors affecting outcomes typically aligned with, or offered possible explanations for, these associations. However, it was not possible to establish the significance of these findings in terms of practice or commissioning decisions. Service users expressed satisfaction with reablement and identified two core impacts: regained independence and, during reablement, companionship. Staff participating in WP3 believed that people with dementia can benefit from reablement, but objectives may differ and expectations for regained independence may be inappropriate. Furthermore, staff believed that flexibility in practice (e.g. duration of home visits) should be incorporated into delivery models and adequate provision made for specialist training of staff. Conclusions The study contributes to our understanding of reablement, and what the impacts are on outcomes and costs. Staff believe that reablement can be appropriate for people with dementia. Findings will be of interest to commissioners and service managers. Future research should further investigate the factors that have an impact on outcomes, and reabling people with dementia. Funding The National Institute for Health Research Health Services and Delivery Research programme.
Article
Full-text available
Background Randomised controlled trials (RCTs) are generally regarded as the “gold standard” for providing quantifiable evidence around the effectiveness and cost-effectiveness of new healthcare technologies. In order to perform the economic evaluations associated with RCTs, there is a need for accessible and good quality resource-use data; for the purpose of discussion here, data that best reflect the care received. Traditionally, researchers have developed questionnaires for resource-use data collection. However, the evolution of routinely collected electronic data within care services provides new opportunities for collecting data without burdening patients or caregivers (e.g. clinicians). This paper describes the potential strengths and limitations of each data collection method and then discusses aspects for consideration before choosing which method to use. Main text We describe electronic data sources (large observational datasets, commissioning data, and raw data extraction) that may be suitable data sources for informing clinical trials and the current status of self-reported instruments for measuring resource-use. We assess the methodological risks and benefits, and compare the two methodologies. We focus on healthcare resource-use; however, many of the considerations have relevance to clinical questions. Patient self-report forms a pragmatic and cheap method that is largely under the control of the researcher. However, there are known issues with the validity of the data collected, loss to follow-up may be high, and questionnaires suffer from missing data. Routinely collected electronic data may be more accurate and more practical if large numbers of patients are involved. However, datasets often incur a cost and researchers are bound by the time for data approval and extraction by the data holders. Conclusions Owing to the issues associated with electronic datasets, self-reported methods may currently be the preferred option. However, electronic hospital data are relatively more accessible, informative, standardised, and reliable. Therefore in trials where secondary care constitutes a major driver of patient care, detailed electronic data may be considered superior to self-reported methods; with the caveat of requiring data sharing agreements with third party providers and potentially time-consuming extraction periods. Self-reported methods will still be required when a ‘societal’ perspective (e.g. quantifying informal care) is desirable for the intended analysis. Electronic supplementary material The online version of this article (10.1186/s12874-018-0649-9) contains supplementary material, which is available to authorized users.
Article
Full-text available
Objectives: Differences between country-specific guidelines for economic evaluations complicate the execution of international economic evaluations. The aim of this study was to develop cross-European recommendations for the identification, measurement and valuation of resource use and lost productivity in economic evaluations using a Delphi procedure. Methods: A comprehensive literature search was conducted to identify European guidelines on the execution of economic evaluations or costing studies as part of economic evaluations. Guideline recommendations were extracted by two independent reviewers and formed the basis for the first round of the Delphi study, which was conducted among European health economic experts. During three written rounds, consensus (agreement of 67% or higher) was sought on items concerning the identification, measurement and valuation of costs. Results: Recommendations from 18 guidelines were extracted. Consensus among 26 panellists from 17 European countries was reached on 61 of 68 items. The recommendations from the Delphi study are to adopt a societal perspective, to use patient report for measuring resource use and lost productivity, to value both constructs with use of country-specific standardized/unit costs and to use country-specific discounting rates. Conclusion: This study provides consensus-based cross-European recommendations on how to measure and value resource use and lost productivity in economic evaluations. These recommendations are expected to support researchers, healthcare professionals, and policymakers in executing and appraising economic evaluations performed in international contexts.
Article
Full-text available
Background Resource use measurement by patient recall is characterized by inconsistent methods and a lack of validation. A validated standardized resource use measure could increase data quality, improve comparability between studies, and reduce research burden. Objectives To identify a minimum set of core resource use items that should be included in a standardized adult instrument for UK health economic evaluation from a provider perspective. Methods Health economists with experience of UK-based economic evaluations were recruited to participate in an electronic Delphi survey. Respondents were asked to rate 60 resource use items (e.g., medication names) on a scale of 1 to 9 according to the importance of the item in a generic context. Items considered less important according to predefined consensus criteria were dropped and a second survey was developed. In the second round, respondents received the median score and their own score from round 1 for each item alongside summarized comments and were asked to rerate items. A final project team meeting was held to determine the recommended core set. Results Forty-five participants completed round 1. Twenty-six items were considered less important and were dropped, 34 items were retained for the second round, and no new items were added. Forty-two respondents (93.3%) completed round 2, and greater consensus was observed. After the final meeting, 10 core items were selected, with further items identified as suitable for “bolt-on” questionnaire modules. Conclusions The consensus on 10 items considered important in a generic context suggests that a standardized instrument for core resource use items is feasible.
Article
Full-text available
Background: Multimorbidity is increasingly being recognized as a serious public health concern. Research into its determinants, prevalence, and management is needed and as the risk of experiencing multiple chronic conditions increases over time, attention should be given to investigating the development of multimorbidity through prospective cohort design studies. Here we examine the baseline patterns of multimorbidity and their association with health outcomes for residents in Yorkshire, England using data from the Yorkshire Health Study. Methods: Baseline data from the Yorkshire Health Study (YHS) was collected from 27,806 patients recruited between 2010 and 2012. A two-stage sampling strategy was implemented which first involved recruiting 43 general practice surgeries and then having them consent to mailing invitations to their patients to complete postal or online questionnaires. The questionnaire collected information on chronic health conditions, demographics, health-related behaviours, healthcare and medication usage, and a range of other health related variables. Descriptive statistics (chi-square and t tests) were used to examine associations between these variables and multimorbidity. Results: In the YHS cohort, 10,332 participants (37.2 %) reported having at least two or more long-term health conditions (multimorbidity). Older age, BMI and deprivation were all positively associated with multimorbidity. Nearly half (45.7 %) of participants from the most deprived areas experienced multimorbidity. Based on the weighted sample, average health-related quality of life decreased with the number of health conditions reported; the mean EQ-5D score for participants with no conditions was 0.945 compared to 0.355 for participants with five or more. The mean number of medications used for those without multimorbidity was 1.81 (range 1-13, SD = 1.25) compared to 3.81 (range 1-14, SD = 2.44) for those with at least two long-term conditions and 7.47 (range 1-37, SD = 7.47) for those with 5+ conditions. Conclusion: Patterns of multimorbidity within the Yorkshire Health Study support research on multimorbidity within previous observational cross-sectional studies. The YHS provides both a facility for participant recruitment to intervention trials, and a large population-based longitudinal cohort for observational research. It is planned to continue to record chronic conditions and other health related behaviours in future waves which will be useful for examining determinants and trends in chronic disease and multimorbidity.
Preprint
Background Survey-based studies are frequently used to describe the economic impact of multiple sclerosis (MS). However, there is no validated health resource survey available, preventing comparison of study results and meaningful conclusions regarding the efficiency of long-term treatments. Objective The aim of this study was to develop and validate a tablet- and paper-based MS health resource utilization survey. Methods We developed and validated the Multiple Sclerosis Health Resource Utilization Survey (MS-HRS), consisting of 24 cost items for paper and tablet users. Data for validation came from two large German observational studies. Survey practicability was assessed according to the response rate. Reliability was described using test-retest reliability as well as Guttman lambda. Construct validity was assessed as convergent and discriminant validity via correlations with associated patient-reported outcomes and known-group analyses. Results In total, 2207 out of 2388 (response rate: 92.4%) patients completed the survey and were included to determine psychometric properties. The test-retest reliability had an intraclass correlation coefficient of 0.828 over a course of 3 months. Convergent validity analyses showed that total costs correlated positively with increased disability (r=0.411, P<.001). For discriminant validity, correlations of total costs with the Treatment Satisfaction Questionnaire for Medication ranged from −0.006 (convenience) to −0.216 (effectiveness). The mean annual cost was €28,203 (SD €14,808) (US 39,203;SDUS39,203; SD US 20,583) with disease-modifying therapies. Conclusions The MS-HRS is a multilingual, reliable, valid, electronically available, and easy-to-administer questionnaire providing a holistic cross-sectional and longitudinal assessment of resource utilization in patients with MS.
Article
Objective: To assess the exchangeability of self-reported and administrative healthcare resource use measurements for cost estimation. Study design and setting: In a systematic review (NHS EED & MEDLINE) reviewers evaluate, in duplicate, the methodological reporting quality of studies comparing the validation evidence of instruments measuring healthcare resource use. The appraisal tool MERQ (Methodological Reporting Quality) is developed by merging aspects form the Guidelines for Reporting Reliability and Agreement Studies (GRASS) and The Standards for Reporting Diagnostic Accuracy (STARD). Results: Out of 173 studies, 35 full-text articles are assessed for eligibility. 16 articles are included in this study. In seven articles, more than 75% of the reporting criteria assessed by MERQ are considered 'good'. The majority of studies score at least 'fair' on most of the reporting quality criteria. In the end, six studies score 'good' on the minimal criteria for reporting. Varying levels of agreement among the different data sources are found, with correlations ranging from 0.14 up to 0.93 and with occurrences of both random and systematic errors. Conclusion: The validation evidence of the small number of studies with adequate methodological reporting quality cautiously supports the exchangeability of both the self-reported and administrative resource use measurement methods.
Article
Background: 'Patient-based outcome measure' is a short-hand term referring to the array of questionnaires, interview schedules and other related methods of assessing health, illness and benefits of health care interventions from the patient's perspective. Patient-based outcome measures, addressing constructs such as health-related quality of life, subjective health status, functional status, are increasingly used as primary or secondary end-points in clinical trials. Objectives: To describe the diversity and reasons for diversity of available patient-based outcome measures. To make clear that criteria investigators should have in mind when they select patient-based outcome measures for use in a clinical trial. Methods: Data sources: Literature was identified by a combination of electronic searches of databases, handsearching of selected journals and retrieval of references cited in available literature. Databases used included MEDLINE, EMBASE, CINAHL, PsychLIT and Sociofile. Study selection: A set of explicit criteria were used for selection of literature. Articles were included if they focused on any methodological aspect of patient-based outcome measures (for example, methods of evaluating such measures, psychometric evaluation of measures, comparative studies of measures, studies reporting validation of measures). Studies were excluded if they only reported use of a measure without evaluation, focused only on cross-cultural issues, focused only on clinician-based outcome measures or discussed economic utility theory only without considering measurement. A total of 5621 abstracts and articles were identified by initial searches as potentially relevant. However, after assessment, 391 key references were selected as useful to the objectives of the review. A further 22 references were incorporated into the final version as a result of comments from external experts and referees. Data synthesis: A first draft synthesising the evidence was produced by the first author of this review (RF) and extensively critiqued by the other three authors. A revised version was then submitted for evaluation to a panel of ten experts recruited to represent a wide range of areas of expertise (including clinical medicine, clinical trials, health economics, health services research, social sciences and statistics). Feedback from this panel was read and discussed by the authors of the review and a third version of the review drafted. The final version is a quasi- consensus view from individuals with a wide range of expertise. Results: Diversity of patient-based outcome measures: Seven major types of instrument can be identified in the literature: disease-specific, site-specific, dimension-specific, generic, summary item, individualised, utility. Concepts, definitions and theories of what such instruments measure are generally not clearly or consistently used. For example, there is little consistency of use or agreement as to the meaning of key terms such 'quality of life' and 'health-related quality of life'. The intended purpose and content of types of instruments vary. There are advantages and disadvantages to each of the different type of instrument when used in a particular clinical trial. Criteria for selecting patient-based outcome measures: There are eight criteria that investigators should apply to evaluate candidate patient-based outcome measures for any specific clinical trial: appropriateness, reliability, validity, responsiveness, precision, interpretability, acceptability, feasibility. These criteria are not consistently defined and the literature associated with the criteria cannot be summarised in clear, explicit and unambiguous terms. It is not possible from available evidence to rank order the relative importance of the eight criteria in relation to decisions about selection of measures to include in a trial. Appropriateness requires that investigators consider the match of an instrument to the specific purpose and questions of a trial. Reliability requires that an instrument is reproducible and internally consistent. Validity is involved in judging whether an instrument measures what it purports to measure. Responsiveness in this context addresses whether an instrument is sensitive to changes of importance to patients. Precision is concerned with the number and accuracy of distinctions made by an instrument. Interpretability is concerned with how meaningful are the scores from an instrument. Acceptability addresses how acceptable is an instrument for respondents to complete. Feasibility is concerned with the extent of effort, burden and disruption to staff and clinical care arising from use of an instrument. Conclusions and recommendations: Investigators need to make their choice of patient-based outcome measures for trials in terms of the criteria identified in this review. Developers of instruments need to make evidence available under the same headings. By means of the above criteria, further primary research and consensus-type processes should be used to evaluate leading instruments in the different fields and specialties of health care to improve use of patient-based outcome measures in research. Primary research is needed either in the form of methodological additions to substantive clinical trials (for example comparing the performance of two or more measures) or studies of leading measures with methodology as the primary rationale.