- Research article
- Open Access
- Open Peer Review
A within-subjects trial to test the equivalence of online and paper outcome measures: the Roland Morris Disability Questionnaire
© Bishop et al; licensee BioMed Central Ltd. 1900
- Received: 15 December 2009
- Accepted: 8 June 2010
- Published: 8 June 2010
Augmenting validated paper versions of existing outcome measures with an equivalent online version may offer substantial research advantages (cost, rapidity and reliability). However, equivalence of online and paper questionnaires cannot be assumed, nor can acceptability to respondents. The aim was to test whether online and written versions of the Roland Morris Disability Questionnaire (RMDQ), a standard measure of functional disability in back pain, are equivalent at both group and individual levels to establish whether they can be used interchangeably.
This is a within-participants equivalence study. 167 participants with back pain fully completed both the paper and online versions of the RMDQ in random order. Participants were recruited from a chiropractic clinic and patient support groups in Southern England. Limits of equivalence were pre-defined as 0.5 RMDQ points, the Bland-Altman range was calculated, and participants' comments were examined using content analysis.
The mean score difference was 0.03 (SD = 1.43), with the 95% Confidence Interval falling entirely within our limits of equivalence (-0.19 to 0.25). The Bland-Altman range was -2.77 to 2.83 RMDQ points. Participants identified unique advantages and disadvantages associated with each version of the RMDQ.
The group and individual level data suggest that online and paper versions of the RMDQ are equivalent and can be used interchangeably. The Bland-Altman range appears to reflect the known measurement properties of the RMDQ. Furthermore, participants' comments confirmed the potential value to be had from offering them the choice of completing the RMDQ online or on paper.
- Back Pain
- Score Difference
- Online Version
- Paper Version
- Roland Morris Disability Questionnaire
Use of the Internet for health care research has grown in the last fifteen years [1, 2]. The internet provides an alternative to postal means of administering self-completed outcome measures. Purported research advantages include improved data input accuracy (no manual data entry, immediate checks for incomplete responses, additional 'meta'-data such as duration/date of response can be logged), reduced costs (no manual data entry, no printing or postage costs), reduced response time, increased response rates, and increased ease of administration [3–7]. There are also potential benefits to participants: online questionnaires can be completed wherever the participant has access to the Internet (work, home, holiday) and there is no need to return them by post.
Trials collecting data using online questionnaires only would however exclude all those people who are not Internet users, which in 2007 was approximately 35% of the UK population . There is also the potential for higher drop-out rates due to the less personal nature of online studies and a subsequent feeling of having less involvement in the research [1, 9, 10]. Some patients may have conditions that make it difficult, painful, or even impossible to use a computer. Researchers could offer participants the choice to complete either online or paper versions of outcome measures. However, it is not safe to assume that online versions of pre-existing outcome measures will be equivalent to written versions [1, 2]. Outcome measures may have different psychometric properties when transferred online and may obtain systematically different responses compared to paper versions. Researchers recommend first testing the online method against its written counterpart and some have documented systematic differences between online and written versions of questionnaires  while others demonstrate equivalence [4, 6, 11].
The equivalence of total scores between online and paper versions of the RMDQ,
The equivalence of missing data rates, and
Participants' perceptions of completing each version.
We used a within-subjects equivalence design, in which all participants completed both versions of the RMDQ on the same day.
Participants and Procedure
Participants were recruited between January 2008 and April 2009 from a local chiropractic clinic and patient support groups (we advertised the study to people with back pain through BackCare and the National Ankylosing Spondylitis Society, both of which provide information and support to people with back pain). Eligible individuals had to have back pain (scoring ≥1 on the RMDQ), be able to access the Internet, and be at least 18 years old. After giving informed consent, participants were given a study pack containing a participant number and instructions as to which RMDQ to complete first, a paper RMDQ, instructions for completing the online RMDQ, and an additional question sheet. Order of completion was randomised using participant number -- before putting together the study packs, the complete list of participant numbers was randomised to receive instructions to complete either the online or the paper RMDQ first. This randomisation was balanced --for each group of 10 participant numbers, five were randomised to complete the online RMDQ first and five were randomised to complete the paper RMDQ first. Participants were instructed to complete both versions of the questionnaire on the same day to ensure the same underlying back pain disability when completing each version (the RMDQ asks respondents to 'think about yourself today' when answering). A laptop computer was available at the clinic for participants to use to complete the online version (but participants were free to complete either one or both versions elsewhere if they wanted to).
We used the original 24-item paper version of the RMDQ  with the response options (Yes/No instead of just Yes) used by Patrick et al. . The online version was developed using online questionnaire software (Psychosurvey) provided by the School of Psychology, University of Southampton. It was designed to mimic the format of the paper version: answers were given by selecting either a 'yes' or 'no' radio button and all questions were displayed on one page. The instructions were repeated after question 12 so they remained visible throughout. For both versions, answering Yes to a question scores 1, answering no scores 0. Items are summed to give a total score (range 0 to 24), where 24 represents the maximum disability score.
We collected standard socio-demographic data and information about the history of participants' back pain. Participants provided written responses to four open-ended questions concerning their perceptions of the advantages and disadvantages of each version of the RMDQ.
Online RMDQ scores were automatically entered into an Excel file and then transferred into SPSS (Version 16.0). Data from the paper questionnaires were entered manually. The equivalence of online and paper RMDQs was tested regarding (1) mean total score difference (2) within-subject score differences and (3) differences in responses to individual questions. These analyses thus took into account (1) individual respondents' score differences and (2) differences in responses to individual questions. Even if mean scores on two versions of an outcome measure are identical, individual respondents may have large score differences and/or there could be differences on individual items.
Analysis of mean score difference
Difference scores were calculated for each participant (difference d = paper RMDQ total score - online RMDQ total score), and shown graphically to approximately follow a Normal distribution. A paired t-test was used to calculate the mean score difference and the associated 95% confidence interval. An Analysis of Variance (ANOVA) model was used with treatment, period and patient effects to account for any differences due to the order of completing the online and paper RMDQs.
To test the hypothesis that the RMDQs were equivalent, a limit of equivalence was defined, i.e. a value for the mean score difference that needed to be exceeded to determine that the two RMDQs were not equivalent. The limit of equivalence is a somewhat arbitrary value that we set at 0.5, to provide a strict test of the equivalence of the two RMDQs: a mean score difference of up to 0.5 cannot be rounded up to a whole RMDQ point. If the entire 95% confidence interval for the mean difference falls between -0.5 and +0.5 then we can be confident that the true difference was no greater than 0.5 in either direction, which is strong evidence of equivalence given the scoring system of the RMDQ. Our sample size calculation showed that 156 participants were required to have 80% power to detect (at a 5% significance level) a mean score difference of 0.1 (assuming standard deviation = 2).
Exploratory analysis of individuals' score differences
The Bland and Altman range (illustrated on Bland-Altman plots) was used to indicate the extent to which scores on the online RMDQ are likely to differ from scores on the paper RMDQ for individual participants . This range is the difference between the lower and upper limits of agreement, where limit of agreement = mean difference ± two standard deviations of the difference. In the Bland and Altman plot, individual participants' score differences were plotted against their mean score, illustrating individual score differences and the effect of the mean score on the score difference.
We repeated our analyses of mean and individual score differences excluding participants scoring less than 5 on either or both versions of the RMDQ (i.e. participants with scores falling within the typical measurement error of the RMDQ ).
Analysis of responses to open-ended questions
Responses to the open-ended questions were entered into MS Word and subjected to a simple content analysis  in which all individual statements were categorised as being positive or negative comments about the online RMDQ or the paper RMDQ, or neutral comments. Similar comments were grouped together inductively within these overarching categories.
Clinical and Demographic Characteristics of Participants.
(n = 167)
Online then paper
(n = 76)a
Paper then online
(n = 83)
18 to 78
19 to 74
18 to 78
Missing data (n, %)
Duration of back pain (yrs)
0.02 to 50
0.04 to 40
0.02 to 37
Missing data (n, %)
Unable to work due to back pain
Missing data (n, %)
There were 15 missing responses on the online RMDQ (equivalent to .004% of items) and 3 on the paper RMDQ (.0007%); they were not concentrated on any one item. Cases with missing data were excluded from the following analyses.
Mean Score Differences
Mean (SD) scores on the RMDQs, mean difference, adjusted mean difference and 95% range of agreement.
Adjusted Mean Difference
Bland Altman 95% range of agreement
(-0.19 to 0.25)
(-0.26 to 0.18)
-2.77 to 2.83
(n = 76)
(n = 83)
Individual-level Score Differences
Excluding all cases with scores below 5 on either RMDQ did not meaningfully alter our findings. The mean score difference increased slightly (0.11, SD = 1.41) but the 95% confidence interval remained acceptable (-0.17 to 0.38); the Bland-Altman range shifted slightly to -2.65 to 2.86.
Typical comments from participants.
"Much easier to change an answer after further consideration if necessary"
"Saves walking to mailbox which can be a problem due to back pain."
"Knew what to expect"
"Can see full overview of questions"
"Sometimes navigating with the mouse was not always easy, i.e. clicking in the yes/no boxes"
"Not keen on using computers"
"I am not a great fan of working online"
"Difficult to change things if you filled in the wrong box"
"Not environmentally friendly"
Our group-level data suggests that mean group scores on the online and paper versions of the RMDQ are equivalent. The 95% Confidence Intervals for the raw mean score and the adjusted mean score difference were both well within our pre-defined limits of equivalence. While there were slightly more missing data points on the online RMDQ than the paper RMDQ the proportion of missing data was very low for both versions, and it would be possible to programme future online versions of the RMDQ to prevent people submitting a questionnaire containing missing data (although such an approach might result in fewer respondents overall). According to our content analysis of responses to open-ended questions, participants were able to identify unique positive and negative features of each version of the RMDQ. They also, in similar numbers, found each version easier and quicker than the other, and a very small number reported finding it physically painful to complete one version or the other. These findings support the value of offering research participants a choice of modality in completing outcome measures. Doing so might enhance patients' experiences of taking part in research and reduce the burden of participation.
Our Bland-Altman analysis suggested that in 95% of individuals who complete the two versions of the RMDQ, total score differences can be expected to be as great as approximately plus or minus three RMDQ points. While this appears to be a substantial difference, it is consistent with the known measurement properties of the RMDQ. The minimum level of detectable change for the RMDQ (i.e. the observable change deemed necessary to be distinguishable from measurement error) is 5 points . The Bland-Altman range identified in our study is thus likely to represent this inherent level of variability (measurement error) in the RMDQ rather than any additional systematic difference between online and paper versions. Equivalence between online and paper versions of established questionnaires is emerging as a common finding in the literature  and our results are consistent with this trend.
Our methodology is limited in some respects. Our findings could be an artefact of our crossover design; participants could have remembered their responses to the first version when completing the second version of the RMDQ. If this occurred then it would have contributed to the consistency we found across online and paper versions. We would recommend future cross-over studies recruit from a known stable population to enable a longer gap between the completion of online and paper questionnaires. Our crossover design enabled us to investigate individual-level equivalence. However, a parallel groups design (randomising participants to complete either the written or the Internet version) would offer a valuable complement, in that using a large, representative sample would control for both known and unknown confounders and eliminate the possible bias due to memory. We recommend that such a study is undertaken to strengthen the evidence base in this area. Further limitations that we acknowledge here and which should be overcome in future research include our lack of control over or monitoring of the time gap between completions of the two versions of the questionnaire, and our lack of independent verification of the order of completion of the two versions of the questionnaire.
This study provides initial evidence that an online version of the RMDQ is equivalent to its paper counterpart, and research participants would appear to value being given the choice of which version to complete. Therefore it should be possible to use the two versions interchangeably to benefit from the research advantages of online administration and to enhance participants' experiences of research. A large scale randomised study is needed to confirm these preliminary conclusions.
This work was supported by a grant from the Southampton Complementary Medicine Research Trust and the Anglo-European College of Chiropractic Research Subvention Fund. Dr Bishop is funded by a Career Development Fellowship from the Arthritis Research Campaign (grant number 18099). Professor Lewith's post is funded by a grant from the Rufford Maurice Laing Foundation. The authors would like to thank everyone who took part in this study and all who facilitated the recruitment process. The funders had no role in study design, in the collection, analysis and interpretation of data, in the writing of the manuscript, or in the decision to submit the manuscript for publication.
- Murray E: Using the internet for research: results at a keystroke. Br J Gen Pract. 2007, 57: 939-940. 10.3399/096016407782605045.View ArticlePubMedPubMed CentralGoogle Scholar
- Buchanan T, Heffernan T, Ling J, Parrott A, Rodgers J, Scholey A: Nonequivalence of on-line and paper-and-pencil psychological test: the case of the prospective memory questionnaire. Behav Res Methods. 2005, 37: 148-154.View ArticlePubMedGoogle Scholar
- Gosling SD, Vazire S, Srivastava S, John OP: Should We Trust Web-Based Studies: A Comparative Analysis of Six Preconceptions About Internet Questionnaires. Am Psychol. 2004, 59: 93-104. 10.1037/0003-066X.59.2.93.View ArticlePubMedGoogle Scholar
- Raat H, Mangunkusumo RT, Moghangoo AD, Juniper EF: Internet and written respiratory questionnaires yield equivalent results for adolescents. Pediatr Pulmonol. 2007, 42: 357-361. 10.1002/ppul.20576.View ArticlePubMedPubMed CentralGoogle Scholar
- Leece P, Bhandari M, Sprague S, Swiontkowski MF, Schemitsch EH, Tornetta P, Devereaux PJ, Guyatt GH: Internet versus mailed questionnaires: a controlled comparison (2). J Med Internet Res. 2004, 29: (6):e39-Google Scholar
- Mangunkusumo RT, Moorman PW, Van Den Berg-De Ruiter AE, Ven Der Lei J, De Koning HJ, Raat H: Internet-administered adolescent health questionnaires compared with a paper version in a randomized study. J Adolesc Health. 2005, 36 (1): 70.e1-70.e6. 10.1016/j.jadohealth.2004.02.020.View ArticleGoogle Scholar
- Bälter KA, Bälter O, Fondell E, Lagerros YT: Web-based and Mailed Questionnaires: A comparison of response rates and compliance. Epidemiology. 2005, 16: 577-579. 10.1097/01.ede.0000164553.16591.4b.View ArticlePubMedGoogle Scholar
- First Release: Internet Access, Households and Individuals 2007. 2007, Office for National Statistics, [http://www.statistics.gov.uk]
- West R, Gilsenan A, Coste F, Zhou X, Brouard R, Nonnemaker J, Curry SJ, Sullivan SD: The ATTEMPT cohort: a multi-national longitudinal study of predictors, patterns and consequences of smoking cessation; introduction and evaluation of internet recruitment and data collection method. Addiction. 2006, 101: 1352-1361. 10.1111/j.1360-0443.2006.01534.x.View ArticlePubMedGoogle Scholar
- Couper M: Web surveys: A review of issues and approaches. Public Opin Q. 2000, 64: 464-494. 10.1086/318641.View ArticlePubMedGoogle Scholar
- Ritter P, Lorig K, Laurent D, Mathews K: Internet versus mailed questionnaires: A randomised comparison. J Med Internet Res. 2004, 6: (3):e29-10.2196/jmir.6.3.e29.View ArticlePubMedPubMed CentralGoogle Scholar
- Maniadakis N, Gray A: The economic burden of back pain in the UK. Pain. 2000, 84: 95-103. 10.1016/S0304-3959(99)00187-6.View ArticlePubMedGoogle Scholar
- Roland M, Morris R: A study of the natural history of back pain. Part I: Development of a reliable and sensitive measure of disability in low-back pain. Spine. 1983, 8: 141-144. 10.1097/00007632-198303000-00004.View ArticlePubMedGoogle Scholar
- Patrick DL, Deyo RA, Atlas SJ, Singer DE, Chapin A, Keller RB: Assessing health-related quality of life in patients with sciatica. Spine. 1995, 20: 1899-1909. 10.1097/00007632-199509000-00011.View ArticlePubMedGoogle Scholar
- Deyo RA, Battie M, Beurskens AJHM, Bombardier C, Croft P, Koes B, Malmivaara A, Roland M, Von Korff M, Waddell G: Outcome measures for low back pain research: A proposal for standardized use. Spine. 1998, 23: 2003-2013. 10.1097/00007632-199809150-00018.View ArticlePubMedGoogle Scholar
- Roland M, Fairbank J: The Roland Morris Disability Questionnaire and the Oswestry Disability Questionnaire. Spine. 2000, 25: 3115-3124. 10.1097/00007632-200012150-00006.View ArticlePubMedGoogle Scholar
- Bombardier C: Outcome assessments in the evaluation of treatment of spinal disorders: summary and general recommendations. Spine. 2000, 25: 3100-3103. 10.1097/00007632-200012150-00003.View ArticlePubMedGoogle Scholar
- UK BEAM trial team: United Kingdom back pain exercise and manipulation (UK BEAM) randomised trial: effectiveness of physical treatments for back pain in primary care. BMJ. 2004Google Scholar
- Bland JM, Altman DG: Statistical methods for assessing agreement between two methods of clinical measurement. The Lancet. 1986, 1: 307-310.View ArticleGoogle Scholar
- Stratford PW, Binkley JM, Solomon P, Finch E, Gill C, Moreland J: Defining the minimum level of detectable change for the Roland-Morris questionnaire. Phys Ther. 1996, 76: 359-365.PubMedGoogle Scholar
- Joffe H, Yardley L: Content and thematic analysis. Research methods for clinical and health psychology. Edited by: Marks DF. 2004, London: Sage, 56-68.Google Scholar
- Gwaltney CJ, Shields AL, Shiffman S: Equivalence of electronic and paper-and-pencil administration of patient-reported outcome measures: A meta-analytic review. Value Health. 2008, 11: 322-333. 10.1111/j.1524-4733.2007.00231.x.View ArticlePubMedGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1471-2474/11/113/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License http://creativecommons.org/licenses/by/2.0, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.