- Research article
- Open Access
- Open Peer Review
Test-retest of computerized health status questionnaires frequently used in the monitoring of knee osteoarthritis: a randomized crossover trial
BMC Musculoskeletal Disorders volume 12, Article number: 190 (2011)
To compare data based on touch screen to data based on traditional paper versions of questionnaires frequently used to examine patient reported outcomes in knee osteoarthritis patients and to examine the impact of patient characteristics on this comparison
Participants were recruited from an ongoing trial (http://ClinicalTrials.Gov Identifier: NCT00655941). 20 female participants, mean age 67 (SD 7), completed KOOS, VAS pain, function and patient global, SF-36, Physical Activity Scale, painDETECT, and the ADL Taxonomy. Patients were randomly assigned to one of two subgroups, completing either the paper or touch screen version first. Mean, mean differences (95% CI), median, median differences and Intraclass Correlation Coefficients (ICCs) were calculated for all questionnaires.
ICCs between data based on computerized and paper versions ranged from 0.86 to 0.99. Analysis revealed a statistically significant difference between versions of the ADL Taxonomy, but not for the remaining questionnaires. Age, computer experience or education-level had no significant impact on the results. The computerized questionnaires were reported to be easier to use.
The computerized questionnaires gave comparable results to answers given on paper. Patient characteristics did not influence results and implementation was feasible.
In the Rheumatology clinic, self-administered Health Status Questionnaires (HSQs) are an important part of the overall evaluation of patients [1, 2]. Several questionnaires are applied in the self-assessment process, and data handling can be tedious, expensive and open to errors, when data are being transferred manually from paper into electronic systems. Implementation of computerized methods of data collection based on touch-screen would be more cost-effective and decrease the risk of error.
Touch screen is a new tool applied in places like libraries and shops, as well as in health care settings [3–5]. However, prior to implementation of data collection based on touch screen, it is crucial to evaluate if questionnaires based on paper and touch screen provide similar information. Furthermore, as clinical studies often include several questionnaires it is relevant to validate multiple computerized questionnaires for the retrieval of patient-reported outcomes (PROs).
Previous studies have examined groups of patients with different diagnoses by a wide variety of computer equipment, but little has been done within the field of knee osteoarthritis (KOA) [1, 2, 6, 7]. As the number of KOA patients are estimated to increase dramatically in the future, there is a huge need for an easy and precise method for retrieval of PROs from this patient category [8, 9].
For the single most common arthritic disease, KOA , a selection of relevant HSQs are: Knee Osteoarthritis Outcome Score (KOOS) , VAS pain, function and patient disability , Medical Outcomes Study 36-Item Short-Form Health Survey (SF-36) , Physical Activity Scale , painDETECT  and Activity of Daily Living (ADL) Taxonomy . Transfer of these from paper to touch screen will, apart from making the whole data collection more cost-effective, prevent missing data, avoid problems and errors during data transfer, and possibly make answering the questions easier [15, 16].
The aim of our study was to compare data based on touch screen to data based on paper version of the above-mentioned PROs to determine if the two versions are comparable. Secondly, we aimed to examine the impact of patient characteristics on differences between questionnaire-versions. Thirdly, we examined the patients' acceptance of computerized questionnaires.
Participants were recruited in March and April 2010 from an ongoing in-house KOA trial (the CAROT-study; http://ClinicalTrials.Gov Identifier: NCT00655941) at The Parker Institute, Frederiksberg Hospital, Denmark. Participants from the CAROT-study were consecutively invited to also participate in this study and recruitment ended when a sufficient number of patients were included. The participants were prior to participation informed of the content of the study.
The study was carried out in a repeated randomized crossover design (Figure 1). That is, patients were randomly assigned to one of two subgroups, completing either the paper or touch screen version first in a waiting-room setting at the hospital (trial profile, see Figure 1). The order of questionnaires was held constant, but patients entered the sequence at different points starting with either the paper or touch screen version. Patients completed both versions (paper and touch screen) of all questionnaires with a 5-minutes interval between versions, and a 5-minutes break between questionnaires. Patients were asked to fill in the paper versions the way they normally would do it. To complete the touch screen versions patients were placed in front of the computer screen and asked to follow the instructions on the screen. No information was given beforehand, but a readily available instructor was present to provide tutoring on demand.
Touch screen data was instantly exported to a specific database, whereas data from paper versions was manually entered into the same database. Time used on each touch screen HSQ was measured in seconds. Staff not involved in the recording session later checked the latter dataset.
In order to assess the feasibility and acceptability of computer-based questionnaires, the participants filled in an additional questionnaire regarding level of education, previous computer experience, and method preferences.
Criteria for selecting questionnaires were: relevance to KOA, designed for self-administration, and together representing a wide variety of questionnaire types. Furthermore, these questionnaires were familiar to the participants as they were also applied in the CAROT-study.
The Knee Injury and Osteoarthritis Outcome Score (KOOS) exploit impairment, disability and handicap within 42 items in 5 domains (activities of daily living, pain, knee-related quality of life, symptoms and sport/recreation). Items are scored from 0-4 and then transformed into a 0-100 scale; 0 representing extreme knee-related problems and 100 representing no knee-related problems [11, 15].
VAS pain, function and patient global scales (0 to 100 mm), which are used in the OMERACT-OARSI responder criteria; tools used for outcomes assessment in KOA research .
The SF-36 questionnaire includes 8 multi item domains (physical function, social function, role-emotional, role-physical, bodily pain, general health, mental health and vitality). These can be combined into 2 summary measures (physical and mental component summary measures). The scales are linearly transformed to a 0 to 100 scale; 0 indicating the least favourable health state and 100 indicating the best state of health .
The Physical Activity Scale assess physical activity in metabolic equivalents (METs) as patients report the amount of time (0 to 24 hours) they spend on 9 different levels of activity on an average weekday; in this trial the total amount of time had to reach 24 hours .
The painDETECT questionnaire is used to evaluate whether chronic pain patients suffer primarily from nociceptive or neuropathic pain . This questionnaire has several subscales and we analyzed subscales as follows: item 1-3 (10-point-Likert scales), item 4 (four figures of pain symptom variation), item 5 (yes/no to symptoms of radiating pain), item 6 (average value of seven 5-point-Likert scales) as well as a total painDETECT score.
The ADL Taxonomy is an instrument that is used to evaluate the patient's ability to perform 47 personal and instrumental activities of daily living (ADL) . We used the Rasch-based questionnaire version (ADL-Q)  and Rasch computer software were employed to convert the ordinal ADL Taxonomy ratings into linear measures of self-reported ADL ability.
The computer-assisted questionnaire solution was freeware-based on Microsoft Visual Studio Professional, using Language Integrated Query (modelling), Visual C# (computer language), Microsoft SQL Server Express (database management and maintenance) and GIMP (freeware image managing software). The underlying database was a Microsoft SQL Server database (saved as an mdf file), built up and modified by the use of SQL Server Management Studio (Microsoft®). The database was built on a relational database design. The entire solution ran on Windows 7. Requirements were Microsoft .NET Framework 3.5, and in order to ensure optimal image quality, the programme was run on a 19-inch Flex Panel PC Wide - Atom touch screen (16:10 Blue Line A/S, Århus N, Denmark, http://www.blue-line.dk).
The screen was wall-mounted and a single question appeared and was answered by tapping with either a finger or a stylus pen. Patients could freely choose the optimal position of the screen (built-in mobility in 3D), in order to ensure that patients had as much privacy as possible, when replying on the screen..
The patient interface was designed to be user-friendly with large visible characters. The questions were answered by placing a bar on a 100 mm horizontal VAS-scale or by marking the relevant squares on e.g. a Likert scale. In addition to these features, the patient was at any time able to tap on: "forward" and "backward". Each questionnaire had to be fully answered before it was possible to continue.
The platform fulfilled all legal requirements regarding protection of patient-sensitive data.
Ethics and consent
Patients gave their consent prior to inclusion. According to Danish law this study did not require approval by the Ethics Committee.
Differences were calculated by subtracting scores from each question on paper versions from the corresponding scores on touch screen versions.
Values for each version, mean differences (95% CI), medians, median differences and Intraclass Correlation Coefficients (ICCs) were calculated for all questionnaires, including relevant subscales. We chose to calculate and display both parametric and non-parametric statistics for all questionnaires as not all data met the requirements of being normally distributed and/or continuous. For the assessment of possible associations between computer skills and differences between questionnaire-versions we calculated the Spearman Correlation Coefficient. The Spearman correlation coefficient was interpreted as follows: < 0.3: none; 0.31-0.5: weak; 0.51-0.7: strong; 0.71-0.9: very strong and > 0.9: excellent. A P-value less than 0.05 (two-tailed) or a 95% confidence interval (CI) not including zero was considered statistically significant. Statistical analyses were performed using SAS version 9.1 for Windows (Chicago, IL, USA).
Rasch computer software WINSTEPS version 3.68.2  was used to generate linear measures of self-reported ADL ability based on the ADL Taxonomy paper and touch-screen data Furthermore, WINSTEPS was used to analyse if any of the patients demonstrated an abnormal response pattern.
In a two-sided tests analysis for additive equivalence of paired means for a given KOOS with bounds -5 and 5 for the mean difference and a significance level of 0.05, assuming a mean difference of 0 KOOS points, a common Standard Deviation of 20, and correlation 0.95, a sample size of 20 pairs yielded a power of 0.922 (> 90%).
A total of 20 female patients with KOA were included in this trial (Figure 1), their age ranged from 54 to 76 years and with a mean of 67 years (Table 1). Approximately 90% of the patients were computer literate and comfortable using computers. With respect to their employment status and level of education, 95% were retired and 25% had only completed elementary school.
An overall comparison of differences between paper and touch-screen versions did not reveal any tendency towards either positive or negative values (Table 2). KOOS revealed high ICCs (0.96-0.98) and mean differences between -1.5 and 0.6. The three VAS scales displayed comparable results with ICCs between 0.88 and 0.95 and mean differences between -8.7 and 2.5 (Table 3). VAS function was significantly different between versions when applying a paired t-test, but due to lack of normal distribution of data, we also analysed this finding with a two-sided paired nonparametric test (Wilcoxon rank sum) and found no significant difference between questionnaire versions (p = 0.24). The two component summary scores in SF-36 revealed ICCs of 0.94 and 0.95 with mean differences < 0.5 (absolute). The Physical Activity Scale had an ICC of 0.93 and a mean difference of 0.4. Results from painDETECT showed a high ICC for all components (0.94-0.99). The ADL Taxonomy ability measures, revealed an ICC of 0.97. There was a mean difference of 0.5, which was tested, in a paired t-test showing a statistically significant difference between questionnaire versions (CI: 0.13; 0.95, p = 0.01). The WINSTEPS analysis revealed one patient who demonstrated an abnormal response pattern on both versions of the ADL Taxonomy. Further analysis of data from the remaining questionnaires revealed that this particular patient contributed notably to the overall variation between versions (data not shown).
Computer skills, age and/or education did not have any impact on differences between questionnaire-versions, we found an unsystematic pattern of non-significant correlations ranging from 0.05 to 0.40 (p-values > 0.05).
When patients answered questions addressing their overall satisfaction with this new questionnaire-modality, we found that 16 out of 20 patients preferred touch screen questionnaires over paper versions. Furthermore, only one patient preferred the paper to the touch-screen version (Figure 2). Among patients who expressed a preference, significantly more stated that the touch screen version was easier and generally preferable.
Time spend on answering questionnaires on touch screen was measured, and results revealed that patients on average spend 6 minutes on SF-36, 0.5 minutes on the VAS scales from the OMERACT-OARSI responder criteria, 2.2 minutes on painDETECT, 9 minutes on the ADL taxonomy and 7 minutes on KOOS.
Also, despite the presence of an instructor during the trial, no patients needed any tutoring related to the touch screen versions of the questionnaires.
In this study we experienced no missing data due to the complete real-time saving of computer data and due to the fact that, clinical staff manually checked all paper versions before patients were allowed to leave the session.
Our aim was to validate touch-screen self-assessment questionnaires for use in the clinic. Comparing paper and touch-screen versions of our selected questionnaires, which covered what is normally used in a Rheumatology setting, our overall finding was a very high agreement between PROs obtained via the paper and the touch-screen versions.
Our results revealed that retired elderly female patients do not experience any problems when using computerized questionnaires, which implies that this method is applicable for the majority of patients in the clinical Rheumatology setting.
Results from our study of KOOS, VAS measurements, and SF-36 are comparable to test-retest results reported in earlier studies [20–24]. While several studies have compared the Physical Activity Scale questionnaire to accelerometers and pedometers , test-retest reliability has not yet been evaluated. Consequently our results may only be compared to other questionnaires assessing METs, which in one case found similar results in patients with hip and/or knee osteoarthritis . The original validation of painDETECT did not include a test-retest evaluation as the authors believed that symptoms of pain would fluctuate so much that such a test would only have limited use . We, therefore, present the first results regarding such a test, and overcome their consideration on fluctuation by having a short period of time between tests. The ADL taxonomy showed that all patients had similar or higher scores on paper compared to touch screen (data not shown) with a mean difference of 0.5 (CI: 0.13; 0.95). Viewing data on a Bland-Altman plot showed that with higher mean values, differences go toward zero. This overall difference may be due the fact that questions addressing easy tasks are presented at the beginning of the questionnaire, and that patients tend to continue the answering of subsequently more difficult questions at the same level when viewing all questions simultaneously in the paper version. A similar observation was done in a previous analysis of differences between questionnaire- and interview-based measures of ADL ability . The touch screen version presents a single question at a time, and may imitate an interview setting that force the patient to a more active consideration of each answer.
Our analyses of the 10-point-Likert scales from painDETECT and the three VAS scores suggest differences in the test-retest results (Table 2 and 3). Direct comparison shows that test-retest ICCs increase and difference diminishes when applying 10-point-Likert scales instead of 100 mm VAS'; a finding that might have implications for future research strategies.
An unforeseen bonus was the patient's positive attitude towards touch-screens. Touch-screen questionnaires were rated preferential and easier to paper versions, independent of level of computer use and skills.
The lack of correlation between previous computer experience and differences between questionnaires has also been reported in other studies, where use of touch-screen questionnaires was reported less stressful and requiring less or no help from staff to understand how to use them . This may be due to only getting one question at the time, and thereby avoiding problems created by interruptions .
Based on our present study, we conclude that our newly developed computer-assisted touch-screen questionnaires for PROs are directly comparable and therefore valid for recording of these data in the clinic as well as in research studies. This is in agreement with other studies comparing paper versions with touch screen for the bath AS questionnaires and the Quebec Scale  for the QOLRAD questionnaire , for WOMAC 3.1 , for RAQol, HAQ and VAS , for short-form McGill and Pain disability Index , for HAQ , for quality of life questionnaires , and for quality of care questionnaire .
Limitations for this study were that most participants were computer literate, and we can therefore not conclude whether or not all patients can use this kind of computer technology. Even so, we know that touch screens are used daily in the collection of data for the DANBIO database; gathering patient reported outcomes from the majority of rheumatologic patients in Denmark [31, 34].
In order to examine the test-retest of the chosen questionnaires, we had to consider several things; patients were not to be excessively tired, the test-retest was to consider possible fluctuation of symptoms and the setup should be so that most patients would accept participation. The chosen setup was a 5-minute interval between versions and questionnaires, as we believed that this offered a reasonable total time use (on average, 60 minutes for answering all questionnaires plus 60 minutes for pauses). Also, for several reasons, we did not believe that recall bias was a major issue in our setup. Many questionnaires were long and time-consuming, the answering of 12 questionnaires does not allow people to memorize a significant part of given answers and the study design (randomization) should level out any significant bias arisen by patients starting with a specific questionnaire. Also, we only included females in this trial, and precautions should be taken when extrapolating results to males; even so, we do not believe that significant differences between genders are to be expected.
With our broad and extended collection of questionnaires, the touch-screens open for further development towards more frequent self-assessment. Another potential of this system is the possibility of transferring answers from self assessment forms to other health institutions, e.g. from a hospital clinic to the GP or to another specialist, and in the electronic form, it will be part of the electronic patient notes in its original form. Data completeness is assured in our software version, as all items must be answered before continuation. The last advantage is the clear marked improvement of data by abolishing key-in errors, as well as the elimination of costs related to entering paper-based data into databases and the manual double-checking of data. As a future perspective, the patients will be able to answer questionnaires from home and may avoid some of the check-up visits, which are a burden of chronic patients.
The overall aim of this study was to investigate the prospect of introducing computerized questionnaires for patients with KOA. The study showed that touch-screen self-assessment questionnaires in the Rheumatology clinic are as reliable as paper questionnaires. The only observed difference, between the two versions of the ADL taxonomy, could be partly accounted for. The patients in general prefer touch screen and further advantages are less need for staff assistance, no errors related to processing of paper versions, and elimination of missing and/or incomplete data.
Kvien TK, Mowinckel P, Heiberg T, Dammann KL, Dale O, Aanerud GJ, Alme TN, Uhlig T: Performance of health status measures with a pen based personal digital assistant1. Ann Rheum Dis. 2005, 64: 1480-1484. 10.1136/ard.2004.030437.
Richter JG, Becker A, Koch T, Nixdorf M, Willers R, Monser R, Schacher B, Alten R, Specker C, Schneider M: Self-assessments of patients via Tablet PC in routine patient care: comparison with standardised paper questionnaires1. Ann Rheum Dis. 2008, 67: 1739-1741. 10.1136/ard.2008.090209.
Harlin SL, Harlin RD, Sherman TI, Rozsas CM, Shafqat MS, Meyers W: Using a structured, computer-administered questionnaire for evaluating health-related quality of life in patients with chronic lower extremity wounds. Ostomy Wound Manage. 2009, 55: 30-39.
Litaker D: New technology in quality of life research: are all computer-assisted approaches created equal?. Qual Life Res. 2003, 12: 387-393. 10.1023/A:1023457927406.
Thekkumpurath P, Venkateswaran C, Kumar M, Newsham A, Bennett MI: Screening for psychological distress in palliative care: performance of touch screen questionnaires compared with semistructured psychiatric interview. J Pain Symptom Manage. 2009, 38: 597-605. 10.1016/j.jpainsymman.2009.01.004.
Bischoff-Ferrari HA, Vondechend M, Bellamy N, Theiler R: Validation and patient acceptance of a computer touch screen version of the WOMAC 3.1 osteoarthritis index1. Ann Rheum Dis. 2005, 64: 80-84. 10.1136/ard.2003.019307.
Wilson AS, Kitas GD, Carruthers DM, Reay C, Skan J, Harris S, Treharne GJ, Young SP, Bacon PA: Computerized information-gathering in specialist rheumatology clinics: an initial evaluation of an electronic version of the Short Form 361. Rheumatology (Oxford). 2002, 41: 268-273. 10.1093/rheumatology/41.3.268.
Bendixen H, Holst C, Sorensen TI, Raben A, Bartels EM, Astrup A: Major increase in prevalence of overweight and obesity between 1987 and 2001 among Danish adults. Obes Res. 2004, 12: 1464-1472. 10.1038/oby.2004.183.
Woolf AD, Pfleger B: Burden of major musculoskeletal conditions. Bull World Health Organ. 2003, 81: 646-656.
Altman RD: Criteria for classification of clinical osteoarthritis. J Rheumatol Suppl. 1991, 27: 10-12.
Roos EM, Roos HP, Ekdahl C, Lohmander LS: Knee injury and Osteoarthritis Outcome Score (KOOS)--validation of a Swedish version1. Scand J Med Sci Sports. 1998, 8: 439-448.
Pham T, van der HD, Altman RD, Anderson JJ, Bellamy N, Hochberg M, Simon L, Strand V, Woodworth T, Dougados M: OMERACT-OARSI initiative: Osteoarthritis Research Society International set of responder criteria for osteoarthritis clinical trials revisited1. Osteoarthritis Cartilage. 2004, 12: 389-399. 10.1016/j.joca.2004.02.001.
Ware JE: SF-36 health survey update. Spine (Phila Pa 1976). 2000, 25: 3130-3139. 10.1097/00007632-200012150-00008.
Aadahl M, Jorgensen T: Validation of a new self-report instrument for measuring physical activity. Med Sci Sports Exerc. 2003, 35: 1196-1202. 10.1249/01.MSS.0000074446.02192.14.
Freynhagen R, Baron R, Gockel U, Tolle TR: painDETECT: a new screening questionnaire to identify neuropathic components in patients with back pain. Curr Med Res Opin. 2006, 22: 1911-1920. 10.1185/030079906X132488.
Bent H, Ratzlaff CR, Goligher EC, Kopec JA, Gillies JH: Computer-administered bath ankylosing spondylitis and Quebec Scale outcome questionnaires for low back pain: agreement with traditional paper format. J Rheumatol. 2005, 32: 669-672.
Törnquist K, Sonn U: Towards an ADL Taxonomy for occupational therapists. Scand J Occup Ther. 1994, 1: 69-76. 10.3109/11038129409106665.
Wæhrens EE: Measuring quality of occupational performance based on self-report and observation. Development and validation of instruments to evaluate ADL task performance. 2010, Umeå, Medical Dissertations
Linacre JM: WINSTEPS® Rasch measurement computer program. 2009, Chicago, IL, 3.68.2
Bekkers JE, de Windt TS, Raijmakers NJ, Dhert WJ, Saris DB: Validation of the Knee Injury and Osteoarthritis Outcome Score (KOOS) for the treatment of focal cartilage lesions. Osteoarthritis Cartilage. 2009, 17: 1434-1439. 10.1016/j.joca.2009.04.019.
Hoher J, Munster A, Klein J, Eypasch E, Tiling T: Validation and application of a subjective knee questionnaire. Knee Surg Sports Traumatol Arthrosc. 1995, 3: 26-33. 10.1007/BF01553522.
Soderman P, Malchau H: Validity and reliability of Swedish WOMAC osteoarthritis index: a self-administered disease-specific questionnaire (WOMAC) versus generic instruments (SF-36 and NHP)1. Acta Orthop Scand. 2000, 71: 39-46. 10.1080/00016470052943874.
Gaston-Johansson F: Measurement of pain: the psychometric properties of the Pain-O-Meter, a simple, inexpensive pain assessment tool that could change health care practices. J Pain Symptom Manage. 1996, 12: 172-181. 10.1016/0885-3924(96)00128-5.
Rat AC, Baumann C, Klein S, Loeuille D, Guillemin F: Effect of order of presentation of a generic and a specific health-related quality of life instrument in knee and hip osteoarthritis: a randomized study. Osteoarthritis Cartilage. 2008, 16: 429-435. 10.1016/j.joca.2007.07.011.
Kwak L: Measuring physical activity in field studies: Comparison of a questionnaire, 24-hour recall and an accelerometer. Eur J Sports Sci. 2007, 7: 193-201. 10.1080/17461390701674088.
Naal FD, Impellizzeri FM, Leunig M: Which is the best activity rating scale for patients undergoing total joint arthroplasty?. Clin Orthop Relat Res. 2009, 467: 958-965. 10.1007/s11999-008-0358-5.
McDonald DD, Fedo J: Older adults' pain communication: the effect of interruption. Pain Manag Nurs. 2009, 10: 149-153. 10.1016/j.pmn.2009.03.003.
Kleinman L, Leidy NK, Crawley J, Bonomi A, Schoenfeld P: A comparative trial of paper-and-pencil versus computer administration of the Quality of Life in Reflux and Dyspepsia (QOLRAD) questionnaire. Med Care. 2001, 39: 181-189. 10.1097/00005650-200102000-00008.
Greenwood MC, Hakim AJ, Carson E, Doyle DV: Touch-screen computer systems in the rheumatology clinic offer a reliable and user-friendly means of collecting quality-of-life and outcome data from patients with rheumatoid arthritis. Rheumatology (Oxford). 2006, 45: 66-71. 10.1093/rheumatology/kei100.
Cook AJ, Roberts DA, Henderson MD, Van Winkle LC, Chastain DC, Hamill-Ruth RJ: Electronic pain questionnaires: a randomized, crossover comparison with paper questionnaires for chronic pain assessment. Pain. 2004, 110: 310-317. 10.1016/j.pain.2004.04.012.
Schefte DB, Hetland ML: An open-source, self-explanatory touch screen in routine care. Validity of filling in the Bath measures on Ankylosing Spondylitis Disease Activity Index, Function Index, the Health Assessment Questionnaire and Visual Analogue Scales in comparison with paper versions. Rheumatology (Oxford). 2010, 49: 99-104. 10.1093/rheumatology/kep333.
Goodhart IM, Ibbotson V, Doane A, Roberts B, Campbell MJ, Ross RJ: Hypopituitary patients prefer a touch-screen to paper quality of life questionnaire. Growth Horm IGF Res. 2005, 15: 384-387. 10.1016/j.ghir.2005.08.002.
Larsson BW: Touch-screen versus paper-and-pen questionnaires: effects on patients' evaluations of quality of care. Int J Health Care Qual Assur Inc Leadersh Health Serv. 2006, 19: 328-338.
Hetland ML: DANBIO--powerful research database and electronic patient record. Rheumatology (Oxford). 2011, 50: 69-77. 10.1093/rheumatology/keq309.
The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1471-2474/12/190/prepub
This study was supported by grants from The Oak Foundation, The Velux Foundation, The Cambridge Health and Weight plan UK, The Danish Rheumatism Association, The Augustinus Foundation, The A.P. Møller Foundation for the Advancement of Medical Science, Hørslev Fonden, Bjarne Jensens Fond and Aase og Ejnar Danielsens fond.
The authors declare that they have no competing interests.
HG drafted the study design, protocol, analysis and interpretation of data as well as the manuscript. The technical solution was developed by PK, who also contributed to the interpretation of data and revised the manuscript. Key persons in the statistical analyses were EW and RC and they also contributed to the design, interpretation of data and revisions of the manuscript. EMB, BDS and HB contributed to the overall design idea, protocol descriptions and revisions the manuscript. All authors approved the final version.