Skip to main content
  • Research article
  • Open access
  • Published:

Are classifications of proximal radius fractures reproducible?

Abstract

Background

Fractures of the proximal radius need to be classified in an appropriate and reproducible manner. The aim of this study was to assess the reliability of the three most widely used classification systems.

Methods

Elbow radiographs images of patients with proximal radius fractures were classified according to Mason, Morrey, and Arbeitsgemeinschaft für osteosynthesefragen/Association for the Study of Internal Fixation (AO/ASIF) classifications by four observers with different experience with this subject to assess their intra- and inter-observer agreement. Each observer analyzed the images on three different occasions on a computer with numerical sequence randomly altered.

Results

We found that intra-observer agreement of Mason and Morrey classifications were satisfactory (κ = 0.582 and 0.554, respectively), while the AO/ASIF classification had poor intra-observer agreement (κ = 0.483). Inter-observer agreement was higher in the Mason (κ = 0.429-0.560) and Morrey (κ = 0.319-0.487) classifications than in the AO/ASIF classification (κ = 0.250-0.478), which showed poor reliability.

Conclusion

Inter- and intra-observer agreement of the Mason and Morey classifications showed overall satisfactory reliability when compared to the AO/ASIF system. The Mason classification is the most reliable system.

Peer Review reports

Background

Fractures of the proximal radius are relatively common injuries, accounting for approximately 1/3 of all elbow fractures and about 1.7 to 5.4% of all fractures in adults [1, 2]. Most of these (85%) occur in adults between 20 and 60 years of age (average of 30 to 40 years) with a male:female ratio of about 2:3 [1, 2]. One third of the lesions are associated with others upper limb injuries, such as carpal bone fractures [3], distal radioulnar joint injuries [4, 5], interosseous membrane injuries [6, 7], capitellar fractures [8], and damage of the medial collateral ligament.

The most common mechanism of injury in these fractures is a fall onto an outstretched hand with an axial load on the radius. In such cases, the radial head and/or neck fracture when they collide with the capitellum, usually with forearm in pronation and the elbow in partial flexion [911].

Fractures of the proximal radius play an important role in injuries of the elbow, not so much by how frequently they occur, but mainly by the potential difficulties in treatment and the complications that can arise, sometimes with serious impairment of function due to pain and loss of mobility. Therefore, proper classification is essential in order to render the proper treatment.

Systems have been developed to help surgeons in classifying fractures into different and clinically useful groups for treatment definition.

In 1954, Mason described the first classification system dividing proximal radius fractures into three types. Type I fractures were nondisplaced or minimally displaced fractures of the head or neck; type II were displaced fractures (more than 2 mm) of the head or neck; and type III were severely comminuted fractures of the proximal radius [12]. In 1962, Johnston expanded the classification of Mason adding type IV, a fracture associated with dislocation of the elbow [13].

In 2008, van Riet and Morrey published a revision of the Mason classification, distinguishing between injuries associated with coronoid fractures, the olecranon fractures, and ligamentous injuries [14].

The AO/ASIF classification was created in 1986 and revised in 2007 [15]. It considers the seriousness of the bone injury and serves as a basis for treatment and prognosis.

The AO/ASIF system specifies three basic types: extra-articular, articular of the radius or ulna, and articular of the radius and ulna. With each group, the fractures are organized in increasing order of severity with regard to morphological complexity, difficulty in treatment, and prognosis. While this system is the most comprehensive, its intra- and inter-observer reliability has shown to be limited [16, 17].

The purpose of a classification system is to name and describe the fractures according to their characteristics providing a hierarchy of those characteristics. A classification system should also guide action or intervention and assist in predicting outcomes of an intervention or treatment [18]. A good system needs to be valid, reliable, and reproducible. The perfect classification system should also standardize the language used to describe the fractures, offer guidelines for treatment, indicate the possibility of complications, and help determine the prognosis. The ideal system should also provide a mechanism to evaluate and compare the results with treatment of similar fractures treated at various centres and reported at different times in the literature [19].

Considering the need to classify the fractures of the proximal radius in an appropriate and reproducible manner, we sought to assess the reproducibility of the three most widely used classification systems. Thus we evaluated the intra- and inter-observer agreement of the Mason modified by Hotchkiss, Morrey and AO/ASIF classifications of proximal radius fractures.

Methods

We analyzed 65 consecutive elbow radiographs performed on patients with fractures of the proximal radius. The patients were treated in the same hospital. Each of the 65 radiographs consisted of two views, anterior-posterior and lateral, and they were numbered, with the patients' names and ages concealed. Radiographs were excluded if the patient had incomplete skeletal development, pathologic fractures or previous elbow surgery.

The image quality was determined by two orthopaedic surgeons. The radiography was accepted only when both of these surgeons considered the radiographs acceptable.

Four observers familiar with the classification systems were selected for analysis. These observers were a second-year resident of orthopaedics (R2), a general orthopaedist (GO), a shoulder and elbow surgery specialist (SES), and a radiologist (RD).

To standardize the information for all observers, each were given self-explanatory diagrams with the classification systems. Each observer classified the 65 images at three different times, according to the three systems (Mason modified by Hotchkiss, Morrey and AO/ASIF). In the first evaluation (T1), the 65 digitized images of radiographs of each patient were viewed on a computer in numerical sequence. Three weeks later, in the second assessment (T2), the sequence of radiographs was randomly altered, as it was in the third assessment (T3), three weeks after T2. This sequence of randomization was known only by a person uninvolved in the assessment of the images.

The data were collected in spreadsheets and kappa coefficients were calculated for analysis according to the method proposed by Fleiss et al [20, 21]. This method not only calculates the agreement expected by chance, as described earlier in the method of Scott and Cohen [22, 23], but also the correlation among more than two observers in the evaluation of nominal variables. The kappa coefficient of agreement indicates the proportion of agreement among observers. The kappa values range from -1 to +1: values between -1 and 0 indicate that observed agreement was lower than that expected by chance, 0 indicates a level of agreement equal to that expected by chance, and +1 indicates total agreement. Overall, kappa values below 0.5 are considered unsatisfactory; values between 0.5 and 0.75 are considered satisfactory and appropriate, and values over 0.75 are considered excellent [24, 25].

This project was approved by the Research Ethics Committee under No. 363/08, on April 4, 2008.

Results

From the 65 initial radiographs six were excluded because of poor technical quality, leaving a sample size to 59 radiographs.

Table 1 summarizes the kappa values for intra-observer comparisons for each observer, at the three time points. The concordance was higher among the different with the Mason and Morrey classifications (mean κ = 0.582 and 0.554, respectively) than with the AO/ASIF classification (mean κ = 0.483).

Table 1 Kappa coefficients for intra-observer comparisons between the three time points (T1, T2 and T3).

Table 2 shows the inter-observer kappa coefficients at each of the times of assessment. Inter-observer agreement was higher in the Mason (κ = 0.429-0.560) and Morrey (κ = 0.319-0.487) classifications than in the AO/ASIF classification (κ = 0.250-0.478).

Table 2 Kappa coefficients for inter-observer comparisons at each time point. (T1, T2, T3)

Discussion

The classification systems for this study were selected because they are the most commonly used and studied for fractures of the proximal radius [14]. Classification systems are of great importance in orthopaedic practice because they are used to describe fractures, guide treatment, and compare treatment outcome within and between studies in the literature. As a result, intra- and inter-observer concordances are essential for any classification system.

In the analysis of intra-observer agreement between three time points, the average kappa coefficient for the AO/ASIF classification was unsatisfactory (κ = 0.483), ranging from 0.305 (for the SES) to 0.676 (for the R2). The broad variability in these results is probably due to the complexity of the classification, and accord with the results of other studies that evaluated the classification for fractures of other bones [17, 26]. Professional experience had no effect on intra-observer agreement with this system, as indicated by the highest kappa coefficient for the R2 and the lowest for the SES. For the Mason and Morrey classifications, intra-observer concordances were satisfactory and similar (κ of 0.692 and 0.644, respectively). The similarity of these coefficients was expected because the Morrey classification is derived from and more complex than the of Mason classification [14].

Regarding inter-observer concordances, the AO/ASIF rating was unsatisfactory at all times, although it improved with time, probably due to the learning curve. A similar pattern occurred with the Morrey classification, but the values were slightly higher but still unsatisfactory. The Mason classification had an unsatisfactory mean kappa coefficient at T1 (0.429). However, the means at T2 and T3 were satisfactory (0.560 and 0.551, respectively). These values may also be explained by the complexity of the classifications.

The AO/ASIF is a more complex system that involves the proximal radius as well as associated ulnar injuries and ligamentous injuries. These added variables will presumably have a negative impact on its reliability.

The results of this study were similar to those reported by Sheps et al. [27], which indicated that the correlation was unsatisfactory for the AO/ASIF classification and better for the system adapted by Hotchkiss from the Mason system. However, even this adapted Mason system was unsatisfactory when considering the lower limit of 95% of the confidence interval.

It is important to mention that the present study was limited to evaluating the agreement between the observers' opinions. This study was unable to measure the accuracy of each observer's opinion. To clarify the accuracy issue, studies in which clinical-radiographic diagnoses made by each observer compared with an examination result or a gold-standard procedure (i.e. one with high sensitivity and specificity) would be needed in order to prove the proposed diagnosis.

We have made the first step to the development of the ideal classification once we studied the reliability of the most widely used classification systems. In the continuing search for this ideal classification there is a need to perform new prospective studies with large number of patients to determine which variables (displacement of the fragments, intra-articular/metaphyseal comminution, associated injuries, patient's age) can guide the ideal treatment and predict the prognosis of such fractures through radiographic examination.

Conclusion

Inter- and intra-observer agreement of the Mason and Morey classifications showed overall satisfactory reliability when compared to the AO/ASIF system.

The Mason classification is the most reliable system.

About the authors

Fabio T Matsunaga 1 , Marcel J S Tamaoki 2 , Eduardo F Cordeiro 3 , Anderson Uehara 4 , Marcos H Ikawa 5 , Marcelo H Matsumoto 6 , João B G dos Santos 7 , João C Belloti 8 .

1Second-year resident in the Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

2Attending physician in Shoulder and Elbow Surgery Sector in the Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

3Second-year resident in the Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

4Attending physician in Shoulder and Elbow Surgery Sector in the Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

5Attending physician in the Department Radiology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

6Head of the Shoulder and Elbow Surgery Sector, Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

7Adjunct professor and head of Hand Surgery Clinic, Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

8Attending physician in Hand Surgery Clinic, Department of Orthopedics and Traumatology, Universidade Federal de São Paulo -- Escola Paulista de Medicina (Unifesp-EPM), São Paulo, Brazil.

References

  1. Van Riet R, Morrey BF, O'Driscoll SW, Van Glabbeek F: Associated injuries complicating radial head fractures: a demographic study. Clin Orthop Relat Res. 2005, 441: 351-355. 10.1097/01.blo.0000180606.30981.78.

    Article  PubMed  Google Scholar 

  2. Kaas L, Van Riet R, Vroemen JPAM, Eygendaal D: The incidence of associated fractures of the upper limb in fractures of the radial head. Strategies Trauma Limb Reconstr. 2008, 3: 71-4. 10.1007/s11751-008-0038-8. Epub 2008 Jul 10

    Article  PubMed  PubMed Central  Google Scholar 

  3. Jupiter JB: The management of fractures in one upper extremity. J Hand Surg. 1986, 11A: 279-82.

    Article  Google Scholar 

  4. Curr J, Coe W: Dislocation if the inferior radio-ulnar joint. Br J Surg. 1946, 34: 74-77. 10.1002/bjs.18003413312.

    Article  CAS  PubMed  Google Scholar 

  5. Edwards G, Jupiter JB: The Essex-Lopresti lesion revised. Clin Orthop. 1988, 234: 61-69.

    PubMed  Google Scholar 

  6. McDougall A, White J: Subluxation of the inferior radio-ulnar joint complicating fracture of the radial head. J Bone Joint Surg. 1957, 39[B]: 278-87.

    Google Scholar 

  7. Taylor TKF, O'Connor BT: The effect upon the inferior radio-ulnar joint and excision of the head of the radius in adults. J Bone Joint Surg. 1964, 46[B]: 83-84.

    Google Scholar 

  8. Osborne G, Cotterill P: Recurrent dislocation of the elbow. J Bone Joint Surg. 1966, 48[B]: 340-46.

    Google Scholar 

  9. Ring D, Jupiter JB: Fracture-dislocation of the elbow. J Bone Joint Surg [Am]. 1998, 80-A: 566-580.

    Google Scholar 

  10. Ring D: Fractures and Dislocations of the Elbow. Rockwood and Green's Fractures in Adults. 2006, Philadelphia: Lippincot Williams & Wilkins, 1: 1011-12. Sixth

    Google Scholar 

  11. Amis AA, Miller JH: Mechanisms of elbow fractures: investigation using impact tests in vitro. Injury. 1995, 26: 163-168. 10.1016/0020-1383(95)93494-3.

    Article  CAS  PubMed  Google Scholar 

  12. Mason ML: Some observations on fractures of the head of the radius with a review of one hundred cases. Br J Surg. 1954, 42: 123-132. 10.1002/bjs.18004217203.

    Article  CAS  PubMed  Google Scholar 

  13. Johnston GW: A follow-up of 100 cases of fracture of the head of the radius with review of the literature. Ulster Med J. 1962, 31: 51-56.

    CAS  PubMed  PubMed Central  Google Scholar 

  14. Van Riet RP, Morrey BF: Documentation of associated injuries ocurring with radial head fractures. Clin Orthop Relat Res. 2008, 466: 130-134. 10.1007/s11999-007-0064-8.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Marsh JL: et al., Fracture and dislocation compendium - 2007: Orthopaedic Trauma Association classifications, database and outcomes comittee. J Orthop Trauma. 2007, 21 (10 Suppl): S1-S133. 10.1097/00005131-200711101-00001.

    Article  CAS  PubMed  Google Scholar 

  16. Burstein AH: Fracture classification systems: do they work and are they useful?. J Bone Joint Surg [Am]. 1993, 75-A: 1743-4.

    Google Scholar 

  17. Kreder HJ, Hanel DP, Mckee M, Jupiter J, Mcgillivary G, Swiontkowski MF: Consistency of AO fracture classification for the distal radius. J Bone Joint Surg [Br]. 1996, 78-B: 726-31.

    Google Scholar 

  18. Dirschl DR, Cannada LK: Classification of Fractures. Rockwood and Green's Fractures in Adults. 2006, Philadelphia: Lippincot Williams & Wilkins, 1: 43-4. Sixth

    Google Scholar 

  19. Martin JS, Marsh JL: Current classification of fractures. Rationale and utility. Radiol Clin North Am. 1997, 35 (3): 491-506.

    CAS  PubMed  Google Scholar 

  20. Fleiss JL: Measuring nominal scale agreement among many raters. Psychological Bulletin. 1971, 76 (5): 378-82. 10.1037/h0031619.

    Article  Google Scholar 

  21. Fleiss JL, Slakter MJ, Fischman SL, Park MH, Chilton NW: Inter-examiner reliability in caries trials. J Dent Res. 1979, 58: 604-609. 10.1177/00220345790580021101.

    Article  CAS  PubMed  Google Scholar 

  22. Scott WA: Reliability of content analysis: the case of nominal scale coding. Public Opinion Quarterly. 1995, 19: 321-5. 10.1086/266577.

    Article  Google Scholar 

  23. Cohen J: A coefficient of agreement for nominal scales. Educ Psychol Meas. 1960, 20 (3): 37-46. 10.1177/001316446002000104.

    Article  Google Scholar 

  24. Martin JS, Marsh JL, Bonar SK, DeCoster TA, Found EM, Brandser EA: Assessment of the AO/ASIF Fracture Classification for distal tibia. J Orthop Trauma. 1997, 11: 477-483. 10.1097/00005131-199710000-00004.

    Article  CAS  PubMed  Google Scholar 

  25. Svanholm H, Starklint H, Gunderson HJ, Fabricius J, Barlebo H, Olsen S: Reproducibility of histomorphological diagnosis with special reference to the kappa statistic. APMIS. 1989, 97: 689-698. 10.1111/j.1699-0463.1989.tb00464.x.

    Article  CAS  PubMed  Google Scholar 

  26. Flinkkila T, Nikkola-Sihto A, Kaarela O, Paakko E, Raatikainen T: Poor interobserver reliability of AO classification of fractures of distal radius. J Bone Joint Surg [Br]. 1998, 80-B: 670-2. 10.1302/0301-620X.80B4.8511.

    Article  Google Scholar 

  27. Sheps D, Hildebrand K, Keifer K: The inter-observer reliability of classifications systems for radial head fractures. J Bone Joint Surg [Br]. 2008, 90-B (supp I): 140-

    Google Scholar 

Pre-publication history

Download references

Acknowledgements

The authors thank Dr Walter Manna Albertoni and Dr Flavio Faloppa for their on-going support and suggestions in the preparation of this manuscript.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Fabio T Matsunaga.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

FTM participated in the design of the study, collected, digitalized and randomized the radiographs, helped to draft the manuscript, revised the bibliography for references and performed the statistical analysis; MJST conceived the study design, analyzed the radiographs and classified the fractures, helped to draft the manuscript and revised the bibliography for references; EFC participated in the design of the study, analyzed the radiographs and classified the fractures and translated the manuscript to English; AU: participated in the design of the study, analyzed the radiographs and classified the fractures; MHI participated in the design of the study, analyzed the radiographs and classified the fractures; MHM conceived the study design, revised the article and gave the final approval of the version to be published; JBGS participated in the design of the study, revised the article and gave the final approval of the version to be published; JCB participated in the design of the study, revised the article; FF revised the article and gave the final approval of the version to be published. All authors read and approved the final manuscript.

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Matsunaga, F.T., Tamaoki, M.J., Cordeiro, E.F. et al. Are classifications of proximal radius fractures reproducible?. BMC Musculoskelet Disord 10, 120 (2009). https://doi.org/10.1186/1471-2474-10-120

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1471-2474-10-120

Keywords