Skip to main content
Log in

Reliability of three radiographic classification systems for knee osteoarthritis among observers of different experience levels

  • Scientific Article
  • Published:
Skeletal Radiology Aims and scope Submit manuscript

Abstract

Objective

To assess the intra- and inter-observer reliability of three commonly referenced radiographic classification systems for knee osteoarthritis in a cohort of arthroplasty candidates.

Materials and methods

Pre-operative radiographs of 112 patients who subsequently underwent primary total knee arthroplasty were evaluated by four independent observers of varying experience. Each x-ray was de-identified, randomised, and classified according to the International Knee Documentation Committee, Kellgren-Lawrence, and Ahlbäck classifications. After a 2-week interval period, each x-ray was again randomised and re-classified.

Results

Regarding inter-observer reliability, the Ahlbäck and Kellgren-Lawrence classifications were shown to have ‘substantial agreement’ (AC 0.79 and 0.85 respectively), and the IKDC was shown to have ‘almost perfect agreement’ (AC 0.97). Regarding intra-observer reliability, the two more experienced observers demonstrated ‘good’ or ‘excellent’ reliability for all classification systems, and the two less experienced observers demonstrated ‘moderate’ intra-observer reliability for all classification systems.

Conclusion

The International Knee Documentation Committee, Kellgren-Lawrence, and Ahlbäck radiographic classifications demonstrated adequate intra- and inter-observer reliability, supporting their potential implementation in surgical practice, or in epidemiological and clinical studies of knee osteoarthritis in a comparable cohort of patients. Clinical experience was positively correlated with intra-observer reliability. Whilst the International Knee Documentation Committee classification demonstrated the greatest reliability, this is likely due to its conservative definitions, and the Ahlbäck and Kellgren-Lawrence classifications are likely more reflective of the spectrum of disease severity encountered in an older patient cohort.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2

Similar content being viewed by others

References

  1. Cross M, Smith E, Hoy D, Nolte S, Ackerman I, Fransen M, et al. The global burden of hip and knee osteoarthritis: estimates from the global burden of disease 2010 study. Ann Rheum Dis. 2014;73(7):1323–30.

    Article  Google Scholar 

  2. Association AO. Hip, Knee & Shoulder Arthroplasty 2019 Annual Report. Australian Orthopaedic Association National Joint Replacement Registry (AOANJRR) 2019.

  3. Ackerman IN, Bohensky MA, Zomer E, Tacey M, Gorelik A, Brand CA, et al. The projected burden of primary total knee and hip replacement for osteoarthritis in Australia to the year 2030. BMC Musculoskelet Disord. 2019;20(1):90.

    Article  Google Scholar 

  4. Kellgren JH, Lawrence JS. Radiological assessment of osteo-arthrosis. Ann Rheum Dis. 1957;16(4):494–502.

    Article  CAS  Google Scholar 

  5. Greco NJ, Anderson AF, Mann BJ, Cole BJ, Farr J, Nissen CW, et al. Responsiveness of the International Knee Documentation Committee Subjective Knee Form in comparison to the Western Ontario and McMaster Universities Osteoarthritis Index, modified Cincinnati Knee Rating System, and Short Form 36 in patients with focal articular cartilage defects. Am J Sports Med. 2010;38(5):891–902.

    Article  Google Scholar 

  6. Ahlback S. Osteoarthrosis of the knee. A radiographic investigation. Acta Radiol Diagn (Stockh). 1968:Suppl 277:277–272.

  7. Merle-Vincent F, Couris CM, Schott AM, Conrozier T, Piperno M, Mathieu P, et al. Factors predicting patient satisfaction 2 years after total knee arthroplasty for osteoarthritis. Joint Bone Spine. 2011;78(4):383–6.

    Article  Google Scholar 

  8. Schnurr C, Jarrous M, Gudden I, Eysel P, Konig DP. Pre-operative arthritis severity as a predictor for total knee arthroplasty patients’ satisfaction. Int Orthop. 2013;37(7):1257–61.

    Article  Google Scholar 

  9. Stone OD, Duckworth AD, Curran DP, Ballantyne JA, Brenkel IJ. Severe arthritis predicts greater improvements in function following total knee arthroplasty. Knee Surg Sports Traumatol Arthrosc. 2017;25(8):2573–9.

    Article  CAS  Google Scholar 

  10. Claessens AA, Schouten JS, van den Ouweland FA, Valkenburg HA. Do clinical findings associate with radiographic osteoarthritis of the knee? Ann Rheum Dis. 1990;49(10):771–4.

    Article  CAS  Google Scholar 

  11. Felson DT. The epidemiology of knee osteoarthritis: results from the Framingham Osteoarthritis Study. Semin Arthritis Rheum. 1990;20(3 Suppl 1):42–50.

    Article  CAS  Google Scholar 

  12. Bedson J, Croft PR. The discordance between clinical and radiographic knee osteoarthritis: a systematic search and summary of the literature. BMC Musculoskelet Disord. 2008;9:116.

    Article  Google Scholar 

  13. Galli M, De Santis V, Tafuro L. Reliability of the Ahlbäck classification of knee osteoarthritis. Osteoarthr Cartil. 2003;11(8):580–4.

    Article  CAS  Google Scholar 

  14. Riddle DL, Jiranek WA, Hull JR. Validity and reliability of radiographic knee osteoarthritis measures by arthroplasty surgeons. Orthopedics. 2013;36(1):e25–32.

    Article  Google Scholar 

  15. Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics. 1977;33(1):159–74.

    Article  CAS  Google Scholar 

  16. Koo TK, Li MY. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med. 2016;15(2):155–63.

    Article  Google Scholar 

  17. Wright RW, Group M. Osteoarthritis classification scales: interobserver reliability and arthroscopic correlation. J Bone Joint Surg Am. 2014;96(14):1145–51.

    Article  Google Scholar 

  18. Weidow J, Cederlund CG, Ranstam J, Karrholm J. Ahlbäck grading of osteoarthritis of the knee: poor reproducibility and validity based on visual inspection of the joint. Acta Orthop. 2006;77(2):262–6.

    Article  Google Scholar 

  19. Fleiss JL, Levin B, Paik MC. Statistical methods for rates and proportions. 3rd ed. J. Wiley: Hoboken, N.J; 2003.

    Book  Google Scholar 

  20. Cicchetti DV, Feinstein AR. High agreement but low kappa: II. Resolving the paradoxes. J Clin Epidemiol 1990; 43(6):551–558.

  21. Gwet KL. Handbook of inter-rater reliability: the definitive guide to measuring the extent of agreement among raters. 4 ed: Advanced Analytics, 2014.

  22. Wongpakaran N, Wongpakaran T, Wedding D, Gwet KL. A comparison of Cohen’s kappa and Gwet’s AC1 when calculating inter-rater reliability coefficients: a study conducted with personality disorder samples. BMC Med Res Methodol. 2013;13:61.

    Article  Google Scholar 

  23. Garbuz DS, Masri BA, Esdaile J. Duncan CP classification systems in orthopaedics. J Am Acad Orthop Surg. 2002;10(4):290–7.

    Article  Google Scholar 

  24. Gunaratne R, Pratt DN, Banda J, Fick DP, Khan RJK, Robertson BW. Patient dissatisfaction following total knee arthroplasty: a systematic review of the literature. J Arthroplast. 2017;32(12):3854–60.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nicholas Wing.

Ethics declarations

Ethical approval was obtained from The Darling Downs Hospital and Health Service Human Research Ethics Committee.

Conflict of interest

The authors declare that they have no conflict of interest.

Informed consent

Informed consent was obtained from all individual participants included in the study.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wing, N., Van Zyl, N., Wing, M. et al. Reliability of three radiographic classification systems for knee osteoarthritis among observers of different experience levels. Skeletal Radiol 50, 399–405 (2021). https://doi.org/10.1007/s00256-020-03551-4

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00256-020-03551-4

Keywords

Navigation