Abstract
Objective
To assess the intra- and inter-observer reliability of three commonly referenced radiographic classification systems for knee osteoarthritis in a cohort of arthroplasty candidates.
Materials and methods
Pre-operative radiographs of 112 patients who subsequently underwent primary total knee arthroplasty were evaluated by four independent observers of varying experience. Each x-ray was de-identified, randomised, and classified according to the International Knee Documentation Committee, Kellgren-Lawrence, and Ahlbäck classifications. After a 2-week interval period, each x-ray was again randomised and re-classified.
Results
Regarding inter-observer reliability, the Ahlbäck and Kellgren-Lawrence classifications were shown to have ‘substantial agreement’ (AC 0.79 and 0.85 respectively), and the IKDC was shown to have ‘almost perfect agreement’ (AC 0.97). Regarding intra-observer reliability, the two more experienced observers demonstrated ‘good’ or ‘excellent’ reliability for all classification systems, and the two less experienced observers demonstrated ‘moderate’ intra-observer reliability for all classification systems.
Conclusion
The International Knee Documentation Committee, Kellgren-Lawrence, and Ahlbäck radiographic classifications demonstrated adequate intra- and inter-observer reliability, supporting their potential implementation in surgical practice, or in epidemiological and clinical studies of knee osteoarthritis in a comparable cohort of patients. Clinical experience was positively correlated with intra-observer reliability. Whilst the International Knee Documentation Committee classification demonstrated the greatest reliability, this is likely due to its conservative definitions, and the Ahlbäck and Kellgren-Lawrence classifications are likely more reflective of the spectrum of disease severity encountered in an older patient cohort.
Similar content being viewed by others
References
Cross M, Smith E, Hoy D, Nolte S, Ackerman I, Fransen M, et al. The global burden of hip and knee osteoarthritis: estimates from the global burden of disease 2010 study. Ann Rheum Dis. 2014;73(7):1323–30.
Association AO. Hip, Knee & Shoulder Arthroplasty 2019 Annual Report. Australian Orthopaedic Association National Joint Replacement Registry (AOANJRR) 2019.
Ackerman IN, Bohensky MA, Zomer E, Tacey M, Gorelik A, Brand CA, et al. The projected burden of primary total knee and hip replacement for osteoarthritis in Australia to the year 2030. BMC Musculoskelet Disord. 2019;20(1):90.
Kellgren JH, Lawrence JS. Radiological assessment of osteo-arthrosis. Ann Rheum Dis. 1957;16(4):494–502.
Greco NJ, Anderson AF, Mann BJ, Cole BJ, Farr J, Nissen CW, et al. Responsiveness of the International Knee Documentation Committee Subjective Knee Form in comparison to the Western Ontario and McMaster Universities Osteoarthritis Index, modified Cincinnati Knee Rating System, and Short Form 36 in patients with focal articular cartilage defects. Am J Sports Med. 2010;38(5):891–902.
Ahlback S. Osteoarthrosis of the knee. A radiographic investigation. Acta Radiol Diagn (Stockh). 1968:Suppl 277:277–272.
Merle-Vincent F, Couris CM, Schott AM, Conrozier T, Piperno M, Mathieu P, et al. Factors predicting patient satisfaction 2 years after total knee arthroplasty for osteoarthritis. Joint Bone Spine. 2011;78(4):383–6.
Schnurr C, Jarrous M, Gudden I, Eysel P, Konig DP. Pre-operative arthritis severity as a predictor for total knee arthroplasty patients’ satisfaction. Int Orthop. 2013;37(7):1257–61.
Stone OD, Duckworth AD, Curran DP, Ballantyne JA, Brenkel IJ. Severe arthritis predicts greater improvements in function following total knee arthroplasty. Knee Surg Sports Traumatol Arthrosc. 2017;25(8):2573–9.
Claessens AA, Schouten JS, van den Ouweland FA, Valkenburg HA. Do clinical findings associate with radiographic osteoarthritis of the knee? Ann Rheum Dis. 1990;49(10):771–4.
Felson DT. The epidemiology of knee osteoarthritis: results from the Framingham Osteoarthritis Study. Semin Arthritis Rheum. 1990;20(3 Suppl 1):42–50.
Bedson J, Croft PR. The discordance between clinical and radiographic knee osteoarthritis: a systematic search and summary of the literature. BMC Musculoskelet Disord. 2008;9:116.
Galli M, De Santis V, Tafuro L. Reliability of the Ahlbäck classification of knee osteoarthritis. Osteoarthr Cartil. 2003;11(8):580–4.
Riddle DL, Jiranek WA, Hull JR. Validity and reliability of radiographic knee osteoarthritis measures by arthroplasty surgeons. Orthopedics. 2013;36(1):e25–32.
Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics. 1977;33(1):159–74.
Koo TK, Li MY. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med. 2016;15(2):155–63.
Wright RW, Group M. Osteoarthritis classification scales: interobserver reliability and arthroscopic correlation. J Bone Joint Surg Am. 2014;96(14):1145–51.
Weidow J, Cederlund CG, Ranstam J, Karrholm J. Ahlbäck grading of osteoarthritis of the knee: poor reproducibility and validity based on visual inspection of the joint. Acta Orthop. 2006;77(2):262–6.
Fleiss JL, Levin B, Paik MC. Statistical methods for rates and proportions. 3rd ed. J. Wiley: Hoboken, N.J; 2003.
Cicchetti DV, Feinstein AR. High agreement but low kappa: II. Resolving the paradoxes. J Clin Epidemiol 1990; 43(6):551–558.
Gwet KL. Handbook of inter-rater reliability: the definitive guide to measuring the extent of agreement among raters. 4 ed: Advanced Analytics, 2014.
Wongpakaran N, Wongpakaran T, Wedding D, Gwet KL. A comparison of Cohen’s kappa and Gwet’s AC1 when calculating inter-rater reliability coefficients: a study conducted with personality disorder samples. BMC Med Res Methodol. 2013;13:61.
Garbuz DS, Masri BA, Esdaile J. Duncan CP classification systems in orthopaedics. J Am Acad Orthop Surg. 2002;10(4):290–7.
Gunaratne R, Pratt DN, Banda J, Fick DP, Khan RJK, Robertson BW. Patient dissatisfaction following total knee arthroplasty: a systematic review of the literature. J Arthroplast. 2017;32(12):3854–60.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Ethical approval was obtained from The Darling Downs Hospital and Health Service Human Research Ethics Committee.
Conflict of interest
The authors declare that they have no conflict of interest.
Informed consent
Informed consent was obtained from all individual participants included in the study.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Wing, N., Van Zyl, N., Wing, M. et al. Reliability of three radiographic classification systems for knee osteoarthritis among observers of different experience levels. Skeletal Radiol 50, 399–405 (2021). https://doi.org/10.1007/s00256-020-03551-4
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00256-020-03551-4