Interrater reliability in neurology objective structured clinical examination across specialties

Med Teach. 2024 Feb;46(2):239-244. doi: 10.1080/0142159X.2023.2244146. Epub 2023 Aug 22.

Abstract

Purpose: To assess interrater reliability and examiners' characteristics, especially specialty, associated with scoring of neurology objective structured clinical examination (OSCE).

Material and methods: During a neurology mock OSCE, five randomly chosen students volunteers were filmed while performing 1 of the 5 stations. Video recordings were scored by physicians from the Lyon and Clermont-Ferrand university teaching hospitals to assess students performance using both a checklist scoring and a global rating scale. Interrater reliability between examiners were assessed using intraclass coefficient correlation. Multivariable linear regression models including video recording as random effect dependent variable were performed to detect factors associated with scoring.

Results: Thirty examiners including 15 (50%) neurologists participated. The intraclass correlation coefficient of checklist scores and global ratings between examiners were 0.71 (CI95% [0.45-0.95]) and 0.54 (CI95% [0.28-0.91]), respectively. In multivariable analyses, no factor was associated with checklist scores, while male gender of examiner was associated with lower global rating (β coefficient = -0.37; CI 95% [-0.62-0.11]).

Conclusions: Our study demonstrated through a video-based scoring method that agreement among examiners was good using checklist scoring while moderate using global rating scale in neurology OSCE. Examiner's specialty did not affect scoring whereas gender was associated with global rating scale.

Keywords: educational measurement; neurology; objective structured clinical examination.

MeSH terms

  • Clinical Competence
  • Educational Measurement / methods
  • Humans
  • Male
  • Medicine*
  • Neurology*
  • Reproducibility of Results
  • Students, Medical*