TY - JOUR
T1 - Use of Generalizability Theory for Exploring Reliability of and Sources of Variance in Assessment of Technical Skills
T2 - A Systematic Review and Meta-Analysis
AU - Andersen, Steven Arild Wuyts
AU - Nayahangan, Leizl Joy
AU - Park, Yoon Soo
AU - Konge, Lars
N1 - Copyright © 2021 by the Association of American Medical Colleges.
PY - 2021/11/1
Y1 - 2021/11/1
N2 - PURPOSE: Competency-based education relies on the validity and reliability of assessment scores. Generalizability (G) theory is well suited to explore the reliability of assessment tools in medical education but has only been applied to a limited extent. This study aimed to systematically review the literature using G-theory to explore the reliability of structured assessment of medical and surgical technical skills and to assess the relative contributions of different factors to variance.METHOD: In June 2020, 11 databases, including PubMed, were searched from inception through May 31, 2020. Eligible studies included the use of G-theory to explore reliability in the context of assessment of medical and surgical technical skills. Descriptive information on study, assessment context, assessment protocol, participants being assessed, and G-analyses was extracted. Data were used to map G-theory and explore variance components analyses. A meta-analysis was conducted to synthesize the extracted data on the sources of variance and reliability.RESULTS: Forty-four studies were included; of these, 39 had sufficient data for meta-analysis. The total pool included 35,284 unique assessments of 31,496 unique performances of 4,154 participants. Person variance had a pooled effect of 44.2% (95% confidence interval [CI], 36.8%-51.5%). Only assessment tool type (Objective Structured Assessment of Technical Skills-type vs task-based checklist-type) had a significant effect on person variance. The pooled reliability (G-coefficient) was 0.65 (95% CI, .59-.70). Most studies included decision studies (39, 88.6%) and generally seemed to have higher ratios of performances to assessors to achieve a sufficiently reliable assessment.CONCLUSIONS: G-theory is increasingly being used to examine reliability of technical skills assessment in medical education, but more rigor in reporting is warranted. Contextual factors can potentially affect variance components and thereby reliability estimates and should be considered, especially in high-stakes assessment. Reliability analysis should be a best practice when developing assessment of technical skills.
AB - PURPOSE: Competency-based education relies on the validity and reliability of assessment scores. Generalizability (G) theory is well suited to explore the reliability of assessment tools in medical education but has only been applied to a limited extent. This study aimed to systematically review the literature using G-theory to explore the reliability of structured assessment of medical and surgical technical skills and to assess the relative contributions of different factors to variance.METHOD: In June 2020, 11 databases, including PubMed, were searched from inception through May 31, 2020. Eligible studies included the use of G-theory to explore reliability in the context of assessment of medical and surgical technical skills. Descriptive information on study, assessment context, assessment protocol, participants being assessed, and G-analyses was extracted. Data were used to map G-theory and explore variance components analyses. A meta-analysis was conducted to synthesize the extracted data on the sources of variance and reliability.RESULTS: Forty-four studies were included; of these, 39 had sufficient data for meta-analysis. The total pool included 35,284 unique assessments of 31,496 unique performances of 4,154 participants. Person variance had a pooled effect of 44.2% (95% confidence interval [CI], 36.8%-51.5%). Only assessment tool type (Objective Structured Assessment of Technical Skills-type vs task-based checklist-type) had a significant effect on person variance. The pooled reliability (G-coefficient) was 0.65 (95% CI, .59-.70). Most studies included decision studies (39, 88.6%) and generally seemed to have higher ratios of performances to assessors to achieve a sufficiently reliable assessment.CONCLUSIONS: G-theory is increasingly being used to examine reliability of technical skills assessment in medical education, but more rigor in reporting is warranted. Contextual factors can potentially affect variance components and thereby reliability estimates and should be considered, especially in high-stakes assessment. Reliability analysis should be a best practice when developing assessment of technical skills.
KW - Analysis of Variance
KW - Checklist/methods
KW - Clinical Competence/statistics & numerical data
KW - Concept Formation
KW - Databases, Factual
KW - Education, Medical/statistics & numerical data
KW - Educational Measurement/methods
KW - Humans
KW - Models, Statistical
KW - Practice Guidelines as Topic/standards
KW - Reproducibility of Results
UR - http://www.scopus.com/inward/record.url?scp=85116816314&partnerID=8YFLogxK
U2 - 10.1097/ACM.0000000000004150
DO - 10.1097/ACM.0000000000004150
M3 - Review
C2 - 33951677
SN - 1040-2446
VL - 96
SP - 1609
EP - 1619
JO - Academic medicine : journal of the Association of American Medical Colleges
JF - Academic medicine : journal of the Association of American Medical Colleges
IS - 11
ER -