TY - JOUR
T1 - Quality of cost evaluations of physician continuous professional development
T2 - Systematic review of reporting and methods
AU - Cook, David A.
AU - Wilkinson, John M.
AU - Foo, Jonathan
N1 - Funding Information:
We thank Barbara L. Baasch-Thomas, BSN, MA, Stephen Maloney, B. Physio, MPH, PhD, and Christopher R. Stephenson, MD, for their efforts in screening studies for inclusion, and Larry J. Prokop, MLIS, for his involvement in creating the search strategy. The study received no external funding.
Publisher Copyright:
© 2022, The Author(s).
PY - 2022/6
Y1 - 2022/6
N2 - Introduction: We sought to evaluate the reporting and methodological quality of cost evaluations of physician continuing professional development (CPD). Methods: We conducted a systematic review, searching MEDLINE, Embase, PsycInfo, and the Cochrane Database for studies comparing the cost of physician CPD (last update 23 April 2020). Two reviewers, working independently, screened all articles for inclusion. Two reviewers extracted information on reporting quality using the Consolidated Health Economic Evaluation Reporting Standards (CHEERS), and on methodological quality using the Medical Education Research Study Quality Instrument (MERSQI) and a published reference case. Results: Of 3338 potentially eligible studies, 62 were included. Operational definitions of methodological and reporting quality elements were iteratively revised. Articles reported mean (SD) 43% (20%) of CHEERS elements for the Title/Abstract, 56% (34%) for Introduction, 66% (19%) for Methods, 61% (17%) for Results, and 66% (30%) for Discussion, with overall reporting index 292 (83) (maximum 500). Valuation methods were reported infrequently (resource selection 10 of 62 [16%], resource quantitation 10 [16%], pricing 26 [42%]), as were descriptions/discussion of the physicians trained (42 [68%]), training setting (42 [68%]), training intervention (40 [65%]), sensitivity analyses of uncertainty (9 [15%]), and generalizability (30 [48%]). MERSQI scores ranged from 6.0 to 16.0 (mean 11.2 [2.4]). Changes over time in reporting index (initial 241 [105], final 321 [52]) and MERSQI scores (initial 9.8 [2.7], final 11.9 [1.9]) were not statistically significant (p ≥ 0.08). Discussion: Methods and reporting of HPE cost evaluations fall short of current standards. Gaps exist in the valuation, analysis, and contextualization of cost outcomes.
AB - Introduction: We sought to evaluate the reporting and methodological quality of cost evaluations of physician continuing professional development (CPD). Methods: We conducted a systematic review, searching MEDLINE, Embase, PsycInfo, and the Cochrane Database for studies comparing the cost of physician CPD (last update 23 April 2020). Two reviewers, working independently, screened all articles for inclusion. Two reviewers extracted information on reporting quality using the Consolidated Health Economic Evaluation Reporting Standards (CHEERS), and on methodological quality using the Medical Education Research Study Quality Instrument (MERSQI) and a published reference case. Results: Of 3338 potentially eligible studies, 62 were included. Operational definitions of methodological and reporting quality elements were iteratively revised. Articles reported mean (SD) 43% (20%) of CHEERS elements for the Title/Abstract, 56% (34%) for Introduction, 66% (19%) for Methods, 61% (17%) for Results, and 66% (30%) for Discussion, with overall reporting index 292 (83) (maximum 500). Valuation methods were reported infrequently (resource selection 10 of 62 [16%], resource quantitation 10 [16%], pricing 26 [42%]), as were descriptions/discussion of the physicians trained (42 [68%]), training setting (42 [68%]), training intervention (40 [65%]), sensitivity analyses of uncertainty (9 [15%]), and generalizability (30 [48%]). MERSQI scores ranged from 6.0 to 16.0 (mean 11.2 [2.4]). Changes over time in reporting index (initial 241 [105], final 321 [52]) and MERSQI scores (initial 9.8 [2.7], final 11.9 [1.9]) were not statistically significant (p ≥ 0.08). Discussion: Methods and reporting of HPE cost evaluations fall short of current standards. Gaps exist in the valuation, analysis, and contextualization of cost outcomes.
KW - Cost effectiveness
KW - Costs and cost analysis
KW - Education, continuing
KW - Education, medical
UR - http://www.scopus.com/inward/record.url?scp=85127418508&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85127418508&partnerID=8YFLogxK
U2 - 10.1007/s40037-022-00705-z
DO - 10.1007/s40037-022-00705-z
M3 - Article
C2 - 35357652
AN - SCOPUS:85127418508
SN - 2212-2761
VL - 11
SP - 156
EP - 164
JO - Perspectives on Medical Education
JF - Perspectives on Medical Education
IS - 3
ER -