Improving Abnormality Detection on Chest Radiography Using Game-Like Reinforcement Mechanics

Despite their increasing prevalence, online textbooks, question banks, and digital references focus primarily on explicit knowledge. Implicit skills such as abnormality detection require repeated practice on clinical service and have few digital substitutes. Using mechanics traditionally deployed in...

Full description

Saved in:
Bibliographic Details
Published inAcademic radiology Vol. 24; no. 11; p. 1428
Main Authors Chen, Po-Hao, Roth, Howard, Galperin-Aizenberg, Maya, Ruutiainen, Alexander T, Gefter, Warren, Cook, Tessa S
Format Journal Article
LanguageEnglish
Published United States 01.11.2017
Subjects
Online AccessGet more information

Cover

Loading…
More Information
Summary:Despite their increasing prevalence, online textbooks, question banks, and digital references focus primarily on explicit knowledge. Implicit skills such as abnormality detection require repeated practice on clinical service and have few digital substitutes. Using mechanics traditionally deployed in video games such as clearly defined goals, rapid-fire levels, and narrow time constraints may be an effective way to teach implicit skills. We created a freely available, online module to evaluate the ability of individuals to differentiate between normal and abnormal chest radiographs by implementing mechanics, including instantaneous feedback, rapid-fire cases, and 15-second timers. Volunteer subjects completed the modules and were separated based on formal experience with chest radiography. Performance between training and testing sets were measured for each group, and a survey was administered after each session. The module contained 74 cases and took approximately 20 minutes to complete. Thirty-two cases were normal radiographs and 56 cases were abnormal. Of the 60 volunteers recruited, 25 were "never trained" and 35 were "previously trained." "Never trained" users scored 21.9 out of 37 during training and 24.0 out of 37 during testing (59.1% vs 64.9%, P value <.001). "Previously trained" users scored 28.0 out of 37 during training and 28.3 out of 37 during testing phases (75.6% vs 76.4%, P value = .56). Survey results showed that 87% of all subjects agreed the module is an efficient way of learning, and 83% agreed the rapid-fire module is valuable for medical students. A gamified online module may improve the abnormality detection rates of novice interpreters of chest radiography, although experienced interpreters are less likely to derive similar benefits. Users reviewed the educational module favorably.
ISSN:1878-4046
DOI:10.1016/j.acra.2017.05.005