Abstract
Medical educators have been unable to produce convincing evidence of the construct validity of written or simulation-based assessments of differential diagnosis (DDX) competencies. In 1987, a team of investigators at our institution introduced preliminary reports regarding the psychometric properties of an artificial intelligence-derived DDX assessment instrument. These investigations produced evidence of the construct validity (experts' DDX performance > novices') of the measures derived from this instrument, a linear, fuzzy set-like expert system.
In this investigation, the authors used a non-linear, “Back Propagation” neural network as a DDX assessment instrument. An Acute Chest Pain knowledge base was acquired from each of twenty-four board certified emergency medicine specialists and seventy-four junior and senior medical students. The neural network used these knowledge bases to simulate and assess each subject's individual DDX performance against twenty Acute Chest Pain/Myocardial Infarction test cases.
Student-t test revealed that the DDX performance of experts was significantly superior to novices (p < .001). This finding provides converging evidence of the validity of DDX performance measures produced by both linear and non-linear, artificial intelligence-derived assessment instruments. These instruments may prove to be a useful and powerful new assessment methodology.
Get full access to this article
View all access options for this article.
