Abstract
Objective
We examine AI trust miscalibration—the discrepancy between an individual’s trust in AI and its actual performance—among university students. We assess how the length of explanations and students’ expertise shape the likelihood of alignment with AI recommendations.
Background
The relationship between explainability and users’ trust in AI systems has been scarcely addressed in the current literature, even though AI-assisted processes increasingly affect all professions and hierarchical levels. Given that human–AI relationships are often formed during education, it is crucial to understand how individual and contextual factors influence students’ assessment of AI outputs.
Method
We conducted in-class experiments with 248 students from multiple universities. Participants solved GMAT questions, then viewed an AI recommendation—sometimes correct, sometimes incorrect—with varying explanation depth and eventually could revise their initial answer; student’s final answer being in line with AI recommendation operationalized our measure of “trust.” We estimated logistic models with control variables, including mixed-effects specifications to account for repeated observations.
Results
Explanation complexity is associated with higher trust on average, but its relevance depends on who reads it and whether AI is correct. Students who previously answered correctly exhibited lower willingness to defer, especially when AI was incorrect; conversely, agreement and consistency effects significantly amplified trust. These behavioral patterns highlight conditions under which AI-generated explanations can foster critical engagement or conversely encourage uncritical acceptance.
Conclusion
Our results point to a “AI knows better” heuristic at work—especially among nonexperts—where polished presentation is easily read as reliability, encouraging uncritical agreement with incorrect recommendations; in parallel, experts benefit more from deeper rationales when AI is accurate, yet still display under-reliance of correct assistance in many cases. Overall, trust calibration is driven less by any single cue than by the alignment of student performance, AI reliability, and explanation design, with prior agreement acting as a powerful amplifier of subsequent alignment.
Application
Our findings imply that instructional approaches should promote independent reasoning before exposure to AI, deploy concise but diagnostically informative explanations, and include brief verification steps before accepting AI recommendations, especially for nonexperts who are more prone to harmful switches. Simple monitoring tools that track helpful versus harmful changes could support a more discerning and productive use of AI tools.
Get full access to this article
View all access options for this article.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
