Back to all papers

A Taxonomy of Machine Hallucination in Radiology.

March 11, 2026pubmed logopapers

Authors

Brooks FJ,Anastasio MA

Affiliations (1)

  • Department of Bioengineering, Grainger College of Engineering, University of Illinois Urbana-Champaign, 1406 W. Green St, 1102 Everitt Lab MC 278, Urbana, IL 61801.

Abstract

Measuring the rate of machine hallucination is critical to assessing the utility and trustworthiness of generative AI deployed in radiology; however, there are multiple widely used notions of what constitutes hallucination. This ambiguity pervades industry, academia, and regulatory discourse, such that even an experienced radiologist cannot be certain that a generative model has been evaluated using a definition of hallucination aligned with clinical use. As a result, the same generative AI system may be characterized as either never hallucinating or always hallucinating, depending on perspective. This article provides a brief, non-technical explanation of the fundamental disparity between two seemingly incongruous notions of machine hallucination. Using terms familiar to radiologists, a taxonomy of machine hallucination is proposed that explicitly delineates output contingencies for deployed AI systems. By clarifying what constitutes hallucination under different interpretations, this framework aims to reduce ambiguity and facilitate clearer communication among users, developers, vendors, and administrators regarding the performance of generative AI in radiology. © RSNA, 2026.

Topics

Journal Article

Ready to Sharpen Your Edge?

Subscribe to join 11k+ peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.