CausalMixNet: A mixed-attention framework for causal intervention in robust medical image diagnosis.

Authors

Zhang Y,Huang YA,Hu Y,Liu R,Wu J,Huang ZA,Tan KC

Affiliations (5)

  • Department of Data Science and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region of China.
  • School of Computer Science, Northwestern Polytechnical University, Xi'an, China.
  • Department of Data Science and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region of China; Department of Computing, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region of China; Research Center on Data Sciences and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region of China.
  • Department of Computer Science, City University of Hong Kong (Dongguan), Dongguan, China. Electronic address: [email protected].
  • Department of Data Science and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region of China; Research Center on Data Sciences and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region of China.

Abstract

Confounding factors inherent in medical images can significantly impact the causal exploration capabilities of deep learning models, resulting in compromised accuracy and diminished generalization performance. In this paper, we present an innovative methodology named CausalMixNet that employs query-mixed intra-attention and key&value-mixed inter-attention to probe causal relationships between input images and labels. For mitigating unobservable confounding factors, CausalMixNet integrates the non-local reasoning module (NLRM) and the key&value-mixed inter-attention (KVMIA) to conduct a front-door adjustment strategy. Furthermore, CausalMixNet incorporates a patch-masked ranking module (PMRM) and query-mixed intra-attention (QMIA) to enhance mediator learning, thereby facilitating causal intervention. The patch mixing mechanism applied to query/(key&value) features within QMIA and KVMIA specifically targets lesion-related feature enhancement and the inference of average causal effect inference. CausalMixNet consistently outperforms existing methods, achieving superior accuracy and F1-scores across in-domain and out-of-domain scenarios on multiple datasets, with an average improvement of 3% over the closest competitor. Demonstrating robustness against noise, gender bias, and attribute bias, CausalMixNet excels in handling unobservable confounders, maintaining stable performance even in challenging conditions.

Topics

Deep LearningImage Interpretation, Computer-AssistedJournal Article

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.