Back to all papers

Lung adenocarcinoma subtype classification based on contrastive learning model with multimodal integration.

Authors

Wang C,Liu L,Fan C,Zhang Y,Mai Z,Li L,Liu Z,Tian Y,Hu J,Elazab A

Affiliations (8)

  • Shenzhen Research Institute of Big Data, Shenzhen, China.
  • National Cancer Center, Chinese Academy of Medical Sciences and Peking Union Medical College, National Clinical Research Center for Cancer, Cancer Hospital and Shenzhen Hospital, Shenzhen, China.
  • Zhejiang University of Finance and Economics, Hangzhou, China.
  • Zhejiang University of Finance and Economics, Hangzhou, China. [email protected].
  • National Cancer Center, Chinese Academy of Medical Sciences and Peking Union Medical College, National Clinical Research Center for Cancer, Cancer Hospital and Shenzhen Hospital, Shenzhen, China. [email protected].
  • The Second Affiliated Hospital, School of Medicine, The Chinese University of Hong Kong, Shenzhen, China.
  • Hongqi Hospital Affiliated to Mudanjiang Medical University, Mudanjiang, China.
  • School of Biomedical Engineering, Shenzhen University, Shenzhen, China.

Abstract

Accurately identifying the stages of lung adenocarcinoma is essential for selecting the most appropriate treatment plans. Nonetheless, this task is complicated due to challenges such as integrating diverse data, similarities among subtypes, and the need to capture contextual features, making precise differentiation difficult. We address these challenges and propose a multimodal deep neural network that integrates computed tomography (CT) images, annotated lesion bounding boxes, and electronic health records. Our model first combines bounding boxes with precise lesion location data and CT scans, generating a richer semantic representation through feature extraction from regions of interest to enhance localization accuracy using a vision transformer module. Beyond imaging data, the model also incorporates clinical information encoded using a fully connected encoder. Features extracted from both CT and clinical data are optimized for cosine similarity using a contrastive language-image pre-training module, ensuring they are cohesively integrated. In addition, we introduce an attention-based feature fusion module that harmonizes these features into a unified representation to fuse features from different modalities further. This integrated feature set is then fed into a classifier that effectively distinguishes among the three types of adenocarcinomas. Finally, we employ focal loss to mitigate the effects of unbalanced classes and contrastive learning loss to enhance feature representation and improve the model's performance. Our experiments on public and proprietary datasets demonstrate the efficiency of our model, achieving a superior validation accuracy of 81.42% and an area under the curve of 0.9120. These results significantly outperform recent multimodal classification approaches. The code is available at https://github.com/fancccc/LungCancerDC .

Topics

Adenocarcinoma of LungLung NeoplasmsJournal Article

Ready to Sharpen Your Edge?

Join hundreds of your peers who rely on RadAI Slice. Get the essential weekly briefing that empowers you to navigate the future of radiology.

We respect your privacy. Unsubscribe at any time.