A non-sub-sampled shearlet transform-based deep learning sub band enhancement and fusion method for multi-modal images.
Authors
Affiliations (6)
Affiliations (6)
- Department of Computer Science and Engineering, PSN College of Engineering and Technology, Tirunelveli, Tamil Nadu, 627152, India. [email protected].
- Department of Computer Science and Engineering, Siddhartha Institute of Technology & Science (SITS), Narapally, Hyderabad, Telangana, 500088, India.
- Department of Computer Science, College of Computers and Information Technology, Taif University, P. O. Box 11099, 21944, Taif, Saudi Arabia.
- Department of Electronics and Communication Engineering, Kuwait College of Science and Technology (KCST), Doha Area, 13060, Kuwait.
- Electrical Engineering Department, University of Business and Technology, 23435, Jeddah, Saudi Arabia. [email protected].
- Engineering Mathematics Department, Faculty of Engineering, Alexandria University, Alexandria, 21544, Egypt. [email protected].
Abstract
Multi-Modal Medical Image Fusion (MMMIF) has become increasingly important in clinical applications, as it enables the integration of complementary information from different imaging modalities to support more accurate diagnosis and treatment planning. The primary objective of Medical Image Fusion (MIF) is to generate a fused image that retains the most informative features from the Source Images (SI), thereby enhancing the reliability of clinical decision-making systems. However, due to inherent limitations in individual imaging modalities-such as poor spatial resolution in functional images or low contrast in anatomical scans-fused images can suffer from information degradation or distortion. To address these limitations, this study proposes a novel fusion framework that integrates the Non-Subsampled Shearlet Transform (NSST) with a Convolutional Neural Network (CNN) for effective sub-band enhancement and image reconstruction. Initially, each source image is decomposed into Low-Frequency Coefficients (LFC) and multiple High-Frequency Coefficients (HFC) using NSST. The proposed Concurrent Denoising and Enhancement Network (CDEN) is then applied to these sub-bands to suppress noise and enhance critical structural details. The enhanced LFCs are fused using an AlexNet-based activity-level fusion model, while the enhanced HFCs are combined using a Pulse Coupled Neural Network (PCNN) guided by a Novel Sum-Modified Laplacian (NSML) metric. Finally, the fused image is reconstructed via Inverse-NSST (I-NSST). Experimental results prove that the proposed method outperforms existing fusion algorithms, achieving approximately 16.5% higher performance in terms of the QAB/F (edge preservation) metric, along with strong results across both subjective visual assessments and objective quality indices.