Recently, medical image fusion has emerged as an impressive technique in merging the medical images of different modalities. Certainly, the fused image assists the physician in disease diagnosis for effective treatment planning. The fusion process combines multi-modal images to incur a single image with excellent quality, retaining the information of original images. This paper proposes a multi-modal medical image fusion through a weighted blending of high-frequency subbands of nonsubsampled shearlet transform (NSST) domain via chaotic grey wolf optimization algorithm. As an initial step, the NSST is applied on source images to decompose into the multi-scale and multi-directional components. The low-frequency bands are fused based on a simple max rule to sustain the energy of an individual. The texture details of input images are preserved by an adaptively weighted combination of high-frequency images using a recent chaotic grey wolf optimization algorithm to minimize the distance between the fused image and source images. The entire process emphasizes on retaining the energy of the low-frequency band and the transferring of texture features from source images to the fused image. Finally, the fused image is formed using inverse NSST of merged low and high-frequency bands. The experiments are carried out on eight different disease datasets obtained from Brain Atlas, which consists of MR-T1 and MR-T2, MR and SPECT, MR and PET, and MR and CT. The effectiveness of the proposed method is validated using more than 100 pairs of images based on the subjective and objective quality assessment. The experimental results confirm that the proposed method performs better in contrast with the current state-of-the-art image fusion techniques in terms of entropy, VIFF, and FMI. Hence, the proposed method will be helpful for disease diagnosis, medical treatment planning, and surgical procedure.
All Science Journal Classification (ASJC) codes
- Computer Science(all)
- Materials Science(all)