TY - GEN
T1 - TCUP-Fusion
T2 - 2024 IEEE Ultrasonics, Ferroelectrics, and Frequency Control Joint Symposium, UFFC-JS 2024
AU - Zhang, Boheng
AU - Zheng, Zelin
AU - Huang, Haorui
AU - Ma, Lingyu
AU - Shen, Yi
AU - Sun, Mingjian
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Multimodal medical imaging provides a superior diagnostic profile compared to single-modality imaging, establishing the importance of multimodal medical image fusion (MMIF) for clinical diagnosis and therapeutic applications. Photoacoustic (PA) and ultrasound (US) imaging have become important diagnostic tools as they possess real-time noninvasive and radiation-free features. However, Traditional and emerging deep learning-based fusion techniques struggle to balance accuracy with real-time performance. We propose an end-to-end unsupervised fusion framework based on Transfomer and Convolutional Neural Networks (CNN) called TCUP-Fusion to solve these problems. The framework achieves efficient extraction of features by designing CNN and Biformer based encoders, while using the proposed SCCSA attention to guide the fusion of features, and finally achieves the representation of the fused image relying on a multi-feature fusion decoder. Finally, we conducted qualitative and quantitative comparison experiments on three different modal PA/US datasets. The results show that TCUPFusion outperforms other state-of-the-art fusion methods on US1-PA fusion datasets. In addition, the algorithm in this paper also performs well on other modal image datasets (CT-MRI, MRI-PET), proving that has good robustness.
AB - Multimodal medical imaging provides a superior diagnostic profile compared to single-modality imaging, establishing the importance of multimodal medical image fusion (MMIF) for clinical diagnosis and therapeutic applications. Photoacoustic (PA) and ultrasound (US) imaging have become important diagnostic tools as they possess real-time noninvasive and radiation-free features. However, Traditional and emerging deep learning-based fusion techniques struggle to balance accuracy with real-time performance. We propose an end-to-end unsupervised fusion framework based on Transfomer and Convolutional Neural Networks (CNN) called TCUP-Fusion to solve these problems. The framework achieves efficient extraction of features by designing CNN and Biformer based encoders, while using the proposed SCCSA attention to guide the fusion of features, and finally achieves the representation of the fused image relying on a multi-feature fusion decoder. Finally, we conducted qualitative and quantitative comparison experiments on three different modal PA/US datasets. The results show that TCUPFusion outperforms other state-of-the-art fusion methods on US1-PA fusion datasets. In addition, the algorithm in this paper also performs well on other modal image datasets (CT-MRI, MRI-PET), proving that has good robustness.
KW - Photoacoustic and ultrasound imaging fusion
KW - Transfomer
KW - deep learning
UR - https://www.scopus.com/pages/publications/85216497104
U2 - 10.1109/UFFC-JS60046.2024.10793458
DO - 10.1109/UFFC-JS60046.2024.10793458
M3 - 会议稿件
AN - SCOPUS:85216497104
T3 - IEEE Ultrasonics, Ferroelectrics, and Frequency Control Joint Symposium, UFFC-JS 2024 - Proceedings
BT - IEEE Ultrasonics, Ferroelectrics, and Frequency Control Joint Symposium, UFFC-JS 2024 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 22 September 2024 through 26 September 2024
ER -