Journal of Southern Medical University ›› 2025, Vol. 45 ›› Issue (6): 1327-1335.doi: 10.12122/j.issn.1673-4254.2025.06.22

Previous Articles    

Incomplete multimodal bone tumor image classification based on feature decoupling and fusion

Qinghai ZENG1,2(), Chuanpu LI1,2, Wei YANG1,2, Liwen SONG3, Yinghua ZHAO3, Yi YANG1,2()   

  1. 1.School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China
    2.Guangdong Provincial Key Laboratory of Medical Image Processing, Guangzhou 510515, China
    3.Department of Radiology, Third Affiliated Hospital of Southern Medical University, Guangzhou 510630, China
  • Received:2024-09-20 Online:2025-06-20 Published:2025-06-27
  • Contact: Yi YANG E-mail:qinghaizeng982@163.com;yiyang20110130@163.com
  • Supported by:
    National Natural Science Foundation of China(82172020)

Abstract:

Objective To construct a bone tumor classification model based on feature decoupling and fusion for processing modality loss and fusing multimodal information to improve classification accuracy. Methods A decoupling completion module was designed to extract local and global bone tumor image features from available modalities. These features were then decomposed into shared and modality-specific features, which were used to complete the missing modality features, thereby reducing completion bias caused by modality differences. To address the challenge of modality differences that hinder multimodal information fusion, a cross-attention-based fusion module was introduced to enhance the model's ability to learn cross-modal information and fully integrate specific features, thereby improving the accuracy of bone tumor classification. Results The experiment was conducted using a bone tumor dataset collected from the Third Affiliated Hospital of Southern Medical University for training and testing. Among the 7 available modality combinations, the proposed method achieved an average AUC, accuracy, and specificity of 0.766, 0.621, and 0.793, respectively, which represent improvements of 2.6%, 3.5%, and 1.7% over existing methods for handling missing modalities. The best performance was observed when all the modalities were available, resulting in an AUC of 0.837, which still reached 0.826 even with MRI alone. Conclusion The proposed method can effectively handle missing modalities and successfully integrate multimodal information, and show robust performance in bone tumor classification under various complex missing modality scenarios.

Key words: bone tumor classification, multimodal imaging, modality missing, feature decoupling, attention fusion