A Multimodal Convolutional Neural Network Based Approach for DICOM Files Classification. , 2025(1), e70107.

Abstract

In this study, we developed a convolutional neural network approach for directly classifying digital imaging and communication in medicine files in medical imaging applications. Existing models require converting this format into other formats like portable network graphics. This conversion leads to metadata loss and classification bias, the developed model processes raw digital imaging and communication in medicine files, thereby preserving both pixel data and embedded metadata. The model was evaluated on chest X-ray images for tuberculosis detection and magnetic resonance imaging scan images for brain tumour classification from the National Institute of Allergy and Infectious Diseases. The X-ray modality achieved a precision of 92.9%, recall of 88.4%, F1-score of 90.6% and accuracy of 90.9%, while the magnetic resonance imaging modality obtained a precision of 80.0%, recall of 79.4%, F1-score of 79.7% and accuracy of 85.5%. These results demonstrate the model’s effectiveness across multiple imaging modalities. A key advantage of this approach is the preservation of diagnostic metadata, enhancing accuracy and reducing classification bias. The study highlights its potential to improve medical imaging and support real-time clinical decision making. Despite the promising results, the study acknowledges limitations in dataset diversity and computational efficiency, with future work focusing on addressing these challenges and further optimising the model for deployment in resource-limited environments.

Description

Keywords

Multimodal convolutional neural network (CNN), DICOM file classification, Raw DICOM processing, Metadata preservation, Image pixel data and metadata fusion

Citation

Vicent, M., Willian, W., & Simon, K. (2025). A Multimodal Convolutional Neural Network Based Approach for DICOM Files Classification. The Journal of Engineering, 2025(1), e70107.