» Articles » PMID: 36586177

Skin Lesion Classification Based on Two-modal Images Using a Multi-scale Fully-shared Fusion Network

Overview
Date 2022 Dec 31
PMID 36586177
Authors
Affiliations
Soon will be listed here.
Abstract

Background And Objective: Due to the complexity of skin lesion features, computer-aided diagnosis of skin diseases based on multi-modal images is considered a challenging task. Dermoscopic images and clinical images are commonly used to diagnose skin diseases in clinical scenarios, and the complementarity of their features promotes the research of multi-modality classification in the computer-aided diagnosis field. Most current methods focus on the fusion between modalities and ignore the complementary information within each of them, which leads to the loss of the intra-modality relation. Multi-modality models for integrating features both within single modalities and across multiple modalities are limited in the literature. Therefore, a multi-modality model based on dermoscopic and clinical images is proposed to address this issue.

Methods: We propose a Multi-scale Fully-shared Fusion Network (MFF-Net) that gathers features of dermoscopic images and clinical images for skin lesion classification. In MFF-Net, the multi-scale fusion structure combines deep and shallow features within individual modalities to reduce the loss of spatial information in high-level feature maps. Then Dermo-Clinical Block (DCB) integrates the feature maps from dermoscopic images and clinical images through channel-wise concatenation and using a fully-shared fusion strategy that explores complementary information at different stages.

Results: We validated our model on a four-class two-modal skin diseases dataset, and proved that the proposed multi-scale structure, the fusion module DCBs, and the fully-shared fusion strategy improve the performance of MFF-Net independently. Our method achieved the highest average accuracy of 72.9% on the 7-point checklist dataset, outperforming the state-of-the-art single-modality and multi-modality methods with an accuracy boost of 7.1% and 3.4%, respectively.

Conclusions: The multi-scale fusion structure demonstrates the significance of intra-modality relations between clinical images and dermoscopic images. The proposed network combined with the multi-scale structure, DCBs, and the fully-shared fusion strategy, can effectively integrate the features of the skin lesions across the two modalities and achieved a promising accuracy among different skin diseases.

Citing Articles

Development, Application and Utility of a Machine Learning Approach for Melanoma and Non-Melanoma Lesion Classification Using Counting Box Fractal Dimension.

Romero-Morelos P, Herrera-Lopez E, Gonzalez-Yebra B Diagnostics (Basel). 2024; 14(11).

PMID: 38893659 PMC: 11171650. DOI: 10.3390/diagnostics14111132.


Color Analysis of Merkel Cell Carcinoma: A Comparative Study with Cherry Angiomas, Hemangiomas, Basal Cell Carcinomas, and Squamous Cell Carcinomas.

Koumaki D, Manios G, Papadakis M, Doxastaki A, Zacharopoulos G, Katoulis A Diagnostics (Basel). 2024; 14(2).

PMID: 38275477 PMC: 10814937. DOI: 10.3390/diagnostics14020230.


STNet: shape and texture joint learning through two-stream network for knowledge-guided image recognition.

Wang X, Han H, Xu M, Li S, Zhang D, Du S Front Neurosci. 2023; 17:1212049.

PMID: 37397450 PMC: 10309034. DOI: 10.3389/fnins.2023.1212049.