There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.
Abstract
What is the accuracy of computer-aided diagnosis of melanoma and how does it translate
to clinical practice? In this meta-analysis of 70 studies, the accuracy of computer-aided
diagnosis is comparable to that of human experts. However, current studies are heterogeneous
and most deviate significantly from real-world scenarios and are prone to biases.
Although computer-aided diagnosis for melanoma appears to be accurate according to
the included studies, more standardized and realistic study settings are required
to explore its full potential in clinical practice. The recent advances in the field
of machine learning have raised expectations that computer-aided diagnosis will become
the standard for the diagnosis of melanoma. To critically review the current literature
and compare the diagnostic accuracy of computer-aided diagnosis with that of human
experts. The MEDLINE, arXiv, and PubMed Central databases were searched to identify
eligible studies published between January 1, 2002, and December 31, 2018. Studies
that reported on the accuracy of automated systems for melanoma were selected. Search
terms included melanoma , diagnosis , detection , computer aided , and artificial
intelligence . Evaluation of the risk of bias was performed using the QUADAS-2 tool,
and quality assessment was based on predefined criteria. Data were analyzed from February
1 to March 10, 2019. Summary estimates of sensitivity and specificity and summary
receiver operating characteristic curves were the primary outcomes. The literature
search yielded 1694 potentially eligible studies, of which 132 were included and 70
offered sufficient information for a quantitative analysis. Most studies came from
the field of computer science. Prospective clinical studies were rare. Combining the
results for automated systems gave a melanoma sensitivity of 0.74 (95% CI, 0.66-0.80)
and a specificity of 0.84 (95% CI, 0.79-0.88). Sensitivity was lower in studies that
used independent test sets than in those that did not (0.51; 95% CI, 0.34-0.69 vs
0.82; 95% CI, 0.77-0.86; P < .001); however, the specificity was similar (0.83;
95% CI, 0.71-0.91 vs 0.85; 95% CI, 0.80-0.88; P = .67). In comparison with dermatologists’
diagnosis, computer-aided diagnosis showed similar sensitivities and a 10 percentage
points lower specificity, but the difference was not statistically significant. Studies
were heterogeneous and substantial risk of bias was found in all but 4 of the 70 studies
included in the quantitative analysis. Although the accuracy of computer-aided diagnosis
for melanoma detection is comparable to that of experts, the real-world applicability
of these systems is unknown and potentially limited owing to overfitting and the risk
of bias of the studies at hand. This meta-analysis evaluates the accuracy of computerized
systems in the diagnosis of melanoma in patients with skin lesions.
Automated melanoma recognition in dermoscopy images is a very challenging task due to the low contrast of skin lesions, the huge intraclass variation of melanomas, the high degree of visual similarity between melanoma and non-melanoma lesions, and the existence of many artifacts in the image. In order to meet these challenges, we propose a novel method for melanoma recognition by leveraging very deep convolutional neural networks (CNNs). Compared with existing methods employing either low-level hand-crafted features or CNNs with shallower architectures, our substantially deeper networks (more than 50 layers) can acquire richer and more discriminative features for more accurate recognition. To take full advantage of very deep networks, we propose a set of schemes to ensure effective training and learning under limited training data. First, we apply the residual learning to cope with the degradation and overfitting problems when a network goes deeper. This technique can ensure that our networks benefit from the performance gains achieved by increasing network depth. Then, we construct a fully convolutional residual network (FCRN) for accurate skin lesion segmentation, and further enhance its capability by incorporating a multi-scale contextual information integration scheme. Finally, we seamlessly integrate the proposed FCRN (for segmentation) and other very deep residual networks (for classification) to form a two-stage framework. This framework enables the classification network to extract more representative and specific features based on segmented results instead of the whole dermoscopy images, further alleviating the insufficiency of training data. The proposed framework is extensively evaluated on ISBI 2016 Skin Lesion Analysis Towards Melanoma Detection Challenge dataset. Experimental results demonstrate the significant performance gains of the proposed framework, ranking the first in classification and the second in segmentation among 25 teams and 28 teams, respectively. This study corroborates that very deep CNNs with effective training mechanisms can be employed to solve complicated medical image analysis tasks, even with limited training data.
Skin lesions are a severe disease globally. Early detection of melanoma in dermoscopy images significantly increases the survival rate. However, the accurate recognition of melanoma is extremely challenging due to the following reasons: low contrast between lesions and skin, visual similarity between melanoma and non-melanoma lesions, etc. Hence, reliable automatic detection of skin tumors is very useful to increase the accuracy and efficiency of pathologists. In this paper, we proposed two deep learning methods to address three main tasks emerging in the area of skin lesion image processing, i.e., lesion segmentation (task 1), lesion dermoscopic feature extraction (task 2) and lesion classification (task 3). A deep learning framework consisting of two fully convolutional residual networks (FCRN) is proposed to simultaneously produce the segmentation result and the coarse classification result. A lesion index calculation unit (LICU) is developed to refine the coarse classification results by calculating the distance heat-map. A straight-forward CNN is proposed for the dermoscopic feature extraction task. The proposed deep learning frameworks were evaluated on the ISIC 2017 dataset. Experimental results show the promising accuracies of our frameworks, i.e., 0.753 for task 1, 0.848 for task 2 and 0.912 for task 3 were achieved.
To measure the performance of smartphone applications that evaluate photographs of skin lesions and provide the user with feedback about the likelihood of malignancy. Case-control diagnostic accuracy study. Academic dermatology department. PARTICIPANTS AND MATERIALS: Digital clinical images of pigmented cutaneous lesions (60 melanoma and 128 benign control lesions) with a histologic diagnosis rendered by a board-certified dermatopathologist, obtained before biopsy from patients undergoing lesion removal as a part of routine care. Sensitivity, specificity, and positive and negative predictive values of 4 smartphone applications designed to aid nonclinician users in determining whether their skin lesion is benign or malignant. Sensitivity of the 4 tested applications ranged from 6.8% to 98.1%; specificity, 30.4% to 93.7%; positive predictive value, 33.3% to 42.1%; and negative predictive value, 65.4% to 97.0%. The highest sensitivity for melanoma diagnosis was observed for an application that sends the image directly to a board-certified dermatologist for analysis; the lowest, for applications that use automated algorithms to analyze images. The performance of smartphone applications in assessing melanoma risk is highly variable, and 3 of 4 smartphone applications incorrectly classified 30% or more of melanomas as unconcerning. Reliance on these applications, which are not subject to regulatory oversight, in lieu of medical consultation can delay the diagnosis of melanoma and harm users.
scite shows how a scientific paper has been cited by providing the context of the citation, a classification describing whether it supports, mentions, or contrasts the cited claim, and a label indicating in which section the citation was made.