People with melanoma often have mole-looking growths on their skin that tend to be irregular in shape and color, and can be hard to tell apart from benign ones, making the disease difficult to diagnose. Even experts can be fooled by this type of skin cancer.
Now, an automated technology that combines imaging with digital analysis and machine learning to help physicians detect melanoma at its early stages has been developed researchers at the Rockefeller University.
In this new approach, images of lesions are processed by a series of computer programs that extract information about the number of colors present in a growth, and other quantitative data. The analysis generates an overall risk score, called a Q-score, which indicates the likelihood that the growth is cancerous.
98 Percent Sensitivity
James Krueger, D. Martin Carter Professor in Clinical Investigation and head of the Laboratory of Investigative Dermatology, said:
“There is a real need for standardization across the field of dermatology in how melanomas are evaluated. Detection through screening saves lives but is very challenging visually, and even when a suspicious lesion is extracted and biopsied, it is confirmed to be melanoma in only about 10 percent of cases.”
A recent study evaluating the tool’s usefulness indicates that the Q-score yields 98 percent sensitivity, meaning it is very likely to correctly identify early melanomas on the skin. The ability of the test to correctly diagnose normal moles was 36 percent, approaching the levels achieved by expert dermatologists performing visual examinations of suspect moles under the microscope.
“The success of the Q-score in predicting melanoma is a marked improvement over competing technologies,”
says Daniel Gareau, first author of the report and instructor in clinical investigation in the Krueger laboratory.
The researchers developed this tool by feeding 60 photos of cancerous melanomas and an equivalent batch of pictures of benign growths into image processing programs. They developed imaging biomarkers to precisely quantify visual features of the growths.
Using multiple computational machine learning approaches, they generated a set of quantitative metrics that differed between the two groups of images, essentially identifying what visual aspects of the lesion mattered most in terms of malignancy, and gave each biomarker a malignancy rating.
By combining the data from each biomarker, they calculated the overall Q-score for each image, a value between zero and one in which a higher number indicates a higher probability of a lesion being cancerous.
As previous studies have shown, the number of colors in a lesion turned out to be the most significant biomarker for determining malignancy. And some biomarkers were significant only if looked at in specific color channels—a finding the researchers say could potentially be exploited to identify additional biomarkers and further improve accuracy.
This work was supported in part by a grant from the National Institutes of Health and in part by the Paul and Irma Milstein Family Foundation and the American Skin Association.
Image: A skin lesion (left) is processed by a new technology to extract quantitative data (right), such as irregularities in the shape of pigmented skin, which could help doctors determine if the growth is cancerous. Credit: Rockefeller University