Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Towards the automatic detection of skin lesion shape asymmetry, color variegation and diameter in dermoscopic images

Abstract

Asymmetry, color variegation and diameter are considered strong indicators of malignant melanoma. The subjectivity inherent in the first two features and the fact that 10% of melanomas tend to be missed in the early diagnosis due to having a diameter less than 6mm, deem it necessary to develop an objective computer vision system to evaluate these criteria and aid in the early detection of melanoma which could eventually lead to a higher 5-year survival rate. This paper proposes an approach for evaluating the three criteria objectively, whereby we develop a measure to find asymmetry with the aid of a decision tree which we train on the extracted asymmetry measures and then use to predict the asymmetry of new skin lesion images. A range of colors that demonstrate the suspicious colors for the color variegation feature have been derived, and Feret’s diameter has been utilized to find the diameter of the skin lesion. The decision tree is 80% accurate in determining the asymmetry of skin lesions, and the number of suspicious colors and diameter values are objectively identified.

Introduction

Melanoma can be recognizable by most physicians and patients when in its advanced stage. However, it is highly curable if diagnosed early and treated timely [12]. With the advent of immunotherapy and targeted therapy, melanoma treatments have improved significantly; the five-year survival rate of early stage melanoma and metastatic melanoma (i.e. stage IV) is 98.4% and 22.5%, respectively [3]. A group of researchers at the New York University found the ABCD rule (Asymmetry, Border irregularity, Color variegation, and Diameter greater than 6mm) in 1985 as a simple framework that physicians, novice dermatologists and non-physicians could use to learn about the features of melanoma in its early curable stage [4]; the rule is being promoted by the American Cancer Society as a method to help in seeking early medical evaluation of any suspicious pigmented lesions. The ABCD rule provides a checklist of premalignant changes in skin lesions and an appraisal of pigmented cutaneous lesions that may need to be further examined by a specialist which might result in further work of dermoscopy or biopsy, or both.

Skin lesion asymmetry is a strong indicator of malignant melanoma [5] such that the degree of asymmetry displayed by a skin lesion is indicative of its malignant potential. As opposed to benign pigmented skin lesions that are usually circular and symmetric, melanomas tend to develop in an uncontrolled fashion and grow at an irregular rate, rendering them to be asymmetric [6]. In layman terms, asymmetry refers to the fact that when drawing a line through the middle of the mole the two halves will not match, meaning that the shape of one half doesn’t resemble the other half (lopsided in shape), providing a warning sign of melanoma. There is no consensus on what asymmetry extent is required before one can tell that the skin lesion is considered asymmetric. It has been shown that the percentage of disagreement between dermatologists on the presence of asymmetry is around 5-10% [4]. A more objective measurement of asymmetry is thus deemed necessary.

Asymmetry evaluation is carried out by separating the lesion into four sectors using orthogonal axes that pass through the lesion centroid and are aligned so that minimum asymmetry (maximum symmetry) is obtained [7]. Different attempts have been made to automatically determine the asymmetry of skin lesions in literature, which can be summarized as shown in Table 1.

thumbnail
Table 1. Determining asymmetry of skin lesions in literature.

https://doi.org/10.1371/journal.pone.0234352.t001

Color variegation is considered the earliest sign of melanoma, and has a high predictability for the diagnosis of the disease. In fact, studies have demonstrated that color variegation might be the most important singular discriminator of melanoma [15]. It refers to the presence of two or more shades of pigment (two or more colors) within the skin lesion border. As opposed to benign lesions which tend to be generally uniform in color, melanoma lesions tend to often contain more than two colors, meaning that the color composition within the skin lesion is inhomogeneous. Melanoma in particular contains one or more of these six suspicious shades of color: white, red, light brown, dark brown, blue-gray, and black.

To characterize the color composition within lesions different approaches have been proposed in literature as depicted in Table 2.

Most early melanomas (i.e. stage 0) tend to be larger than 6mm in diameter (i.e. size of a pencil eraser). Some studies argue that smaller diameters can exist in melanoma, which makes this criterion not absolute especially that 10% of melanomas tend to be missed in the early diagnosis (i.e. diameter <6mm) if the diagnosis was based only on diameter, making it preferable to use a computer vision system when evaluating diameter [21]. Table 3 lists different studies used to find the skin lesion diameter.

The aim of this paper is to present an automated approach to determining skin lesion asymmetry, color variegation, and diameter in dermoscopic images, where different methods and measures are proposed for carrying out this task. Border irregularity (i.e. the B feature in the ABCD rule) has been discussed in our other work in [25].

Methods

Image segmentation

In this section we describe the segmentation method utilized in the paper. In particular, we propose an improved version of Otsu’s method [26] for skin lesion segmentation coupled with pre-processing and post-processing stages as described below.

Say we have two classes: lesion (L) and skin (S), the variance of the pixels in those two classes can be defined as: (1) (2) where k is the graylevel value, [0, L − 1] is the range of graylevel (intensity) levels, pi is the number of times pixel (graylevel) i occurred in the image which can be obtained from the image histogram. The histogram is normalized and perceived as a probability distribution, that is: (3) where ni is the number of pixels at graylevel i, and N = (n0+ n1+ …nL − 1) represents the total number of pixels in the image.

μL and μS represent the lesion class mean and skin class mean, respectively, which are defined as: (4) (5) PL and PS represent the probabilities of class occurrence of the lesion and skin, respectively, defined as: (6) (7)

The within class variance which Otsu’s method attempts to minimize by finding an optimal threshold is defined as: (8)

The between class variance on the other hand, which Otsu’s method attempts to maximize, is defined as: (9) where μT(k) is the total mean, defined as: (10)

This is equivalent to Eq 11, that is subtracting the within-class variance from the total variance (σ2) of the combined distribution. (11)

The threshold k with the maximum between-class variance has also the minimum within-class variance.

The class separability η is: (12)

In our proposed improvement, which we refer to here as Otsu-II, instead of using PL(k) and PS(k) we use the new measures DL(k) and DS(k) defined as: (13) (14) (15) (16) where σ′ is the normalized standard deviation (). and can be rewritten as: (17) (18)

Thus, how spread out the pixel intensities are in each class is taken into account, such that when the pixel intensities are spread apart D will be large and when they are tightly bunched together D will be small. Standard deviation is used since it expresses the statistical distribution of each class more accurately than variance, especially that the dispersion of classes are measured as the distance between the mean of a class and any intensity value, which is proportional to the standard deviation rather than variance. This makes the optimal threshold less biased towards the larger variance among two class variances (i.e. larger dispersion of two classes) [27]. Segmentation performance of using the new measures will be demonstrated in terms of the Dice coefficient [28] in the results section.

To improve the segmentation process further, we introduce pre-processing and post-processing operations which together form the proposed approach depicted in Fig 1. The approach starts by locating the important parts of the image (i.e. lesion). Otsu-II is then applied on the salient object to obtain a binary image that will be used as a groundtruth in the trimap generation step in which each image pixel will be assigned to either the lesion, skin, or as a mixture of both the foreground (lesion) and background (skin). The alpha matte is finally created to distinguish more accurately between the skin and lesion, eventually finding the final segmentation.

Saliency detection is the process of automatically locating the important parts of an image, where saliency refers to the unique features (i.e. pixels) of the image. The output of the saliency detection step is a map where the intensity of each pixel represents the probability of the pixel belonging to a salient object. In this work, we use a Discriminative Regional Feature Integration approach (DRFI) for saliency detection [29, 30] as it is considered one of the most efficient algorithms for saliency detection [30].

After detecting the salient objects, we apply our improved Otsu method (i.e. Otsu-II) on the objects to create binary images that will serve as image masks (i.e. groundtruth) in the trimap generation step, in which each pixel in the image is assigned to three possible values: definite foreground (lesion), definite background (skin) and uncertain (a mixture of foreground and background pixels). To create the trimap, erosion and dilation morphological operations are applied on the binary image mask of the skin lesions (i.e. salient objects). The trimap can be generated using the following formula [31]: (19) where M(.) refers to a set of pixels in the image, and F(.) and B(.) are the functions that extract the foreground and background pixels, respectively. R, E, D and G denote the trimap, eroded image, dilated image, and the gap between the foreground and background resulting from the morphological operations, respectively. F(E) represents the foreground pixels (white), B(D) represents the background pixels (black), and M(G) represents the uncertain pixels (gray).

The final step in the proposed approach is matting, that is finding (creating) the alpha matte/channel αn which is used to accurately distinguish between the foreground and background, rendering the final segmentation. In this step, we utilize KNN (K-nearest Neighbors) matting [32] to find αn, which is represented as follows: (20) where the inputs In and Tn are the original input image and the corresponding trimap image, respectively.

As we are going to measure asymmetry (especially when using Scale-invariant feature transform (SIFT)) and color variegation using the original color image, we would like to focus only on the extracted skin lesion rather than the background and any accompanying artefacts. This can be achieved by merging the original image (i.e. color image) and its corresponding alpha matte described above. Fig 2 shows some examples on original images and their extracted lesions using this process.

thumbnail
Fig 2. The original color image and its extracted lesion.

https://doi.org/10.1371/journal.pone.0234352.g002

The proposed segmentation approach was applied on 204 randomly chosen skin lesion images (nevus: 175, melanoma: 29) extracted from the “ISIC 2018: Skin Lesion Analysis Towards Melanoma Detection grand challenge datasets” [33, 34]. The Dice coefficient [28] is used to measure the similarity between two images by finding the spatial overlap between two binary images, resulting in a value that lies between 0 (no overlap) and 1 (agree perfectly). The Dice coefficient can be defined as follows: (21) where A is the algorithm output and G is the ground truth.

Asymmetry

For measuring asymmetry we build a vector of three measurements that will be subsequently used to train and test a classifier (i.e. decision tree) as will be explained in more detail in the coming sections. Decision trees offer a structured way of decision making in pattern recognition and are characterized by an order of set nodes, such that each of the internal nodes is associated with a decision variable of one or more features [35]. The decision tree algorithm used in this paper is CART (Classification and Regression Trees) [36], which is represented as a binary tree (i.e. two branches). Tree nodes are expanded (i.e. depth) until all leaves are considered pure, that is when no further splits can be made.

SIFT based similarity.

SIFT [37] is used as a sparse local descriptor where interest keypoints are detected in an image to describe invariant features (invariant to image scaling, rotation, and translation) in a local patch, such that an image will be represented as a collection of local feature vectors (shape descriptors) provided that a feature is a 128-dimensional vector representing a local region in the image. The goal of shape descriptors is to uniquely characterize the shape of the object to enable comparison of 2D object silhouettes. The group of feature vectors created by SIFT would thus represent the shape of the image. Using SIFT, we are detecting stable feature points in an image and then for each point a set of features that describe a small region around the point are provided, meaning that we are eventually extracting local information from digital images. Those extracted features are then used to match objects between different scenes.

We split the extracted lesions (Fig 2) vertically and horizontally across the centre into four equal halves, and use SIFT to measure the image similarity (showing correspondences) between each opposite half (top vs. bottom and right vs. left) using the 128-dimensional local feature vectors. The total similarity score is measured as vs + hs, where vs and hs are the vertical and horizontal similarity, respectively. The greater the value the more similar the two halves of the skin lesion, and vice versa. Asymmetry means that the two halves are not similar. In measuring vs and hs, we identify the best two matches for each keypoint using OpenCV’s brute-force matcher, where a feature descriptor on one image half will be matched with all feature descriptors on the other half and the best (closest) two matches will be returned based on the Euclidean distance. This is followed by the ratio test technique to remove any outliers (false positives) resulting from the brute-force matching step. Since we select the best two matches for each keypoint descriptor, ratio test considers a match to be good if the distance ratio between the first and second match is smaller than a specific empirical value (we chose it to be 0.7 which is a typical value for Lowe’s ratio). It should be emphasized that we only investigate the asymmetry of shape in this work, whilst dermatologists in general consider asymmetry in shape, color and structural (border) distribution.

Fig 3(a) shows an example of an extracted skin lesion and its four halves (vertical and horizontal halves), Fig 3(b) shows the keypoints of each half (in red), and Fig 3(c) shows the lines connecting the matching keypoints between each half. The similarity evaluates to 23 between the left and right halves and evaluates to 22 between the top and bottom halves. The total similarity in this skin lesion is the sum of the two similarity values (i.e. 23 + 22 = 45), making it tend to be more asymmetric and the likelihood to be a melanoma lesion.

thumbnail
Fig 3. (a) Extracted skin lesion and its four halves along the vertical and horizontal axes (b) Keypoints (in red) of each half (c) Matching keypoints between each half (top figure: Left and right halves, bottom figure: Top and bottom halves).

https://doi.org/10.1371/journal.pone.0234352.g003

Projection profiles.

Projection profiles [38] are data structures that are used to store the number of foreground pixels when the image is projected over the X-Y axes. They are one-dimensional representations of a two-dimensional image content, and are considered as compact representations of images since many useful information is retained in projections. In this approach, symmetry is measured by projecting the segmented skin lesion in the x and y directions and then comparing their histograms. Assume we have a binary image of size M × N (M: height, N: width), the projection of the image onto a line can be obtained by partitioning the line into bins and finding the number of 1 valued pixels that are perpendicular to the bin. Horizontal and vertical projections can then be obtained by counting the number of 1 pixels from each bin in the horizontal and vertical directions, respectively.

The horizontal projection is the number of foreground (skin lesion) pixels in each row, and is defined as: (22) where B[i, j] is the pixel value at (i, j), and H[i] is the number of foreground pixels in the ith horizontal row. Thus, for each horizontal line of pixels the number of foreground pixels are computed.

On the other hand, the vertical projection represents the number of foreground pixels in each column, and is defined as: (23) where V[j] is the number of foreground pixels in the jth vertical column. The vertical projection of each column is thus computed.

The horizontal and vertical projection profiles can be represented as a histogram (Fig 4). The values of each histogram represent the density distribution of the skin lesion.

thumbnail
Fig 4. Histograms of horizontal and vertical projections.

https://doi.org/10.1371/journal.pone.0234352.g004

After finding the two histograms, they are compared using a correlation method. Let I denote the pixel intensity, H1(I) is the histogram which represents the number of pixels in the first image having pixel intensity I, and H2(I) is the histogram which represents the number of pixels in the second image having pixel intensity I. Using the correlation method, the two histograms are compared based on the following equations: (24) where, (25) is the mean value of each pixel in image k, and N is the total number of histogram bins.

Symmetrical shapes (i.e. circle) evaluate to a correlation value of 1. The more asymmetrical the shape the less the correlation value.

Skewness.

Image moments are useful in describing objects after the segmentation is carried out. They are scalar quantities that are used to capture the image’s signficant features. A measure of asymmetry in an image can be given by its skewness, which is a statistical measure of a distribution’s degree of deviation of the respective projection from symmetry. If the projection is symmetric with respect to the mean (origin), the corresponding skewness evaluates to zero. The degree of skewness can be determined using two third order moments: M30 and M03.

To map from the image domain to the momenta domain, the uniqueness theorem of the momenta [39] can be used which states that the momenta sequence (general moment) Mpq is determined by the density distribution function (i.e. skin lesion image). A general/standard moment Mpq of an image f(x, y) is defined as: (26) where f(x, y) are the graylevels of individual pixels, p and q are positive integers, and r = p + q is the order of the moment.

The third order moment is the skewness of distances between the pixels in the image and its geometrical centre, measuring the bias of the distribution of pixels. The direction of the skewness can be obtained from the sign of the result of the moment, such that when the moment is negative the distribution will bias towards the left of the centre, and when positive will bias towards the right of the centre [40].

The skewness of the horizontal and vertical projections can be defined as shown in Eqs 27 and 28, respectively. (27) (28)

Color variegation

The identification of colors in skin lesions is considered a subjective task even for experienced dermatologists, which deems it necessary to develop an automatic objective approach to identifying colors in skin lesions. In this paper, we use the same rationale as in [19, 20] to determine the suspicious colors present in each skin lesion. However, we use the CIELab color space which is more representable of the human perception than the RGB color space. Moreover, we derive the suspicious colors CIELab values based on the color distribution of our dataset, making it more accurate in determining the suspicious colors. In addition, it is not clear how the colors in [19, 20] are derived (apart from the white, black and red colors in [20] where the standard RGB values are used), as the RGB value for the white color in [19] does not represent the actual color (another color is produced rather than the color of correspondence). Authors of those two studies used only one representative value for each suspicious color which might not be absolute (not representative enough) especially that we can have different levels (shades) of the same color (i.e. light brown). As opposed to those two studies, we use Minkowski distance instead of Euclidean distance.

Before attempting to find color variegation (The C feature in the ABCD rule) in the skin lesion, we convert the RGB image to the CIELab (or CIEL*a*b*) color space [41] since the RGB color space does not closely match the human visual perception, whereas the CIELab color space is designed to approximate/model the human vision (i.e. the L component closely matches the human perception of lightness) and contains in theory every single color the human eye can perceive, allowing it to exploit the characteristics of the human visual pereception better. Moreover, CIELab is considered more precise since the distance between colors using this color space corresponds to the perceived color distance. CIELab is device independent, meaning that the color model is based on the perception of the human eye and is designed to describe what colors look like regardless of what device they are displayed on.

To determine the range of color shades that represent the six suspicious colors of melanoma based on our image dataset, we extract and analyze the color palettes of all our images based on which we are able to determine the range of shades that would determine dark brown, light brown, … etc. In particular, each image in the dataset is represented in terms of its dominant colors which constitute the image’s palette, such that those dominant colors would be the best possible colors by which we can display the image with the least amount of error. Clusters of dominant colors (group of pixels) are formed using k-means clustering [42, 43], where each pixel in the image has a CIELab value associated with it. In this paper we set the number of clusters to 7 (6 suspicious colors + pure black) since we omit the pure black color (i.e. CIELab = [0, 0, 0]) in our measurement of the number of suspicious colors as it most likely belongs to the background (skin).

The process consists of applying k-means with a specific number of clusters (i.e. dominant colors) which is equal to the number of colors the color palette will be composed of. Each pixel color is then affected to the nearest cluster centroid according to the Euclidean distance. K-means minimizes the within-cluster sum of squared distances (i.e. Euclidean distance) between the centroid and the other pixels in the cluster. Using Euclidean distance in the CIELab space is uniform with difference perceived by the eye.

K-means finds the dominant colors in an image through an iterative corrective process, where colors can be thought of as points in the color space cloud that we aim to cluster around some mean (dominant color). K-means starts with a random palette of seven dominant colors {1, 2, …, k} as the starting point, where k is the number of clusters representing the dominant colors. Each pixel is then assigned a color label of the nearest dominant color. Image pixels in this case are thus grouped by their dominant color value. New averages are then computed to update the cluster centers. If image pixels belong to the same clusters for two successive iterations, the process is considered finished and the final color palette is formed. Fig 5 shows a skin lesion image and its 3D scatter plot representing the different pixel colors in CIELab color space, in addition to the lesion’s color palette resulting from k-means clustering with k = 7. The CIELab values of the color palette from left to right are: [0, 0, 0], [59.263, 6.519, -1.826], [47.051, 9.465, 2.307], [67.222, 3.355, -4.661], [22.237, 0.779, -3.284], [40.018, 10.696, 4.107], [53.075, 8.206, -0.027].

thumbnail
Fig 5. (a) 3D scatter plot of skin lesion image pixels in CIELab color space (b) Skin lesion color palette showing the 7 dominant colors in the image (the first color which represents the pure black color will be omitted from our color variegation measurement).

https://doi.org/10.1371/journal.pone.0234352.g005

Representing the image in terms of its dominant colors is a more realistic approach when working on color variegation; for instance the image shown in Fig 5 is composed of 1480 unique colors, which makes it very difficult in determining the range of color shades that would represent the 6 suspicious colors of melanoma.

Analyzing the images used in our experiments, the suspicious CIELab color ranges (i.e. [min] − [max]) are determined as shown in Table 4. For the suspicious colors white, red and blue-gray, since they are not expressed in our dataset we use their standard CIELab color values. Skin lesions which have any of their pixel CIELab color values belonging to the range of black, dark brown, or light brown color values are considered to possess those colors. For the colors of white, red and blue-gray, a skin lesion is considered to possess one of those colors if the (i) Minkowski distance between the pixel color and any of the aforementioned colors is less than a threshold T, measured as being the half of the Minkowski distance between the two extremes of colors (white and black), which evaluates to 50 (ii) pixels that belong to the suspicious color represent more the 5% of the skin lesion pixels.

Minkowski distance is considered a generalization of the Euclidean and Manhattan distances [44] and is defined as [45]: (29) where p ≥ 1 is a real number. The distance represents the Manhattan distance and the Euclidean distance when p = 1 and p = 2, respectively.

The advantage of using Minkowski distance is that mathematical results can be shown for the whole class of distance functions, and users can adapt the distance function to suit the needs of the application by modifying the Minkowski parameter p, which is set to p = 3 in this paper. Based on the above, the image in Fig 5 is composed of 3 suspicious colors.

Diameter

To measure the diameter of the skin lesion we utilize Feret’s diameter [46, 47], which is the distance between two parallel tangents at the contour of the object (i.e. skin lesion) that are located on opposite sides of the object at an arbitrary selected angle. The maximum Feret diameter of an object is the distance between its two furthest points measured in a given direction. The average value over many orientations can also be used, meaning that Feret’s diameter can be referred to as the average distance between two tangents in the opposite sides of the object parallel to some fixed direction. In this paper we use the maximum Feret diameter as our diameter measure. Finding the Feret’s diameter of some object shape is a commonly used measure in shape analysis.

Feret’s diameter is also called caliper diameter since the measurement involves placing the object for which we want to find the diameter inside the jaws of a caliper, with the caliper oriented at some specified angle (i.e. 0°, 45°, 90°, 135°). The jaws are then closed on the object tightly while maintaining the angle. The distance between the jaws is the Feret diameter at angle (direction) θ. In terms of digital images, this is made by isolating the corner pixels of the object’s perimeter and taking the maximum distance between each corner pixel to all other corner pixels, meaning that Feret’s diameter of an object in direction θ is the projection of the object on the axis oriented in direction θ. It should be emphasized that Feret’s diameter is based on the binary 2D image of the object. The main advantage of using Feret’s diameter over other measures is due to its correspondence with the real physical diameter of the object. In other words, it corresponds to the length that would be measured if we handle the object between the teeth of a caliper.

As the skin lesion diameter in the real world is measured in millimeters (mm) and our diameter results are returned in pixels, we need to represent our diameter results in terms of the standard unit (mm); this can be made using spatial calibration (geometric correction) which involves calibrating the image against a known value (i.e. mm) and then applying such calibration to the uncalibrated image (i.e. in pixels). The idea is thus to represent the diameter in units rather than pixels. An image produced in units is called a spatially calibrated image. However, to conduct such calibration one needs to know the original measure in real world and then map that to pixels. Since we do not have the original real world measures available, we used an image in our dataset (i.e. ISIC dataset) that has a ruler displayed (Fig 6(a)) to get an estimate on the skin lesion measure in millimeters and deduce from that how many pixels would be in 1 mm. The image was also zoomed in to better reflect the sizes of the skin lesions in our test images. Doing such calibration, we found that for our 256 × 256 images we had 29.7 pixels/mm. Fig 6(b) shows samples of skin lesion images, their segmentations and corresponding Feret’s diameter values.

thumbnail
Fig 6. (a) Skin lesion image with a ruler used to spatially calibrate our test images and deduce the pixels/mm value (b) Skin lesions’ Feret’s diameter (in pixels and millimeters).

https://doi.org/10.1371/journal.pone.0234352.g006

After applying the methods explained above, we use the extracted features to derive a decision on the asymmetry, color variegation (i.e. number of suspicious colors), and diameter on the skin lesion images fed to our system.

Results and discussion

Applying Otsu-II (the improved Otsu thresholding version we propose in this paper) and the original Otsu approach on the 204 images lead to a Dice similarity coefficient value of 87.7% and 82.5%, respectively. Fig 7(a) shows some examples on segmentation results obtained by Otsu and Otsu-II methods, along with the groundtruth of each corresponding image. As can be noticed, Otsu-II is able to improve the area coverage of the skin lesion.

thumbnail
Fig 7. (a) Skin lesion images segmented using Otsu and Otsu-II methods, along with their corresponding groundtruth (b) Skin lesion images segmented using our proposed segmentation approach and U-Net, along with their corresponding groundtruth.

https://doi.org/10.1371/journal.pone.0234352.g007

It should be emphasized that we tested another adaptation of Otsu’s method where we kept PL(k) and PS(k) from Eqs 6 and 7 in Eqs 17 and 18 (i.e. Otsu-II), resulting in the new within and between class variances shown in Eqs 30 and 31, respectively, which we refer to here as Otsu-II′: (30) (31)

Having three Otsu related methods, we expanded our evaluation on two more types of datasets. One dataset contained skin lesions that cover most of the image (i.e. large lesions) and another set where the skin lesion comprises a small region of the image. The dataset used with Otsu and Otsu-II methods above contained a mix of such lesions (i.e. mixed). The datasets that contained images with large and small skin lesions were composed of 129 and 162 images, respectively.

The Dice similarity of the three Otsu related methods on the different datasets is summarized in Table 5. The poor performance of Otsu-II′ on small images was due to the fact that it was not robust to artifacts (i.e. hair, ruler) that were more apparent in images with small skin lesions.

thumbnail
Table 5. Dice similarity of Otsu related methods on three datasets.

https://doi.org/10.1371/journal.pone.0234352.t005

To evaluate the performance of our segmentation approach (Fig 1) against state-of-the-art methods, we compare our approach with U-Net [48], an end-to-end encoder-decoder network firstly used in medical image segmentation and has also been utilized in skin lesion segmentation in dermoscopic images [49, 50]. U-Net is trained on 1935 dermoscopy images along with their corresponding groundtruth response masks. Images used to train U-Net were resized to 256 × 256 pixels, and the model was trained for 20 epochs on a Tesla P100 GPU. Training the model took 27.1 minutes and testing it on the 204 images took 18.1 seconds. Fig 7(b) shows some samples of test images, their corresponding groundtruth, and the results using our proposed segmentation approach and U-Net. The average Dice similarity of the 204 test images evaluates to 88% and 76.2% for our proposed segmentation approach and U-Net, respectively. Visual results (Fig 7(b)) show that our approach is able to detect the fine structures of skin lesion borders—a crucial factor when detecting skin lesion border irregularity (i.e. B feature in the ABCD rule)—better than U-Net, a feature that U-Net seems to lack. The results of U-Net as can be noticed are a bit blurry. Such blurriness can be removed by applying a thresholding technique, but we kept the original results of U-Net as-is for comparison purposes.

204 skin lesion images were sent to a dermatologist (Dr.Sally O’Shea) to label as symmetric or asymmetric based on shape (most of the skin lesions contained 1 suspicious color and a diameter larger than 150 pixels—5.1 mm). Of the 204 images labeled by the dermatologist, 35 images were classified as being symmetric and 169 images as asymmetric. Fig 8 shows some samples of symmetric and asymmetric images. The extracted asymmetry features/measurements (SIFT based similarity, projection profiles, and skewness) have been used to train and test a decision tree on an 80: 20 ratio. That is, 80% of the data was used for training the decision tree and 20% of the data was used for testing the decision tree. 136 asymmetric and 28 symmetric images were used for training the decision tree. 33 asymmetric and 7 symmetric images were used for testing the decision tree. After training the decision tree, 30 and 2 asymmetric and symmetric images were predicted correctly, respectively, meaning that an 80% accuracy has been obtained. The decision tree as can be noticed performed better on asymmetric images; this can be due to having more asymmetric images in the training data. But, in general, we believe that more data samples which are also balanced (number of asymmetric and symmetric samples are the same) could improve the results significantly. This requires much labor work (i.e. labeling by dermatologists) and is a topic of interest we would be willing to explore in future studies.

thumbnail
Fig 8. Samples of symmetric and asymmetric images labeled by the dermatologist.

https://doi.org/10.1371/journal.pone.0234352.g008

To determine the asymmetry, color variegation, and diameter of an unknown (test) sample, the asymmetry features (SIFT based similarity, projection profiles, and skewness) are extracted and fed to a decision tree which is used to predict whether the skin lesion is symmetric or asymmetric based on the extracted measures. The number of suspicious colors in the skin lesion are then determined based on the values shown in Table 4, and the diameter is then measured based on Feret’s diameter explained above. Fig 9 shows a skin lesion and its extracted features using the proposed approach.

thumbnail
Fig 9. Skin lesion asymmetry, color variegation, and diameter features extracted using the proposed approach.

https://doi.org/10.1371/journal.pone.0234352.g009

Conclusion

To overcome the subjectivity in measuring asymmetry and color variegation, and to avoid missing melanoma in the early diagnosis due to possessing a diameter less than 6mm, an objective approach has been proposed where the extracted asymmetry measurements were used to train a decision tree which was then utilized for predicting the asymmetry of new skin lesion images. The range of the CIELab color space colors that demonstrate the suspicious colors of melanoma have been derived, and the diameter has been found using the Feret’s diameter method. The approach is able to predict asymmetry with 80% accuracy and measure color vriegation and diameter objectively. The proposed Otsu improvement (Otsu-II) outperforms the original both Otsu and Otsu-II′ in skin lesion segmentation, and when combined with the pre/post processing steps to represent our segmentation approach, this provides better segmentation results than U-Net. In summary, the contribution of our work involves proposing a segmentation approach that outperforms standard segmentation approaches, an asymmetry measure composed of a vector of SIFT based similarity, projection profiles, and skewness which can then be automatically determined using a trained decision tree, and objective approaches to determine the number of suspicious colors in a skin lesion along with its diameter. The measures can be incorporated in a desktop computer vision system where clinicians can feed the system with dermoscopic images and obtain an objective evaluation on those measures. In future work we aim at using a more balanced dataset with more symmetric and asymmetric samples (requires more laborious work), as this could improve the decision tree prediction accuracy. Enhancing Otsu-II and making it more robust to different artifacts will also be investigated. Moreover, different machine learning approaches will be analyzed.

References

  1. 1. “NIH Consensus Conference: Diagnosis and treatment of early melanoma,” JAMA, vol.268, 1993.
  2. 2. M. Hintz-Madsen, “Probabilistic framework for classification of dermatoscopic images,” PhD thesis, Technical University of Denmark, 1998.
  3. 3. Melanoma Survival Rates. https://www.curemelanoma.org/about-melanoma/melanoma-staging/melanoma-survival-rates/. Last accessed 29 March 2020.
  4. 4. Stoecker W.V., Li W.W., and Moss R.H., “Automatic detection of asymmetry in skin tumors,” Computerized Med. Imag. Graph., vol.16, no.3, pp. 191–197, May, 1992.
  5. 5. Nachbar F., Stolz W., Merkle T., Cognetta A.B., Vogt T., Landthaler M., et al, “The abcd rule of dermatoscopy,” Journal of the American Academy of Dermatology, 30(4):551–559, 1994. pmid:8157780
  6. 6. Neville B., Damm D., Allen C., and Chi A., “Oral and maxillofacial pathology,” Elsevier Canada, 4th edition, 2015.
  7. 7. Braun R., Rabinovitz H., Oliviero M., et al, “Dermoscopy of pigmented skin lesions,” Journal American Academy of Dermatology, vol.52, pp.109–121, 2005.
  8. 8. V. Ng and D. Cheung, “Measuring asymmetries of skin lesions,” In: Proc. IEEE Int. Conf. systems, man, and cybernetics (SMC), vol. 5. Piscataway, NJ: IEEE Press; pp.4211-6, 1997.
  9. 9. Ng V., Fung B., and Lee T., “Determining the asymmetry of skin lesion with fuzzy borders,” Comput. Biol. Med. 35, 103–120. 2005. pmid:15567181
  10. 10. Seidenari S., Pellacani G., and Giannetti A., “Digital videomicroscopy and image analysis with automatic classification for detection of thin melanomas,” Melanoma Research, vol.9, pp.163–171, 1999.
  11. 11. Andreassi L., Perotti R., Rubegni P., et al., “Digital dermoscopy analysis for the differentiation of atypical nevi and early melanoma,” Arch Dermatol, vol.135, pp. 1459–1465, 1999.
  12. 12. M. Amico, M. Ferri, I. Stanganelli, “Qualitative asymmetry measure for melanoma detection,” In Proc. IEEE Int. Symp. Biomed. Imag.: Nano Macro, vol.2, pp.1155–1158, 2004.
  13. 13. L. Ma, A. Guo, S. Zou, and W. Xu, “Irregularity and asymmetry analysis of skin lesions based on multi-scale local fractal distributions”, 2nd International Congress on Image and Signal Processing, 2009.
  14. 14. K. Clawson, P. Morrow, B. Scotney, D. McKenna, and O. Dolan, “Determination of optimal axes for skin lesion asymmetry quantification,” In IEEE International Conference on Image Processing (ICIP 2007), vol.2, pp.453-456, 2007.
  15. 15. Dice R., “Measures of the Amount of Ecologic Association Between Species,” Ecology. 26 (3): 297–302, 1945.
  16. 16. Kalwa U., Legner C., Kong T., and Pandey S., “Skin cancer diagnostics with an all-inclusive smartphone application,” Symmetry, 11(6), 2019.
  17. 17. Tenenhaus A., Nkengne A., Horn J., Serruys C., Giron A., and Fertil B., “Detection of melanoma from dermoscopic images of naevi acquired under uncontrolled conditions,” Skin Research and Technology, 16(1):85–97, 2010. pmid:20384887
  18. 18. Umbaugh S., Moss R., and Stoecker W., “Automatic color segmentation of images with applications in detection of variegated coloring in skin tumors,” IEEE Eng Med. Biol. 8, 43–52, 1989.
  19. 19. Kasmi R. and Mokrani K.,“Classification of malignant melanoma and benign skin lesions: implementation of automatic ABCD rule,” IET Image Proc. 10(6),448–455, 2016.
  20. 20. G. Grammatikopoulos, A. Hatzigaidas, A. Papastergiou, P. Lazaridis, Z. Zaharis, D. Kampitaki, et al, “Automated malignant melanoma detection using Matlab,” Proc. Fifth Int. Conf. on Data Networks, Communications and Computers, Bucharest, Romania, 2006.
  21. 21. Seidenari S., Fabiano A., Jalbout Al S., Bassoli S., Borsari S., Magnoni C., et al, “Relationship between histological and computer-based assessment of melanoma diameter and thickness in head versus trunk-limbs melanomas,” Head Neck Oncol. 06;5(3):32, 2013.
  22. 22. Grammatikopoulos G., Hatzigaidas A., Papastergiou A. et al., “Simple matlab tool for automated malignant melanoma diagnosis,” WSEAS Trans. on Information Science and Applications, vol. 3, no. 4, pp. 460–465, 2007.
  23. 23. Garnavi R., Aldeen M., and Bailey J., “Computer-aided diagnosis of melanoma using border and wavelet-based texture analysis,” In: IEEE Trans. Inf. Technol. Biomed., vol. 16, no. 6, 2012.
  24. 24. Zaqout I., “Diagnosis of skin lesions based on dermoscopic images using image processing techniques,” In International Journal of Signal Processing, Image Processing and Pattern Recognition, 9(9):189–204, 2016.
  25. 25. Ali A., Li J., Yang G., and O’Shea S.J., “A machine learning approach to automatic idetection of irregularity in skin lesion border using dermoscopic images,” PeerJ Computer Science, 2020.
  26. 26. N. Otsu, “A threshold selection method from gray-level histogram,” IEEE Trans. Syst. Man Cybern., vol. 9, pp. 62–66, Jan. 1979.
  27. 27. J.M. Sung, H.G. Ha, B.Y. Choi, and Y.H. Ha, “Image thresholding based on within-class standard deviation using standard deviation,” In Proceedings of SPIE—The International Society for Optical Engineering, 2014.
  28. 28. Cohn B. A., “In consideration of the E in the melanoma ABCDE mnemonic,” Arch Dermatol 142: 528–529, 2006. pmid:16618882
  29. 29. Wang J., Jiang H., Yuan Z., Cheng M.-M., Hu X., and Zheng N., “Salient object detection: A discriminative regional feature integration approach,” Int. J. Comput. Vision, vol. 123, no. 2, pp. 251–268, 2017.
  30. 30. Jahanifar M., Zamani Tajeddin N., Mohammadzadeh Asl B., and Gooya A., “Supervised saliency map driven segmentation of lesions in dermoscopic images,” IEEE Journal of Biomedical and Health Informatics, 2018. pmid:29994323
  31. 31. Sun Y., Liu M., Meng MQ.H., “Active perception for foreground segmentation: an RGB-D data-based background modeling method,” IEEE Trans Autom Sci Eng., 2019.
  32. 32. Chen Q., Li D., and Tang C., “KNN matting,” IEEE transactions on patern analysis and machine intelligence 35, 9, pp. 2175–2188, 2013.
  33. 33. P. Tschandl, C. Rosendahl, and H. Kittler, “The ham10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions,” arXiv preprint arXiv:1803.10417, 2018.
  34. 34. Codella N., Gutman D., Emre Celebi M., Helba B., Marchetti M., Dusza S., et al, “Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (isbi),” hosted by the international skin imaging collaboration (isic). arXiv:1710.05006, 2017.
  35. 35. Sethi I.K., “Entropy nets: from decision trees to neural networks,” Proc. IEEE 78 (10), 1605–1613, 1990.
  36. 36. Breiman L., Friedman J.H., Olshen R.A., and Stone C.J., “Classification and regression trees,” Wadsworth Inc, 1984.
  37. 37. Lowe D. G., “Distinctive image features from scale-invariant keypoints,” IJCV, 60(2):91–110, 2004.
  38. 38. D.S. Bloomberg, G.E. Kopec, and L. Dasari, “Measuring document image skew and orientation,” Document Recognition II (SPIE vol. 2422), San Jose, CA, February 6-7, 302-316, 1995.
  39. 39. Hu M.K., “Visual pattern recognition by moment invariants,” IRE Transactions on Information Theory 8, 179–187, 1962.
  40. 40. Li C., “Content-based microscopic image analysis,” Logos Verlag, 2016.
  41. 41. C.I.E., “Recommendations on uniform color spaces, color difference equations, psychometric color terms,” Supplement No.2 to CIE publication No.15 (E.-1.3.1) 1971/(TC-1.3.), 1978.
  42. 42. Sebestyen G. S., “Decision making processes in pattern recognition,” New York: Macmillan, 1962.
  43. 43. J. MacQueen, “Some methods of classification and analysis of multivariate observations,” In L. M. Le Cam & J. Neyman (Eds.), Proceedings of the fifth Berkeley symposium on mathematical statistics and probability (Vol. 1, pp. 281-297). Berkeley, CA: University of California Press, 1967.
  44. 44. P.J. Groenen, U. Kaymak, and J. van Rosmalen, “Fuzzy clustering with minkowski distance functions,” In Advances in Fuzzy Clustering and its Applications (eds J. Valente de Oliveira and W. Pedrycz), 2007.
  45. 45. Butenko S., Chaovalitwongse W., and Pardalos P.,“Clustering challenges in biological networks,” World Scientific, 2009.
  46. 46. L. R. Feret, “Assoc. Internat. pour l‘Essai des Mat.,” Zurich, 2, Group D., 1931.
  47. 47. W.H. Walton, “Feret’s statistical diameter as a measure of particle size,” Nature, London, 162,329-330, 1948.
  48. 48. Ronneberger O., Fischer F., and Brox T., “U-net: Convolutional networks for biomedical image segmentation,” MICCAI, 2015.
  49. 49. A. Ali, J. Li, S. O’Shea, G. Yang, T. Trappenberg, and X. Ye, “A deep learning based approach to skin lesion border extraction with a novel edge detector in dermoscopy images,” The International Joint Conference on Neural Networks, 2019.
  50. 50. A. Ali, J. Li, and T. Trappenberg, “Supervised versus unsupervised deep learning based methods for skin lesion segmentation in dermoscopy images,” Canadian Conference on Artificial Intelligence, pages 373–379, 2019.