Академический Документы
Профессиональный Документы
Культура Документы
Abstract—Melanoma is one of the deadliest forms of cancer; Nowadays, a technique used by dermatologists to diagnose skin
hence, great effort has been put into the development of diag- lesions and, consequently, to detect melanomas is dermoscopy.
nosis methods for this disease. This paper addresses two dif- This is a noninvasive procedure used for in vivo observation
ferent systems for the detection of melanomas in dermoscopy
images. The first system uses global methods to classify skin of skin lesions. The physician places gel on the skin lesion
lesions, whereas the second system uses local features and the and inspects it with a magnification instrument (dermatoscope,
bag-of-features classifier. This paper aims at determining the best stereomicroscope, or a digital imaging system), which amplifies
system for skin lesion classification. The other objective is to the lesion 6–100×, depending on the instrument used [2]. This
compare the role of color and texture features in lesion classifica- magnification allows the recognition of several surface and
tion and determine which set of features is more discriminative.
It is concluded that color features outperform texture features subsurface structures, which are not visible to the naked eye,
when used alone and that both methods achieve very good re- and that can be used to diagnose a skin lesion using one of the
sults, i.e., Sensitivity = 96% and Specificity = 80% for global several medical diagnostic algorithms, e.g., the ABCD rule [3],
methods against Sensitivity = 100% and Specificity = 75% for Menzies method [4], and the seven-point checklist [5]. These
local methods. The classification results were obtained on a data three algorithms have a common first step, i.e., the identification
set of 176 dermoscopy images from Hospital Pedro Hispano,
Matosinhos. of the inspected lesion regarding its origin as melanocytic or
nonmelanocytic; melanoma is a melanocytic lesion, since it
Index Terms—Bag of features (BoF), color, computer-aided is a malignant proliferation of melanocytes. In their second
diagnosis, dermoscopy, melanoma, texture.
step, they distinguish between benign melanocytic lesions and
melanoma using different approaches. The ABCD rule assigns
I. I NTRODUCTION
a score to a lesion. This score is a combination of the scores for
BARATA et al.: TWO SYSTEMS FOR THE DETECTION OF MELANOMAS IN DERMOSCOPY IMAGES 3
Fig. 2. Block diagram of two melanoma detection systems, using (a) global and (b) local features.
classification: global or local? Second, it aims at comparing the a local patch centered at the keypoint. Since the number of
performance of two classes of features (color and texture) in keypoints and local features varies from image to image, we
melanoma identification and determining which feature is the cannot directly feed a classifier with these data. Instead, all local
best discriminator in this specific problem. features associated with all the training images are gathered and
used to compute a smaller set of prototypes (centroids) denoted
III. S YSTEM OVERVIEW as visual words. Then, the local features of each dermoscopy
image are assigned to the nearest visual word, and a histogram
The analysis of dermoscopy images is an object recognition is computed. The histogram counts the number of times each
problem under very specific conditions. Several approaches visual word was selected. A statistical classifier is then trained
have been proposed to deal with this kind of problems. Some to discriminate melanoma lesions from nonmelanoma ones,
of them characterize the object by a set of global features (e.g., using the histogram of visual words as input.
color moments and gradient histograms) and use these features This paper tries to determine which of these strategies per-
to discriminate the object from the background. However, these forms best in melanoma detection. The answer is not simple. If
techniques fail to recognize objects in more complex settings, a skin lesion is homogeneous, the first approach will probably
e.g., when the object boundary is unknown and cannot be easily be the best since it is able to describe the lesion by a set of
found in the image or when the object shape, color, and texture global features that use all the available information. However,
exhibit severe changes and cannot be easily described by global since skin lesions often have differential structures (pigment
measurements extracted from the image. A recent trend consists network, vascular network, dots, streaks, etc.) that are localized
in selecting local patches (small regions) in the image and and appear in specific regions, these structures may not be well
describing each patch by a set of local features. This approach described by global features, and local methods may represent
does not require a global model for the object and can easily them better.
cope with changes in shape, color, and texture, achieving good Both systems assume that the dermoscopy image is seg-
results in very difficult problems [30]. mented and that the lesion region is separated from healthy
In this paper, we will compare two different strategies for skin. Several segmentation methods were proposed for this
melanoma detection and build a recognition system based on problem [35]–[39]. A comparison among several methods can
each of them. The first system describes the dermoscopy image be found in [40]. However, there is no consensus about which
by a set of global features and uses a classifier to discriminate one performs best. We adopted a simple adaptive thresholding
melanomas from nonmelanoma lesions [see the block diagram technique (see the description of the method in the Appendix),
in Fig. 2(a)]. This is a supervised system since the classifier and the output of the segmentation step was verified by an
learns to detect the melanoma lesions using a training set of expert and edited when necessary. This procedure was adopted
images, which is labeled by an expert. Each training image, in order to make the system performance independent of the
i.e., I (k) , k ∈ 1, . . . , L, is characterized by feature vector xk ∈ specific algorithm used for segmentation. Otherwise, it would
Rn and by binary label yk ∈ {0, 1}. The classifier is trained be difficult to interpret the system decisions since errors might
to discriminate both types of images using the training data be caused by insufficient information provided by the features
(xk , yk ), k = 1, . . . , L. or by segmentation errors.
The second system characterizes the dermoscopy image by
using a BoF approach [see Fig. 2(b)] [29], [31]–[34]. First,
IV. G LOBAL L ESION C LASSIFICATION S YSTEM
a set of keypoints is selected inside the lesion region. Then,
each keypoint is characterized by a vector of local features. The global classification system consists of three processing
This feature vector represents color and texture properties in steps [see Fig. 2(a)]. First, the skin lesion is segmented. Then, a
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
BARATA et al.: TWO SYSTEMS FOR THE DETECTION OF MELANOMAS IN DERMOSCOPY IMAGES 5
Fig. 4. Histograms of (top) amplitude and (bottom) orientation for (left) melanoma and (right) nonmelanoma lesions.
The mixture coefficients are denoted as color components. V. D ETECTION W ITH L OCAL F EATURES
However, this color space has a number of drawbacks: it is not
BoF methods have recently been used in complex image
perceptually uniform, it depends on the acquisition setup, and
analysis problems, e.g., image retrieval and object-class recog-
it exhibits a high correlation among the three color channels
nition in large databases [53], [54]. Since complex objects can-
[49]. To overcome these difficulties, other color representations
not be represented by a global model, BoF methods represent
have been proposed, e.g., biologically inspired color spaces
them by a collection of local models, e.g., image patches. In
such as the opponent color space (Opp) [50], [51], color spaces
a first stage, a set of keypoints, i.e., pi ∈ R2 , is defined in
related to the human description of color such as hue satura-
the image. In the case of dermoscopy images, we assume that
tion and brightness (HSV and HSI) [49], [52], or perceptu-
keypoints are nodes of a regular grid of size Δ × Δ defined
ally uniform color spaces such as the CIE La∗ b∗ and L∗ uv
in the image domain. Second, each keypoint is described by
[49]. Although there is a 1-1 map between all pairs of color
a feature vector xi ∈ Rn , which represents local information
spaces, the choice of the color space modifies the recogni-
conveyed in the image patch of size Δ × Δ centered at the
tion results since their components present different statistics
ith keypoint [see Fig. 3(d)]. The size of the patches is equal to
(e.g., some are more correlated than others) and different
the size of the grid to avoid overlapping. Patches for which the
histograms.
area is more than 50% outside the lesion are discarded. Given
In this paper, we have used all six color spaces mentioned
an input image I, we will denote the family of local features
above. We characterize the color distribution in lesion region
associated to I by
R (or R1 , R2 ) using a set of three color histograms, each of
them with Mc bins. The color histogram associated to the color F = {x1 , . . . , xN }, xi ∈ Rn (7)
channel, i.e., Ic (x), c ∈ {1, 2, 3}, is given by
where N is the number of keypoints inside the lesion (see
1 Fig. 3, bottom right).
hc (i) = bc (Ic (x)) i = 1, . . . , Mc (6)
N
x∈R
A. Training
where N = #R denotes the number of pixels in region R, and
i is the histogram bin. The bins are defined by splitting the color The training phase aims to learn the parameters of a BoF
component range (which depends on the color space) into Mc classifier using a set of labeled dermoscopy images. Let T =
subintervals with the same length. {I (1) , . . . , I (L) } be a set of L images used to train the classifier
Fig. 5 shows the concatenation of the histograms hc for four and let
different color spaces and splitting region R into R1 and R2 .
L
The differences between melanoma and nonmelanoma features F= F (k) (8)
are more noticeable in these cases. k=1
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
Fig. 5. Concatenated histograms of (first row) RGB, (second row) L∗ uv, (third row) Opp, and (fourth row) HSV for (left) melanoma and (right) nonmelanoma
lesions, computed in R1 and R2 .
be the set of all local features detected in the training set. clustering algorithm (e.g., k-means algorithm). The prototypes
First, we will approximate the feature vectors in F by a set are known as visual words to emphasize the similarity with the
of prototypes (centroids) c1 , . . . , cK . This is done using a bag-of-words method used in text analysis. After obtaining a
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
BARATA et al.: TWO SYSTEMS FOR THE DETECTION OF MELANOMAS IN DERMOSCOPY IMAGES 7
(k)
is assigned to each feature vector xi .
Then, a histogram that counts the occurrence of each of the
prototypes is computed for each of the training images I k .
Therefore
1 (k)
(k)
N
h (k)
(l) = (k) δ li − l (10)
N i=1
TABLE II TABLE IV
B EST PARAMETERS FOR THE kNN C LASSIFIER B EST E XPERIMENTAL R ESULTS U SING G LOBAL F EATURES
AND AN A DA B OOST C LASSIFIER
TABLE III
B EST E XPERIMENTAL R ESULTS U SING G LOBAL F EATURES
AND AN RBF-SVM C LASSIFIER
TABLE V
B EST PARAMETERS FOR THE A DA B OOST C LASSIFIER
BARATA et al.: TWO SYSTEMS FOR THE DETECTION OF MELANOMAS IN DERMOSCOPY IMAGES 9
Fig. 6. Variation of SE and SP for (first row) R and (second row) R1 + R2 , using W ∈ [2, 100].
We sought to determine the impact that increasing the num- more discriminative information. The classifier selects infor-
ber of weak classifiers (number of selected bins) would have mation from R1 and R2 , which suggests that both regions are
on the classification results. To do so, we varied the value of informative. We have computed the mean and median values
W , while maintaining nbins constant and equal to the one of for the two most selected bins for each of the three best color
the best configuration for each single feature (see Table V). features (see Table VI). The most selected bin is labeled as “1”
The curves obtained are shown in Fig. 6. It is clear that the and the second most selected bin as “2” (recall Fig. 7, to identify
performance of the classifier depends on the number of features the corresponding bins for each color feature). These bins have
used. This reinforces the idea that most of the bins contain noisy values considerably different for the two classification classes,
information and have a marginal influence on the classifier which demonstrates that these bins can be used to distinguish
output. between melanomas and nonmelanomas.
The characteristics of AdaBoost allows to further explore Fig. 8 shows four examples classified by the proposed
the data set. We performed tests with the best single color system using Adaboost, C6 , and the best combination of
features (C2 , C4 , and C6 ) and their combinations with the parameters for this feature. These examples correspond to
best texture feature (T1 ) to determine if some histogram bins the four possible cases, i.e., true negative (TN)—correctly
are more selected than others. This would mean that these classified nonmelanoma, false positive (FP)—incorrectly clas-
bins are very discriminative. The results are shown in Fig. 7. sified nonmelanoma, false negative (FN)—incorrectly classi-
All the tests were performed using the best configuration (see fied melanoma, and true positive (TP)—correctly classified
Table V) and considering the two regions R1 and R2 , since it is melanoma.
clear that extracting features separately from inner and border Although we do not use the same features and data sets as
improves the classification results. By inspecting the results, it the other works found in the literature, a tentative comparison
is possible to notice that some bins are indeed selected most can be performed. Faziloglu et al. [26] reported the following
of the time. These bins correspond to specific channels of each values: SE = 84.3% and SP = 83.0%, and Stanley et al. [23]
color space, e.g., for C6 , the two most selected bins correspond reported the following values: SE = 87.7% and SP = 74.9%.
to bins of the histogram of the second channel (O2 ), and for Both groups use only color features, and their data sets are
C2 and C4 , the most selected bin is from the histogram of larger than ours (258 and 226 images, respectively), possessing
the second channel [Saturation (S)], whereas the two other an equal proportion of melanomas and nonmelanomas. With
correspond to the third channel [Value (V ) and Intensity (I), our best configurations and classifier, we achieve results similar
respectively]. An interesting result is that when a fusion of two to the ones reported by these two groups, i.e., SE = 96.0% and
best features is performed, the classifier only selects bins from SP = 80.0% using C2 /C4 and SE = 92.0% and SP = 85%
color histograms, which demonstrates that color features have using C6 . Our classification results are also similar to the ones
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
Fig. 7. Average frequency of selection of the most selected histogram bins for (a) C2 , (b) C4 , (c) C6 , (d) T1 + C2 , (e) T1 + C4 , and (f) T1 + C6 .
TABLE VI
M EAN AND M EDIAN VALUES FOR THE T WO M OST S ELECTED B INS (T HE M OST S ELECTED B IN IS “1”
AND THE S ECOND M OST S ELECTED IS “2”) FOR C2 , C4 , AND C6
BARATA et al.: TWO SYSTEMS FOR THE DETECTION OF MELANOMAS IN DERMOSCOPY IMAGES 11
Fig. 9. Variation of SE and SP using (first row) K ∈ {100, 200, 300}, (second row) Δ ∈ {20, 40, . . . , 100}, (third row) nbins ∈ {15, 25, 35, 45}, and the
(fourth row) three tested distances.
and global features. Most previous works use global features A second driving idea associated with this paper is the
(texture, shape, and color) associated to the whole lesion fol- evaluation of the role played by color and texture features in the
lowed by a binary classifier trained from the data. However, decision. Most works consider both types of features together,
local features are gaining increasing importance in many image but they do not attempt to clarify if one of these features plays a
analysis problems (e.g., image retrieval and object recognition) more relevant role. This question has been explicitly addressed
and may be seen as a promising alternative. in this paper.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
BARATA et al.: TWO SYSTEMS FOR THE DETECTION OF MELANOMAS IN DERMOSCOPY IMAGES 13
Fig. 10. Threshold computation from the histogram in the case of (right) two
peaks.
A PPENDIX
L ESION S EGMENTATION Fig. 11. Threshold computation from the histogram in the case of (right) two
peaks.
Lesion segmentation is obtained using an adaptive thresh-
olding algorithm, which comprises three steps: 1) histogram
computation; 2) peak detection; and 3) threshold estimation. detected (see Fig. 11). This involves two steps. First, close
These three steps are described in the sequel. peaks are merged and replaced by a single peak, i.e., the largest
First, the input image is converted into a gray-level image by one. We consider that two peaks located at i1 and i2 are close
selecting the channel with the highest entropy [40]. Then, the if |i1 − i2 | < Δi (Δi = 8). If the number of peaks is larger
histogram of intensity, i.e., h(i), i = 0, . . . , 255, is computed than 2 after this step, the highest peak is selected. Then, all the
using all the pixels inside a rectangular window, which is remaining ones are tested. The second selected peak is the one
located at the center of the image. The histogram is then low- that corresponds to the largest depth of the valley between two
pass filtered with a Gaussian filter. The filter impulse response histogram peaks.
is G(i) = C exp{−i2 /(2σ 2 )}, where σ = 5, and C is a nor-
malization constant.
We assume that the analysis intersects the lesion, and one of R EFERENCES
the histogram modes is associated to the lesion. We therefore [1] “Cancer facts and figures 2012,” American Cancer Society, Atlanta, GA,
USA, 2012, Tech. Rep.
extract the most significant peaks of the histogram (with a [2] Dermoscopy Tutorial 2003. [Online]. Available: http://www.dermoscopy.
maximum 2). If the histogram has a single peak located at org/atlas/base.htm
i1 , the threshold is defined by T = i1 + ΔT . The ΔT offset [3] W. Stolz, A. Riemann, and A. B. Cognetta, “ABCD rule of dermatoscopy:
A new practical method for early recognition of malignant melanoma,”
was empirically tuned and set to 15. If two significant peaks Eur. J. Dermatol., vol. 4, no. 7, pp. 521–527, 1994.
are extracted, we choose the threshold that corresponds to the [4] S. Menzies, C. Ingvar, K. Crotty, and W. McCarthy, “Frequency and mor-
maximum valley. The valley is defined as follows: Given two phologic characteristics of invasive melanomas lacking specific surface
microscopic features,” Arch. Dermatol., vol. 132, no. 10, pp. 1178–1182,
peaks located at i1 , i2 , (i2 > i1 ), we define the valley at i as the Oct. 1996.
difference between the straight line defined by the peaks and [5] G. Argenziano, G. Fabbrocini, P. Carli, V. De Giorgi, E. Sammarco, and
the histogram amplitude at i (see Fig. 10). M. Delfino, “Epiluminescence microscopy for the diagnosis of doubtful
melanocytic skin lesions. Comparison of the ABCD rule of dermatoscopy
The only information missing is how to select the most and a new 7-point checklist based on pattern analysis,” Arch. Dermatol.,
significant peaks in a histogram if more than two peaks are vol. 134, no. 12, pp. 1563–1570, Dec. 1998.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
[6] J. Mayer, “Systematic review of the diagnostic accuracy of dermatoscopy [26] Y. Faziloglu, R. Stanley, R. Moss, and W. S. R. McLean, “Colour his-
in detecting malignant melanoma,” Med. J. Aust., vol. 167, no. 4, pp. 206– togram analysis for melanoma discrimination in clinical images,” Skin
210, Aug. 1997. Res. Technol., vol. 9, no. 2, pp. 147–156, May 2003.
[7] M. Binder, M. Puespoeck-Schwartz, A. Steiner, H. Kittler, K. Wolff, [27] X. Yuan, Z. Yang, and G. Zouridakis, “SVM-based texture classification
M. Muellner, and H. Pehamberger, “Epiluminescence microscopy of and application to early melanoma detection,” in Proc. 28th IEEE EMBS
small pigmented skin lesions: Short-term formal training improves the Annu. Int. Conf., 2006, pp. 4775–4778.
diagnosis performance of dermatologists,” J. Amer. Acad. Dermatol., [28] M. Sheha, M. S. Mabrouk, and A. Sharawy, “Automatic detection of
vol. 36, no. 2, pp. 197–202, Feb. 1997. melanoma skin cancer using texture analysis,” Int. J. Comput. Appl.,
[8] G. Di Leo, A. Paolillo, P. Sommella, and C. Liguori, “An improved vol. 42, no. 20, pp. 22–26, Mar. 2012.
procedure for the automatic detection of dermoscopic structures in digital [29] J. Sivic and A. Zisserman, “Video google: A text retrieval approach to
elm images of skin lesions,” in Proc. 2008 IEEE Comput. Soc. VECIMS, object matching in videos,” in Proc. 9th IEEE Int. Conf. Comput. Vis.,
2008, pp. 190–194. 2003, pp. 1470–1477.
[9] M. Sadeghi, M. Razmara, P. Wighton, T. K. Lee, and M. S. Atkins, “A [30] J. Philbin, O. Chum, M. Isard, J. Sivic, and A. Zisserman, “Object retrieval
novel method for detection of pigment network in dermoscopic images with large vocabularies and fast spatial matching,” in Proc. IEEE Conf.
using graphs,” Comput. Med. Imaging Graph., vol. 35, no. 2, pp. 137– Comput. Vis. Patter Recog., 2007, pp. 1–8.
143, Mar. 2011. [31] G. Qiu, “Indexing chromatic and achromatic patterns for content-based
[10] C. Barata, J. Marques, and J. Rozeira, “Detecting the pigment network in colour image retrieval,” Pattern Recog., vol. 35, no. 8, pp. 1675–1686,
dermoscopy images: A directional approach,” in Proc. 33rd IEEE EMBS Aug. 2002.
Annu. Int. Conf., Sep. 2011, pp. 5120–5123. [32] C. Walravem, B. Caputo, and A. Graf, “Recognition with local features:
[11] G. Fabbrocini, G. Betta, G. Di Leo, C. Liguori, A. Paolillo, A. Pietrosanto, The kernel recipe,” in Proc. 9th IEEE Int. Conf. Comput. Vis., 2003,
P. Sommella, O. Rescigno, S. Cacciapuoti, F. Pastore, V. Vita, I. Mordente, pp. 257–264.
and F. Ayala, “Epilluminescence image processing for melanocytic skin [33] S. Lazebnik, C. Schmid, and J. Ponce, “Beyond bags of features:
lesion diagnosis based on 7-point check list: A preliminary discussion on Spatial pyramid matching for recognizing natural scene categories,” in
three parameters,” Open Dermatol. J., vol. 4, pp. 110–115, 2010. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recog., 2006,
[12] W. Stoecker, M. Wronkiewicz, R. Chowdhury, R. J. Stanley, J. Xu, pp. 2169–2178.
A. Bangert, B. Shrestha, D. A. Calcara, H. S. Rabinovitz, M. Oliviero, [34] F. S. Khan, J. van de Weijer, and M. Vanrell, “Top-down color attention
F. Ahmed, L. A. Perry, and R. Drugge, “Detection of granularity in for object recognition,” in Proc. IEEE 12th Int. Conf. Comput. Vis., 2009,
dermoscopy images of malignant melanoma using color and texture pp. 979–986.
features,” Comput. Med. Imaging Graph., vol. 35, no. 2, pp. 144–147, [35] M. E. Celebi, G. Schaefer, and H. Iyatomi, “Objective evaluation of
Mar. 2011. methods for border detection in dermoscopy images,” in Proc. 30th IEEE
[13] M. E. Celebi, H. Iyatomi, W. Stoecker, R. Moss, H. Rabinovitz, EMBS Annu. Int. Conf., 2008, pp. 3056–3059.
G. Argenziano, and H. Soyer, “Automatic detection of blue-white veil and [36] M. E. Celebi, H. A. Kingravi, H. Iyatomi, Y. A. Aslandogan,
related structures in dermoscopy images,” Comput. Med. Imaging Graph., W. V. Stoecker, and R. H. Moss, “Border detection in dermoscopy im-
vol. 32, no. 8, pp. 670–677, Dec. 2008. ages using statistical region merging,” Skin Res. Technol., vol. 14, no. 3,
[14] W. Stoecker, K. Gupta, R. Stanley, R. Moss, and B. Shrestha, “Detection pp. 347–353, Aug. 2008.
of asymmetric blotches in dermoscopy images of malignant melanomas [37] M. Burroni, L. Alparone, and F. Argenti, “Comments on ‘a new algo-
using relative color,” Skin Res. Technol., vol. 11, no. 3, pp. 179–184, rithm for border description of polarized light surface microscopic images
Aug. 2005. of pigmented skin lesions’,” IEEE Trans. Med. Imag., vol. 25, no. 12,
[15] G. Di Leo, A. Paolillo, P. Sommella, and G. Fabbrocini, “Automatic diag- pp. 1655–1656, Dec. 2006.
nosis of melanoma: A software system based on the 7-point checklist,” in [38] H. Iyatomi, H. Oka, M. Saito, A. Miyake, M. Kimoto, J. Yamagami,
Proc. 43rd Hawaii Int. Conf. Syst. Sci., 2010, pp. 1818–1823. S. Kobayashi, A. Tanikawa, M. Hagiwara, K. Ogawa, G. Argenziano,
[16] H. Ganster, A. Pinz, E. Wildling, M. Binder, and H. Kittler, “Auto- H. P. Soyer, and M. Tanaka, “Quantitative assessment of tumour area
mated melanoma recognition,” IEEE Trans. Med. Imaging, vol. 20, no. 3, extraction from dermoscopy images and evaluation of the computer-based
pp. 233–239, Mar. 2001. methods for automatic melanoma diagnostic system,” Melanoma Res.,
[17] M. Elbaum, “Computer-aided melanoma diagnosis,” Dermatol. Clin., vol. 16, no. 2, pp. 183–190, Apr. 2006.
vol. 20, no. 4, pp. 735–747, Oct. 2002. [39] M. E. Celebi, Q. Wen, S. Hwang, H. Iyatomi, and G. Schaefer, “Le-
[18] P. Rubegni, G. Cevenini, M. Burroni, R. Perotti, G. Dell’Eva, P. Sbano, sion border detection in dermoscopy images using ensembles of thresh-
and C. Miracco, “Automated diagnosis of pigment skin lesions,” Int. J. olding methods,” Skin Res. Technol., vol. 19, no. 1, pp. e252–e258,
Cancer, vol. 101, no. 6, pp. 576–580, Oct. 2002. Feb. 2013.
[19] A. Blum, H. Luedtke, U. Ellwanger, R. Schwabe, G. Rassner, and [40] M. Silveira, J. Nascimento, J. Marques, A. Marçal, T. Mendonça,
C. Garbe, “Digital image analysis for diagnosis of cutaneous melanoma. S. Yamauchi, and J. Maeda, “Comparison of segmentation methods for
Development of a highly effective computer algorithm based on analysis melanoma diagnosis in dermoscopy images,” IEEE J. Sel. Topics Signal
of 837 melanocytic lesions,” Brit. J. Dermatol., vol. 151, no. 5, pp. 1029– Processing., vol. 3, no. 1, pp. 35–45, Feb. 2009.
1038, Nov. 2004. [41] Y. Freund and R. Schapire, “A decision-theoretic generalization of online
[20] M. Burroni, P. Sbano, G. Cevenini, M. Risulo, G. Dell’eva, P. Barbini, learning and an application to boosting,” J. Comput. Syst. Sci., vol. 55,
C. Miracco, L. Fimiani, M. Andreassi, and P. Rubegni, “Dysplastic naevus no. 1, pp. 119–139, Aug. 1997.
versus in situ melanoma: Digital dermoscopy analysis,” Brit. J. Dermatol., [42] C. J. Burges, “A tutorial on support vector machines for pattern recog-
vol. 152, no. 4, pp. 679–684, Apr. 2005. nition,” Data Mining Knowl. Discovery, vol. 2, no. 2, pp. 121–167,
[21] S. Seidenari, G. Pellacani, and C. Grana, “Pigment distribution in Jun. 1998.
melanocytic lesion images: A digital parameter to be employed for [43] R. M. Haralick, K. Shanmugam, and I. Dinstein, “Textural features for
computer-aided diagnosis,” Skin Res. Technol., vol. 11, no. 4, pp. 236– image classification,” IEEE Trans. Syst., Man Cybern., vol. SNC-3, no. 6,
241, Nov. 2005. pp. 610–621, Nov. 1973.
[22] M. E. Celebi, H. A. Kingravi, B. Uddin, H. Iyatomi, Y. Aslandogan, [44] T. Ojala, M. Pietikäinen, and D. Harwood, “A comparative study of tex-
W. Stoecker, and R. Moss, “A methodological approach to the clas- ture measures with classification based on feature distributions,” Pattern
sification of dermoscopy images,” Computerized Medical Imaging and Recog., vol. 29, no. 1, pp. 51–59, Jan. 1996.
Graphics, vol. 31, no. 6, pp. 362–373, Sep. 2007. [45] R. Azencott, J. Wang, and L. Younes, “Texture classification using win-
[23] R. Stanley, W. Stoecker, and R. Moss, “A relative color approach to color dowed Fourier filters,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 19,
discrimination for malignant melanoma detection in dermoscopy images,” no. 2, pp. 148–153, Feb. 1997.
Skin Res. Technol., vol. 13, no. 1, pp. 67–72, Feb. 2007. [46] S. Arivazhagana, L. Ganesanb, and S. Priyala, “Texture
[24] H. Iyatomi, H. Oka, M. E. Celebi, M. Hashimoto, M. Hagiwara, classification using Gabor wavelets based rotation invariant
M. Tanaka, and K. Ogawa, “An improved Internet-based melanoma features,” Pattern Recog. Lett., vol. 27, no. 16, pp. 1976–1982,
screening system with dermatologist-like tumor area extraction algo- Dec. 2006.
rithm,” Comput. Med. Imaging Graph., vol. 32, no. 7, pp. 566–579, [47] D. Heeger and J. Bergen, “Pyramid-based texture analysis/synthesis,” in
Oct. 2008. Proc. SIGGRAPH, 1995, pp. 229–238.
[25] N. Situ, X. Yuan, G. Chen, and J. Zouridakis, “Malignant melanoma [48] T. Randen and J. Husoy, “Filtering for texture classification: A com-
detection by bag-of-features classification,” in Proc. 30th IEEE EMBS parative study,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 21, no. 4,
Annu. Int. Conf., 2008, pp. 3110–3113. pp. 291–310, Apr. 1999.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
BARATA et al.: TWO SYSTEMS FOR THE DETECTION OF MELANOMAS IN DERMOSCOPY IMAGES 15
[49] M. Tkalcic and J. Tasicl, “Colour spaces: Perceptual, historical and appli- Mariana Francisco received the E. Biomedical and
cational background,” in Proc. IEEE Region 8 EUROCON_Computer as M.Sc. degrees from Technical University of Lisbon,
a Tool, 2003, vol. 1, pp. 304–308. Lisbon, Portugal, in 2010 and 2012, respectively.
[50] M. Bratkova, S. Boulos, and P. Shirley, “orgb: A practical opponent color She is currently with the Institute for Systems and
space for computer graphics,” IEEE Comput. Graph. Appl., vol. 29, no. 1, Robotics, Instituto Superior Técnico, Lisbon.
pp. 42–55, Jan./Feb. 2009.
[51] K. van de Sande, T. Gevers, and C. Snoek, “Evaluating color descriptors
for object and scene recognition,” IEEE Trans. Pattern Anal. Mach. Intell.,
vol. 32, no. 9, pp. 1582–1593, Sep. 2010.
[52] P. Kakumanu, S. Makrogiannis, and N. Bourbakis, “A survey of skin-
color modeling and detection methods,” Pattern Recog., vol. 40, no. 3,
pp. 1106–1122, Mar. 2007.
[53] T. Leung and J. Malik, “Representing and recognizing the visual appear-
ance of materials using three-dimensional textons,” Int. J. Comput. Vis., Teresa Mendonça received the Applied Math and
vol. 43, no. 1, pp. 29–44, Jun. 2001. Ph.D. degrees from the University of Porto, Porto,
[54] R. Fergus, P. Perona, and A. Zisserman, “Object class recognition by Portugal, in 1980 and 1993, respectively.
unsupervised scale-invariant learning,” in Proc. IEEE Int. Conf. Comput. She is currently an Assistant Professor with the
Vis. Pattern Recog., 2003, pp. II-264–II-271. Department of Mathematics, Faculty of Sciences,
[55] P. Viola and M. J. Jones, “Robust real-time face detection,” Int. J. Comput. University of Porto, where she is a Researcher with
Vis., vol. 57, no. 2, pp. 137–154, May 2004. the Center for Research and Development in Mathe-
matics and Applications. She has authored or coau-
thored more than 120 papers in international journals
Catarina Barata received the E. Biomedical and and conferences. Her current research interests in-
M.Sc. degrees from Technical University of Lisbon, clude systems modeling and linear control, biomedi-
Lisbon, Portugal, in 2009 and 2011, respectively. She cal applications, and statistical image processing.
is currently working toward the Ph.D. degree in the
Institute for Systems and Robotics, Instituto Superior
Técnico, Lisbon. Jorge S. Marques received the E.E. and Ph.D. de-
Her research interests include image processing, grees and the Aggregation Title from Technical Uni-
pattern recognition, and dermoscopy. versity of Lisbon, Lisbon, Portugal, in 1981, 1990,
and 2002, respectively.
He is currently an Associate Professor with the
Department of Electrical and Computer Engineer-
ing, Instituto Superior Técnico, Lisbon, where he is
also a Researcher with the Institute for Systems and
Margarida Ruela received the E. Biomedical and
Robotics. He has authored or coauthored over 150
M.Sc. degrees from Technical University of Lisbon, papers in international journals and conferences and
Lisbon, Portugal, in 2010 and 2012, respectively.
has authored a book entitled Pattern Recognition:
She is currently a Researcher with the Institute for
Statistical and Neural Methods (IST Press, 2005, 2nd ed., in Portuguese). His
Systems and Robotics, Instituto Superior Técnico,
current research interests include statistical image processing, shape analysis,
Lisbon. Her research interests include biomedical
and pattern recognition.
image and signal processing and pattern recognition.
Dr. Marques was the Cochairman of the IAPR Conference IbPRIA 2005,
the President of the Portuguese Association for Pattern Recognition from 2001
to 2003, and an Associate Editor of the Statistics and Computing Journal
(Springer).