- Systematic Review
- Open access
- Published:
Skin image analysis for detection and quantitative assessment of dermatitis, vitiligo and alopecia areata lesions: a systematic literature review
BMC Medical Informatics and Decision Making volume 25, Article number: 10 (2025)
Abstract
Vitiligo, alopecia areata, atopic, and stasis dermatitis are common skin conditions that pose diagnostic and assessment challenges. Skin image analysis is a promising noninvasive approach for objective and automated detection as well as quantitative assessment of skin diseases. This review provides a systematic literature search regarding the analysis of computer vision techniques applied to these benign skin conditions, following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses guidelines. The review examines deep learning architectures and image processing algorithms for segmentation, feature extraction, and classification tasks employed for disease detection. It also focuses on practical applications, emphasizing quantitative disease assessment, and the performance of various computer vision approaches for each condition while highlighting their strengths and limitations. Finally, the review denotes the need for disease-specific datasets with curated annotations and suggests future directions toward unsupervised or self-supervised approaches. Additionally, the findings underscore the importance of developing accurate, automated tools for disease severity score calculation to improve ML-based monitoring and diagnosis in dermatology.
Trial registration
Not applicable.
Introduction
In dermatological clinical trials, numerous digital images are captured to evaluate treatment effects. Automated image analysis, lesion detection, and feature extraction can help experts assess results [1]. Furthermore, machine learning (ML) can be used to identify statistical trends and biases across trials. Integrating state-of-the-art (SOTA) computer vision (CV) techniques into electronic clinical outcome assessments (eCOAs) can transform clinical treatment development [2]. Beyond controlled trials, applying CV to less standardized images from physicians or patients offers benefits for reliable and faster skin condition detection, despite challenges such as variability in image quality and capture standards [3]. Advances in CV are increasingly addressing these limitations, providing promising tools for stakeholders.
Currently, ML applications in dermoscopy focus predominantly on malignant skin lesion tasks like classification, detection, and segmentation [4]. Benign skin diseases, while impactful on quality of life, are less explored. These include Atopic Dermatitis (AD), Stasis Dermatitis (SD), Vitiligo (VI), and Alopecia Areata (AA), which affect millions worldwide:
-
AD is a chronic inflammatory condition affecting 20% of children [5]. The symptoms include recurrent lesions, itching, and dryness, as well as acute or chronic manifestations like erythema, oozing, and lichenification [5, 6].
-
SD, linked to chronic venous insufficiency, affects the lower extremities and presents with discoloration, itching, redness, swelling, and pain due to elevated venous pressure [7].
-
VI, the most common depigmenting disorder, affects 0.1–2% of the global population. Characterized by lighter patches from melanocyte loss, it has psychological but not life expectancy impacts, with origins tied to genetics and stress [8].
-
AA, an autoimmune condition affecting 2% of people globally, causes hair loss in patches or universally, affecting all demographic groups [9].
Applying ML and CV techniques to these diseases could enhance the efficiency and reliability of diagnosis and monitoring through automated systems, thereby providing repeatable and trustworthy outcomes. A systematic screening of the literature concerning the application of ML and CV approaches for skin medical image analysis focusing on VI, AA, SD, and AD aims to reveal the relevant potentials, limitations, challenges, and opportunities. In this systematic review, a detailed description of this search is provided, focusing on methods that quantify skin visual patterns to guide useful knowledge extraction for downstream tasks.
Methods
To provide consistent feedback on the research work conducted in recent years in the field of ML and CV techniques concerning the four diseases, a systematic search is completed to provide a list of eligible papers and their key findings. The entire procedure is reported in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines, and the respective checklist is provided in Appendix B. The inclusion criteria required for a study to:
-
Focus on one or more of the four skin diseases of interest or include one or more of the four skin diseases among other diseases in the multiclass setting. In the case of dermatitis-related work, the paper must explicitly discuss SD or AD.
-
Propose an automatic solution to classify, detect, and/or segment the query class that is AI-based or hand-engineered.
-
Report evaluation results on open-source datasets or reasonably described in-house datasets.
The search was conducted on the Scopus database and included all combinations of words that refer to skin diseases and the names of ‘vitiligo’, ‘alopecia areata’, ‘atopic dermatitis’, and ‘stasis dermatitis’, in conjunction with AI- and CV-related terms. The alternative terms ‘venous’, ‘gravitational’, ‘congestion’, and ‘varicose’ for stasis and ‘eczema’ for dermatitis, among others, commonly found in the literature, were also taken into consideration. The term ‘leukoderma’ differs from VI, but it has been included in the search because this practice may uncover interesting CV works transferable to the VI domain due to the similarity of the two diseases. The search time frame spans from January 2004 to December 2024 (the specific end date is regulated by the Scopus database update and consultation time points). It captures emerging methodologies since AI and CV were radically evolving. The specific search queries used, including keywords, language and publication limitations applied, and dates of the last Scopus consultation, can be found in Appendix A. The manual review of all retrieved works, 441 in total, has led to the extraction of useful information about the methods used, and in some cases promising results. Specifically, the screening process comprised two screening rounds, prior to which an automated deduplication step identified 7 duplicate entries, reshaping the total paper count to 434. The first round screened the titles, abstracts and, if necessary, introductions of the 434 papers, providing a high-level view of each paper’s contents and leading to the rejection of 351 ineligible entries. The second round consisted of the meticulous study of the 83 retrieved papers, 46 of which were finally judged as eligible for inclusion (Fig. 1). The selected worksare scanned to identify methods related to image classification,
Lesion segmentation, lesion detection, and image processing techniques for image preprocessing and augmentation. A special focus has also been placed on discovering efforts for automated skin disease severity quantification, using measures such as the Vitiligo Area Scoring Index (VASI) and Severity of Alopecia Tool (SALT) scores, as well as other relevant metrics used by researchers. Both the screening and data collection parts of this review were conducted by 3 researchers working independently, without the assistance of automation software. Even though in some studies, information about the method was incomplete and the evaluation pipeline returned no strong proof of the robustness and performance of the proposed methodology (criticized by N. van Geel et al. [10]), a critical decision was made to include such studies in the review to extract additional useful information.
Results
The reviewed studies on skin disease lesions are categorized into four main skin conditions AA, AD, SD, VI and classified by downstream tasks: classification, object detection, segmentation, and severity score calculation. Additionally, some studies include data augmentation techniques to address data scarcity and overfitting. Most studies focus on VI, whereas eczema-related research helps compensate for the limited research on SD and AD because these conditions share visual characteristics. Among the downstream tasks, classification is the most common, followed by segmentation. The severity score calculation ranks lower, often integrating segmentation as part of the process. Object detection is the least emphasized because segmentation provides a more granular pixel-level analysis, making it preferable for detailed skin lesion assessment. The publication timeline is from August 2012 to April 2024. As shown in Figs. 2 and 3, the most productive year refers to 2023 with 21 works, followed by 2022 with 12 works. Works related to AA witnessed a sudden peak in the year 2023, possibly due to the release of high-quality datasets, whereas vitiligo publications are spread over all years in contrast to articles on other diseases that appear sporadically.
Machine learning applications for Alopecia Areata
Most existing studies in the field of computer vision for AA refer to the task of image classification [11,12,13,14,15,16,17,18,19]. The existing approaches are based on the typical machine-learning pipeline shown in Fig. 4. In each machine learning subtask, namely preprocessing, feature extraction, feature selection, and classification, the ML/CV algorithms for the existing works are depicted. In the case of neural networks [11, 12, 14,15,16], the subtask choices are fewer since the network architecture handles the feature extraction and classification subtasks entirely. However, there exists an approach by Mittal et al. that breaks down the total process into subtasks, even when using neural networks [12]. The feature extraction process is managed by a pretrained VGG16 network, and the classification is performed by a Support Vector Machine (SVM) classifier. As shown in Fig. 4, the use of histogram equalization techniques for contrast enhancement is a popular choice globally or locally (Contrast Limited Adaptive Histogram.
Overview of CV and ML approaches for AA in the relevant literature. GLCM stands for Gray-Level Co-occurrence matrix, LBP for Local Binary Patterns, AAA for Artificial Algae Algorithm, WNN for Wavelet Neural Network, and MELM for Modified Extreme Learning Machine. The dotted frame, labeled Neural Network is an optional choice in the pipeline that can replace the handcrafted feature extraction and classification process
Equalization - CLAHE). The team of Saraswathi and Pushpa presents significant contributions in the field [16,17,18,19,20,21], accounting for 54% of the relevant literature. Their most promising results are reported in [18] showcasing 96.94% accuracy in a four-class AA classification task. The above results are obtained when enhancing the Attention-based Balanced Multi-Tasking Ensembling Deep (AB-MTEDeep) model, which combines feature extraction from various scales and residual connections with the Generative Adversarial (GAN) -generated training data. A common throughout all classificationstudies is the utilization of the Figaro1k [22] and DermNet datasets [23].
Concerning the image segmentation task, the literature review reveals three existing works, two supervised and one unsupervised technique. Lee et al. [24] tackle a dual segmentation task to segment both scalp and hair loss areas employing a U-Net [25], whereas in [26], Bernardis and Castello-Soccio assign a cluster to each pixel of the image as per its hair density using the K-means clustering algorithm to differentiate between bald, low-density, and normal hair scalp regions in scalp images. Due to its simplicity, the system’s inference time and ability to differentiate between AA and other causes of low hair density areas are low, as the authors acknowledge. The works described by Lee et al. [24] and Gudobba et al. [27], although referring to segmentation, are described in the following subsection due to their main objective being closely related to the score (Severity of Alopecia Tool) calculation task.
Severity of alopecia tool quantitative analysis for Alopecia Areata
The segmentation techniques discussed earlier are foundational for the quantitative analysis of AA images. A widely adopted metric for assessing AA severity is the SALT score [28, 29], which evaluates scalp hair loss across four regions—vertex, right profile, left profile, and posterior—each assigned a weight factor and a score from 0 to 1. The SALT score, ranging from 0 (no hair loss) to 100 (complete hair loss), is calculated by summing the weighted scores for each region. It is used to monitor treatment efficacy by comparing baseline (BL) and follow-up (F/U) scores, with percentage changes computed using a specific formula (Eq. 1) [30]. Despite its usefulness, SALT has limitations, such as ignoring factors like the duration and recurrence of hair loss, psychological impact, and involvement of other areas (e.g., eyebrows, eyelashes) [31]. Additionally, the manual nature of the calculation introduces potential subjectivity and errors, thereby making the calculation time-intensive [31].
Automated machine learning techniques for calculating SALT have been extensively studied to address time and subjectivity challenges. These methods primarily rely on segmentation approaches to differentiate hair from scalp areas, which are categorized into one unsupervised and two supervised techniques.
In the supervised domain, Lee et al.‘s segmentation method [24] employs the AloNet model, a CNN based on U-Net [25], to classify each pixel as “hair loss” or “scalp area.” This functionality is embedded in a web application to measure hair loss in AA patients. Similarly, Gudobba et al. [27] developed the HairComb algorithm, which uses two encoder-decoder branches based on UNet and ResNet50 to automatically calculate hair loss percentage across alopecia subtypes. HairComb is integrated into Trichy, a web-based tool, which provides user guidance for image capturing. HairComb reported 7% absolute error in calculating affected area percentages, which is comparable to the state-of-the-art SALT algorithm in [24].
The unsupervised approach proposed by Bernardis et al. [26] leverages pixel intensity of neighboring pixels to create a visual vocabulary for encoding images into vectors, which are clustered into labels—scalp, low density, and normal hair—for distinguishing hair and scalp in AA images. Diverging from these works, Seol et al. [32] explored a two-dimensional planimetric method for calculating the actual surface area of AA as a means of validating SALT scoring.
Machine learning applications for atopic and stasis dermatitis
Studies specifically focusing on AD and SD are relatively scarce. To address this, the search is broadened to include studies from the overarching domain of eczema, with many categorizing eczema diseases generically [33,34,35,36,37,38,39,40]. However, some works focus on specific eczema diseases, particularly the erythematous-squamous (ES) class, which excludes AD and SD [41,42,43]. Others include seborrheic dermatitis and psoriasis within multiclass settings or classify diverse dermatitis cases [44,45,46,47], offering valuable insights into the automatic processing of AD and SD images due to their potential visual similarities in CV models. For example, Zhou et al. [45] found that using a green background improved classification performance for lesions with black and red colors.
In terms of feature extraction mechanisms in classification tasks, Nourin et al. [48] used handcrafted features, such as Histogram of Gradients (HOG) and Gray Level Co-occurrence Matrix (GLCM) to classify images of eczema, hemangioma, melanoma, and SD, achieving 95.3% accuracy with GLCM and 78% with HOG. In contrast, learned features extracted from deep CNN architectures [49,50,51] report higher accuracy (96.04–97.5%) on datasets that include SD images. Hybrid features combining handcrafted and learned approaches are examined in [52], where the ReliefF technique refines the feature set, which is then fed into classifiers, such as SVM, K-Nearest Neighbor (KNN), and DT, to achieve 97% accuracy. Gradient -based Class Activation Maps (Grad-CAM) [53] visual explanations accompany these results.
Srivastava et al. [54] developed an image-processing method for detecting eczema-affected regions. Their approach begins with image preprocessing, including noise reduction, contrast enhancement, and quality improvement, followed by segmentation using K-means in the Lab color space, Otsu thresholding, and morphological operations. Features such as color, border, and texture are subsequently extracted. Similarly, two studies [55, 56] address segmentation tasks for psoriasis lesions using simple CNNs combined with optimization techniques, specifically the Adaptive Chimp Optimization Algorithm (AChOA) and the Adaptive Golden Eagle Optimization (IGEO). These methods achieve high segmentation accuracy (97%) and may inform segmentation approaches for AD and SD given the common CV challenges between psoriasis and eczema. One of these studies utilized a private clinical dataset of 7,000 images, including 4,200 images of psoriasis and 2,800 images of healthy skin [55]. In contrast, Srivastava et al. [54] employed an unsupervised approach based on K-means clustering and Otsu thresholding [57] to segment eczema-affected regions in digital images.
In addition to these methods, Rajathi et al. [50] applied machine learning techniques to classify digital images into varicose ulcer stages (five stages) and tissue types (four classes), while also performing lesion segmentation and wound area calculation. Furthermore, the segmentation methods discussed in [58,59,60] are described in the following section as part of severity score calculation techniques.
Eczema area and severity index calculation
The Eczema Area and Severity Index (EASI) [61] was developed in 1998 and later validated to meet the demands of investigators in need of a standardized evaluation tool for the severity of AD signs in clinical studies [62]. The formula for the EASI score includes visual estimation in four body regions (head and neck, upper extremities, trunk, and lower extremities), where each region is assigned an area score. Next, each region is assessed separately for the following four signs: erythema, edema/papulation, excoriation, and lichenification. Each sign is assigned an intensity score ranging from 0 to 3: 0, absent; 1, mild; 2, moderate; and 3 being severe [63]. The SCORAD (Severity Scoring of Atopic Dermatitis) index is also validated but combines subjective assessment of patients’ symptoms with observation of signs [62]. Regarding the EASI calculation, a modified formula was used (Eq. 2):
Area Score is the percentage of eczema/total skin region, while the Intensity Score can range from 4 for mild eczema (setting 1 point for each of redness, thickness, scratching, and lichenification) to 12 (3 points for each) for severe eczema. The Region Score is the percentage of skin affected by eczema in each of the following four body regions: head (including neck), trunk, upper limbs, and lower limbs.
Machine learning researchers have developed automated systems for calculating the EASI score to help dermatologists achieve more consistent and reliable results. Alam et al. [58] proposed an automated eczema detection and severity measurement model by using 85 web-acquired images. Their pipeline includes: (a) a skin-region detection module leveraging the YCbCr color space, (b) an eczema-region detection module using k-means clustering in the LAB color space and morphological operations, (c) a feature extraction mechanism for color, texture (using GLCM), and border attributes, and (d) a classification system comprising two binary SVMs. One SVM distinguished between healthy and eczema skin, while the other classified eczema severity as mild or severe.
Bang et al. [64] trained four CNN architectures (ResNet V1-2, GoogleNet, and VGG-Net) to determine the optimal encoder for calculating individual EASI components. The accuracy rates were 90.63% for erythema, 89.06% for induration/papulation, 87.50% for excoriation, and 85.94% for lichenification.
Attar et al. [59] developed “EczemaNet2,” an enhanced version of their earlier “EczemaNet” model [60], which integrates U-Net to detect and segment AD regions. EczemaNet initially used an R-CNN-based approach, where cropped regions of interest (ROIs) were fed into seven classifiers, each producing a severity score (0–3) for specific disease signs. These scores were averaged across ROIs to calculate the EASI score, with additional support for the TISS [48] and SASSAD [49] scores. EczemaNet achieved a Root Mean Square Error (RMSE) of 1.929 ± 0.019 for EASI. EczemaNet2 replaced the R-CNN segmentation stage with two U-Nets, one for skin segmentation and another for AD segmentation. The postprocessing steps merge segmented regionsto generate square image crops for classifier input. Data augmentation techniques, including the pix2pix network, resulted in 25% and 40% improvements in segmentation and eczema detection performance, respectively, compared to the original EczemaNet pipeline.
Machine learning applications for vitiligo
Starting with the classification task for VI images, the literature review presents a significant number of works. The tasks vary in different taxonomies relative to the number of classes, the feature extraction mechanism type, the base versus ensemble classification techniques, and the application of data augmentation and transfer learning (TL) approaches. The simpler approaches are based on the plain feature extraction-classifier pipeline, excluding the use of elaborated preprocessing steps or techniques.
Concerning hand-crafted features, the authors of [65] propose the use of Mel Frequency Cepstral Coefficients (MFCC)—features often used for audio-related tasks—and i-Vectors as feature vectors paired with SVM and MLP classifiers, resulting in the best-performing out of four feature extraction-classifier combinations. Nosseir et al. [66] classify warts, hemangiomas, and VI using first- and second-order statistical features (GLCM) from pixel values. Related to learned features, Sharma et al. [67] perform feature extraction using Inception-V3, and various ML and DL algorithms are tested as classifiers, including a simple Naive Bayes and a CNN, all of which perform well. Bashar et al. [68] test four CNNs as feature extractors and four classifiers using similar methodology to obtain comparable results. A custom autoencoder CNN, which is an architecture commonly used for generative tasks, is defined and trained as a classifier [69], with the authors reporting 90% accuracy on the validation set. In a multiclass context, Algudah et al. [70] classify VI and five skin diseases using a short custom CNN. The model is simple, and the results are satisfactory and comparable to the work in [71]. Agrawal et al. [72] classify melanoma, VI, and vascular tumor images using fine-tuned InceptionV3. However, a noticeable 17% difference between training and test accuracy is observed.
Although more complex and prone to overfitting, ensemble models are often used by researchers to improve classification performance [73,74,75]. Liu et al. [73] employ three identical ResNet50 CNNs trained on different color spaces (RGB, HSV, and YCrCb) in an ensemble to identify skin images affected by VI. Saini et al. [74] define a two-model voting classifier trained on GLCM features. Dodia et al. [75] manually construct a small five-class dataset using a VGG-16-based feature extractor and a tree ensemble classifier trained with XGBoost [76]. These ensemble methods confirm that ensemble models outperform base classifiers. However, the use of private datasets in these studies hinders the comparison of state-of-the-art methods.
TL approaches are prevalent for utilizing pre-existing knowledge in a more specific domain [68, 77,78,79]. Mishra et al. [77] propose deep supervision for skin classification using activation mapping to create an image mask and targeting the network layer whose effective receptive field aligns best with the activation mask. An auxiliary loss function is fused with the standard loss function during training, thereby improving the performance of the VI datasets. Zhang et al. [78] utilize an in-house dataset al.ong with public datasets, offering an alternative open VI dataset. Their three TL-based models demonstrate superior performance to that of human experts. Bashar et al. [68] employ TL with four different DL architectures on the dataset used in earlier work [78].
In a different direction, generative models enhance datasets with synthetic samples. Luo et al. [79] propose a Cycle-Consistent Generative GAN-based augmentation procedure followed by super-resolution of the generated images. This method improves classification accuracy by 9.3% for a ResNet50 model compared to the non-augmented TL. Similarly, Mondal et al. [80] train a Wasserstein GAN with Gradient Penalty (WGAN-GP) to augment their datasetand then used CNNs to classify normal skin, leprosy, tinea versicolor, and VI. These models achieve accuracy in the 0.81–0.94 range. However, the datasetsize raises validity concerns. Liang et al. [81] introduce a novel method, Multi-hierarchy Contrastive Learning with Pareto Optimality (MHC-PO), which jointly trains models to learn data representations and perform classification tasks.
For VI lesion segmentation, preprocessing often involves classifying pixels into skin and background [82, 83]. Nugroho et al. [84] use independent component analysis (ICA) to represent skin disorders, followed by a region-growing algorithm for segmentation. Weakly supervised methods constitute a small but significant share of the literature [85, 86]. Bian et al. [85] use activation maps [87] of binary classifiers and the SLIC superpixel algorithm to delineate VI-affected areas, improving Intersection over Union (IoU). Low et al. [88] apply face recognition technology to correct image angles. Semi-supervised approaches like the mean-teacher learning framework proposed by Wang et al. [89] address labeled data scarcity by training a student model using pseudo-labels assigned by a teacher model.
U-Nets are widely used in medical segmentation tasks. For VI, U-Nets and their variations have been implemented [88, 90,91,92]. Low et al. [88] optimize U-Net’s encoding path with different CNN architectures, achieving the best results with InceptionResNetV2. Gou et al. [90] train three segmentation models—PSPNet [93], U-Net, and Unet++—on a large, annotated dataset, with Unet + + [94] performing best on in-house data but underperformed on open-source data. Li et al. [91] developed a U-Net-inspired model for facial VI cases, incorporating augmentation methods based on lesion color similarity between databases and target images.
Unsupervised methods show promise, particularly in scenarios with limited or subjective data labeling. Khatibi et al. [83] propose a stacked ensemble approach combining color-space-specific fully connected networks and clustering algorithms. Mehmood et al. [95] employ a simpler unsupervised technique to classify pixels based on color values, whereas Anthal et al. [96] use learning vector quantization networks for pixel classification. Nurhudatiana et al. [82] utilize Fuzzy C-means clustering for skin-background and VI segmentation based on YCbCr and RGB color spaces. In addition, Geel et al. [97] analyze ImageJ thresholding functions for VI lesion delineation.
Finally, bounding-box detection methods have been explored. Sorour et al. [98] employ a sequential configuration of YOLO-v5 models to predict VI-affected areas, including melanoma.
Vitiligo area scoring index calculation
The VASI is a widely used standard for assessing the extent of vitiligo (VI) lesions, providing repeatable but subjective insights into disease progression. The VASI score is calculated (Eq. 3) as the product of the affected area, measured in “hand units” (each equal to 1% of total body surface area), and the degree of depigmentation, expressed as a percentage from 0 to 100 [99] (Fig. 5). Accurate assessment involves registering the total affected body surface area and the extent of depigmentation for each lesion. While VASI relies on manual evaluation, techniques like superpixels [100] and level-set segmentation with SIFT and RANSAC [101] have been explored to aid in quantifying VI-affected areas, though their methods and outcomes vary. For instance, the superpixel approach lacks detail on calculating body surface area, whereas the level-set method yields percentage scores representing changes over time but
Visual representation of various degrees of depigmentation [99], licensed under CC BY 3.0
Not direct VASI scores. Alternative metrics, such as Facial VASI (F-VASI), Total VASI (T-VASI) [102], Vitiligo European Task Force (VETF) [103], and Vitiligo Extent Score (VES) [104], address limitations of the standard VASI by including facial surfaces or focusing on other aspects of disease assessment. These variations provide alternative approaches for quantifying and monitoring VI progression.
Data augmentation in the scope of medical skin images
Data augmentation requires an initial seed of images on which the technique can be based to generate new samples by applying a well-established transformation of the initial image (rotation, jittering, blurring, contrast enhancement). Alternatives refer to mixing two original images or copying and pasting parts of the initial image to a target sample [89, 91]. In the field of dermoscopy, image data augmentation has been applied in several cases, with the main objective of enhancing the training results of machine learning algorithms. As shown in Fig. 6, the data augmentation works related to the diseases in question are divided into four categories: (a) Geometric Transformations, (b) Kernel Filters [12, 52, 59, 88], (c) GAN-based [18, 52, 59, 79, 81, 91, 98], and (d) Mixing images [89, 91]. Starting from the case of basic image manipulations and their application to dermoscopy images, data augmentation techniques provide a copy of the original image by applying a simple transformation. In deep learning approaches, neural style transfer and GANs are frequently employed to solve the task of dermoscopy image augmentation. Such approaches require a wealth of training samples to produce reliable new samples ; therefore, they can not be deemed useful in cases where a sparsity of data samples is encountered. Luo et al. [79], use a Cycle-Consistent Generative Adversarial Network [105], which is followed by a super-resolution module to make amends for the absence of wood lamp images. The discrimination of vitiligo patterns in such images is more effective resulting in rather promising results. Mondal et al. [80]employed a Wasserstein GAN with a Gradient Penalty to generate synthetic VI images and increase the robustness of the overall classification scheme. An interesting example of the augmentation of macroscopy skin images captured by mobile devices is described by Andrade et al. [106]. A cycle-consistent adversarial network is used for the described objective, yielding effective quantitative metrics in the form of the Fréchet Inception Distance, while qualitative evaluation returns promising results in some cases. The technique could be used for different skin diseases in cases of comparing different endpoints (dermoscopy vs. macroscopy images from mobile phones) of the lesion between timestamps. In [52] the authors employ a GAN to generate the generation of synthetic skin lesion images. However, no samples or evaluation of the generative network is provided. Abdelhalim et al. [107] propose a progressively growing GAN for generating skin cancer images. The adversarial network exploits the gradually increasing image generation to address common training inconsistencies and improve image quality at higher resolution. Although these techniques have been rarely applied to images depicting areas of AA, AD/SD, and VI, success in the generic setting of skin lesions reveals the potential for their utilization for the increase of image samples for the diseases in question.
Mixing approaches have also been proposed in the relevant literature for increasing the number of image samples. A promising notion with different developed variations is the copy-and-paste procedure of the instance in question [108]. The copy-paste approach has been proven to significantly improve the results of segmentation algorithms compared to plain image transformations. It can be briefly described as the detection of objects in question in images and the overlay of these objects on a target in a systematic or random manner, meaning that the positions in the cuts are overlayed on the target images, are selected by a strategy concerning the context, or are placed at random locations of the image. Along with the image, the corresponding segmentation mask is modified to adhere to the given variations. More effective techniques for exploiting copy-paste have been proposed in [109,110,111] where contextual information is utilized for the detection of more effective locations and the alleviation of presented artifacts and noise mainly due to the direct operation of copying and pasting parts of the image.
Wang et al. [89], use the copy and paste augmentation to address data shortage in VI images for the corresponding skin lesion segmentation. The editing of the VI images in the proposed methodology is conducted using a poison-blending technique [112] to paste the object into the target image and applying an improvement of the original method based on a mixed gradient mode modification. The proposed technique is applied as a preprocessing step to generate additional samples and their corresponding masks to train a semi-supervised Mean Teacher segmentation scheme. The results demonstrate that the proposed augmentation technique significantly increases the segmentation metrics and represents a promising path when annotated VI samples are scarce.
Another interesting approach that attempts to address this issue of data scarcity is proposed in [91]. Dedicated to enhancing the performance of deep segmentation architectures, this approach generates patches of VI lesions on target images by exploiting the potential of Progressive Histogram Colour Transfer (PHCT), as originally proposed by Pouli and Reinhard [113]. The authors detect the most suitable VI color transfer using a similarity metric and introduce patches of VI-shaped regions in the target image. Thus, a significant increase in the segmentation performance mechanism is achieved.
Datasets for the surveyed diseases
A list of the available public datasets containing dermoscopy and on-the-wild skin lesion images for the four diseases is provided in Table 1. In terms of image quantity, the presented numbers are significantly less because most of the images in the dataset are transformed duplicates of the original images. An example is provided for the Vtigo Dataset2, which is reported to contain 1,187 images in total; however, the actual number of images is reduced to the duplicates that derive from horizontal flips.
Discussion
Through this review, a wider perspective on the diseases under examination is gained. More specifically, the review reports the effects on the skin that can be captured in the form of digital images, the proposed diagnostic methods and treatment plans, and the corresponding improvements that attempt to automatically quantify therapies and the depicted visual patterns. The reader can be led to the formation of significant insights regarding best practices that have resulted in efficient ML and CV schemes along with the limitations that have hindered previous attempts and open issues that require to be addressed through new approaches. Although most of the works in digital dermoscopy refer to skin cancer [139,140,141,142], this review is an attempt to turn the interest of dedicated researchers to other skin diseases that affect quality of life.
In general, data scarcity creates a considerable challenge, and the four skin diseases bear no exception. Although several publicly available datasets have been found with corresponding annotations to some extent, the purpose of capturing such images was not intended for ML applications [143]. Images of different resolutions, shapes, and capturing conditions constitute an arduous operational field for CV and AI algorithms to perform and extract useful outcomes. In addition to the abovementioned shortcomings, many samples include artifacts or regions of interest that are enlarged on top of existing morphological findings. Training skin-condition predictive models requires a large number of images taken under specific conditions (e.g., distance, point of view, areas of body parts). These datasets are usually collected through clinical examinations and are not provided along with the publication of the related results, either due to ethical reasons or due to purposes of commercializing the produced models. An additional obstacle, even for extensive clinical image datasets, is variations in the annotations and respective scores among the qualified dermatologists [144], even from the same expert at different times of re-evaluation. Therefore, it should be noted that the creation of publicly available curated datasets for each disease would greatly assist the development of more effective relevant ML algorithms and their fairer quantitative evaluation [145]. In skin disease research, GANs can model the distribution of skin disease images and generate synthetic samples resembling real manifestations. This review highlights GAN-based approaches as promising for generative AI in addressing dataset scarcity. However, effectively training GANs is challenging, and requires large datasets, typically thousands of images, to achieve satisfactory results. Even then, this approach might not improve the robustness of the detection model, asreported in a recent study of AD [42]. Nevertheless, synthetic image generation is a domain rapidly evolving and might be able to overcome the limitations of previous models, such as the one used in that study. Diffusion models [146] can generate high-fidelity, high-resolution synthetic skin disease images. They can capture intricate details and nuances in images, making them useful for generating diverse synthetic images that represent various skin diseases, textures, colors, and manifestations. This approach provides an innovative way to augment datasets, understand disease progression, and enhance the robustness of ML models in dermatology research and medical imaging. However, it is important to ensure that the generated synthetic images accurately reflect real-life skin conditions and are used responsibly alongside authentic clinical data. In addition, advanced techniques like elaborate copy-paste methods can further increase sample diversity and improve the representation of each lesion. Li et al. [91], apart from proposing a novel technique for increasing the number of VI image samples, point out the necessity of considering that different types of VI representations depict different visual patterns, and, therefore, may be treated diversely. Apart from the obvious disadvantages due to data scarcity, it should be taken into consideration that most of the presented works use privately owned datasets, and even when publicly available datasets are employed, the experimental setup greatly differs in each case. This diversity hinders the extraction of useful conclusions about the effectiveness of the proposed approaches, the SOTA results, and future directions.
The calculation of severity scores (SALT, EASI, and VASI), is important for providing objective and accurate quantification of each disease and is therefore a key objective of ML approaches. However, most approaches fail to develop pipelines that can effectively calculate existing scores, which results in approximations. With reference to AA and SALT, the technological advancements extend to the exploitation of ML techniques through web [24] and mobile [27] applications, thusproviding solutions to practical issues in real life and facilitating their use far from the supervised environment of a healthcare clinic or a research lab. Research on SD and EASI falls within the broader field of dermatitis and eczema lesions. Challenges such as the complexity of the visual characteristics of diseases, which can be confused with other skin conditions, as well as research prioritization and data scarcity, hinder the development of ML and CV algorithms. Regarding the quantification of visual patterns for VASI scores, the literature presents efforts using ML to extract information from vitiligo-affected skin areas, focusing on segmentation or calculating differentiation between endpoints rather than providing an actual VASI score. In a relevant literature review [10], few studies dealt with the automated calculation of VASI scores. Moreover, in most studies, the evaluation process fails to provide robust evidence of effectiveness, resulting in unreliable quantifiable results, if any exist. The severity scores (SALT, EASI, and VASI) for each disease can create a bridge between health and machine experts to research new approaches for the unbiased and reliable monitoring of skin disease progression. Although the systematic review of existing works highlights the potential of automated methods to accurately segment VI lesions, the lack of an automated tool that is specifically targeted at the calculation of VASI score, as proposed by the relevant experts, is compelling.
Following the latest regulations in the US and Europe for the development of responsible, trustworthy, transparent, and reliable AI, the enhancement of ML models with interpretability properties should be pursued in alignment with the requirement of delivering models that can support their results with reasoning for the integration of research projects, into routine clinical workflows. Indications of explainability approaches for the developed AI models for the four diseases were reported in [52, 85, 86]. Instead of utilizing visual explanations to deliver reasoning concerning the classification results, the visualizations in [85] and [86] are employed as a form of weakly-supervised segmentation mask that guides the process to focus on important regions for each skin disease.
Conclusions
This report presents a systematic review of the research literature concerning benign skin diseases. The report dives into the applications of CV and ML techniques for the extraction of knowledge from images to focus on the specifics of Atopic Dermatitis, Stasis Dermatitis, Alopecia Areata, and Vitiligo diseases. Apart from demonstrating works that refer to the SOTA on downstream tasks such as classification, detection, and segmentation, the report contains references to the severity indices applied by relevant experts to assess the depicted lesions, the data augmentation issue, and existing datasets. The shortcomings of previous implementations and the latest advancements from other fields of medical imaging that can contribute to the tasks in question have been identified to a large extent.
Although a significant number of publicly available datasets are presented herein, the qualitative issues and the actual quantity of unique samples reveal the need for disease-specific datasets with curated annotations. The poor exploitation of these datasets is demonstrated by the extensive usage of privately owned datasets and data augmentation techniques reported in the literature. On the other hand, the interrater variability [147, 148] in the annotations of skin lesions suggests future directions toward unsupervised or self-supervised approaches.
The review discusses the integration of severity scores, such as SALT, EASI, and VASI into ML approaches for monitoring skin diseases. These scores help connect healthcare and ML experts to develop unbiased, reliable methods for tracking disease progression. For AA, SALT has been effectively used in ML-based web and mobile applications, enabling monitoring outside clinical settings. However, ML and CV advancements in dermatitis and eczema are challenged by the complexity of symptoms and data scarcity. Regarding vitiligo, attempts to use ML for VASI score calculation are limited and lack a robust, reliable method. The need for automated tools specifically designed for accurate VASI score calculation is underscored, highlighting the gap in current research. Emphasis should be placed on developing accurate automated VASI tools, improving data collection, refining ML algorithms for complex conditions, integrating multi-modal data, standardizing ML approaches, and creating user-friendly applications for non-clinical use.
Incorporating skin image analysis in clinical workflows and web/mobile applications can facilitate rapid and precise diagnosis, which is crucial for early intervention and improved treatment outcomes. By reducing the time required for accurate diagnosis, clinicians can initiate appropriate treatments promptly, thereby improving patient management and validating healing progression. The assessment procedure will greatly benefit from descriptive and comparable visualizations to support and justify their reports.
References
Akbarialiabad H, Pasdar A, Murrell DF. Digital twins in dermatology, current status, and the road ahead. NPJ Digit Med. 2024;7:228. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41746-024-01220-7.
Shamout FE, Zhu T, Clifton DA. Machine learning for clinical outcome prediction. IEEE Rev Biomed Eng. 2020;14:116–26. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/RBME.2020.3007816.
Combalia M, Codella NC, Rotemberg VM, Helba B, Vilaplana V, Reiter O, Halpern AC, Puig S, Malvehy J. BCN20000: Dermoscopic Lesions in the Wild. Scientific Data 2019 11.
Hasan MK, Ahamad MA, Yap CH, Yang G. A survey, review, and future trends of skin lesion segmentation and classification. Comput Biol Med. 2023;155:106624. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2023.106624.
Weidinger S, Novak N. Atopic dermatitis. Lancet. 2016;387(10023):1109–22. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/S0140-6736(15)00149-X.
Raimondo A, Lembo S. Atopic dermatitis: epidemiology and clinical phenotypes. Dermatol Pract Concept. 2021;11(4):e2021146. https://doiorg.publicaciones.saludcastillayleon.es/10.5826/dpc.1104a146.
Yosipovitch G, Nedorost ST, Silverberg JI, Friedman AJ, Canosa JM, Cha A. Stasis dermatitis: an overview of its clinical presentation, pathogenesis, and management. Am J Clin Dermatol. 2023;24(2):2. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s40257-022-00753-5.
Bergqvist C, Ezzedine K. Vitiligo: a focus on pathogenesis and its therapeutic implications. J Dermatol. 2021;48:252.
Zhou C, Li X, Wang C, Zhang J. Alopecia Areata: an update on etiopathogenesis, diagnosis, and management. Clin Rev Allergy Immunol. 2021;61(3):403. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s12016-021-08883-0.
van Geel N, et al. Image analysis systems to calculate the surface area of vitiligo lesions: a systematic review of measurement properties. Pigment Cell Melanoma Res. 2022;35(5):480. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/pcmr.13056.
Shakeel CS, Khan SJ, Chaudhry B, Aijaz SF, Hassan U. Classification framework for healthy hairs and alopecia areata: a Machine Learning (ML) approach. Comput Math Methods Med. 2021;2021:1–10. https://doiorg.publicaciones.saludcastillayleon.es/10.1155/2021/1102083.
An Analysis of Alopecia Areata Classification Framework for Human Hair Loss Based on VGG-SVM Approach,. pnr. 2022;13. https://doiorg.publicaciones.saludcastillayleon.es/10.47750/pnr.2022.13.S01.02.
Mittal A, Biswas DB, Karthikeyan U, Prediction of Alopecia Areata using CNN, in. 2023 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC), Salem, India: IEEE, May 2023, pp. 141–144. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ICAAIC56838.2023.10140778
Roy M, Protity AT. Hair and scalp disease detection using machine learning and image processing. 2023;30. https://doiorg.publicaciones.saludcastillayleon.es/10.24018/ejcompute.2023.3.1.85.
Ibrahim S, Azmy ZAN, Mangshor NNA, Sabri N, Fadzil AFA, Ahmad Z. Pre-trained classification of scalp conditions using image processing. Indonesian J Electr Eng Comput Sci. 2020;20(1):138–44. https://doiorg.publicaciones.saludcastillayleon.es/10.11591/ijeecs.v20.i1.pp138-144.
Saraswathi C, Pushpa B, FRCNN based Deep Learning for Identification and Classification of Alopecia Areata, in. 2023 Fifth International Conference on Electrical, Computer and Communication Technologies (ICECCT), Erode, India: IEEE, Feb. 2023, pp. 1–7. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ICECCT56650.2023.10179804
Saraswathi C, Pushpa B. Machine Learning Algorithm for Classification of Alopecia Areata from Human Scalp Hair Images, in Computational Vision and Bio-Inspired Computing, S. Smys, J. M. R. S. Tavares, and F. Shi, Eds., in Advances in Intelligent Systems and Computing. Singapore: Springer Nature, 2023, pp. 269–288. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-981-19-9819-5_21
Saraswathi C, Pushpa B. AB-MTEDeep Classifier Trained with AAGAN for the Identification and Classification of Alopecia Areata, Engineering, Technology & Applied Science Research, vol. 13, no. 3, pp. 10895–10900, Jun. 2023, https://doiorg.publicaciones.saludcastillayleon.es/10.48084/etasr.5852
Saraswathi C, Pushpa B. Multi-class support vector machine classification for detecting alopecia areata and scalp diseases. Int J Recent Innov Trends Comput Commun. 2023;11(11):63–70. https://doiorg.publicaciones.saludcastillayleon.es/10.17762/ijritcc.v11i11s.8071.
Research, Scholar, Department of Computer and Information Science, Faculty of Science, University, Nadu A, India C, Saraswathi C, Pushpa B. Attention balanced multi-dimension multi-task deep learning for alopecia recognition. IJST. 2023;16(18):1365–73. https://doiorg.publicaciones.saludcastillayleon.es/10.17485/IJST/v16i18.29.
Saraswathi C, Pushpa B. Ensemble of pre-learned deep learning model and an optimized LSTM for Alopecia Areata classification. J Intel Fuzzy Syst. 2023;45(6):11369–80.https://doiorg.publicaciones.saludcastillayleon.es/10.3233/JIFS-232172
Muhammad UR, Svanera M, Leonardi R, Benini S. Hair detection, segmentation, and hairstyle classification in the wild. Image Vis Comput. 2018;71:25–37. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.imavis.2018.02.001.
Dermnet.com. We are currently Redesigning Dermnet Skin disease Atlas, Dermnet.com. Available: https://dermnet.com/Accessed 14 Aug 2024
Lee S, et al. Clinically applicable deep learning framework for measurement of the extent of hair loss in patients with Alopecia Areata. JAMA Dermatol. 2020;156(9):1018–20. https://doiorg.publicaciones.saludcastillayleon.es/10.1001/jamadermatol.2020.2188.
Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: Navab N, Hornegger J, Wells WM, Frangi AF, editors. in Medical Image Computing and Computer-assisted intervention – MICCAI 2015. Lecture Notes in Computer Science. Cham: Springer International Publishing; 2015. pp. 234–41. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-319-24574-4_28.
Bernardis E, Castelo-Soccio L. Quantifying Alopecia Areata via texture analysis to automate the SALT score computation. J Investig Dermatol Symp Proc. 2018;19(1):S34–40. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jisp.2017.10.010.
Gudobba C, et al. Automating hair loss labels for universally scoring Alopecia from images: rethinking Alopecia scores. JAMA Dermatol. 2023;159(2):143. https://doiorg.publicaciones.saludcastillayleon.es/10.1001/jamadermatol.2022.5415.
Olsen EA, et al. Alopecia Areata investigational assessment guidelines–part II. J Am Acad Dermatol. 2004;51(3):440–7. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jaad.2003.09.032.
Sibbald C. Alopecia Areata: an updated review for 2023. J Cutan Med Surg. 2023;27(3):241–59. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/12034754231168839.
Olsen EA. Investigative guidelines for Alopecia Areata. Dermatol Ther. 2011;24(3):311–9. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/j.1529-8019.2011.01415.x.
Determining Alopecia Areata Severity, National Alopecia Areata Foundation | NAAF. Available: https://www.naaf.org/determining-alopecia-areata-severity/Accessed 22 Jan 2024
Seol JE, Hong SM, Ahn SW, Jang SH, Kim H. Two-dimensional planimetry for Alopecia Areata severity evaluation compared with severity of alopecia tool: a pilot study. Skin Res Technol. 2023;29(9):e13440. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/srt.13440.
Melbin K, Raj YJV. Automated detection and classification of skin diseases using diverse features and improved gray wolf-based multiple-layer perceptron neural network. Int J Imaging Syst Tech. 2021;31(3):1317–33. https://doiorg.publicaciones.saludcastillayleon.es/10.1002/ima.22524.
Md Hossen N, Panneerselvam V, Koundal D, Ahmed K, Bui FM, Ibrahim SM. Federated machine learning for detection of skin diseases and enhancement of Internet of Medical Things (IoMT) security. IEEE J Biomed Health Informa. 2023;27(2):835–41. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/JBHI.2022.3149288.
Setiawan NB, Natalia F, Ferdinand FV, Sudirman S, Ko CS. Classification of skin diseases and disorders using convolutional neural network on a Mobile Application. ICIC Int 学会. 2021. https://doiorg.publicaciones.saludcastillayleon.es/10.24507/icicelb.12.08.715.
Saifan R, Jubair F. Six skin diseases classification using deep convolutional neural network. IJECE. 2022;12(3):3072. https://doiorg.publicaciones.saludcastillayleon.es/10.11591/ijece.v12i3.pp3072-3082.
Hameed N, Hameed F, Shabut A, Khan S, Cirstea S, Hossain A. An Intelligent computer-aided Scheme for classifying multiple skin lesions. Computers. 2019;8(3):62. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/computers8030062.
Hameed N, Shabut AM, Ghosh MK, Hossain MA. Multi-class multi-level classification algorithm for skin lesions classification using machine learning techniques. Expert Syst Appl. 2020;141:112961. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.eswa.2019.112961.
Bakiyalakshmi R, Umamaheswari S, Anbu, Rohit, Saranraj. Dermatology Disease Detection Model using image Processing and recurrent neural networks. J Adv Res Dynamic Control Syst. 2019;11(06–Special Issue):656–60.
Rajendran Dr Ebenezer Abishek GS, Dr B, Vijayalakshmi A. Dr V., Analysis And Diagnosis Using Deep -Learning Algorithm On Erythemato-Squamous Disease, International Journal of Engineering Trends and Technology - IJETT. Available: https://ijettjournal.org/Accessed 18 Apr 2024.
Menai MEB. Random forests for automatic differential diagnosis of erythemato-squamous diseases. IJMEI. 2015;7(2):124. https://doiorg.publicaciones.saludcastillayleon.es/10.1504/IJMEI.2015.068506.
Ubeyli ED, Güler I. Automatic detection of erthemato-squamous diseases using adaptive neuro- fuzzy inference systems. Comput Biol Med. 2005;35(5):421–33. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2004.03.003.
Design of Automatic Detection of Erythemato-squamous Diseases Through Threshold-based ABC-FELM Algorithm. Available: https://scialert.net/abstract/?doi=jai.2013.245.256Accessed 18 Apr 2024
Goceri E. Diagnosis of skin diseases in the era of deep learning and mobile technology. Comput Biol Med. 2021;134:104458. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2021.104458.
Zhou J, Wu Z, Jiang Z, Huang K, Guo K, Zhao S. Background selection schema on deep learning-based classification of dermatological disease. Comput Biol Med. 2022;149:105966. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2022.105966.
Goceri E. Deep learning based classification of facial dermatological disorders. Comput Biol Med. 2021;128:104118. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2020.104118.
Junayed MS, Sakib ANM, Anjum N, Islam MB, Jeny AA. EczemaNet: A Deep CNN-based Eczema Diseases Classification, in 2020 IEEE 4th International Conference on Image Processing, Applications and Systems (IPAS), 2020, pp. 174–179. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/IPAS50080.2020.9334929
Nourin N, Kundu P, Saima S, Rahman MA, GLCM and, Feature-Based HOG. Skin Disease Detection Using Artificial Neural Network, in Proceedings of International Conference on Information and Communication Technology for Development, M. Ahmad, M. S. Uddin, and Y. M. Jang, Eds., in Studies in Autonomic, Data-driven and Industrial Computing. Singapore: Springer Nature, 2023, pp. 355–364. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-981-19-7528-8_28
Evwiekpaefe AE, Amrevuawho of, classification of dermatologic manifestations of cardiovascular disease using, Mar. efficientnetv2 CNN model., International Journal of Intelligent Computing and Information Sciences, vol. 23, no. 1, pp. 115–127, 2023, https://doiorg.publicaciones.saludcastillayleon.es/10.21608/ijicis.2023.184311.1242
Rajathi V, Chinnasamy A, Selvakumari P, Net DUTC. A novel deep ulcer tissue classification network with stage prediction and treatment plan recommendation. Biomed Signal Process Control. 2024;90:105855. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.bspc.2023.105855.
Hammad M, Pławiak P, ElAffendi M, El-Latif AAA, Latif AAA. Enhanced deep learning approach for accurate eczema and psoriasis skin detection. Sensors. 2023;23(16):7295. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/s23167295.
Rasheed A, Umar AI, Shirazi SH, Khan Z, Nawaz S, Shahzad M. Automatic eczema classification in clinical images based on hybrid deep neural network. Comput Biol Med. 2022;147:105807. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2022.105807.
Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization, in 2017 IEEE International Conference on Computer Vision (ICCV), Oct. 2017, pp. 618–626. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ICCV.2017.74
Srivastava S, Singh A, Gupta R. (2018). Automatic Detection of Eczema Using Image Processing. In: Woungang, I., Dhurandher, S, editors International Conference on Wireless, Intelligent, and Distributed Environment for Communication. WIDECOM 2018. Lecture Notes on Data Engineering and Communications Technologies, vol 18. Springer, Cham. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-319-75626-4_13
Mohan S, Kasthuri N. Automatic Segmentation of Psoriasis Skin Images Using Adaptive Chimp Optimization Algorithm-Based CNN, J Digit Imaging, vol. 36, no. 3, pp. 1123–1136, Jun. 2023, https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s10278-022-00765-x
Panneerselvam K, Nayudu PP. Improved golden eagle optimization based CNN for automatic segmentation of psoriasis skin images. Wirel Pers Commun. 2023;131(3):1817–31. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s11277-023-10522-0.
Otsu N. A threshold selection method from gray-level histograms. IEEE Trans Syst Man Cybern. 1979;9(1):62–6. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/TSMC.1979.4310076.
Alam MN, Munia TTK, Tavakolian K, Vasefi F, MacKinnon N, Fazel-Rezai R. Automatic detection and severity measurement of eczema using image processing. Annu Int Conf IEEE Eng Med Biol Soc. 2016;2016:1365–8. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/EMBC.2016.7590961.
Attar R, et al. Reliable detection of eczema areas for fully automated assessment of eczema severity from digital camera images. JID Innov. 2023;3(5):100213. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.xjidi.2023.100213.
Pan K, Hurault G, Arulkumaran K, Williams HC, Tanaka RJ. EczemaNet: automating detection and Severity Assessment of atopic dermatitis. In: Liu M, Yan P, Lian C, Cao X, editors. in Machine learning in Medical Imaging. Lecture Notes in Computer Science. Cham: Springer International Publishing; 2020. pp. 220–30. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-030-59861-7_23.
Tofte S, Graeber M, Cherill R, Omoto M, Thurston M, Hanifin JM. Eczema area and severity index (EASI): a new tool to evaluate atopic dermatitis. J Eur Acad Dermatol Venereol. 1998;11:S. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/S0926-9959(98)95291-6.
Hanifin JM, et al. The eczema area and severity index (EASI): assessment of reliability in atopic dermatitis. Exp Dermatol. 2001;10(1):11–8. https://doiorg.publicaciones.saludcastillayleon.es/10.1034/j.1600-0625.2001.100102.x.
Hanifin JM, Baghoomian W, Grinich E, Leshem YA, Jacobson M, Simpson EL. The Eczema Area and Severity Index—A practical guide. Dermatitis. 2022;33(3):187–92. https://doiorg.publicaciones.saludcastillayleon.es/10.1097/DER.0000000000000895.
Bang CH, et al. Automated severity scoring of atopic dermatitis patients by a deep neural network. Sci Rep. 2021;11(1):6049. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41598-021-85489-8.
Salamea C, Chica JF. Vitiligo Detection Using Cepstral Coefficients: 1st International Conference on Advances in Emerging Trends and Technologies, ICAETT 2019, Advances in Emerging Trends and Technologies 2020 1;389–398. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-030-32022-5_36
Nosseir A, Shawky MA. Automatic Classifier for Skin Disease Using k-NN and SVM, Proceedings of the 2019 8th International Conference on Software and Information Engineering 2019 pp. 259–262. https://doiorg.publicaciones.saludcastillayleon.es/10.1145/3328833.3328862
Sharma S, Guleria K, Kumar S, Tiwari S. Deep learning based model for detection of vitiligo skin disease using pre-trained inception V3. Int J Math Eng Manag Sci. 2023;8(5):1024–39. https://doiorg.publicaciones.saludcastillayleon.es/10.33889/IJMEMS.2023.8.5.059.
Bashar N, Suliman MA. Vitiligo image classification using pre-trained Convolutional Neural Network Architectures, and its economic impact on health care. 2022. Available: https://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-313749Accessed 24 Jan 2024.
Agrawal N, Aurelia S, Classification of Vitiligo using CNN Autoencoder, in. 2022 International Conference on Applied Artificial Intelligence and Computing (ICAAIC), May 2022, pp. 174–176. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ICAAIC53929.2022.9792650
Alqudah AM, Alquraan H, Qasmieh IA. Segmented and non-segmented skin lesions classification using transfer learning and adaptive moment learning rate technique using pretrained convolutional neural network. J Biomimetics Biomaterials Biomedical Eng. 2019;42:67–78. https://doiorg.publicaciones.saludcastillayleon.es/10.4028/www.scientific.net/JBBBE.42.67.
Sandeep R, Vishal KP, Shamanth MS, Chethan K. Diagnosis of Visible Diseases Using CNNs, in Proceedings of International Conference on Communication and Artificial Intelligence, V. Goyal, M. Gupta, S. Mirjalili, and A. Trivedi, Eds., in Lecture Notes in Networks and Systems. Singapore: Springer Nature, 2022, pp. 459–468. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-981-19-0976-4_38
Agrawal N, Aurelia S. Corroboration of skin Diseases: Melanoma, Vitiligo & Vascular Tumor using Transfer Learning, 2021 7th International Conference on Electrical Energy Systems (ICEES) 2021 pp. 590–592. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ICEES51510.2021.9383682
Liu J, Yan J, Chen J, Sun G, Luo W. Classification of Vitiligo based on convolutional neural network. In: Sun X, Pan Z, Bertino E, editors. Artificial Intelligence and Security. Lecture Notes in Computer Science. Cham: Springer International Publishing; 2019. pp. 214–23. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-030-24265-7_19.
Saini K, Singh S. Oct., Vitiligo disease prediction using K-mean, GLCM and voting classification, 2555, p. 020013, 2022, https://doiorg.publicaciones.saludcastillayleon.es/10.1063/5.0109172
Dodia D, Jakharia H, Soni R, Borade S, Jain N. Human Skin Disease Detection using MLXG Model, in Joint Proceedings of the Workshop on Computer Vision and Machine Learning for Healthcare (CVMLH 2022) and the Workshop on Technological Innovations in Education and Knowledge Dissemination (WTEK 2022), L. Garg, G. Ganesan, G. Ganesan, and G. Kumar, Eds., in CEUR Workshop Proceedings, vol. 3338. Hybrid Event, Chennai: CEUR, Apr. 2022, pp. 01–14. Available: https://ceur-ws.org/Vol-3338/#ICCS_CVMLH_01. Accessed 30 Jan 2024.
Chen T, Guestrin C. XGBoost: A Scalable Tree Boosting System, in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, in KDD ’16. New York, NY, USA: Association for Computing Machinery 2016, pp. 785–794. https://doiorg.publicaciones.saludcastillayleon.es/10.1145/2939672.2939785
Mishra S, Zhang Y, Zhang L, Zhang T, Hu XS, Chen D. (2022). Data-Driven Deep Supervision for Skin Lesion Classification. International Conference on Medical Image Computing and Computer-Assisted Intervention.
Zhang L, et al. Design and assessment of convolutional neural network based methods for vitiligo diagnosis. Front Med (Lausanne). 2021;8:754202. https://doiorg.publicaciones.saludcastillayleon.es/10.3389/fmed.2021.754202.
Luo W, Liu J, Huang Y, Zhao N. An effective vitiligo intelligent classification system. J Ambient Intell Hum Comput. 2023;14(5):5479–88. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s12652-020-02357-5.
Mondal B, Das N, Santosh KC, Nasipuri M. Improved Skin Disease Classification Using Generative Adversarial Network, in 2020 IEEE 33rd International Symposium on Computer-Based Medical Systems (CBMS) 2020, pp. 520–525. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/CBMS49503.2020.00104
Liang S, Tian S, Kang X, Zhang D, Wu W, Yu L. Skin lesion classification base on multi-hierarchy contrastive learning with pareto optimality. Biomed Signal Process Control. 2023;86:105187. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.bspc.2023.105187.
Nurhudatiana A, Computer-Aided A. Diagnosis System for Vitiligo Assessment: A Segmentation Algorithm, in Intelligence in the Era of Big Data, R. Intan, C.-H. Chi, H. N. Palit, and L. W. Santoso, Eds., in Communications in Computer and Information Science. Berlin, Heidelberg: Springer, 2015, pp. 323–331. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/978-3-662-46742-8_30
Khatibi T, Rezaei N, Ataei Fashtami L, Totonchi M. Proposing a novel unsupervised stack ensemble of deep and conventional image segmentation (SEDCIS) method for localizing vitiligo lesions in skin images. Skin Res Technol. 2021;27(2):126. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/srt.12920.
Nugroho H, Ahmad Fadzil MH, Shamsudin N, Hussein SH. Computerised image analysis of vitiligo lesion: evaluation using manually defined lesion areas. Skin Res Technol. 2013;19(1):e72. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/j.1600-0846.2011.00610.x.
Bian Z, Xia S, Xia C, Shao M. Weakly Supervised Vitiligo Segmentation in Skin Image through Saliency Propagation, in 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 2019, pp. 931–934. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/BIBM47256.2019.8983145
Shih C, Lin C-H, Weng Y-C, Jian JY, Lin Y-C, Implementation of Weakly Supervised Vitiligo Treatment Evaluation System, in. 2022 IEEE 4th Eurasia Conference on Biomedical Engineering, Healthcare and Sustainability (ECBIOS), 2022, pp. 44–47. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ECBIOS54627.2022.9945053
Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning Deep Features for Discriminative Localization, in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA: IEEE, Jun. 2016, pp. 2921–2929. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/CVPR.2016.319
Low M, Huang V, Raina P. Automating Vitiligo Skin Lesion Segmentation Using Convolutional Neural Networks, in 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), 2020, pp. 1–4. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ISBI45749.2020.9098682
Wang J, Ding X, Xiao J. Poisson-based image editing for semi-supervised vitiligo lesion segmentation with limited annotations. Comput Biol Med. 2023;165:107320. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.compbiomed.2023.107320.
Guo L, et al. A deep learning-based hybrid artificial intelligence model for the detection and severity assessment of vitiligo lesions. Ann Transl Med. 2022;10(10):590. https://doiorg.publicaciones.saludcastillayleon.es/10.21037/atm-22-1738.
Li Y, Kong AW-K, Thng S. Segmenting Vitiligo on clinical face images using CNN trained on synthetic and internet images. IEEE J Biomed Health Inform. 2021;25(8):3082–93. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/JBHI.2021.3055213.
Li Y, Thng S, Kong AW-K. Bridging the gap between Vitiligo segmentation and clinical scores. IEEE J Biomed Health Inf. 2023;1–12. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/JBHI.2023.3342069.
Zhao H, Shi J, Qi X, Wang X, Jia J. Pyramid Scene Parsing Network 27, 2017, arXiv: arXiv:1612.01105. https://doiorg.publicaciones.saludcastillayleon.es/10.48550/arXiv.1612.01105
Zhou Z, Siddiquee MMR, Tajbakhsh N, Liang J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation, Jul. 18, 2018, arXiv: arXiv:1807.10165. https://doiorg.publicaciones.saludcastillayleon.es/10.48550/arXiv.1807.10165
Mehmood N, Khan SJ, Rashid M. K-means Clustering-based Color Segmentation on Vitiligo Skin Lesion, in 2022 International Conference on Emerging Trends in Smart Technologies (ICETST). 2022, pp. 1–5. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ICETST55735.2022.9922940
Anthal J, Upadhyay A, Gupta A. Detection of Vitiligo Skin Disease using LVQ Neural Network, in 2017 International Conference on Current Trends in Computer, Electrical, Electronics and Communication (CTCEEC) 2017, pp. 922–925. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/CTCEEC.2017.8455029
Geel N, et al. Reference method for digital surface measurement of target lesions in vitiligo: a comparative analysis. Br J Dermatol. 2019;180(5):1198–205. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/bjd.17190.
Sorour SE, Hany AA, Elredeny MS, Sedik A, Hussien RM. An Automatic Dermatology Detection System based on Deep Learning and Computer Vision. IEEE Access. 2023;11:137769–78. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ACCESS.2023.3340735.
Kawakami T, Hashimoto T. Disease severity indexes and treatment evaluation criteria in vitiligo. Dermatol Res Pract. 2011;2011:750342. https://doiorg.publicaciones.saludcastillayleon.es/10.1155/2011/750342.
Toh JJH, et al. Automated scoring of vitiligo using superpixel-generated computerized digital image analysis of clinical photographs: a novel and consistent way to score vitiligo. Br J Dermatol. 2018;179(1):220. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/bjd.16563.
Sheth VM, Rithe R, Pandya AG, Chandrakasan A. A pilot study to determine vitiligo target size using a computer-based image analysis program. J Am Acad Dermatol. 2015;73(2):342. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jaad.2015.04.035.
Kitchen H, et al. Meaningful Changes in what matters to individuals with Vitiligo: content validity and meaningful change thresholds of the Vitiligo Area Scoring Index (VASI). Dermatol Ther (Heidelb). 2022;12(7):1623–37. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s13555-022-00752-8.
Taïeb A, Picardo M, on behalf of the other VETF members. The definition and assessment of vitiligo: a consensus report of the Vitiligo European Task Force. Pigment Cell Res. 2007;20(1):27–35. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/j.1600-0749.2006.00355.x.
Van Geel N, et al. Development and validation of the Vitiligo Extent score (VES): an International Collaborative Initiative. J Invest Dermatol. 2016;136(5):978–84. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jid.2015.12.040.
Zhu J-Y, Park T, Isola P, Efros AA. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks, in 2017 IEEE International Conference on Computer Vision (ICCV), 2017, pp. 2242–2251. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ICCV.2017.244
Andrade C, Teixeira LF, Vasconcelos MJM, Rosado L. Data augmentation using adversarial image-to-image translation for the segmentation of mobile-acquired dermatological images. J Imaging. 2020;7(1):2. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/jimaging7010002.
Abdelhalim ISA, Mohamed MF, Mahdy YB. Data augmentation for skin lesion using self-attention based progressive generative adversarial network. Expert Syst Appl. 2021;165:113922. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.eswa.2020.113922.
Ghiasi G et al. Simple Copy-Paste is a Strong Data Augmentation Method for Instance Segmentation 2021arXiv: arXiv:2012.07177. Available: http://arxiv.org/abs/2012.07177Accessed 10 Jan
Dvornik N, Mairal J, Schmid C. Modeling Visual Context is Key to Augmenting Object Detection Datasets 2018arXiv: arXiv:1807.07428. Available: http://arxiv.org/abs/1807.07428Accessed 10 Jan 2024
Dwibedi D, Misra I, Hebert M. Cut, Paste and Learn: Surprisingly Easy Synthesis for Instance Detection, in 2017 IEEE International Conference on Computer Vision (ICCV), Venice: IEEE, Oct. 2017, pp. 1310–1319. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ICCV.2017.146
Fang H-S, Sun J, Wang R, Gou M, Li Y-L, Lu C. InstaBoost: Boosting Instance Segmentation via Probability Map Guided Copy-Pasting, 2019 arXiv: arXiv:1908.07801. Available: http://arxiv.org/abs/1908.07801Accessed 10 Jan 2024
Di Martino JM, Facciolo G, Meinhardt-Llopis E. Poisson image editing. Image Process Line. 2016;6:300–25. https://doiorg.publicaciones.saludcastillayleon.es/10.5201/ipol.2016.163.
Pouli T, Reinhard E. Progressive histogram reshaping for creative color transfer and tone reproduction, in Proceedings of the 8th International Symposium on Non-Photorealistic Animation and Rendering - NPAR 10, Annecy, France: ACM Press, 2010, pp. 81–90. https://doiorg.publicaciones.saludcastillayleon.es/10.1145/1809939.1809949
Kaggle. (2024). Skin datasets. Kaggle. https://www.kaggle.com/datasets/mohamedabdalgwad/skindatasets
Kaggle. (2024). DermNet. Kaggle. https://www.kaggle.com/datasets/shubhamgoel27/dermnet
Kaggle. (2024). Vitmon21. Kaggle. https://www.kaggle.com/datasets/khulankhalzaa/vitmon21
Kaggle. (2024). Vitiligo. Kaggle. https://www.kaggle.com/datasets/shinynose/vitiligo
Roboflow. (2024). Vitiligo detection. Roboflow. https://universe.roboflow.com/vitiligo/vitiligo-detection01/dataset/10
Roboflow. (2024). Vitiligo detection dataset. Roboflow. https://universe.roboflow.com/sedki/vitiligo-amw4r
Roboflow. (2024). Hair segmentation dataset. Roboflow. https://universe.roboflow.com/haircosys/hair-8kwjn/dataset/1
Roboflow. (2024). Vitiligo dataset. Roboflow. https://universe.roboflow.com/new-workspace-pdikd/vitiligodataset-yxvpr/dataset/4
Roboflow. (2024). Vitiligo segmentation dataset. Roboflow. https://universe.roboflow.com/vitiligo/vitiligo-seg01
Mendeley Data. (2024). Skin disease segmentation dataset. Mendeley Data. https://data.mendeley.com/datasets/75k76546ms/1
Roboflow. (2024). Segmentation dataset. Roboflow. https://universe.roboflow.com/project-1wzee/-rjwm3/dataset/1
Roboflow. (2024). Hair segmentation dataset. Roboflow. https://universe.roboflow.com/haircosys/hair-8kwjn
Roboflow. (2024). Eczema disease classification using EfficientNet architecture. Roboflow. https://universe.roboflow.com/mini-project-g4im3/eczema-disease-classification-using-efficientnet-architecture
Roboflow. (2024). Eczema object detection. Roboflow. https://universe.roboflow.com/neelesh-kumar-6ccll/eczema-gw7re
Roboflow. (2024). Disease area detection. Roboflow. https://universe.roboflow.com/tharindu-kumara-c1xoz/disease-area-detection
Roboflow. (2024). Object detection dataset. Roboflow. https://universe.roboflow.com/biopsyai/new_uae
Roboflow. (2024). Atopic dermatitis classification. Roboflow. https://universe.roboflow.com/nam-hai-xqnqq/atopicdermatitis-ejphb
Roboflow. (2024). Skin disease object detection. Roboflow. https://universe.roboflow.com/skin-disease/skin-disease-vrvtv
Roboflow. (2024). Atopic dermatitis segmentation. Roboflow. https://universe.roboflow.com/fyp-l8uct/3.-atopic-dermatitis
Roboflow. (2024). Nummular preprocessing. Roboflow. https://universe.roboflow.com/new-workspace-bamsf/nummular-preprocessing
Roboflow. (2024). Skin disease classification. Roboflow. https://universe.roboflow.com/kelixo/skin_disease_ak
Roboflow. (2024). Eczema segmentation. Roboflow. https://universe.roboflow.com/fyp-l8uct/fyp-eczema
Kaggle. (2024). Skin disease classification image dataset. Kaggle. https://www.kaggle.com/datasets/riyaelizashaju/skin-disease-classification-image-dataset/data
Kaggle. (2024). 20 skin diseases dataset. Kaggle. https://www.kaggle.com/datasets/haroonalam16/20-skin-diseases-dataset/data
Kaggle. (2024). Skin diseases image dataset. Kaggle. https://www.kaggle.com/datasets/ismailpromus/skin-diseases-image-dataset
Maglogiannis I, Delibasis KK. Enhancing classification accuracy utilizing globules and dots features in digital dermoscopy. Comput Methods Progr Biomed. 2015;118(2):124–33. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.cmpb.2014.12.001.
Patel RH, Foltz EA, Witkowski A, Ludzik J. Analysis of Artificial Intelligence-based approaches Applied to Non-invasive Imaging for Early Detection of Melanoma: a systematic review. Cancers. 2023;15(19):4694. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/cancers15194694.
Marka A, Carter JB, Toto E, Hassanpour S. Automated detection of nonmelanoma skin cancer using digital images: a systematic review. BMC Med Imaging. 2019;19(1):21. https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12880-019-0307-7.
Maragoudakis M, Maglogiannis I. Skin lesion diagnosis from images using novel ensemble classification techniques, in Proceedings of the 10th IEEE International Conference on Information Technology and Applications in Biomedicine, Corfu, Greece: IEEE, 2010, pp. 1–5. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ITAB.2010.5687620
Sun J, Yao K, Huang G, Zhang C, Leach M, Huang K, Yang X. Machine learning methods in skin Disease Recognition. A Systematic Review. Processes; 2023.
Bożek A, Reich A. Assessment of Intra- and inter-rater reliability of three methods for measuring atopic dermatitis severity: EASI, Objective SCORAD, and IGA. Dermatology. 2017;233:16–22.
Wen D, Soltan AA, Trucco E, Matin RN. From data to diagnosis: skin Cancer image datasets for Artificial Intelligence. Clinical and experimental dermatology; 2024.
Liu Z, Ma C, She W, Xie M. Biomedical image segmentation using denoising diffusion probabilistic models: a comprehensive review and analysis. Appl Sci. 2024;14(2):632. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/app14020632.
Fortina AB, Peserico E, Silletti A, Zattra E. Where’s the naevus? Inter-operator variability in the localization of melanocytic lesion border. Skin Res Technol. 2012;18(3):311–5. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/j.1600-0846.2011.00572.x.
Hurault G, et al. Detecting Eczema areas in Digital images: an impossible Task? JID Innov. 2022;2(5):100133. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.xjidi.2022.100133.
Acknowledgements
Not applicable.
Funding
This research was conducted in collaboration with the University of Piraeus and Pfizer. Pfizer is the research sponsor.
Author information
Authors and Affiliations
Contributions
A.K., K.M. and A.Z. prepared the manuscript. A.K., K.M. and A.Z. performed the investigation. A.K., K.M., and A.Z. wrote and edited the manuscript. A.K. and A.Z performed the visualizations. I.M. performed supervision. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Kallipolitis, A., Moutselos, K., Zafeiriou, A. et al. Skin image analysis for detection and quantitative assessment of dermatitis, vitiligo and alopecia areata lesions: a systematic literature review. BMC Med Inform Decis Mak 25, 10 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-024-02843-2
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-024-02843-2