Introduction

Psoriasis, usually characterized by rapid skin cell buildup, manifests as red patches with white scales. In fact, this disease is a major threat to human health due to varied chronic conditions, higher prevalence rate, increased mental depression and anxiety, treatment challenges, etc. Dermatologists face diagnostic challenges due to variability and subjectivity, hindering efficient clinical practice. Its prevalence rate ranges from 0.91 to 8.5% in adults and 0.0–2.1% amongst children1. Global prevalence is around 2–3% of world’s population, according to the world psoriasis day consortium2. In northern India, the prevalence rate of psoriasis amongst adults varies from 0.44 to 2.8%. A comprehensive study conducted across various medical institutions in Dibrugarh, Calcutta, Patna, Darbhanga, Lucknow, New Delhi, and Amritsar revealed that the overall incidence of psoriasis among skin patients is 1.02%, with observed rates ranging from 0.44 to 2.2%3. Chronic plaque psoriasis is the most prevalent form of the disease, accounting for over 90% of cases worldwide4. In 2016, psoriasis was associated with an estimated 5.6 million disability-adjusted life years (DALYs), which is at least three times higher than that of inflammatory bowel disease, according to the Global Burden of Disease Study5. Most prevalence data stems from hospital-based research, with a lack of large-scale population studies to accurately assess the community prevalence of this condition6,7,8,9. Artificial intelligence (AI)-enabled image recognition presents a promising solution for automating diagnostics and improving clinical efficiency. Psoriasis has several treatments available, however there is currently no known cure for the condition, effective management improves patients’ quality of life, especially at early stages. AI has transformed multiple fields by integrating human-like capabilities such as learning and reasoning into software systems, enabling computers to perform tasks typically executed by humans. Advances in computing power, extensive datasets, and innovative AI algorithms special deep transfer learning models enhanced the scopes of biomedical image-guided problem solving in areas like finger vein recognition10, diabetic retinopathy detection11,12,13, lung nodule identification14, RNA engineering15,16, bio-mathematical challenges17, smart agriculture18, and various cancer detection (viz., skin, breast and prostate)19,20,21,22,23,24,25,26,27 etc.

The integration of AI in clinical settings raises important ethical aspects, particularly concerning patient privacy, data security, and potential biases in diagnostics. Protecting patient confidentiality necessitates stringent data protection measures to prevent unauthorized access, while robust security protocols are essential for safeguarding sensitive health information. Additionally, AI models can inherit biases from their training data, potentially resulting in disparities in healthcare delivery; ensuring diverse and representative datasets can help address these issues. It is vital to connect the technical performance of AI models with their clinical impact, illustrating how improved diagnostic accuracy and efficiency can enhance patient outcomes, streamline clinical processes, and assist healthcare providers in making informed decisions.

Deep learning (DL) has shown exceptional success in analyzing various image types—such as magnetic resonance images, optical images, digital images, and dermoscopic images—enabling effective prediction and classification of a range of disorders, including diabetic retinopathy, cancer, cardiac arrest, breast cancer, brain tumors, and fatty liver disease28,29,30. Leveraging digital image processing and machine learning techniques, researchers implemented automated disease detection under low-resource computational frameworks11,31. Studies on psoriasis detection have explored various approaches, including transfer learning models and support vector machines (SVM), achieving accuracies of 92.2% and 90%, respectively32,33. Researchers also utilized the capabilities of DL, particularly convolutional neural networks (CNN) and long short-term memory (LSTM) models, to achieve notable accuracies of 84.2% and 72.3% in classifying various types of psoriasis34. DL models have further been instrumental in automated psoriasis diagnosis, showcasing promising accuracy in multiple investigations35,36,37,38,39,40. An AI-based dermatology diagnostic assistant, powered by the EfficientNet-B4 algorithm, exhibited an impressive overall accuracy of 95.80%, in addressing both two-class and four-class psoriasis classification problems41,42. Nieniewski et al. (2023) demonstrated transfer learning models with SVM-RBF classifiers with an accuracy of 80.08% for detecting psoriasis43. Another notable implementation of a smart-assisted psoriasis identification system was employed CNN models to achieve an outstanding accuracy of 98.1%44. Fink et al. (2018)45 adopted intra-class-correlation-coefficients (ICCCs) and mean absolute difference (MAD), achieving a mean ICCC of 0.877 and MAD of 2.2. Additionally, Raj et al. (2021)46proposed a transfer learning model for psoriasis segmentation, resulting in Dice-Coefficient and Jaccard scores of 97.43% and 96.05% respectively. Hsieh et al. (2022) introduced a nail psoriasis severity index utilizing the mask R-CNN architecture47, while Shrivastava et al. (2016) developed a computer-aided system for psoriasis detection using principal component analysis with 540 aggregated images48. Furthermore, Shtanko et al. (2022)49 developed a CNN architecture using the Dermnetdataset, achieving 98% prediction accuracy in psoriasis detection with respect to other skin diseases. A clinical decision support system was developed using DL model specifically to psoriasis treatment strategies50. Okamoto et al. (2022) utilized a customized dataset for psoriasis severity analysis with a transfer learning model51. In addition, the severity of psoriasis was evaluated using dictionary learning and cascaded deep learning methods based on skin images52,53,54,55,56. Table 1 summarized the literature describing DL models for psoriasis recognition.

Table 1 Summary of literature cited deep learning models using psoriasis imaging modality.

According to the literature, it can be observed that automated detection of psoriasis still remains a challenging task. Its varied nature manifests with different types viz., plaque, guttate, inverse, pustular and erythrodermic psoriasis etc., that affect human population physically and mentally both. This work aimed at developing an efficient deep transfer leaning framework for screening of psoriatic infection with more accuracy over the existing studies, where the patch-based image selection from the raw clinical photographs improved the model efficiency. Literature survey revealed that transfer learning models were found to efficiently implement in many of the automated disease detection processes with higher accuracy. This work focused on employing an efficient framework for psoriasis screening using standard deep transfer learning models viz., VGG-16, VGG-19, ResNet-50, and MobileNetV1. It is well known that training of these models requires large number of images, but the availability of real diagnostic and annotated images is a challenging task. Under such circumstances, our proposed framework included patch-generation module to generate more sub-images from raw images prior for efficient training. Moreover, such module enabled the pre-trained models to capture finer information about the infection, which may be missed during image sampling. The technical contribution can be thought of exploring adaptive moment (Adam) and adaptive gradient (AdaGrad) optimizers for hyperparameter tuning for psoriasis detection, aiming to achieve the best possible model performance. The performance of these transfer learning models was evaluated and compared in respect to sensitivity, specificity and receiver operating characteristic (ROC) curve.

Proposed methodology

Image acquisition

Under the guidance of the dermatologist, a customized digital image dataset was developed by capturing 175 photos of psoriasis diseased skin and 150 images of healthy skin at a reputable state hospital in West Bengal, India. The digital photos of human skins were captured following relevant guidelines and regulations as per Institutional ethical approval [Ref. No. 09/IEC/RNLKWC/2024; Institutional Ethics Committee (IEC)/Institutional Review Board (IRB), Raja Narendralal Khan Women’s College (Autonomous), Midnapur – 7210102, India] and no human samples (tissue / serum) were used in this study. Informed consent was obtained from all the participating subjects. The original images were saved in JPEG format, each with dimensions of 2532 × 1170 pixels (see Fig. 1). A digital single-lens mobile camera was utilized to capture high-resolution human skin images. These images exclusively focus on facial skin, taken under the illuminating fluorescence light source. Our dataset comprises of a total of 60 subjects, ranging in age from 20 to 45 years old. Given the nature of dermatological conditions such as psoriasis, we encountered significant variability in the raw images. This variability includes background noise, portions of normal skin, and blurry areas, especially in the context of the region of interest, which is the psoriasis-affected skin. To optimize the image data used for training transfer learning models, a meticulous process of image preparation was undertaken (such as patch creation, resizing and scaling etc.). We made a deliberate choice to focus on picture patches that contain only the signatures of psoriasis [class I] and healthy skin [class II] (see Fig. 2). This approach allows our models to concentrate on the key features of interest while filtering out irrelevant noise and distractions.

Figure 1
figure 1

Sample representations of psoriasis and normal skin images.

Figure 2
figure 2

Patch generation and selection process from raw captured digital images.

Using the image patches, the proposed methodology predicting psoriasis disease has been graphically described in the above figure.

Image preprocessing

  1. A)

    Resizing image.

One of the key tasks of pre-processing for transfer learning involves resizing images. This step involves adjusting the dimensions of the input images to fit the specific requirements of the chosen transfer learning models (VGG16, VGG19, MobileNetV1, ResNet-50). These models have been pre-trained on datasets (ImageNet) where images are of size 224 × 224 pixels with 3 color channels (red, green, blue). So, the selected patch images were resized from approximately 240 × 240 × 3 to 224 × 224 × 3 pixels for compatibility of the model. This ensures consistency in input dimensions across all images, making it easier for the model to learn patterns and features.

  1. B)

    Normalizing pixel values.

Normalization involves adjusting the pixel values of the images. Pixel values in images typically range from 0 to 255 (8-bit representation), with 0 being black and 255 being white. Normalization standardizes these values to a range between 0 and 1. Each pixel value in the image matrix is divided by 255. This operation scales down the pixel values so they fall within the range of 0 to 1.

  1. C)

    Encoding labels into binary format (1s and 0s).

This step involves transforming categorical labels into a numerical format suitable for training transfer learning models. First, we identify the unique categories in the provided labels. We then assign a binary representation to each unique category. Each original label is then converted into a binary vector based on its binary representation. The resulting binary vectors serve as the ground truth labels during model training. For instance, when the model predicts an image, it outputs a vector of probabilities for each class. This vector is then compared to the ground truth binary label vector using a loss function like categorical cross-entropy to update the model’s weights during training.

Psoriasis screening using transfer learning

Transfer learning adapts for new tasks, leveraging knowledge from one task for another. It’s like training a model basic ability before honing it for a particular task, which saves time and frequently yields better results, especially when there is shortage of data. Four popular pre-trained models (VGG16, VGG19, MobileNetV1, ResNet-50) were used for psoriasis diagnosis. The VGG16 and VGG19 models are renowned for their simplicity, elegance, and exceptional performance in image classification tasks. These were developed by the Visual Geometry Group (VGG) at the University of Oxford. These models stand out for their deep design, which includes 16 or 19 weight layers, depending on the model. Each layer uses 3 × 3 convolutional filters, and a 2 × 2 max-pooling layer comes after it. These models were able to capture complex information in pictures while retaining a consistent architecture because to the use of tiny convolutional filters, which makes them simpler to comprehend and train.

MobileNetV1 is unique in that it emphasizes low memory use and model efficacy. This is accomplished by using depthwise separable convolutions, which significantly lower the network’s parameter count while maintaining competitive accuracy. In response to the growing need for efficient deep neural networks that can run on low-resource devices, such smartphones, Google developed MobileNet. We have used the V1 version of this specific MobileNet architecture.

ResNet-50 introduced the idea of residual blocks, where shortcut connections omit one or more levels and contains 50 layers. The training of very deep networks is facilitated by this skip-connection design. Microsoft Research created ResNets to solve the vanishing gradient issue in extremely deep networks. ResNet-50 has achieved records on several benchmark datasets, including ImageNet, and performs superbly in image classification tasks.

After running all these models in the context of psoriasis detection, we observed that MobileNetV1 used to perform better as compared to other transfer learning models. Apart from the performance matrix, MobileNetV1 was found to be the lightest and simple algorithm due to its ability to perform efficiently under low computing resource-based framework. The proposed methodology has been depicted in the below flowchart (see Fig. 3).

Figure 3
figure 3

Flowchart of the proposed deep transfer learning framework for psoriasis image classification.

Proposed architecture

  1. A)

    MobileNetV1 framework.

MobileNetV1 model, architecture reviewed by google with the application of depth wise separable convolution layer the model was able reduce the size and complexity compared to the other state of the art models. In these depth wise separable layers, a depthwise convolution and a pointwise convolution were used. Based on these two operations the cost function became

$$\:{C}_{cost}=({D}_{k}\times\:\:{D}_{k}\times\:M\times\:\:{D}_{F}\times\:\:{D}_{F})+(M\times\:N\times\:\:{D}_{F}\times\:\:{D}_{F})$$
(1)

A width multiplier (α) was introduced to control the number of channels in depth wise convolutional cost. A detailed summary of Eqs. (1 & 2) notations with their interpretations were given in the Table 2.

Table 2 Summary table of the model notations with interpretation.

Now the cost function was updated (\(\:{D}_{k}\times\:\:{D}_{k}\times\:\alpha\:M\times\:\:{D}_{F}\times\:\:{D}_{F}+\alpha\:M\times\:\alpha\:N\times\:\:{D}_{F}\times\:\:{D}_{F})\) where range of α lies between 0 and 1. After a width multiplier a resolution multiplier was also added to control the input image resolution of the network so, the final version of the cost function came as follows.

$$\:{{C}_{cost}=(D}_{k}\times\:\:{D}_{k}\times\:\alpha\:M\times\:\:{\rho\:D}_{F}\times\:\:{\rho\:D}_{F})+(\alpha\:M\times\:\alpha\:N\times\:\:{\rho\:D}_{F}\times\:\:{\rho\:D}_{F})$$
(2)

Here ρ value lies in the range of 0 to 1. The develop cost function makes the main difference between MobileNetV1 and other architectures in context to the diagnostic classification. The proposed architecture, designed for 3-channel color skin images (RGB), utilized 224 × 224 image-patches as input. The initial segment included a standard convolutional layer with 32 filters, followed by batch normalization and ReLU activation to address the vanishing gradient problem, maintaining spatial dimensions at 112 × 112 × 32. The second part focused on depthwise separable convolutional layers (see Eqs. 1&2), enhancing feature mapping capacity with 288 and 2048 parameters. This pattern was repeated with zero-padding layers for feature extraction. The last output layer of the MobileNetV1 model was removed, an extra set layer had been added to make the model more accurate in the following task. After getting the output from the last depthwise separable convolutional layer a flatten layer was used to convert the 3D output to 1D array input. Next a dense layer with ReLU activation was included to learn complex patterns in the data. One dropout layer was also added to counter the overfitting scenario by reducing the reliance on specific neurons. Lastly an output layer with a single neuron and a sigmoid activation function was added to make the final classification. For visual understanding and detailed architecture of the proposed methodology using MobileNetV1 model, graphical and tabular representations in Fig. 4; Table 3 respectively.

Figure 4
figure 4

Overall architecture of fine-tuned MobileNetV1 model for psoriasis.

Table 3 Detailed architecture of the fined-tuned mobilenetv1.

The pseudo code of proposed architecture was given below:

figure a
  1. B)

    Hyperparameter tuning.

In order to minimize the loss function, which was determined using binary cross-entropy, two optimizers and a learning rate were taken into consideration throughout the MobileNetV1 model’s parameter tweaking process. Due to the improvement in performance shown in the training and testing process compared to the other models, we decided to modify the parameters of the MobileNetV1.

Learning rate

It establishes the extent to which the model changes throughout training. A low learning rate might result in a lengthy training process or a local minimum, whereas a high learning rate could force the model to skip optimum solutions. In our case, we take a learning rate in the range [0.0001, 0.001] to test the model on different learning thresholds57. To enhance its performance, or, to put it another way, to boost its accuracy, we have slightly altered its parameters. Table 3 includes the fine-tuned version’s architecture in detail, and Fig. 5 represents the selection of the learning rate for our model.

Figure 5
figure 5

Training and Validation accuracy with changing learning rate of our proposed fine-tuned MobileNetV1with Adam optimizer.

Optimization functions

In the process of parameter tweaking the MobileNetV1 model, we explore four different optimizers (SGD, SGD with moments, Adam, AdaGrad). Among them, two optimizers, viz., adaptive moment (Adam) with a leaning rate of [0.0001, 0.001] and adaptive gradient (AdaGrad) with an auto-selected leaning rate, were found to minimize the loss, computed by the binary cross-entropy function. It automatically selected the learning rate of the layers and optimized the model’s performance accordingly. This specific optimizer, which has previously been explored in this area of medical image analysis, performed well in the task of classification40,58. Hence, the governing Eqs. (37) of the loss function and optimizers were mathematically represented below.

Suppose we have total ‘n’ no. of samples where \(y_i\)  and \({\widehat y}_i\)  denote the actual and predicted outputs. Then, the loss due to the training of the network is computed by binary cross-entropy function as

$$\:L\left({y}_{i},{\widehat{y}}_{i}\right)=-\frac{1}{n}\sum\:_{i=0}^{n}{y}_{i}\text{log}\left({\widehat{y}}_{i}\right)+\left(1-{y}_{i}\right)\text{log}\left(1-{\widehat{y}}_{i}\right)$$
(3)

Now the following optimizers were used as follows:

  1. (a)

    Adaptive moment estimation (known as Adam).

$$\:{w}_{t+1}={w}_{t}-\gamma\:{m}_{t}$$
(4)
$$\:{m}_{t}=\beta\:{m}_{t-1}+\left(1-\beta\:\right)\left[\frac{\delta\:L\left({y}_{i},{\widehat{y}}_{i}\right)}{\partial\:{w}_{t}}\right]$$

where \(\:\frac{\delta\:L\left({y}_{i},{\widehat{y}}_{i}\right)}{\partial\:{w}_{t}}\) denoted gradient at loss function.

  1. (b)

    Adaptive Gradient Algorithm (known as AdaGrad).

To automatically assign the learning rate, an advanced optimizer was adopted with an expectation to provide better performance in classification task.

$$\:{w}_{t+1}={w}_{t}-\frac{\eta\:}{\sqrt{{G}_{ii}+\in\:}}\:\times\:\:\:\:\frac{\delta\:L\left({y}_{i},{\widehat{y}}_{i}\right)}{\partial\:{w}_{t}}\:$$
(5)

Where \(\:\frac{\delta\:L\left({y}_{i},{\widehat{y}}_{i}\right)}{\partial\:{w}_{t}}\) signifies the gradient of the loss function with respect to the parameters at the current time step t.

  1. (c)

    Stochastic Gradient Descent (SGD).

SGD optimizer is a variant of gradient descent algorithm, which uses one or a small number of training instances at a time to update the model parameters more frequently than batch gradient descent.

$$\:{\theta\:}_{t+1}={\theta\:}_{t}-\eta\:{\nabla\:}_{\theta\:}L\left({\theta\:}_{t}\right)$$
(6)

Where \(\:{\theta\:}_{t}\), \(\:\eta\:\) and \(\:{\nabla\:}_{\theta\:}L\left({\theta\:}_{t}\right)\) indicate the model parameters at ‘t’ iteration, learning rate, and the gradient (partial derivative) of the loss function L(θ) respectively.

  1. (d)

    Stochastic Gradient Descent with Moments.

The formula was updated for SGD with momentum, introduced a velocity term, which turns into a running average of past gradients to maintain the direction of optimization.

$$\:{v}_{t+1}={\gamma\:v}_{t}-\eta\:{\nabla\:}_{\theta\:}L\left({\theta\:}_{t}\right)$$
$$\:{\theta\:}_{t+1}={\theta\:}_{t}-{v}_{t+1}$$
(7)

where \(\:{\theta\:}_{t}\) model parameters (weights) at iteration t., \(\:\eta\:\): Learning rate, a small constant that controls the step size of the update, \(\:{\nabla\:}_{\theta\:}L\left({\theta\:}_{t}\right)\): Gradient (partial derivative) of the loss function L(θ) with respect to parameters \(\:{\theta\:}_{t}\), \(\:{v}_{t}\)​: Velocity or momentum term at iteration t, which accumulates the previous gradients, \(\:\gamma\:\): Momentum coefficient.

  • E) Performance metric

Recall/Sensitivity

The capacity of a model to accurately identify all positive instances within a dataset is measured by recall, also known as true positive rate or sensitivity. It determines the proportion of genuine positives to all real positives. High recall means the model captures the majority of positive examples, which is crucial because missing a positive instance might have serious repercussions.

$$\:\text{S}\text{e}\text{n}\text{s}\text{i}\text{t}\text{i}\text{v}\text{i}\text{t}\text{y}/\text{R}\text{e}\text{c}\text{a}\text{l}\text{l}\left(\%\right)=\frac{TP}{\left(TP+FN\right)}\times\:100$$
(8)

where TP- True positive; FN- False negative.

Specificity

Specificity is a statistic that measures how well a model can recognize real negatives among all other possible negatives. It determines the proportion of real negatives to all real negatives.

$$\:\text{S}\text{p}\text{e}\text{c}\text{i}\text{f}\text{i}\text{c}\text{i}\text{t}\text{y}\left(\%\right)=\frac{TN}{\left(TN+FP\right)}\times\:100$$
(9)

where TN- True negative; FP- False positive.

Precision

In the context of a classification system, precision measures the percentage of correctly identified positive items compared to the total number of objects categorized as positive. It illustrates how well the model predicts positive events and reduces the likelihood that negative events will be mistaken for positive ones. The number of true positive findings divided by the total of true positive and false positive results is the formula used to calculate precision. It draws attention to how trustworthy a model’s positive predictions are, which is crucial in situations where misclassifying negatives as positives might have serious repercussions.

$$\:\text{P}\text{r}\text{e}\text{c}\text{i}\text{s}\text{i}\text{o}\text{n}\left(\%\right)=\frac{TN}{\left(TN+FP\right)}\times\:100$$
(10)

Accuracy

Accuracy is defined as the proportion of correctly predicted instances (including true positives and true negatives) to all occurrences in the dataset. It is a frequently employed indicator for assessing a classification model’s performance.

$$\:\text{A}\text{c}\text{c}\text{u}\text{r}\text{a}\text{c}\text{y}\:\left(\text{\%}\right)=\frac{TP+TN}{\left(TP+FP+TN+FN\right)}\times\:100$$
(11)

Where TP- True positive; TN- True negative; FP- False positive; FN- False negative.

Receiver operating characteristic (ROC)

The performance of a model at various discriminating thresholds is depicted graphically by the ROC curve. At various threshold values, it shows the true positive rate (sensitivity) versus the false positive rate (1 - specificity). The model’s overall performance is summarized by the area under the ROC curve (AUC), with a greater AUC denoting better discrimination between positive and negative cases.

Dice coefficient

A statistical method for determining how similar two samples are is the dice coefficient, sometimes called the dice similarity coefficient (DSC). When a confusion matrix is used for classification tasks, particularly when dealing with binary classification or picture segmentation issues. To compare the MobileNetV1 performance with a non-ML base diagnostic method (i.e., clinician’s decision), the dice coefficient was obtained using the following formula.

$$\:DC=\frac{2\times\:TP}{(2\times\:TP+FP+FN)}$$
(12)

Results and discussion

Computational setup

The proposed skin disease recognition model underwent experiments using two systems: PC (8GB memory, 2GB GPU) for image pre-processing and Google Colab (12.69 GB RAM, 15 GB GPU) for model fitting and validation.

Psoriasis image-patch processing and classification

The skin images were initially processed into 240\(\:\times\:\)240 patches, resulting in 497 patches representing psoriasis and 396 patches of healthy skin. These patches were then resized to 224\(\:\times\:\)224 pixels to fit the requirements of pre-trained convolutional neural network models, specifically VGG16, VGG19, ResNet-50, and MobileNetV1. Each of these models was trained using 80% of the data, aiming to fine-tune the models for skin lesion classification. Amongst these, MobileNetV1 model was found to perform better over other models to distinguish between psoriasis and healthy skin patches. Furthermore, the model fine-tuned with the Adam and AdaGard optimizers led to improve the model’s performance (using Eqs. 811). Delving into the architecture of MobileNetV1 used in this study, it can be noted that the model was structured with specific layer groupings for different levels of feature map generation. The first 37 layers filters were responsible for generating low-level feature maps, capturing basic patterns and textures in the input image patches. Following this, layers filters 38 to 75 were designed to extract mid-level features, such as more complex shapes and structures within the skin lesions. Lastly, layers filters 76 to 86 focused on high-level feature extraction, capturing intricate details and unique characteristics specific to psoriasis lesions. A sample of feature map of the different layers was given below in Fig. 6.

Figure. 6
figure 6

(a) Visualization of the Input Image; (b) Visualization of low, mid and high-level feature map as generated by the convolutional layers of the fine-tuned MobileNetV1 architecture.

In the classification module of the proposed architecture, we experimentally adopted four optimizers using Eqs. (47) for minimizing the loss function using Eq. (3). From Table 4, it can be observed that the among all the optimizer AdaGrad optimizer provided the maximum accuracy as compared to the other optimizers.

Table 4 Experimental verification of the proposed model using various optimizers.

Accordingly, the MobileNetV1 model fine-tuned with AdaGrad optimizer and other hyper-parameters was found to achieve overall accuracy of 99.13%. In Table 5, the performance of the proposed methodology was compared to the state-of-the-art models, as given below.

Table 5 Model performance table (statistical measures and information about the models).

In Table 5, it can be observed that the fine-tuned MobileNetV1-based methodology achieved 94.25% sensitivity, 96.42% specificity, and 99.13% accuracy, respectively. Therefore, the achieved sensitivity of 94.25% indicated the model’s ability to correctly identify a true positive case of psoriasis from the dataset. The specificity of 96.42% signifies the model’s ability to accurately identify true negative cases of healthy skin.

Figure 7
figure 7

Visualization of the Model performance parameters (Accuracy, Sensitivity, Specificity).

Figure 8
figure 8

ROC curves showing comparative evaluation of performances of the proposed model with others.

Figure 7 presents the performance matrix as a heatmap for a better understanding of the proposed methodology. The model accuracy of 99.13% represents the percentage of correct classification of both psoriasis and healthy skin patches. In this context, a ROC graph was provided below in Fig. 8 to compare the performance of the state-of-the-art models with the proposed methodology graphically.

Figure 9
figure 9

(a) Visualization of training and test loss over the epochs; (b) Visualize the training and test accuracy over the epochs for train and test data.

In addition to the quantitative performance evaluation, the proposed fine-tuned MobileNetV1 model’s efficiency was assessed based on medical expert’s opinion using Eq. (12), as follows:

$$\:DC=\frac{2\times\:76}{2\times\:76+1+2}=0.98$$

where the DC value (= 0.98) demonstrates very close agreement between the medical expert’s diagnostic decision and the predicted outcome of our proposed model. Therefore, it can be thought of as a reliable AI-assisted tool for practical implementation at least where there is a scarcity of specialized clinicians.

Table 6 Comparative analysis of the proposed fine-tuned mobilenetv1 model with state-of-art works.

In Table 6, the methodology proposed, dataset used, and accuracy achieved by the state-of-the-art literatures were summarized and compared with our proposed fine-tuned transfer learning methodology towards psoriasis disease detection. So, it can be well observed that the proposed model achieved significant accuracy for screening psoriasis from a healthy skin image. As the model required a large number of psoriasis images for training purposes, primarily, an image augmentation process in terms of image-patch generation was followed from raw psoriasis images that included resizing all patches into 224 × 224 pixels. Further, the pixel values were normalized into the range [0, 1], which improved the classification task by ensuring consistent input image dimensions and enhancing model compatibility. Such image augmentation impacted the model performance to achieve higher accuracy, from 88 to 96%, as visualized in the ROC plot (see Fig. 10).

Figure 10
figure 10

Visualization of best fitted model performance with respect to input images.

At the same time, the runtime complexities of the transfer learning models were computed and compared with respect to the system configuration, as mentioned below in Table 7. It can be observed that VGG16 has a runtime of 42.75 s and an accuracy of 89.43, while its more advanced counterpart, VGG19, takes slightly longer at 43.64 s, yielding an improved accuracy of 90.67. MobileNetV1 demonstrates a significant performa nce leap, running in 41.17 s with an accuracy of 97.24. ResNet-50 is the fastest among the conventional models, with a runtime of 28.27 s, but it falls short in performance with an accuracy of 86.96.

Table 7 Computation of models’ runtime.

The standout performer is the proposed MobileNetV1 model, which not only achieves a superior runtime of 28.12 s but also delivers an impressive accuracy of 98.02. An additional configuration of this proposed model pushes the accuracy even higher to 99.13, with a runtime of 28.18 s. Among all the models, the proposed MobileNetV1 is the best due to its optimal balance of speed and performance. It not only demonstrates the shortest runtime but also achieves the highest accuracy, indicating exceptional efficiency and accuracy in classification tasks. This makes it the most suitable model for real-time applications where both quick processing and high accuracy are crucial.

The proposed MobileNetV1 utilized depthwise separable convolutions, which split the standard convolution operation into depthwise and pointwise convolutions. This design significantly reduces the number of parameters and the computational load. The lower number of parameters made the proposed framework suitable for deployment on mobile and edge devices with limited computational resources. Though it requires more runtime than ResNet-50, the accuracy of the MobileNetV1 bits of every other model of the study, along with ResNet-50.

Conclusion

This work proposed a fine-tuned transfer learning model (MobileNetV1) for automated detection of psoriasis disease using normal optical images of human skin. Though this proposed methodology achieved the maximum accuracy, this study has some limitations: (a) patch selection from raw images by the experienced medical expert guided; (b) unavailability of the large dataset. The model’s performance is highly dependent on the quality of the input images, poor lighting, low resolution, or obstructions can negatively impact accuracy. (c) The model may not generalize well to all patient demographics or variations in psoriasis presentation. Training on diverse datasets is essential to improve its robustness.

The proposed MobileNetV1 model, optimized for hyper-parameter tuning, demonstrates superior performance in psoriasis detection, achieving a remarkable 99.13% accuracy and a runtime of 28.18 s. This efficiency makes it ideal for real-time clinical use. The model can be integrated into dermatology clinics to enhance diagnostic accuracy, reduce healthcare professionals’ workload, and support telemedicine, improving access to dermatological care for remote patients. Additionally, it aids in personalized treatment planning, disease monitoring, and large-scale data analysis, offering potential applications for other skin conditions and contributing to research on new treatments.