COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Comparative Analysis Of Convolutional Neural Network Models For Digital Image-Based Melanoma Classification Ana Kurniawati *1. Aniqoh Hana Haura2 Information System. Faculty of Computer Science and Information Technology. Universitas Gunadarma. Jl. Margonda Raya No. Depok 16424. Jawa Barat e-mail: *1ana@staff. id, 2hana. haura06@gmail. Abstract Melanoma is one of the most malignant forms of skin cancer, with an incidence rate of 9% in Indonesia. Traditional biopsy-based diagnosis, though crucial, is invasive and timeconsuming, creating barriers for early detection. To address this issue, this research compares two Convolutional Neural Network (CNN) models for digital image-based melanoma The study utilized a publicly available dataset from Kaggle, consisting of 17,805 images . elanoma and non-melanom. , which were divided into training, validation, and testing The models were trained using the Adamax and SGD optimizers for 100 epochs. The performance of the models was evaluated based on accuracy, loss, precision, recall, and F1score. The CNN model with the best architecture, which consisted of two fully connected layers, achieved an accuracy of 93. 18% and a loss of 0. 1636, outperforming the alternative model. These results confirm the effectiveness of CNN models in classifying melanoma images and support the development of a web-based platform that allows users to upload or capture images for rapid and non-invasive detection. KeywordsAi Convolutional Neural Network. Melanoma Skin Cancer. Website INTRODUCTION According to Global Burden Cancer (Globoca. data in 2020, the number of skin cancer cases in Indonesia reached 18,000 with approximately 3,000 deaths . In Indonesia, the three most common types of skin cancer are basal cell skin cancer . 5%), squamous cell skin cancer . %), and melanoma skin cancer . 9%) . Melanoma is a skin cancer that develops in melanocytes, the skin pigment cells that produce melanin. Cancer-affected melanocytes can still form melanin, so melanoma usually appears brown or black. However, some melanocytes can no longer form melanin, causing melanoma to appear pink or white . Early symptoms of melanoma are marked by the appearance of new moles or changes in the shape, size, and color of existing moles, accompanied by itching and possible bleeding. Body parts that often become locations for melanoma are the face, hands, back, and feet . Melanoma is the most malignant type of skin cancer due to its rapid spread capability, even to internal organs. Additionally, early symptoms resembling ordinary moles often make melanoma difficult for patients to recognize or detect directly with the naked eye. A skin cancer diagnosis can be performed through direct examination of the patient's skin, laboratory tissue testing . , or skin sampling . Doctors typically perform biopsy examinations, but this procedure is time-consuming and invasive . Computer vision technology can assist in rapid and non-invasive melanoma detection. Convolutional Neural Network (CNN) is a deep learning algorithm utilized to address computer vision problems such as classification, segmentation, and detection . CNN enables computers to learn independently from data without explicit programming and extract features automatically . CNN has high accuracy because it consists of layered networks that allow it to identify hidden features in image data that are not directly visible to the human eye. CNN is a development of 258 ISSN: 1978-1520 multilayer perceptron specifically designed to process data with grid topology, such as twodimensional images . Several studies have utilized Convolutional Neural Networks (CNN. for melanoma skin cancer detection, each employing different architectures. In 2020. Jasman Pardede and Dwi Adi Lenggana Putra utilized the DenseNet121 architecture, which is known for its dense connectivity between layers, allowing for efficient feature extraction and reduced parameter numbers. However, one limitation of DenseNet121 is its computational complexity, which can lead to longer training times and higher resource consumption . In 2022. Reynaldi Rio Saputro. Apri Junaidi, and Wahyu Andi Saputra performed melanoma classification using a CNN-based Their model showed promising results but lacked a detailed exploration of hyperparameter optimization, which could potentially improve model performance . In 2023. Muhammad Faris Fahru Rozi. Sri Mulyono, and Ghufron employed the MobileNet v2 architecture for Android-based melanoma detection, focusing on mobile devices. MobileNet v2 is known for its efficiency in mobile environments, offering a lightweight model with reduced computational overhead. However, a limitation of this architecture is its lower accuracy compared to more complex models like DenseNet121, especially when dealing with small datasets or highly varied input images . These studies highlight the trade-offs between model complexity, computational efficiency, and accuracy in melanoma detection, which inform the approach taken in the present research. In this research, two CNN models with different architectures were developed to obtain the best model. In recent years. Convolutional Neural Networks (CNN. have been widely used for melanoma detection due to their ability to automatically extract complex features from images. This study compares two CNN models with different architectures to identify the most effective approach for melanoma classification. Both models were trained for 100 epochs, with the first model using an SGD optimizer and the second model employing an Adamax optimizer. The models were evaluated and tested on a dataset to select the best performer, based on performance metrics such as accuracy and loss. The model with the highest accuracy and lowest loss was implemented on the website. The website allows users to upload or capture images directly for melanoma detection, providing immediate prediction results and their associated probabilities. The development of the CNN model and the web interface was completed using Python. TensorFlow, and Streamlit. To ensure widespread accessibility. Streamlit Cloud was employed for the deployment process. While this approach demonstrates the effectiveness of CNN models for melanoma classification, it is important to acknowledge the limitations. One key limitation is the dependency on high-quality input images, as the modelAos performance can degrade with poor image quality or lighting conditions. Additionally, while the CNN model performs well on the dataset used, the generalization to more diverse and larger datasets remains a challenge. The current state of the art in melanoma detection has seen the use of more complex architectures such as ResNet. VGG, and DenseNet, which offer improved accuracy but at the cost of increased computational resources. The novelty of this work lies in the combination of a CNN-based detection model with an accessible, user-friendly web interface, making early melanoma detection faster and noninvasive, while also prioritizing efficiency for mobile and web deployment. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 RESEARCH METHODS Research Design This research follows a structured 5-stage process, illustrated in Figure 1. To provide better clarity, the stages have been reorganized into distinct sections. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Figure 1. Research Design Dataset and Preprocessing The first stage involves determining the object and classification method. This study focuses on comparing two Convolutional Neural Network (CNN) models for digital image-based melanoma detection. The data used in this research is secondary data obtained from the Kaggle platform, which consists of 17,805 melanoma and nonmelanoma images. The preprocessing stage standardizes pixel values and image sizes to ensure consistency across the dataset. This is achieved by rescaling and resizing images, which helps improve the model's learning efficiency. Each sub-dataset consists of two classes: melanoma class and not melanoma class, as shown in Fig. Figure 2. Sample Data A ACNN Model Architectures The second stage is the development of two distinct CNN architectures. The two CNN models are designed to compare different architectural approaches and determine the best model for melanoma classification. The first model, referred to as Model CNN1FC, consists of 4 convolutional layers, 4 max-pooling layers, 1 flattened layer, 1 dropout layer, and 1 fully connected layer. The second model, referred to as Model CNN-2FC, adds one more fully connected layer, resulting in a total of 2 fully connected layers. These models are built to vary in depth and complexity to assess their impact on classification The CNN architecture shown in Fig. 3 consists of two main parts: feature learning . r feature extractio. and classification. Feature learning is responsible for extracting features from input data, while classification functions classify input data into specific classes based on features learned by the model . The feature learning section consists of convolutional layers, pooling layers, and ReLU activation functions. Meanwhile, the classification section includes flattening layers, fully connected layers, and softmax activation functions . COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Figure 3. CNN Architecture The convolutional layer functions extract features from images through convolution operations, which are dot product multiplications between image matrices and kernels. The result of the convolution operation is a two-dimensional matrix called an activation map or feature map, which represents features extracted from the image . The pooling layer is a layer that performs downsampling operations or dimension reduction on the feature map. Its purpose is to speed up computation and reduce the risk of overfitting in the model. There are several pooling methods, including max pooling, which selects the largest value in a window, and average pooling, which calculates the average value in a window. The flattening layer functions transform feature maps from multidimensional matrices into one-dimensional vectors. This process is done so that image data can be processed in the fully connected layer . The fully connected layer or dense layer is typically located at the end of the CNN architecture, functioning to classify input data into specific classes based on features extracted by previous layers. The final result of the fully connected layer is a vector containing probabilities for each class in the classified image . Like the convolutional layer, the fully connected layer also performs dot product operations. The difference is that in the fully connected layer, each neuron is connected to all neurons in the previous layer . Softmax is an activation function commonly used in output layers. This function plays a role in calculating probabilities for multi-class classification by selecting the class with the highest probability value. The output produced by the softmax activation function is probability values ranging from 0 to 1 . A ATraining Configuration and Optimizer Choices The third stage involves training the two CNN models using the preprocessed dataset. The Model CNN-1FC is trained using the Stochastic Gradient Descent (SGD) optimizer, while the Model CNN-2FC uses the Adamax optimizer. Both models are trained for 100 The optimizer choices are made based on their suitability for image classification The training process ensures that the models can learn from the data and generalize well to new, unseen data. A AEvaluation Metrics In the fourth stage, the modelsAo performance is evaluated using various metrics, including accuracy, loss, precision, recall, and F1-score . These metrics are crucial for assessing the models' ability to classify melanoma images correctly. The evaluation process helps identify the best-performing model, which will be implemented for melanoma detection. A ASystem Deployment The final stages of the research involve deploying the selected CNN model onto a webbased platform. A website is developed to allow users to upload or capture images for melanoma detection. The deployment process is completed using Streamlit Cloud, ensuring that the platform is widely accessible to users. After developing the website, it undergoes local testing to ensure its functionality before being made publicly available. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 RESULT AND DISCUSSION Dataset and Data Collection In this research, a secondary melanoma dataset obtained from the Kaggle platform . ttps://w. com/datasets/drscarlat/melanom. is utilized for training, validation, and testing purposes. The dataset consists of 17,805 images, divided into three distinct subsets: training, validation, and testing. Each subset includes two classes: melanoma and non-melanoma. The training dataset contains 10,682 images, equally divided between melanoma . ,341 image. and non-melanoma . ,341 image. The validation dataset consists of 3,562 images, with 1,781 images for each class. The test dataset has 3,561 images, with 1,781 melanoma images and 1,780 non-melanoma images, as summarized in Table 2. This balanced division ensures that each model can be trained, validated, and tested on a representative sample, minimizing the risk of bias in the classification process. Table 2. Dataset Training Dataset Validation Dataset Testing Dataset Melanoma Class Total Not Melanoma Class Total Preprocessing Data Before training the models, the image data undergoes essential preprocessing steps to ensure consistency, uniformity, and improved model performance. Preprocessing helps standardize the input data, which is crucial for CNNs to perform effectively. The preprocessing steps include rescaling and resizing. Rescaling: This step standardizes pixel values across images to a fixed range, typically between 0 and 1. This is important because neural networks, especially CNNs, are sensitive to the scale of input data. Without rescaling, the network might struggle to converge during training, as it would require more time to adjust weights for inputs with varied Resizing: All images are resized to a uniform spatial dimension of 224x224 pixels. This ensures that the model receives input images of the same size, which is required for CNNs. resizing, we standardize the spatial dimensions of the images, preventing the model from being biased by images with different sizes or aspect ratios. These preprocessing steps are implemented using the ImageDataGenerator class from the TensorFlow library. ImageDataGenerator allows for real-time data augmentation and modification during the training process without altering the original data, enabling more efficient and scalable training. The . flow_from_directory() method is used for automatic data labeling, as the images are stored in subdirectories where each folder represents a different class . elanoma or non-melanom. The target_size parameter is set to . , . to resize images to the required dimensions, and class_mode is set to 'categorical' for one-hot encoding of class labels. The batch_size is set to 64, meaning that the model processes 64 images at a time. For the test dataset, the shuffle parameter is set to False to maintain the original order of images, ensuring consistency in model evaluation and prediction. Impact of Preprocessing: Improved Convergence: By rescaling the pixel values and resizing images to a fixed size, the model trains more efficiently. This reduces the risk of issues like vanishing or exploding gradients and speeds up convergence during the training process. Consistency Across the Dataset: Preprocessing ensures that all images are standardized in terms of pixel range and size, which is critical for the CNN to learn features effectively across all images. Enhanced Performance: Data preprocessing, especially rescaling and resizing, has been shown to improve the performance of CNN models in tasks like image classification. By providing a 262 ISSN: 1978-1520 uniform input format, the model can focus more effectively on learning meaningful patterns and features, rather than being distracted by irrelevant variations in the input data. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 Compared to other techniques, such as manual feature extraction or unmodified raw images, preprocessing significantly enhances model generalization and accuracy. Without preprocessing, the model would face difficulty in processing images with varying sizes and pixel values, leading to poor performance and slower training times. Developing Two Distinct CNN Architectures This research develops two CNN models with different architectures to obtain the best. Model CNN-1FC, consists of 4 convolutional layers, 4 max-pooling layers, 1 flattened layer, 1 dropout layer, and 1 fully connected layer, illustrated in Fig. Figure 4. The Architecture of the Model CNN-1FC Meanwhile. Model CNN-2FC, adds one more fully connected layer, resulting in a total of 2 fully connected layers. , as in Fig. Figure 5. The Architecture of ModelCNN-2FC The difference in the number of layers used in Model CNN-1FC and Model CNN-2FC affects the total features learned by both models. Model CNN-1FC learns 134,306 features in the data, while the Model CNN-2FC learns 1,277,282 features in the data, as shown in Fig. Figure 6. Model Summary 263 ISSN: 1978-1520 Training the Two CNN Models The next stage is training the two CNN models using the processed data. Before starting the training process, the models are compiled to configure training parameters, which include selecting optimization functions, loss functions, and model performance evaluation metrics. The optimization function plays a role in directing the model learning process toward optimal values. The loss function is used to measure the difference between values predicted by the models and actual values. Meanwhile, evaluation metrics are used to monitor and evaluate the models' performance during the training process. This research applies different optimization functions to both models: Stochastic Gradient Descent (SGD) for the first model and Adamax for the second model. SGD was chosen for its simplicity and common usage. Adamax was chosen for its flexible adaptation mechanism, which potentially accelerates and stabilizes the learning process. Both models use the categorical crossentropy loss function, suitable for classification problems with more than two classes, and accuracy metrics. To control the learning rate during training and address overfitting risks, the learning_rate_reduction variable is implemented using the ReduceLROnPlateau class from the Keras library. This class automatically adjusts the learning rate based on validation accuracy metric monitoring. If validation accuracy shows no improvement for two consecutive iterations, the learning rate will be reduced by 50% from its previous value. Both CNN models are trained for 100 epochs using 10,682 training data and 3,562 validation data that have been processed. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 Evaluating the Training Results of the Two CNN Models The evaluation of the two CNN models' training results, as presented in Table 3, is performed by analyzing accuracy and loss values on both the training and validation datasets. The learning curve is a graph showing changes in model accuracy and loss during the training process, consisting of accuracy graphs and loss graphs . Learning curves also function to identify whether the model is in an underfitting, good fit, or overfitting condition . In addition, learning curves are used to visually observe how accuracy and loss change as the number of training epochs This evaluation helps measure each modelAos ability to learn patterns in the data and its capacity to generalize to unseen data. While this study focuses on comparing two CNN models with simple architectures, it is important to acknowledge the potential benefits of experimenting with more complex, state-ofthe-art CNN architectures. Popular CNN models such as LeNet. AlexNet. VGG. GoogLeNet, and ResNet have been extensively used in image classification tasks, including melanoma detection, and could provide significant improvements in performance. These architectures typically offer deeper networks, better feature extraction capabilities, and improved generalization compared to simpler CNN models. Given their proven success in image classification, the use of these more advanced models could lead to more accurate and robust melanoma detection. Future work could explore the performance of these models in comparison to the simpler architectures used in this study. This would provide a more comprehensive evaluation of the CNN models' performance and could potentially enhance the results for melanoma detection. In this study, however, the focus was to compare two basic CNN models to establish a baseline performance for melanoma detection. More complex architectures could be explored in subsequent experiments to better understand their impact on model accuracy, training time, and computational efficiency. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Table 3. Training Results of Model CNN-1FC Epoch Training Accuracy Validation Accuracy Training Loss Validation Loss Figure 7. Learning Curve of The Model CNN-1FC The Model CNN-1FC training results as in Fig. 7 show the model can learn patterns in training data well and demonstrates good generalization ability toward validation data. At epoch 100, the model achieves a fairly high accuracy of 0. 8457 on training data and 0. 8450 on validation Additionally, the learning curve shows accuracy improvement in both training and validation data, as well as decreased loss values in both types of data. This pattern indicates that the model has stable performance in training and validation processes. Table 4. Training Results of The Model CNN-2FC Epoch Training Accuracy Validation Accuracy Training Loss Validation Loss COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Figure 8. Learning Curve of The Model CNN-2FC Model CNN-2FC in Fig. 8 also shows good model ability in learning patterns in training data and generalizing validation data well. At epoch 100, the model achieves a very high accuracy 9522 on training data and 0. 9410 on validation data. Additionally, the learning curve shows significant accuracy improvement in training and validation data, accompanied by a significant decrease in loss values. This pattern indicates that the model does not experience underfitting or overfitting problems. Therefore, in this research, no further adjustments to model hyperparameters were made. Testing the Two CNN Models The next stage is testing the two CNN models using 3,561 processed test data. This testing stage is conducted to evaluate the extent of the model's performance in predicting new data and representing their performance in real-world applications. Comparing the Two CNN Models' Performance Fig. 9 shows that the Model CNN-1FC successfully classified 1,504 melanoma data into the melanoma category and 1,436 non-melanoma data into the non-melanoma category correctly. However, the model misclassified 344 melanoma data into the non-melanoma category and 277 non-melanoma data into the melanoma category. Based on these classification results, other evaluation metrics were calculated. The precision value obtained for the melanoma class is 0. and for the non-melanoma class is 0. The recall value produced by the model is 0. 84 for the melanoma class and 0. 81 for the non-melanoma class. Furthermore, the calculated F1-score is 83 for the melanoma class and 0. 82 for the non-melanoma class. These results indicate that the first model performs reasonably well, though there is still room for improvement, particularly in reducing false positives and false negatives. Figure 9. Testing Results of the First CNN Model ISSN: 1978-1520 In contrast, the Model CNN-2FC achieved a better classification performance. successfully classified 1,607 melanoma data into the melanoma category and 1,711 nonmelanoma data into the non-melanoma category, as shown in Fig. However, it misclassified only 69 melanoma data as non-melanoma and 174 non-melanoma data as melanoma. Based on these classification results, other evaluation metrics were also calculated. The precision value for the melanoma class was 0. 96, and for the non-melanoma class was 0. The recall value for the melanoma class was 0. 90, and for the non-melanoma class, it was 0. The calculated F1-score for both classes was 0. These results indicate that the second model outperforms the first model in terms of both precision and recall, suggesting a better ability to generalize to new data. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 Figure 10. Testing Results of the Second CNN Model The improved performance of the second model can likely be attributed to two main Addition of an Extra Fully Connected Layer: The second model includes an additional fully connected layer compared to the first model. This deeper architecture allows the model to learn more complex patterns and interactions within the data, which can lead to improved classification accuracy. The increased number of parameters enables the model to capture more detailed features from the input images, resulting in a better ability to distinguish between melanoma and non-melanoma classes. Optimizer Choice (Adama. : The second model used the Adamax optimizer, which is a variant of the Adam optimizer that adapts learning rates based on the weight of each parameter. This optimizer can sometimes offer faster convergence and better generalization, especially in cases where the model has more complex architectures. In this case, the Adamax optimizer likely contributed to the better performance by effectively adjusting the learning rates during training, leading to improved accuracy and reduced loss. The following is a performance comparison chart of two CNN model architectures, namely CNN-1FC and CNN-2FC. This chart presents a visualization of the accuracy and loss of each model during the training process, aiming to illustrate the learning effectiveness and overall performance of both models. The blue and green lines represent accuracy . eft axi. : Model CNN2FC consistently shows higher accuracy compared to CNN-1FC. The red and dashed orange lines represent loss . ight axi. : Model CNN-2FC has lower loss values, indicating more effective Overall. Model CNN-2FC has proven to deliver better performance in terms of both accuracy and loss efficiency compared to CNN-1FC illustrated in Fig. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Figure 11. Accuracy and Loss Values of Two CNN Models Drawing Conclusions from the Comparison Results Based on a comprehensive evaluation comparing the two CNN models' performance, it can be concluded that both models have a good ability to predict new data. The second CNN model shows better performance compared to the first model, achieving 0. 9318 accuracies and 1636 loss. Based on these results, the second CNN model was chosen to be implemented in the The implementation process requires saving the model so that weight and bias parameters, which are important information for the model, can be maintained and reused. In this research, the model is saved in . h5 file format. Developing a Website And Implementing the Best CNN Model Before developing the website and implementing the best CNN model, a thorough planning phase is essential. This stage ensures that the system built will not only meet user needs but also maintain high reliability and efficiency. The website is designed to provide an intuitive user interface, enabling users to upload or capture images for melanoma detection. The model is implemented using the best-performing CNN architecture identified in the previous stages, ensuring optimal accuracy in predictions. To ensure smooth operation, the system was developed using Python. TensorFlow, and Streamlit. The deployment of the website was facilitated through Streamlit Cloud, allowing the model to be accessed and used on various devices. The focus of this development is on creating a user-friendly, efficient, and accessible platform for melanoma detection, leveraging the power of deep learning models while prioritizing ease of use. Figure 12. Activity Diagram ISSN: 1978-1520 In this research, system design is carried out using UML diagrams, specifically activity diagrams, which are used to model system activity flows systematically and clearly . The activity diagram in Fig. 12 illustrates interactions between users, the system, and other internal components to facilitate an understanding of image-based melanoma skin cancer detection system After accessing the website, users are given options to choose between an image upload feature or a direct image capture feature. The website allows users to either upload or capture images for melanoma detection. Users can upload an image from their device, and the system will validate the file format before processing it. Alternatively, users can take a picture using the websiteAos camera functionality, with appropriate permissions requested for camera Once an image is selected or captured, it undergoes preprocessing, feature extraction, and classification using the trained CNN model. Subsequently, on the same web page, the system will display the model's prediction results along with their probability values. In this research, the website is built with Python programming language with Streamlit framework. The website interface design is shown in Fig 13. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 Figure 13. Website Interface Design Testing the Website in a Local Environment After developing the website and implementing the CNN model on the web, the next stage is functionality verification in the local environment. This process aims to ensure the performance of all features meets specifications and fulfills previously established objectives before being used by the public. Deploying the Website The final stage is deploying the website so it can be accessed by users generally through an appropriate platform. This research utilizes Streamlit Cloud as the deployment medium. The website display on a mobile device is shown in Fig 14 and on a desktop device is shown in Fig COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Figure14. Website Application Display on Mobile Figure 15. Website Application Display on Desktop CONCLUSION This research demonstrates the potential of Convolutional Neural Networks (CNN. for digital image-based melanoma detection, with the second CNN model outperforming the first in terms of accuracy and loss. This model has been successfully integrated into a web-based platform, providing a user-friendly interface for melanoma detection through image upload or The platform, hosted on Streamlit Cloud, offers accessibility across various devices, making it a valuable tool for early melanoma detection. The results of this study highlight the promising role of deep learning in clinical settings, particularly for assisting with early skin cancer diagnosis. However, there are some limitations, including the reliance on high-quality images and the challenge of generalizing the model to diverse real-world datasets. Future research can focus on expanding the dataset to include more diverse images, improving model robustness, and integrating additional features such as saving prediction results and providing referrals to dermatologists or nearby healthcare facilities. Further investigation into combining CNNs with other diagnostic tools and exploring the modelAos performance in real-world clinical environments could enhance its clinical value and applicability in assisting healthcare professionals in melanoma diagnosis. REFERENCES