COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1 Hyperparameter Tuning Exploration to Maximize MobileNet Performance in Classification Kidney Tumor Sandy Putra Siregar*1. Safii2. Sundari Retno Andani3 Department of Informatics MagisterAos. STIKOM Tunas Bangsa. Pematangsiantar. Indonesia e-mail: *1sandy@amiktunasbangsa. id, 2m. safii@amiktunasbangsa. ra@amiktunasbangsa. 1,2,3 Abstract The main focus of this research is on developing the MobileNet architecture in order to produce a kidney tumor classification model that is accurate, resistant to overfitting, and remains consistent with variations in datasets and training parameters. This study aims to develop a MobileNet architecture to produce a software model that can precisely identify with high accuracy, perform kidney tumor classification, and avoid failure in generalizing new data, called overfitting, as well as evaluate the difference in accuracy generated from several variations of datasets and parameters. The method used in this study is MobileNet with hyperparameter tuning and fine-tuning, and it was compared with the MobileNet Baseline method. The dataset consists of 12,446 images classified as Normal. Cyst. Stone, and Tumor, collected from Kaggle. The findings of this study on the division of the 80:10:10 ratio of the proposed method image data resulted in 100% accuracy, 100% precision, 100% recall, and 100% F1-Score. This study is expected to produce architecture modifications that can classify kidney tumors with high accuracy, so that the hypothesis is achieved. In addition, various approaches in medical image analysis using deep learning have shown better results in identifying various tumors, especially in this research, in the classification and detection of kidney tumors. KeywordsAi Kidney Tumor. Classification. MobileNet. Hyperparameter. Deep Learning INTRODUCTION The kidney is a vital organ that plays an important role in maintaining the body's homeostatic balance through blood filtration, fluid and electrolyte regulation, and excretion of metabolic waste . One of the most important organs in the body, it is susceptible to acute damage in sepsis, known as Sepsis-Associated Acute Kidney Injury (SA-AKI), which significantly increases the risk of developing chronic kidney disease, cardiovascular disorders, and death . 2020 marked a significant challenge in kidney transplantation, with a decrease in the overall number of transplants due to a decline in transplants from living donors, although transplants from deceased donors increased. The COVID-19 pandemic has had a significant impact on the transplant process, including fewer new candidates, increased deaths on the waiting list, and a high rate of unutilized organs. In addition, living donor transplantation remains low in both adult and pediatric populations, with marked disparities in access by race . Tumors or neoplasms are one of the primary causes of morbidity . llness and healt. and mortality . eath from specific disease. among kidney post-transplant patients, with the total risk of cancer nearly double that of the general population. This study found that Kaposi's sarcoma and hematological cancers are the most common tumor types, particularly in the first three years following transplantation, and that the majority of cases are linked to the effects of post-transplant These findings highlight the significance of short and long-term oncologic monitoring in managing renal transplant patients . Kidney tumors are one of the most frequent diseases worldwide, and early detection is key to treatment success . , . Malignant tumors can damage the kidneys and spread to other organs, resulting in lethal malignancies. TAMs are immune cells that play a dual role. , . they play a vital role in tumor microenvironmental 448 ISSN: 1978-1520 dynamics by encouraging angiogenesis and metastasis while also altering immune responses that promote tumor growth. Recognizing chronic kidney disease (CKD) as a priority in national health policies is an important step toward improving renal healthcare globally. , given that CKD remains a significant health burden, with a consistent trend of rising incidence, mortality, and disability from 1990 to 2021, particularly in countries with low socio-demographic indices. There is a need to develop therapies to improve toxic protein clearance to maintain renal function by confirming that UMOD mutase-induced uromodulin aggregate formation causes endoplasmic reticulum stress and immunological reactions that directly contribute to progressive renal tubular damage in ADTKD. This method opens new avenues for targeted molecular therapies in hereditary kidney disease. , . FaslerAos study demonstrated that juvenile kidney tumor cells respond to diverse cytokines by activating the STAT and NF-kB pathways. This supports the in vitro model's potential for mapping cytokine signaling, which is pertinent to creating more effective immunotherapies. TAMs are an important target in cancer immunotherapy due to their overwhelming polarization towards the M2 phenotype. Clinical efforts to restrict TAM recruitment or convert them to the M1 phenotype are ongoing. This method is very promising despite significant obstacles in terms of efficacy and widespread adoption in therapy. AI-powered diagnostic devices are urgently required to detect early kidney illnesses such as stones, cysts, and malignancies. The AI model based on Vision Transformers and CNN is a possible technique for clinical integration to increase the accuracy of radiology diagnosis. demonstrates accurate and consistent segmentation performance in detecting kidney and renal tumors on CT-Scan scans, particularly for tiny lesions . Moreover, its performance remains consistent when evaluated using independent test data, indicating robust generalization. According to Abdelrahman and Viriri's study, deep learning in kidney tumor segmentation, specifically convolutional neural networks, has significantly improved diagnostic accuracy and detection effectiveness and is now an essential part of modern radiology and treatment . Applying the lightweight, quick, and efficient MobileNet architecture in skin picture categorization produces excellent results, with a 90% accuracy on the HAM1000 dataset. The model can handle class imbalance and increase generalization by altering the architecture and employing data augmentation techniques. These findings suggest that the MobileNet model, which informed the decision-making process in learning algorithms, is suitable for early skin disease detection systems, particularly under low-power conditions. The fundamental challenge is improving the model's performance through effective hyperparameter change. Andry Luiz's statistical hyperparameter tuning research, using analysis of variance and Tukey's test, discovered the optimal and optimized CNN architecture configuration for visual categorization of construction machinery. This technique gives a more in-depth understanding of parameter interactions and allows for the most accurate model configuration selection. Consequently, the recommended combination of DenseNet and Adagrad has the highest average accuracy in binary testing . 0%) and multiclass classification . 8%). Using modern image processing methods, transfer learning approaches, and ideally tuned hyperparameter optimization, this work accurately classifies renal CT images with high accuracy, allowing for better radiological diagnosis and early detection of kidney illnesses. In Amini's research, the hyperparameter tuning strategy improves the model development component, resulting in prediction performance that is 10%-25% greater than that of the traditional univariate The combination of hyperparameter optimization, ensemble approaches, and multivariate inputs shows that model parameter refinement is required to address the complexity of rainfall dynamics under changing climatic circumstancesAithe significance of autonomous hyperparameter tuning in increasing the accuracy of rainfall forecast models. Deep neural networks (DNN. performed significantly better after utilizing methods such as random search and the Parzen tree structure estimator. Parvathi and Jonnadula suggested a two-stage method for segmenting kidney tumors using a Custom U-Net and Mask R-CNN. This work offered accurate segmentation findings on COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1 the TCGA-KIRC dataset, with an accuracy value of 91% and an IoU of 91%. The model can identify complex tumor shapes, fluctuations in image lighting, and overlapping objects. Although segmentation algorithms such as U-Net and Mask R-CNN produce good spatial detail results . % accurac. MobileNet-based classification approaches with hyperparameter-tuning can increase accuracy even further with limited computational resources. Furthermore, the proposed CNN method for kidney tumor identification is versatile and At epoch-19, the model in this investigation achieved 94. 94% accuracy, 95. validation accuracy, and 0. 1270 validation loss . The key disadvantage of this model is that it does not employ pre-trained data, which puts the model's performance in danger of overfitting and can enhance accuracy. The technique or model can be developed by adjusting the architecture or hyperparameter-tuning to optimize the architecture further, resulting in higher accuracy for more accurate identification. Thus, the employment of approaches involving architecture change or hyperparameter tuning is urgently required to improve accuracy and efficiency in In Addition . Pimpalkar's novel renal CT image categorization method employs transfer learning-based deep learning models such as VGG16. ResNet50. AlexNet, and InceptionV3, adjusted meticulously through a structured hyperparameter configuration process. This strategy combines image augmentation techniques, watershed segmentation. Otsu thresholding, and relief method-based feature selection to detect better kidney diseases, such as cysts, stones, tumors, and normal kidneys. The results showed that the suggested model attained the greatest classification accuracy of 99. 96%, making it a high-performance solution in CT scan image-based automated diagnosis systems. It could help with early detection and more accurate clinical decision-making. Although this study obtained an accuracy of 99. 96%, greater configuration and optimization of the CNN architecture are required to avoid model overfitting and underfitting. Based on the aforementioned developments, findings show that deep learning-based methods in medical image analysis have demonstrated increased effectiveness in performing disease classification, including kidney tumors. One promising architecture is MobileNet, which provides lightweight and efficient computational performance. Classification requires a model that accurately identifies the presence of tumors and is lightweight enough to be widely implemented with hardware limitations. This research centers on developing a baseline MobileNet-based model into a proposed method for renal tumor classification with various dataset configurations and parameters, emphasizing achieving accurate results and strong generalization ability. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 RESEARCH METHODS This research begins with an initial data handling process used as the basis for developing deep learning models. After the data is acquired, a data preprocessing stage is performed, where the data is divided into two parts: training data, validation . nitial testin. , and testing data. The training process was conducted over 13 epochs, the length of time or number of iterations chosen to enable stable model learning while preventing excessive iterations that could cause overfitting in the deep learning architecture training workflow. The model is then built using the MobileNet CNN architecture as the baseline and compared with the optimized proposed method. In the final stage, the performance of each model is evaluated using the confusion matrix. The overall stages of this research can be seen in the methodology diagram presented in Figure 1. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Initial Data Handling Pre-Processing Data Train. Val Modelling MobileNet Baseline Proposed Method Data Test Training MobileNet Baseline Proposed Method Evaluasi Figure 1. Research Stage Initial Dataset Handling The dataset used in this study was obtained from the Kaggle CT KIDNEY DATASET: Normal-Cyst-Tumor and Stone . and consists of 12,446 wood images. The dataset is divided into four classes, namely: normal . ,077 image. , cyst . ,709 image. , stone . ,377 image. , and tumor . ,283 image. The data is divided into three main parts: training, validation, and testing. In this study, the dataset was split into training, validation, and test sets using a standard split. Although the class composition was not perfectly balanced, random data splitting helped ensure that each dataset split had a relatively similar class distribution. In this way, the model still obtained a picture that represented the actual data conditions. Since there were no changes to the class composition, the evaluation results could be considered to reflect the model's ability to process the class distribution in the original data, so the interpretation of accuracy and other metrics still reflected the model's performance against natural class variations. This composition aims to evaluate the performance of the model on various proportions of training and test data. representative sample of the dataset is illustrated in Figure 2. Normal COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1 Cyst . Stone . Tumor Figure 2. An illustrative subset of the dataset is shown to provide an overview: . Normal, . Cyst, . Stone, and . Tumor The data samples illustrated in Figure 2, which are provided in JPG format, are used for model training, validation, and testing. The JPG format was chosen due to its storage efficiency and compatibility with various deep learning architectures. Architecture Model Model architecture in deep learning refers to a layered structure of computational units designed to learn a representation of data progressively. Each layer has a specific role in extracting and transforming features from the input, resulting in a goal-driven final output, such as classification, segmentation, or prediction. The suggested approach uses a hyperparameter-tuned version of the MobileNet architecture with an input image resolution of 224y224 pixels. accordance with conventional techniques for training on large-scale datasets, a batch size of 32 was chosen. Figure 3 shows a structural comparison between the proposed technique and the baseline MobileNet model. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Input Feature Output MobileNet Baseline Sotfmax Loss Center Loss . Model Baseline Input Feature Output BatchNormalization Dropout . Dense . ReLu. L2, 1e-. BatchNormalization Dense . ReLu. L2, 1e-. Proposed Method Sotfmax Loss Center Loss . Proposed Method The Convolutional Layer The Average Pool Layer The Depthwise Separable Convolution Layer Architectural Component Development The Channel_Attention The Fully Connected Layer The Spatial_Attention Figure 3. Model Architecture . MobileNet Baseline, . Proposed Method Figure 3 presents the core components of the improved convolutional neural network The model comprises several essential building blocks, beginning with convolutional layers that capture low-level spatial features from the input image. To improve computational efficiency, the architecture adopts depthwise separable convolutions, which separate spatial filtering from channel-wise processingAisignificantly reducing the number of parameters while maintaining high performance. Furthermore, the model integrates channel attention and spatial attention mechanisms to enhance its ability to focus on the most relevant information across both feature dimensions. The extracted features are then condensed through average pooling and subsequently passed to the fully connected layers for final classification. The hyperparameters utilized in the model configuration are detailed in Table 1. Table 1. Hyperparameter Tuning Values Comparative Value Activation Regularization Dense ReLu BatchNormalization Dropout Dense ReLu BatchNormalization Layer ISSN: 1978-1 Table 1 above shows the hyperparameter tuning values for the architecture added to the baseline MobileNet model, called the proposed method. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 Model Evaluation In this evaluation stage, the trained MobileNet model was tested using separate datasets consisting of three kidney tumour disease data classes and one class of normal kidney. The testing process involves making predictions and evaluating them using a confusion matrix to determine the actual classification performance of the model. The model's performance was assessed using four primary metrics: accuracy, precision, recall, and F1 Score. Accuracy measures a model's classification skill by comparing the proportion of correctly predicted examples to the total number of samples. Precision is the number of correct positives divided by the total number of anticipated positives. Recall . lso known as sensitivit. refers to the proportion of actual positive cases properly detected by the model. Finally, the F1 score indicates the harmonic mean of precision and recall, making it a useful metric when dealing with unbalanced data, particularly when false positives and false negatives are equally relevant. RESULT AND DISCUSSION The dataset used in this study was collected from Kaggle (CT KIDNEY DATASET: NormalCyst-Tumor and Ston. and included 12,446 photos of kidney tumours and normals. The collection is organized into four categories: normal . ,077 photo. , cyst . ,709 image. , stone . ,377 image. , and tumour . ,283 image. The data was separated into three main sections: training, validation, and testing. Table 2 illustrates the data-splitting strategy. The image to be trained has numerous size variations, including 705x569 pixels, 512x512 pixels, 804x651 pixels, 652x528 pixels, and 798x646 pixels. Model comparison is configured with Adam's optimizer, input size 224, epoch 13, and batch size 32. Table 2 displays the data framework that will be used to train the baseline MobileNet model and the proposed method. Table 2. Dataset Splitting Framework Splitting Data Class Name Training Data Validation Data Testing Data Normal 80-10-10 Cyst Stone Tumor Normal 70-15-15 Cyst Stone Tumor Normal 60-20-20 Cyst Stone Tumor Table 2 provides information on training the baseline MobileNet model and the proposed model using datasets divided 80-10-10 . % train data, 10% valid data, and 10% test dat. , 7015-15, and 60-20-20. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 MobileNet Baseline This architecture utilizes depthwise separable convolution to significantly reduce computational complexity compared to standard convolution, while maintaining competitive accuracy on benchmark datasets such as ImageNet. This baseline MobileNet model was trained using dataset splits of 80-10-10 . % train data, 10% valid data, and 10% test dat. , 70-15-15, and 60-20-20. Model Baseline Dataset 80-10-10 . Model Baseline Dataset 70-15-15 . Model Baseline Dataset 60-20-20 Figure 4. The training process shows the performance of the model . Dataset 80-10-10, . Dataset 70-15-15, and . Dataset 60-20-20 As depicted in Figure 4, the training process was closely monitored by observing the accuracy and loss graphs. The figure presents the outcomes of the MobileNet Baseline without fine-tuning and hyperparameter tuning. The resulting training curve, shown in Figure 4 . , illustrates the model's performance over time. The outcomes indicate a training accuracy of 33% and a corresponding loss value of 0. Figure 4 . The outcomes indicate a training 455 ISSN: 1978-1 accuracy of 89. 67% and a corresponding loss value of 0. Figure 4 . The outcomes indicate a training accuracy of 88. 36% and a corresponding loss value of 0. Figure 5 shows a detailed performance evaluation of the model using a confusion matrix. This matrix provides granular insight into categorization results by showing the distribution of true positives, false positives, true negatives, and false negatives. It identifies distinct misclassification patterns and illustrates the model's ability to discriminate across classes. The results support the model's reliability in practical classification tasks, particularly when assessed alongside precision, recall, and F1-score metrics. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 Metrics Dataset 80-10-10 . Metrics Dataset 70-15-15 . Metrics Dataset 60-20-20 COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Figure 5. Confusion Metrics Derived from Dataset Evaluation . Dataset 80-10-10, . Dataset 70-15-15, and . Dataset 60-20-20 Figure 5 MobileNet Baseline Confusion Matrix . illustrates the classification outcomes across four classes. Within the Cyst category, the model accurately identified 347 instances, while 24 were misclassified. For the Normal class, a total of 493 images were correctly predicted, and 15 were incorrectly classified. In the case of the Stone category, the model achieved 94 correct predictions and 44 misclassifications. Lastly, for the Tumor class, 179 instances were correctly classified, whereas 50 samples were incorrectly labeled. For the confusion matrix with dataset . 70-15-15, there are 527 correct cyst images and 30 incorrect images, 727 correct normal images and 35 incorrect images, 156 correct stone images and 51 incorrect images, and 286 correct tumor images and 77 incorrect images. Dataset . 60-20-20, there are 664 correct cyst images and 78 incorrect images, 952 correct normal images and 64 incorrect images, 187 correct stone images and 81 incorrect images, and 390 correct tumor images and 67 incorrect images. Proposed Method This design uses depthwise separable convolution to drastically reduce computational costs compared to traditional convolution, while maintaining comparable accuracy on test datasets such as ImageNet. The output of the proposed approach is illustrated through modifications to the CNN architecture, specifically by replacing the MobileNet Baseline, and adjusting several hyperparameters such as dropout application, regularization techniques, batch normalization, and other configurations. The base MobileNet model was trained on datasets split into 80-10-10, 70-15-15, and 60-20-20. The outcomes of the training phase. Figure 6 below: Model with Dataset 80-10-10 . Model with 70-15-15 COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1 Model with Dataset 60-20-20 Figure 6. The Training Process Shows the Performance of the Proposed Model . Dataset 80-10-10, . Dataset 70-15-15, and . Dataset 60-20-20 As shown in Figure 6, the training process is carried out systematically, emphasizing accuracy and loss graphs. This training provides the results of the proposed method. The resulting training curve, shown in Figure 6 . , illustrates the model's performance over time. The outcomes indicate a training accuracy of 100% and a corresponding loss value of 0. The outcomes indicate a training accuracy of 99,57% and a corresponding loss value of 0. The outcomes indicate a training accuracy of 99,04% and a corresponding loss value of 0. Figure 7 depicts the performance evaluation of the model using a confusion matrix. Can be seen in Figure 7 below: Proposed Method Dataset 80-10-10 Proposed Method Dataset Variation 70-15-15 COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 ISSN: 1978-1520 Proposed Method Dataset Variation 60-20-20 Figure 7. Confusion Metrics Derived from Dataset Evaluation . Dataset 80-10-10, . Dataset 70-15-15, and . Dataset 60-20-20 Figure 7 presents the confusion matrix generated from the proposed method, offering a comprehensive view of the classification results across four distinct categories. Within the Cyst category, the model accurately identified 272 instances, while 0 were misclassified. For the Normal class, a total of 509 images were correctly predicted, and 0 were incorrectly classified. the case of the Stone category, the model achieved 139 correct predictions and 0 Lastly, for the Tumor class, 229 instances were correctly classified, whereas 0 samples were incorrectly labeled. For the confusion matrix with dataset . 70-15-15, there are 555 correct cyst images and 2 incorrect images, 763 correct normal images and 0 incorrect images, 208 correct stone images and 0 incorrect images, and 337 correct tumor images and 6 incorrect Confusion matrix with dataset . 60-20-20, there are 731 correct cyst images and 12 incorrect images, 1015 correct normal images and 1 incorrect image, 276 correct stone images and 0 incorrect images, and 447 correct tumor images and 11 incorrect images. Empirical Evaluation Results The results confirm the capability of the MobileNet-based approach in managing intricate datasets while effectively learning meaningful feature representations. Using depthwise separable convolutions. MobileNet substantially minimizes computational complexity and parameter size while maintaining a high level of accuracy. MobileNet outperforms other picture classification algorithms, despite its small size. The evaluation, as shown in Table 3, comprises performance measures such as precision, recall. F1-score, and total test correctness. Table 3. Dataset Splitting Framework Model MobileNet Baseline Proposed Method Variation Dataset Precision Recall F1Score Accuracy Set 80-10-10 84,25% 86,50% 89,33% Set 70-15-15 89,25% 85,75% 87,25% 89,67% Set 60-20-20 84,75% 88,04% Set 80-10-10 Set 70-15-15 99,50% 99,50% 99,75% 99,57% Set 60-20-20 98,25% 98,50% 99,04% Table 3 shows, based on the classification results. MobileNet Baseline follows with the accuracy result on the 80-10-10 dataset reaching 89,67%, the 70-15-15 dataset reaching 89,33%, and the accuracy of the dataset reaching 88,04. Meanwhile, the proposed method achieved the 459 ISSN: 1978-1 highest performance, reaching 100% in terms of precision, recall. F1-Score, and accuracy beyond all other models in accuracy identifying samples on the 80-10-10 dataset. These findings collectively highlight the superiority of the proposed method in image classification tasks, confirming the effectiveness of the study. The following Figure 8 comparison of classification results can be seen as a comparison chart. COGITO Smart Journal Ae Vol. No. December 2025. P-ISSN: 2541-2221. E-ISSN: 2477-8079 Comparison Model Variation Dataset Set 80-10-10 Set 70-15-15 Set 60-20-20 Set 80-10-10 Set 70-15-15 Set 60-20-20 MobileNet Baseline Precision Proposed Method Recall F1-Score Accuracy Figure 8. Chart Dataset Evaluation . Dataset 80-10-10, . Dataset 70-15-15, and . Dataset 60-20-20 CONCLUSION This study emphasizes the significance of proper hyperparameter tuning in a convolutional neural network (CNN) architecture, particularly in diagnosing kidney tumor A comparison was conducted between the MobileNet Baseline model and the improved Proposed Method using dataset splits of 80-10-10, 70-15-15, and 60-20-20. The Proposed Method outperformed the others with the 80-10-10 dataset, achieving 100% precision, recall. F1-score, and accuracy, thereby demonstrating high reliability and accuracy in managing the challenges inherent to the classification process. Experimental results reveal that the baseline MobileNet architecture provides efficient and computationally lightweight classification performance but still falls short of obtaining ideal accuracy on complex datasets. In contrast, the proposed strategy, which applies hyperparameter tuning to MobileNet, successfully enhances the model's overall This gain shows that parameter optimization is crucial for improving the model's generalization capacity in image classification tasks. This paper makes a substantial contribution to medical imaging by offering a robust and high-performance CNN-based classification framework specifically customized to effectively identify and categorize the visual characteristics of kidney cancer. Subsequent studies may examine the model's applicability across different imaging modalities and associated diagnostic problems and incorporate interpretability elements to further its clinical integration potential. ACKNOWLEDGMENTS Thank you very much for the knowledge and guidance provided by the STIKOM Tunas Bangsa Informatics lecturers in completing this paper. REFERENCES