a Jurnal Pendidikan Teknologi Informasi dan Komunikasi JENTIK JENTIK https://ejournal. com/index. php/jentik Predicting Student Depression Using the Naive Bayes Model on the Student Depression Dataset from Kaggle Rebina Putri Sonjaya1*. Andre Rangga Gintara2. Lala Septem Riza3. Eki Nugraha5. Didin Wahyudin6 1,2,3,4,5,6 Article Information: Received June 15, 2025 Revised June 22, 2025 Accepted June 23, 2025 Keywords: Dataset Kaggle. Depression Detection. Machine Learning. Nayve Bayes. Student Mental Health Muhammad Nursalman4. Universitas Pendidikan Indonesia Bandung. Indonesia rebinaputris@upi. Abstract Background of Study: The increasing prevalence of depression among college students highlights the urgent need for effective early detection strategies to promote mental well-being within higher education environments. Aims and Scope of Paper: This study aims to develop a predictive model for student depression using the Naive Bayes classification algorithm, with a focus on identifying key contributing factors from student-related data. Methods: The research utilizes the Student Depression dataset from Kaggle, containing structured survey data on academic stress, sleep duration, financial stress. GPA, and family mental health history. Data preprocessing included feature selection, handling of missing values, and normalization. The dataset was split into training and testing sets at a 75:25 ratio. Model training was conducted using the R programming language with the application of Laplace smoothing. Result: The Naive Bayes model achieved an accuracy of 77. a specificity of 84. 21%, and a sensitivity of 68. 42%, indicating strong predictive performance, particularly in identifying depressive cases. Financial and academic stress were identified as the most influential factors. Conclusion: Despite its simplicity, the Naive Bayes algorithm proves to be an effective tool for initial screening of students at risk of depression, offering valuable support for educational institutions in delivering timely mental health interventions. Introduction Today, global concern over the high prevalence of depression among college students is increasing. report from the WHO shows that almost all psychiatrists experienced in dealing with college students agree that about 10 % of members of higher education institutions are likely to experience an emotional disturbance each year, which seriously hinders their academic or work activities. Previous research has also stated that depression has an impact on a large number of college students (Lorentzen et al. , 2. (Liu et , 2. , and the prevalence is higher compared to the general population (Campbell et al. , 2. (Adlaf et al. , 2. Early detection of depression is essential for timely intervention, appropriate support, and reducing the serious impacts that can arise (Moon et al. , 2. (Windarwati et al. , 2. (Sayed et al. However, many college students tend to delay seeking psychological help until their depressive symptoms reach moderate to severe levels. This is often due to low mental health literacy and the perception that their problems are not serious enough to require professional help (Kustimah et al. , 2. To address these challenges, the use of technology-based self-detection tools is becoming increasingly important so that universities can identify at-risk students and proactively intervene. Advances in machine How to Cite ISSN Published by : Sonjaya. Gintara. Riza. Nursalman. Nugraha. , & Wahyudin. Predicting Student Depression Using the Naive Bayes Model on the Student Depression Dataset from Kaggle. JENTIK : Jurnal Pendidikan Teknologi Informasi Dan Komunikasi, 4. https://doi. org/10. 58723/jentik. : 2963-1963 : CV Media Inti Teknologi ABDIGERMAS JENTIK Sonjaya. Gintara. Riza. Nursalman. Nugraha. , & Wahyudin. learning (ML) have opened up great opportunities in developing data-driven prediction systems, which are able to process complex relationships between variables with high accuracy (Lee et al. , 2. (Hatton et , 2. (Shatte et al. , 2. (Gil et al. , 2. A number of ML algorithms have been used in depression prediction, including Support Vector Machine. Decision Tree. Neural Network. Latent Dirichlet Allocation, and Clustering (Shatte et al. , 2. However, most research is still limited to the use of complex models that are often less transparent and difficult to One of the main challenges in depression detection research is limited access to relevant and structured In this study, the Student Depression dataset from Kaggle was used, which contained survey data related to symptoms of depression, stress, anxiety, as well as demographic factors and student lifestyle. This structured, survey-based dataset allows for clearer feature interpretation and reduces the need for complex preprocessing required in social mediaAebased approaches. Although there have been many studies of machine learningAebased depression predictions, most of them focus on the general population or use data from social media that require complex text extraction. This kind of structured survey-based approach to students is still relatively rare, even though it can be more easily implemented by educational institutions. In this context, the Naive Bayes Classifier algorithm offers a simple yet effective alternative, assuming conditional independence between features making it suitable for limited datasets and quick predictions (Cruz et al. , 2. Its computational efficiency, ease of parameter estimation, and interpretability make it ideal for deployment by non-technical users such as campus counsellors (Stiglic et al. , 2. Although often considered a baseline. Naive Bayes shows competitive performance, especially on structured survey datasets: for example, (Cruz et al. , 2. 03 % accuracy in predicting depression among 519 university students, and (Haque et al. , 2. reported up to 85 % accuracy using survey-derived features. Therefore, this study leverages Naive Bayes for its practicality and transparency, aiming to develop a depression prediction model in college students using actual survey datasets from Kaggle. With a focus on early detection, method efficiency, interpretability, and practical application in educational settings, the study is expected to contribute to efforts to prevent and treat depression in higher education environments. Research Methods Figure 1. Computing Model The computational model in this study is divided into 5 processes: data collection, data pre-processing, modeling, model improvement and final results. A diagram of the computational model can be seen in Figure 1. Jurnal Pendidikan Teknologi Informasi dan Komunikasi Vol 4. No 1, pp 9-20 2025 ABDIGERMAS JENTIK Predicting Student Depression Using the Naive Bayes Model on the Student Depression Dataset from Kaggle Data Collection Data was collected from the Kaggle platform with a dataset titled Student Depression, which contains various features related to students' mental states. Furthermore, feature validation is performed by an expert . n this context it can be a psychology studen. to ensure only the relevant features are retained. The expert conducted validation based on psychological relevance, such as whether the feature reflected emotional, behavioral, or cognitive indicators commonly associated with depressive symptoms. If there is a feature that is considered invalid or irrelevant, it is removed before the data is saved in CSV format. This expertdriven feature screening helps maintain the clinical validity of the dataset prior to preprocessing and modeling (Shatte et al. , 2. Data Preprocessing This stage aims to prepare the dataset for effective model training by ensuring data consistency, completeness, and proper formatting. The steps involved are as follows: Loading Dat The dataset Student_Depression_Dataset. csv read into R using the read. csv() function. The stringsAsFactors = FALSE argument is used to ensure that text fields are not automatically converted into factors, giving more control at a later stage. df <- read. csv("Student_Depression_Dataset. csv", stringsAsFactors = FALSE) Column Name Cleanup (Sub step 3a is code. The column names in the dataset are cleaned using the clean_names() function of the janitor package. This process changes the name of the column to a standard format . , all lowercase, spaces replaced with an underscore _), such as academic_pressure, work_study_hours, etc. This makes it easy to write code and be df <- clean_names. Column Selection (Sub-step 3b in the cod. Based on the initial instructions, the relevant columns are selected: Categorical Predictor Features: academic_pressure, sleep_duration, degree, work_study_hours, financial_stress, family_history_of_mental_illness. Features of Numerical Predictor: cgpa. Identity columns . aved but not for model. : gender, age. Variabel Target: depression. Other columns that are not included in this list are ignored for the modeling process. The R code checks to make sure all the required fields are in the dataset. Handling Missing Values (Sub-step 3c in the cod. Iterations are carried out on each column that has been selected. If a column is character-typed, the value of an empty string ("") or string "na" . o matter the case cas. is converted to a standard NA value of R. Tabel 1. Example of Missing Value Handling Raw Entry (Befor. Processed Value (Afte. "na" "NA" After that, all rows containing at least one NA value in the selected columns are removed from the dataset using na. omit(). The code also checks if there is any data left after this process. Jurnal Pendidikan Teknologi Informasi dan Komunikasi Vol 4. No 1, pp 9-20 2025 ABDIGERMAS JENTIK Sonjaya. Gintara. Riza. Nursalman. Nugraha. , & Wahyudin. Data Transformation . d sub-steps in cod. Conversion of Categorical Columns to Factors: Columns that are categorical . ender, academic_pressure, sleep_duration, degree, work_study_hours, financial_stress, family_history_of_mental_illness, and depression as target. are converted to data factor types in R. Prior to conversion, text values are cleaned of spaces at the beginning/end . rimws()) and converted to uppercase . oupper()) to ensure consistency . , "Yes", "yes", "YES" are treated the sam. The code checks whether the depression target variable has at least two levels required for classification. Tabel 2. Example of Categorical Value Transformation Original Transformed " less than 5 hours " LESS THAN 5 HOURS "yes" YES CGPA Scale Conversion: The cgpa column that originally had a scale of 0-10 is converted to a scale of 04 to match the GPA standard. First, it is ensured that the cgpa column is numerical type. If there is a comma as a decimal separator, it is converted to a period. Rows with invalid cgpa values . an't be changed to numerica. will be deleted. The numerical cgpa value is then divided by 2. Clipping is performed, which is the cgpa value that after conversion exceeds 4 will be set to 4, and those less than 0 will be set to 0. This cgpa column is retained as a numerical feature for the Nayve Bayes model. Tabel 3. Example of CGPA Scale Conversion . Ae10 to 0Ae. Original CGPA Converted . Ae4 scal. Preparation of Final Data for Modeling: A new dataframe . odeling_d. is created that contains only the final predictor features . odel_features_final which includes categorical features and numerical cgp. and target depression variables. All categorical features and target variables are ensured to be factor-type, and factor levels that have no observations are removed using droplevels(). Data Splitting Datasets that have gone through pre-processing . odeling_d. are then divided into two subsets: Training Set: Used to "teach" or train the Nayve Bayes model. Test Set: Used to test how well a trained model can make predictions on new data that it has never The partitioning process is done using the sample. split() function from the caTools package. The split ratio is set to 75% for the training data and 25% for the test data (SplitRatio = 0. This 75:25 division is a commonly accepted standard in machine learning experiments and offers a good balance between training efficiency and testing reliability (Shin et al. , 2. Stratification based on the target variable . is applied to ensure that both subsets maintain the same class distribution, which is especially important when working with potentially imbalanced data. This helps prevent biased model evaluation and improves generalization to unseen data. is used before partitioning to ensure that the random process in the data selection is reproducible, so that if the code is re-run, the result of the data partition will remain the same. This practice ensures the experiment is reliable and can be repeated or audited consistently by other researchers (Berrar. Checks were carried out to ensure that the training data and test data were not empty and both still had representations of both target classes . This step is essential to confirm that no class is excluded from either subset, which could otherwise invalidate the model training or testing process. Jurnal Pendidikan Teknologi Informasi dan Komunikasi Vol 4. No 1, pp 9-20 2025 ABDIGERMAS JENTIK Predicting Student Depression Using the Naive Bayes Model on the Student Depression Dataset from Kaggle Building the Nayve Bayes Model The Nayve Bayes model is constructed using the training data . rain_dat. as follows: This section not only outlines the technical construction of the model but also includes critical enhancements such as Laplace smoothing and robust error handling with tryCatch, which improve stability and generalizability in practical applications. Model Formula: The formula used is depression . This means that we want to predict the depression variable using all the other variables present in the train_data as predictor features. The formula depression . instructs the model to predict the depression target using all other variables. Model Training: The naiveBayes() function of the e1071 package is used to train the model. A data = train_data: Specifies the data used for training. A laplace = 1: Apply Laplace smoothing with parameter 1. This is a technique to address the zero probability problem that can occur if there is a combination of features and classes that never appear in the training data, thus making the model more stable. A The numerically typed cgpa feature will be automatically handled by the naiveBayes function assuming that its values are Gaussian-distributed . for each target class. The model will calculate the average and standard deviation of the cgpa for each class from the training data. TryCatch blocks are used to handle potential errors during the model training process and provide a warning message if there is a factor feature with only one unique level, which would make it unsuitable for training. The results of the trained model are stored in naive_bayes_model object and then printed to the console to display the model summary. formula_nb <- depression . naive_bayes_model <- NULL tryCatch({ naive_bayes_model <- naiveBayes. ormula = formula_nb, data = train_data, laplace = . }, error = function. { print. aste("Error Naive Bayes:", e$messag. ) for. ol_name in setdiff. rain_dat. , "depression")){ if. rain_data[. ol_nam. ]) && nlevels. rain_data[. ol_nam. ]) < 2 && nrow. rain_dat. > . { print. aste("Peringatan: Fitur faktor", col_name, "hanya memiliki 1 level unik. ")) aive_bayes_mode. ) stop("Gagal melatih model Naive Bayes. print("Output Model Naive Bayes:") print. aive_bayes_mode. Laplace smoothing is particularly important in classification problems with small datasets or imbalanced class distributions. In practice, it avoids assigning a zero probability to any category simply because it does not appear in the training data. For example, if a certain category of a predictor variable is present in the test data but was not seen during training for a particular class, the model without Laplace smoothing would assign zero probability to the entire prediction. By adding a small value . n this case, . to all feature counts. Laplace smoothing ensures that every possible feature-class combination receives a non-zero probability, making the model more reliable and generalizable. Jurnal Pendidikan Teknologi Informasi dan Komunikasi Vol 4. No 1, pp 9-20 2025 ABDIGERMAS JENTIK Sonjaya. Gintara. Riza. Nursalman. Nugraha. , & Wahyudin. In addition, the use of the tryCatch block ensures that errors during model training . uch as the presence of factor variables with only one unique leve. do not cause the entire process to fail. Instead, the code explicitly checks each factor predictor and prints a warning if it is found to be unsuitable for training. This proactive handling of potential data quality issues enhances model stability and provides clear feedback to the developer, facilitating debugging and dataset refinement. This approach ensures both model robustness and interpretability, even in the presence of imperfect or imbalanced data. Making Predictions Once the model has been successfully trained, it is used to make predictions on the test data . est_dat. The predict() function is used with naive_bayes_model and newdata = test_data to generate a depression class prediction for each observation in the test data. The prediction results are stored in the predictions predictions <- predict. aive_bayes_model, newdata = test_dat. Model Evaluation The performance of the model was evaluated by comparing the predictions with the actual depression values of the test data: Factor Level Adjustment: To ensure consistency during evaluation, the factor level of the actual target variable . est_data$depressio. and the prediction variable is equalized using all_possible_levels taken from train_data$depression. This is important for confusionMatrix() to function correctly. This step is especially crucial in R, where mismatched factor levels between prediction and reference values can cause errors or misalignment in evaluation metrics. Handling of NA on Factors . f an. : Checks are carried out to see if any NA values appear after adjustment of the factor level. If there are, rows with those NA values will be removed from the evaluation, although this should not occur if the data is already clean. This check helps prevent misleading metrics due to invalid rows being included in the evaluation. Confusion Matrix: The confusionMatrix() function from the caret package is used to create a contingency table . onfusion matri. that compares the modelAos predictions with the actual values. This function also accounts for unbalanced classes by providing balanced accuracy, a robust metric in such contexts. Performance Metrics: The output of confusionMatrix includes various important metrics namely Accuracy, 95% CI (Confidence Interva. for Accuracy. No Information Rate (NIR). P-Value [Acc > NIR]. Kappa. Sensitivity (Recal. Specificity. Post Pred Value (Precisio. Neg Pred Value. Balanced Accuracy: The average of Sensitivity and Specificity. The accuracy of the model is specifically extracted and printed to the console. These comprehensive steps for evaluating classification performance help ensure the reliability and validity of the model in detecting depression cases accuratel. Results and Discussion Results The Nayve Bayes model has been trained using 75% of the data from Student_Depression_Dataset. datasets that have gone through the pre-processing stage. Features used in modeling include academic_pressure, sleep_duration, work_study_hours, financial_stress, family_history_of_mental_illness . s a categorical featur. , and cgpa . s a numerical feature after scaling conversion to 0-. Laplace smoothing with a value of 1 is applied during model training to improve Summary of the Nayve Bayes Model A-priori probability of Target Class . Jurnal Pendidikan Teknologi Informasi dan Komunikasi Vol 4. No 1, pp 9-20 2025 ABDIGERMAS JENTIK Predicting Student Depression Using the Naive Bayes Model on the Student Depression Dataset from Kaggle Figure 2. A priori probability. Class '0' (Non-Depresse. : 0. Class '1' (Depressio. : 0. 55%) This indicates that in the training data, there are slightly more cases classified as 'Depression' . lass '1'). Conditional Probability of Features: The model calculates the probability of each category/value of a feature appearing, given a specific depression state . lass '0' or '1'). Some of the important findings of conditional probability are: A Academic_Pressure A Figure 2. Probability academic_pressure. Higher levels of academic stress . , level '5') are more commonly associated with class '1' (Depressio. with a probability of 0. 332, compared to class '0' (Non-Depressio. which has a probability of 0. 078 for the same level. In contrast, lower levels of academic stress . , level '1') were more often associated with class '0' (P=0. than class '1' (P=0. Sleep_Duration A Figure 3. Probability sleep_duration The sleep duration of "LESS THAN 5 HOURS" was more often found in class '1' (Depression. P=0. than in class '0' (Non-Depression. P=0. Financial_Stress Figure 4. Probability financian_stress A Financial stress at the highest level . , level '5') was significantly more common in class '1' (Depression. P=0. than in class '0' (Non-Depressive. P=0. In contrast, the lowest level financial stress . , level '1') was more common in class '0' (P=0. than class '1' (P=0. Family_History_of_Mental_Illness A Figure 5. Probability family_history_of_mental_illness The presence of a family history of mental illness (YES) is slightly more often observed in class '1' (Depression. P=0. than in class '0' (Non-Depression. P=0. CGPA Jurnal Pendidikan Teknologi Informasi dan Komunikasi Vol 4. No 1, pp 9-20 2025 ABDIGERMAS JENTIK Sonjaya. Gintara. Riza. Nursalman. Nugraha. , & Wahyudin. Figure 6. CGPA Probability The average GPA . cale 0-. for class '0' (Non-Depressio. tandard deviation . , while for class '1' (Depressio. it was 3. tandard deviation 0. The difference in average GPA between these two groups is relatively small. Evaluation Results on Test Data . rom print. onfusion_matrix_result. ) Confusion Matrix: Figure 7. Confusion Matrix Results (Assuming '0' = Not Depression, '1' = Depressio. A True Negative (TN): 1978 A False Positive (FP): 645 A False Negative (FN): 913 A True Positive (TP): 3439 Overall Model Accuracy: 77. Figure 8. Model Accuracy Results A 95% Confidence Interval for Accuracy: . 7667, 0. A No Information Rate (NIR): 0. 5855 (Model accuracy well above NIR) A P-Value [Acc > NIR]: < 2. 2e-16 (Statistically significant model accurac. Statistics Kappa Figure 9. Kappa Statistics Results A 0. 5335 (Indicates moderate agreemen. Performance Metrics Per Class . ith '0' as the Positive Class for this metri. A A A A A Figure 10. Performance metrics per class Sensitivity (Recall for class '0' - No Depressio. : 0. Specificity (Recall for class '1' - Depressio. : 0. Post Pred Value (Precision for class '0' - Not Depresse. : 0. Neg Pred Value (Precision for class '1' - Depressio. : 0. Balanced Accuracy: 0. Discussion The Nayve Bayes model developed to predict the depression status of students based on the Student_Depression_Dataset. csv showed a fairly good performance with an overall accuracy of 77. 66% on the test data. This accuracy is significantly higher than the random guessing rate (No Information Rate 5. , indicating that the model is successfully capturing relevant patterns in the data. Jurnal Pendidikan Teknologi Informasi dan Komunikasi Vol 4. No 1, pp 9-20 2025 ABDIGERMAS JENTIK Predicting Student Depression Using the Naive Bayes Model on the Student Depression Dataset from Kaggle From the conditional probability analysis generated by the model, several factors show a stronger association with depressive status: Financial Stress and Academic Stress: These two variables demonstrate the most notable differences between depressed and non-depressed students. Higher levels of both stressors are more frequently found in students predicted as depressed. This supports findings from prior research, such as (Jin, 2. , who also used this Kaggle dataset . ith a Decision Tree mode. , and from (Danahy et al. , who highlighted financial burdens like debt and limited access to social activities as stress Sleep Duration: Sleep deprivation . ess than 5 hour. showed a strong correlation with depression, aligning with findings by (Alqurashi et al. , 2. and (Dinis & Braganya, 2. who noted that both insufficient and excessive sleep are linked to poor mental health. Family History of Mental Illness: While not as strong as stress variables, this factor still contributes to increased likelihood of depression, in line with the known genetic and environmental risks of GPA: After normalization to a 0Ae4 scale. GPA did not show a significant predictive difference between depressed and non-depressed students, indicating it may be less useful than psychosocial or lifestyle indicators. The confusion matrix revealed that the modelAos Specificity . 21%) is higher than its Sensitivity . 42%), suggesting it is more effective in identifying true positive depression cases than avoiding false positives. This is suitable for early screening purposes, where missing a case of depression is more harmful than a false alarm. Similar to (Haque et al. , 2. , model performance can vary depending on the dataset's structure and the modelAos clinical objectives. A Kappa value of 0. 5335 indicates a moderate degree of agreement between the model's prediction and the actual condition, indicating that the model provides better results than a chance deal. 1 Implications Nayve Bayes model provides a useful basis for understanding the factors associated with depression in college students in the context of the datasets used. While its accuracy is not perfect, it can be an early tool in identifying students who may be at risk and need further attention. The limitations of the Nayve Bayes model, such as the assumption of independence between features, may limit its ability to capture more complex interactions between variables. Future research may explore other machine learning algorithms or ensemble techniques for the potential for improved accuracy and deeper understanding. 2 Research Contribution This study contributes to the growing body of research on mental health prediction among college students by validating the utility of a simple, interpretable algorithm Nayve Bayes on a structured survey dataset. Unlike many machine learning studies that rely on unstructured social media text or black-box models, this research emphasizes practical implementation using interpretable features. By showing that Nayve Bayes can achieve reasonable accuracy, particularly in detecting true depression cases, the study supports its potential as a lightweight and low-resource screening tool in academic institutions. 3 Limitations Although the Nayve Bayes algorithm performed adequately in this study, it has inherent limitations. Its core assumption of feature independence may oversimplify the complex interrelationships that often exist between psychological, social, and behavioral variables. Additionally, the dataset used was cross-sectional and self-reported, which could introduce bias or limit the generalizability of findings over time and across different cultural contexts. The lack of temporal tracking also prevents the model from distinguishing between short-term emotional responses and more persistent depressive symptoms. Another limitation lies in the exclusive use of structured categorical and numeric survey data, which may not capture the nuances of students' mental health experiences, unlike multimodal data such as sensor data, diaries, or open 4 Suggestions Future research is encouraged to address these limitations by incorporating more complex or hybrid machine learning models that can learn interactions and hierarchies among variablesAisuch as Random Forests. Gradient Boosting Machines, or Deep Learning approaches (Agarwal et al. , 2. (Rony et al. Jurnal Pendidikan Teknologi Informasi dan Komunikasi Vol 4. No 1, pp 9-20 2025 ABDIGERMAS JENTIK Sonjaya. Gintara. Riza. Nursalman. Nugraha. , & Wahyudin. Incorporating Explainable AI (XAI) may also increase transparency and trust in automated predictions, especially in educational and clinical settings (Atlam et al. , 2. Researchers should consider utilizing longitudinal data to monitor depressive symptoms over time, which would allow the development of predictive models that are sensitive to early warning signs (Chikersal et , 2. Lastly, ethical considerations such as student privacy, informed consent, and cultural adaptation of AI models must be taken into account to ensure the modelAos responsible deployment in different educational environments, including Indonesia (Saeidnia et al. , 2. (Alavi et al. , 2. Conclusion This study aims to develop and evaluate a depression prediction model in college students using the Nayve Bayes algorithm by utilizing Student_Depression_Dataset. csv dataset. After going through a series of stages including pre-processing of data, model training, and evaluation, several important conclusions can be drawn: Nayve Bayes Model Performance: The developed model showed quite good performance, with an accuracy of 77. 66%, a specificity of 84. 21%, and a sensitivity of 68. A Kappa value of 0. indicates a moderate degree of agreement between the prediction and the actual conditions, indicating that the model performs better than random predictions. Factors Associated with Depression: Based on the conditional probability analysis of the Nayve Bayes model, several factors show a significant association with the depressive status of college students are Financial stress and academic stress being the dominant factors that appear most often. Short sleep duration . ess than 5 hour. is also an important indicator that leads to a depressive Family history of mental illness contributed, although not as strong as the previous two GPA (CGPA) shows a non-significant difference between the depressed and nondepressed groups, so its contribution as a differentiator is relatively low. Practical Implications: These findings confirm that although the Nayve Bayes algorithm is relatively simple, it can be used as an initial tool to detect the risk of depression in college students. With an understanding of significant factors such as financial stress, academic stress, and sleep duration, educational institutions can develop more targeted and preventive mental intervention and support strategies. Overall, this study confirms the potential use of machine learning, particularly Nayve Bayes algorithms, in an effort to understand and predict depression among college students, which can be the basis for the development of more effective prevention and intervention strategies in higher education settings. Acknowledgment The authors would like to express their sincere gratitude to the Department of Computer Science Education. Universitas Pendidikan Indonesia, for providing the academic support and research facilities used in this Special thanks are extended to the supervising lecturersAiLSR. MN. EN, and DWAifor their invaluable guidance, encouragement, and constructive feedback throughout the research process. The authors also acknowledge the contributions of colleagues and classmates who provided insight and support during the preparation of this manuscript. Author Contribution Statement RPS and ARG contributed to data collection, preprocessing, and initial model implementation. LSR. MN, EN, and DW provided supervision, guidance on research design, and validation of results. All authors contributed to the discussion of results, manuscript revision, and approved the final version of the References