Jurnal Kecerdasan Buatan dan Teknologi Informasi Vol. No. May 2025, pp. e-ISSN: 2964-2922, p-ISSN: 2963-6191 DOI : https://doi. org/10. 69916/jkbti. MACHINE LEARNING-BASED CLASSIFICATION OF SPACE TRAVEL ELIGIBILITY USING SUPPORT VECTOR MACHINE. RANDOM FOREST. AND XGBOOST Teguh Rizali Zahroni1. Bahtiar Imran2. Muhammad Tahrir*3. Muh. Akshar4. Zahrotul IstiAoanah Marroh5 1,3,4,5 Rekayasa Kayu. Politeknik Pertanian Negeri Samarinda. Indonesia Rekayasa Sistem Komputer. Fakultas Teknologi Informasi dan Komunikasi. Universitas Teknologi Mataram. Indonesia Email: 1teguhrizalizahroni@gmail. com, 2bahtiarimranlombok@gmail. com, 3mtahrir26@politanisamarinda. akshar@gmail. com, 5hamzahzhie@gmail. (Received: April 12, 2025. Revised: April 30, 2025. Accepted: May 5, 2. Abstract This study applies machine learning classification techniques to predict passenger displacement events based on corrupted data retrieved from a hypothetical interstellar spacecraft mission. Using a cleaned and preprocessed dataset containing demographic, behavioral, and exposure-related features, we compare the performance of three classification models: Random Forest. Support Vector Machine (SVM), and XGBoost. Each model is trained on 80% of the data and evaluated on the remaining 20% using training accuracy, cross-validation accuracy, and test accuracy metrics. After feature selection. SVM shows significant improvement, with test accuracy increasing from 97% to 72. 40%, and cross-validation accuracy improving from 68. 12% to 71. Random Forest maintains consistent performance with test accuracy improving from 79. 53% to 80. 05%, and cross-validation accuracy at XGBoost achieves the most stable results, with test accuracy rising from 79. 76% to 79. 93% and cross-validation accuracy from 79. 15% to 79. Feature importance analysis further enhances model interpretability, particularly in ensemble-based methods. The comparative analysis demonstrates that Random Forest and XGBoost are more effective in handling high-dimensional, partially incomplete datasets, making them suitable for complex predictive tasks in uncertain data environments. Keywords: classification models, interdimensional displacement, machine learning, passenger classification, space anomaly prediction. INTRODUCTION With advancements in space exploration and theoretical physics, the prospect of interstellar travel has been a subject of extensive scientific inquiry. By the 30th century, as envisioned by futurists and astrophysicists, humanity may have developed the capability to explore and potentially inhabit exoplanets beyond the solar system. One of the conceptualized interstellar missions is the launch of Spaceship. , a theoretical passenger liner designed to transport nearly 13,000 emigrants to three identified exoplanets within habitable zones. A potential challenge in deep-space travel involves the interaction between spacecraft and unexplored astrophysical phenomena. Hypothetically, while traversing near Alpha Centauri en route to 55 Cancri E, the vessel might encounter an unanticipated spacetime anomalyAiakin to those postulated in general relativity and quantum gravity modelsAiconcealed within an interstellar dust cloud. Such an event could theoretically lead to a highenergy perturbation affecting the fabric of spacetime, resulting in an unforeseen displacement of nearly half of the passengers into an alternate dimensional framework, if such dimensions exist as suggested by certain interpretations of string theory. This study explores a machine learning-based approach to analyzing passenger data and identifying factors that may correlate with this hypothesized phenomenon. By leveraging classification models within data mining, we aim to assess the feasibility of predicting interdimensional transport events, drawing parallels to anomaly detection techniques in contemporary aerospace engineering and astrophysical research. Identifying which passengers were affected by this incident poses a significant challenge due to the limited and corrupted data retrieved from the shipAos damaged computer system. The complexity of interdimensional displacement necessitates a robust analytical approach, making machine learning classification models a viable solution. Data mining techniques can extract meaningful patterns from the available passenger records, enabling the prediction of which individuals were transported by the anomaly. Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost Previous studies have demonstrated the effectiveness of machine learning algorithms in predicting complex phenomena based on historical data. Research on predictive classification models has been widely applied in various fields, including medical diagnostics, financial risk assessment, and space mission anomaly detection. In. , machine learning models such as Random Forest and Support Vector Machines (SVM) were successfully used to classify space mission failures with high accuracy. Similarly, in. Optimized boosting algorithms, such as XGBoost, were utilized to investigate the influence of cosmic radiation and solar activity on climatic patterns, yielding high-performance predictive results. Several studies have explored predictive modeling in extreme environments. In . Artificial intelligence (AI)based classification approaches were utilized in space medicine applications, though several challenges persist in ensuring their effectiveness under spaceflight conditions. , with neural networks demonstrating superior accuracy compared to traditional statistical approaches. In another study . , deep learning techniques were applied to classify gravitational wave anomalies. The effectiveness of machine learning in analyzing complex space-related datasets highlights its potential application in predicting interdimensional displacement events. Despite the numerous classification models applied in previous research, no studies have specifically addressed the classification of passengers transported to alternate dimensions due to spacetime anomalies. Therefore, this study aims to classify affected passengers using data mining techniques, focusing on Random Forest. Support Vector Machines (SVM), and XGBoost models. These models were selected due to their strong predictive performance in handling high-dimensional data . , their efficiency in decision-making, and their robustness in classification tasks . The dataset used in this study consists of passenger records extracted from the Spaceship TitanicAos damaged system, including features such as demographic details, cabin information, onboard activities, and anomalous energy exposure readings. The dataset is split into 80% training data and 20% testing data to evaluate model performance. The results of this study are expected to contribute to the development of predictive frameworks for space missions and enhance safety protocols for future interstellar travel. METHODOLOGY The methodology used in this research is illustrated in figure 1. Dataset Pre-processing Dataset Splitting Training and Testing Support Vector Machine (SVM) Random Forest XGBoost Evaluation Feature Selection Figure 1. Research Flow Dataset The dataset for this study was sourced from a dataset-sharing platform. Kaggle. com, accessible via the link: https://w. com/competitions/spaceship-titanic. The data is provided in CSV format and consists of 8,693 records with 14 attributes. An example of the dataset is presented in Figure 2, while the attributes utilized in this study are detailed in Table 1. Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost Table 1. The Attribute of Dataset Number Attribute PassengerId HomePlanet CryoSleep Cabin Destination Age VIP RoomService FoodCourt ShoppingMall Spa VRDeck Name Transported Figure 2. Dataset in CSV Format Pre-processing In the pre-processing stage, the initial steps of data handling are performed, including selecting important features, cleaning null or missing data, and removing redundant data. This process is illustrated in Figure 3. Figure 3. Dataset processed with JupyterLab Dataset Splitting In the dateset splitting stage, the dataset is split into two sets: the training data and the testing data, with a proportion of 80% for training and 20% for testing. The goal is to prevent bias in predictions, meaning that the model should not perform well only on the training data but also on the actual test data. Specifically, for classification Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost algorithms, the stratified shuffle split technique is used to ensure that both datasets maintain the distribution of the two categories. The splitting result obtained 6,954 rows for the training data and 1,739 rows for the testing data. the target column, the predicted "Transported" data is assigned the number 1, while the non-transported data is assigned the number 0. Training and Testing During the training and testing phase, the process is conducted using three algorithms: Support Vector Machine (SVM). Random Forest, and XGBoost. Prior to training and testing, feature scaling is applied using the Min-Max Scaler method. Additionally, hyperparameter tuning is performed simultaneously to optimize the parameters for each The tuning method used in this study is RandomSearchCV. Furthermore, the training process incorporates K-Fold Cross-Validation to prevent the model's performance from being influenced by luck or bias in the test data selection. Feature Selection In this stage, data quality and model performance were enhanced by identifying and selecting the most relevant features using a feature importance technique. Rather than relying on traditional methods such as Mean Loss Decrease, this study adopted the Mean Score Decrease approach, which is more versatile and applicable across different classification models. The Mean Decrease in Accuracy (MDA) method, commonly used for evaluating feature importance in machine learning models, follows a structured procedure to assess the contribution of each feature to model performance. First, the values of each feature . r colum. are individually shuffled to disrupt any existing relationship with the target variable, effectively removing the feature's informational value. Second, the model is re-evaluated using the modified dataset, keeping all other parameters constant. Finally, the drop in performanceAitypically measured by the decrease in accuracy or other relevant metricsAiis calculated. A greater decrease in accuracy indicates a more important feature. This method provides an intuitive and model-agnostic approach to understanding which features most significantly influence predictive performance. A feature that, when randomized, leads to a significant drop in performance is considered important. Conversely, features that cause little to no decrease in performance can be considered redundant or uninformative. These unimportant features were removed from the dataset to reduce noise, simplify the model, and improve generalization. This data-driven selection process not only helped improve accuracy but also enhanced model interpretability and reduced computational complexity during training and tuning phases. Evaluation The performance of the classification models was evaluated using numerical analysis methods named classification report includes four key performance metrics: Precision. Recall. F1-score, and Support. Each of these metrics offers different insights into the modelAos performance: Precision: Measures the proportion of positive identifications that were actually correct. ycNycycyce ycEycuycycnycycnycyceyc ycEycyceycaycnycycnycuycu = ycNycycyce ycEycuycycnycycnycyceyc yaycaycoycyce ycEycuycycnycycnycyceyc A high precision score indicates that when the model predicts a passenger will be transported, it is usually Recall: Measures the proportion of actual positives that were correctly identified. ycNycycyce ycEycuycycnycycnycyceyc ycIyceycaycaycoyco = ycNycycyce ycEycuycycnycycnycyceyc yaycaycoycyce ycAyceyciycaycycnycyceyc High recall means that most of the transported passengers were successfully identified by the model. F1-Score: The harmonic mean of precision and recall. ya1 Oe ycIycaycuycyce = 2 y ycEycyceycaycnycycnycuycu y ycIyceycaycayco ycEycyceycaycnycycnycuycu ycIyceycaycaycoyco This score balances precision and recall and is particularly useful when the classes are imbalanced. Support: Indicates the number of actual occurrences of each class in the test dataset. While not a performance metric, it provides context for interpreting the above scores. Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost The combination of visual plots and numerical scores enabled a comprehensive assessment of model It also helped identify any bias toward a particular class and informed further model adjustments where necessary. RESULTS AND DISCUSSION Data Testing Before Feature Selection To visually interpret the learning process and generalization behavior of the Random Forest model, a graphical approach was employed. Figure 4 illustrates the concept of Bootstrap Aggregating (Baggin. used by the Random Forest, in which each individual decision tree is trained on a different bootstrapped subset of the training data. The output of each tree contributes to the final prediction through majority voting . or classificatio. or averaging . or This ensemble approach reduces variance and prevents overfitting by averaging out the noise of individual Figure 4. Bagging Illustration The training process for the Random Forest classifier began with a hyperparameter tuning stage using the RandomizedSearchCV method combined with 5-fold cross-validation. In this process, the model was evaluated using a defined search space consisting of parameters such as the number of estimators . _estimator. , tree depth . ax_dept. , the fraction of features to consider at each split . ax_feature. , and the minimum number of samples required at a leaf node . in_samples_lea. A total of 10 candidate parameter combinations were randomly sampled from this space and evaluated across 50 training iterations. From the tuning process, the best-performing parameter combination was found to be 133 trees . _estimators = . , a maximum depth of 47 . ax_depth = . , a feature usage fraction of approximately 0. ax_features = 4. , and a minimum of 11 samples required per leaf . in_samples_leaf = . This configuration was then used to train the model using the full training set and evaluated against the testing set. The resulting performance metrics demonstrate a consistent and balanced outcome. The model achieved an accuracy of 0. 8263 on the training dataset, while the average cross-validation accuracy across the 5 folds was 0. When evaluated on the testing data, the model reached an accuracy of 0. These scores suggest that the model generalizes well, as evidenced by the relatively small difference between training and testing accuracy. Furthermore, the high cross-validation score supports the robustness of the selected hyperparameters and indicates that the model is not overfitting the training data. Figure 5. Learning Curve for Random Forest In figure 5, the blue line represents the training accuracy, which starts below 0. 78 and increases before stabilizing. The green line represents the test accuracy, which initially decreases from 0. 85 before stabilizing around 0. Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost The increasing training accuracy followed by stabilization suggests the model is learning and improving as more data is used. However, the slight decline in test accuracy indicates that the model might be starting to overfit, as it performs slightly better on the training data than on the test data. The stabilization of both accuracies suggests that the model has reached an optimal level of complexity. The training process for the Support Vector Machine (SVM) classifier was carried out using hyperparameter tuning via RandomizedSearchCV, with a 5-fold cross-validation approach. The hyperparameters optimized in this process were the regularization parameter C and the kernel coefficient gamma, both searched within a logarithmic scale ranging from 10AA to 10A, defined using a log-uniform distribution. This setting allows the search algorithm to explore a wide range of values, which is crucial for models like SVM that are sensitive to the choice of these After conducting the randomized search, the best-performing combination of hyperparameters was found to be C = 0. 201 and gamma = 0. With this configuration, the model achieved an accuracy of 0. 5037 on the training data, a cross-validation mean score of 0. 6812 across the five folds, and a test set accuracy of 0. The results indicate that while the SVM model reached a relatively good cross-validation score, its performance on the training and especially testing datasets was suboptimal. This significant drop in accuracy on the test set suggests that the model may not have generalized well and might have been affected by the complexity of the data or by the limitations in how the current feature set aligns with the SVM decision boundaries. Additional exploration, such as feature engineering or kernel adjustments, may be necessary to improve SVM performance in future iterations. Figure 6. SVM Decision Boundary with Margin Figure 6 illustrates the decision boundary . produced by the SVM classifier in a two-dimensional feature space. The solid black line represents the optimal hyperplane that separates the two classes. The dashed lines indicate the margins on either side of the hyperplane, which define the boundary of the maximum margin region. The data points shown in different colors correspond to the two classes, while the support vectorsAicritical points that lie on the margin boundariesAiare highlighted with larger, unfilled markers. This visualization helps to demonstrate how the SVM identifies the hyperplane that maximizes the margin between the classes, ensuring good generalization to unseen data. The XGBoost classifier was trained using hyperparameter tuning with RandomizedSearchCV and 5-fold crossvalidation. The best parameters obtained were: max_depth = 5, learning_rate = 0. 0266, n_estimators = 101, subsample = 0. 586, colsample_bytree = 0. 958, gamma = 2, reg_alpha = 1. 13, and reg_lambda = 0. Figure 7. XGBoost Model Performance Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost With this configuration, which is shown in figure 7. the model achieved a training accuracy of 0. 8041, a mean cross-validation score of 0. 7915, and a test accuracy of 0. These results indicate consistent and stable performance across datasets, suggesting that the model generalizes well and is not significantly overfitting. Feature Selection Using Mean Score Decrease (MSD) As part of the model improvement strategy through improvement by data, a feature selection process was conducted to identify the most significant predictors contributing to model performance. The method applied in this study is Mean Score Decrease (MSD), which was chosen for its versatility and model-agnostic nature, making it suitable for various classification algorithms used in this research, including Random Forest. SVM, and XGBoost. MSD operates by systematically shuffling the values of each feature in the dataset while keeping all other features The trained model is then re-evaluated on this modified dataset, and the corresponding drop in performance . , accuracy or another relevant metri. is recorded. If the removal of a particular feature results in a significant decrease in model performance, the feature is deemed important. Conversely, if the perturbation of a feature does not substantially affect the modelAos accuracy, the feature is considered less influential or potentially The magnitude of this performance drop represents the mean score decrease and serves as the basis for ranking feature importance. The results of the MSD analysis in this study revealed that the VIP feature had the lowest mean score decrease among all features. This indicates that shuffling the values of VIP did not significantly impact the modelAos accuracy, suggesting that it contributes minimally to the classification task. Based on this finding, the decision was made to remove the VIP feature from the dataset in subsequent training stages. This refinement aims to simplify the model, reduce the risk of overfitting, and enhance generalization by eliminating non-informative features. Figure 8 is a graph illustrating the mean score decrease for all analyzed features: Figure 8. Mean Score Decrease Diagram Data Testing After Feature Selection After removing the VIP feature as a result of feature importance analysis, the three machine learning modelsAi Random Forest. Support Vector Machine (SVM), and XGBoostAiwere re-evaluated to assess the impact of this selection on model performance. The comparative results reveal varied but generally positive changes across the For the Random Forest model, the selected hyperparameters after tuning were max_depth = 23, max_features OO 592, min_samples_leaf = 20, and n_estimators = 190. The training accuracy experienced a slight decline from 8263 to 0. 8181, while cross-validation accuracy remained nearly unchanged . 7967 before, 0. 7964 afte. Interestingly, test accuracy improved modestly from 0. 7953 to 0. 8005, suggesting better generalization and robustness when the less informative VIP feature was removed. The SVM model showed the most substantial improvement across all metrics. With tuned hyperparameters C OO 0033 and gamma OO 0. 0046, training accuracy increased significantly from 0. 5037 to 0. Cross-validation and test accuracy also rose markedly, from 0. 6812 to 0. 7141 and 0. 4997 to 0. 7240, respectively. These results imply that the presence of the VIP feature may have introduced noise or variance that the SVM model struggled to interpret. Its removal allowed the model to learn a more meaningful decision boundary. In the case of XGBoost, the post-selection tuning yielded max_depth = 10, n_estimators = 162, and other refined parameters, including learning rate and regularization factors. Performance improved slightly but consistently, with training accuracy moving from 0. 8041 to 0. 8139, cross-validation from 0. 7915 to 0. 7931, and test accuracy from Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost 7976 to 0. While the increase was modest, the consistency across all metrics indicates a minor but positive effect of removing the VIP feature. In summary, the feature selection processAiguided by the mean score decrease methodAihelped improve model performance, especially for SVM, by eliminating a non-contributive feature. These findings reinforce the importance of careful feature evaluation in machine learning workflows to avoid overfitting and enhance generalization A comparative analysis between the model performance before and after feature selection can be observed in Table 2. This table highlights the key evaluation metricsAitraining accuracy, cross-validation accuracy, and test accuracyAifor each algorithm (Random Forest. SVM, and XGBoos. , clearly illustrating the impact of removing the VIP feature on model performance. Table 2. Comparation of model Performance befor and after Feature Selection Model Condition Train Accuracy CV Accuracy Test Accuracy Random Forest Before Feature Selection After Feature Selection SVM Before Feature Selection After Feature Selection XGBoost Before Feature Selection After Feature Selection Evaluation After Feature Selection The Random Forest model performs consistently well, achieving 82% accuracy on the training data and 80% on the test data. Precision, recall, and F1-scores are balanced across both classes, indicating good generalization with no signs of overfitting. The classification performance can be visually observed in Figure 9. Figure 9. Classification Report for Random Forest The Support Vector Machine (SVM) model shows moderate performance, with 73% accuracy on the training set and 72% on the test set. The model appears slightly biased toward the "False" class, achieving higher recall in that category, while struggling more with the "True" class. This indicates a potential imbalance in class prediction. Overall, the SVM performs reasonably well, but not as balanced as the Random Forest model. The classification performance is illustrated in Figure 10. Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost Figure 10. Classification Report for Support Vector Machine The XGBoost model demonstrates consistent and well-balanced performance, achieving 81% accuracy on the training set and 80% on the test set. Precision and recall values are relatively even across both classes, suggesting that the model handles both "True" and "False" categories effectively. This balance indicates good generalization and reliable classification capability. The detailed results are presented in Figure 11. Figure 11. Classification Report for XGBoost Discussion The results obtained from the classification experiments show that feature selection has a notable impact on model performance, especially in the case of the Support Vector Machine (SVM). After removing the VIP featureAi identified as the least important based on the Mean Score Decrease (MSD) methodAithe SVM model experienced a considerable increase in accuracy, from 0. 50 to 0. 72 on the test data. This improvement suggests that the VIP feature was contributing noise or irrelevant information, thereby hindering the model's ability to generalize. In contrast, both Random Forest and XGBoost models showed relatively stable performance before and after feature selection, with only marginal differences in their accuracy scores. Random Forest achieved the highest test accuracy after feature selection at 0. 80, while XGBoost closely followed with 0. These results suggest that both ensemble models are robust and less sensitive to the inclusion or exclusion of the VIP feature, likely due to their inherent ability to handle feature importance internally through techniques such as bootstrapping and regularization. Among the three classifiers evaluatedAiSVM. Random Forest, and XGBoostAithe Random Forest model emerges as the best-performing model after feature selection. It not only achieved the highest test accuracy . but also maintained a strong balance between precision and recall across both classes. Additionally, its macro and weighted F1-scores were consistently higher compared to the other models. These findings indicate that Random Forest offers the most reliable performance in this classification task, effectively handling both bias and variance, and generalizing well to unseen data. Zahroni, et. machine learning-based classification of space travel eligibility using support vector machine, random forest, and xgboost Therefore, while SVM showed the most improvement post-feature selection. Random Forest remains the most optimal model overall based on its consistent and superior performance across all evaluation metrics. The comparison of model accuracy after feature selection is shown in Figure 12. Figure 12. Comparison of Model Accuracy after Feature Seleection CONCLUSION Based on the overall analysis, it can be concluded that feature selection has a noticeable impact on model The removal of the VIP feature led to improvements in classification accuracy, particularly for the SVM model, which showed the most significant gain. Among all evaluated models, the Random Forest classifier achieved the highest accuracy with a value of 0. 8005 on the test dataset, making it the best-performing model in this study. The XGBoost model followed closely with an accuracy of 0. 7993, while the SVM model showed the lowest accuracy of These results demonstrate that careful feature selection, combined with appropriate model tuning, can significantly enhance predictive performance in classification tasks. The Random Forest model stands out as the most reliable classifier for this dataset, delivering superior accuracy compared to the other algorithms evaluated. DAFTAR PUSTAKA