METHOMIKA: Jurnal Manajemen Informatika & Komputerisasi Akuntansi Vol. 9 No. 1 (April 2. ISSN: 2598-8565 . edia ceta. ISSN: 2620-4339 . edia onlin. LEAF-TYPE IMAGE CLASSIFICATION USING DEEP LEARNING METHOD CONVOLUTION NEURAL NETWORK Mikael Reichi Sopany*. Teny Handhayani Fakultas Teknologi Informasi. Universitas Tarumanagara. Jakarta. Indonesia Email: mikael. 535210053@stu. DOI: https://doi. org/10. 46880/jmika. Vol9No1. ABSTRACT One of the most important parts of an ecosystem is plants. Plant life has given us many benefits, including food, oxygen, and medicine. There are many species of plants, each with its unique benefits and utilities. In this paper, we aim to identify plants by their leaves using deep learning. For this research, we utilized the convolutional neural network architecture Xception to classify five different types of leaves. We used 1,075 images of leaves that can be classified into five different types. The classification model achieved an overall accuracy score of 74%. hope that the results of our research will help people's lives by enabling them to identify the plants they have so that they can use them for their benefit. Keyword: Convolutional Neural Network. Classification. Deep Learning. Supervised Learning. INTRODUCTION Plant life is a fundamental part of the history of our planet. Plants have provided resources, created habitats for animals, and influenced the climate on a global scale (Morris, et al. , 2. There are approximately 374,000 accepted species of plants, each with its own characteristics (Christenhusz & Byng. However, identifying these species remains a significant challenge. Studies show that up to 10Ae20% of plant species have yet to be scientifically classified, while many known species still lack sufficient identification data (Pimm & Joppa, 2. Additionally, the Global Strategy for Plant Conservation (GSPC) has highlighted that a significant percentage of plant species are at risk of extinction due to habitat loss, underscoring the urgency of developing efficient identification tools. Because there are so many species of plants, identifying them is very difficult. That is why we need a tool to help us identify them so that we can use them to their fullest potential. There has been a growing interest in classifying and identifying plants in recent years, notably through the examination of leaf traits. For botanical studies, crop management, and environmental monitoring, leaves are an invaluable source of data and play a key role in plant taxonomy. The ability to precisely categorize leaf types has significant implications for various industries, such as agriculture, forestry, and biodiversity preservation. Despite the advancements in botany, traditional plant identification methods are often labor-intensive, require expert knowledge, and can be prone to human error. There has been a growing interest in classifying and identifying plants in recent years, notably through the examination of leaf traits. For botanical studies, crop management, and environmental monitoring, leaves are an invaluable source of data and play a key role in plant taxonomy. The ability to precisely categorize leaf types has significant implications for various industries, such as agriculture, forestry, and biodiversity preservation. The introduction of deep learning has significantly improved image classification tasks, notably with Convolutional Neural Networks (CNN. (Yu, 2. CNNs have demonstrated remarkable performance in many visual identification tasks, including object detection, facial recognition, and image classification. Researchers have successfully classified images of leaves using CNNs, paving the way for promising automated plant identification Several studies have compared different CNN architectures for plant classification. For example, (Liu. Yang. Cheng, & Song, 2. , constructed a ten-layer CNN for plant leaf classification and applied sample augmentation to improve accuracy. Their experimental results on the Flavia dataset, which consists of 4,800 leaf images across 32 species, showed that their proposed method achieved an accuracy of 87. These findings suggest that deep learning, specifically CNNs, offers a significant improvement over traditional classification techniques, reinforcing the relevance of this research. The purpose of this research is to explore the application of deep learning, specifically a Halaman 86 METHOMIKA: Jurnal Manajemen Informatika & Komputerisasi Akuntansi Vol. 9 No. 1 (April 2. Convolutional Neural Network, for leaf-type image Deep learning is a subset of machine learning that uses complex algorithms and vast amounts of data to train a model (Sarker, 2. create a reliable and effective model that can accurately categorize leaf types, we will leverage CNNs' built-in ability to learn hierarchical features from images. The developed model could be incorporated into mobile apps or web-based systems, allowing users to recognize plant species by taking pictures of their For this research, we will use a deep learning method called a convolutional neural network with the Xception architecture. In traditional machine learning, programmers must carefully define specific features that the computer should focus on. This process, known as feature extraction, can be time-consuming and requires The computer's success rate depends on the programmer's ability to define a feature set accurately. This process is called supervised learning. Deep learning has an advantage over traditional machine learning because the program can build the feature set on its own without the programmerAos supervision. This process is called unsupervised learning. Unsupervised learning is usually faster and more accurate, as it eliminates the need for manually engineering features (Diallo, 2. Convolutional neural networks (CNN. are a type of artificial neural network that can learn and extract features from data automatically using a convolution process. Unlike traditional methods that require manual feature engineering. CNNs eliminate the need for human intervention. CNN architecture is inspired by the human visual system. Each artificial neuron in a CNN corresponds to a biological neuron, while the kernels in a CNN represent different receptors that can detect various features. The activation functions in CNNs mimic the biological process in which only neural signals that exceed a certain threshold are transmitted to the next neuron (Li. Liu. Yang. Peng, & Zhou, 2. Xception is a deep learning architecture composed of a linear sequence of depthwise separable convolution layers with residual connections. This simplified design makes it easy to create and modify using high-level libraries like Keras or TensorFlowSlim. It requires only 30Ae40 lines of code, unlike more complex architectures such as Inception V2 or V3. publicly available implementation of Xception using Keras and TensorFlow is included in the Keras ISSN: 2598-8565 . edia ceta. ISSN: 2620-4339 . edia onlin. Applications module under the MIT license (Chollet. METHODS Algorithm Xception is different from Inception because Xception has a modified depthwise separable As shown in Figure 1 The original depthwise separable convolution is comprised of a depthwise convolution followed by a pointwise Unlike traditional convolution, it avoids performing convolution across all channels, reducing the number of connection and resulting in a smaller, more efficient model. Figure 1. Original Depthwise Separable Convolution As shown in Figure 2 The modified depthwise separable convolution is a technique that involves applying a pointwise convolution followed by a depthwise convolution. This approach is inspired by the Inception-v3 architecture, where 1y1 convolutions are used to reduce the number of channels before performing nyn spatial convolutions. This modification is slightly different from the original, which typically uses the depthwise convolution before the pointwise Figure 2. Modified Depthwise Separable Convolution used as an Inception Module in Xception Halaman 87 METHOMIKA: Jurnal Manajemen Informatika & Komputerisasi Akuntansi Vol. 9 No. 1 (April 2. For this program, we will be using hyperparameters such as ReLU activation with 1024 units and SoftMax activation with 5 units. The optimizer for the model will be the Adam optimizer with a 0. 001 learning rate. We will be using 50 epochs, with each epoch having 5 steps. These hyperparameters were chosen based on their effectiveness in previous deep learning studies and their ability to enhance model performance while preventing overfitting. ReLU (Rectified Linear Uni. Activation are multiple layers of activation that the CNN model uses. Simplicity and the reliability of ReLu activation cause an increase in performance compared to other activation models (Ramachandran. Zoph, & Le, 2. ReLu function activation performs Thresholding of the zero value on image input pixels. ReLU activation creates value pixels that have a value less than zero then it will be set to 0 in the image. SoftMax activation or SoftMax Classifier is another type of activation function used in CNN. SoftMax activation is a combination of sigmoid However. SoftMax activation unlike sigmoid activation that is used for binary classification. SoftMax activation usually counts probability in each target category/class of all possible target classes. used in multi-classification models. SoftMax activation will return each class and the target class which will have the highest probability (Nwankpa. Ijomah. Gachagan, & Marshall, 2. The Adam optimizer was chosen due to its adaptive learning rate properties, which help achieve faster convergence compared to traditional optimization algorithms such as Stochastic Gradient Descent (SGD). The learning rate of 0. 001 was selected based on prior research indicating that it provides a good balance between model stability and convergence Additionally, 50 epochs were chosen to allow sufficient learning while avoiding overfitting, and each epoch consists of 5 steps to ensure effective weight These hyperparameters were fine-tuned to optimize classification performance while maintaining computational efficiency. Evaluation The Evaluation method for this research is by using Confusion Matrix. To do the evaluation we used classification report from the scikit-learn metrics the result of classification report is a table composed of precision, recall, f1-score, and accuracy. Precision is the metric that give you the proportion of true positives to the number of total positives that the ISSN: 2598-8565 . edia ceta. ISSN: 2620-4339 . edia onlin. model predicts. Recall is the metric that focuses on how good the model is at finding the positives. F1 Score is the metric that combines recall and precision. Accuracy measures how many predictions were true out of all the predictions that were made. Here is the equation to calculate the number of each result in a classification report (Foody, 2. ycEycyceycaycnycycnycuycu = !" $ %" . ycIyceycaycaycoyco = !" $ %& . ya1 = ' ( ")* ,-,. / ( 0* 122 ")* ,-,. /$ 0* 122 !" $ !& yaycaycaycycycaycayc = !" $ !& $ %" $ %& . Data For this research, the dataset consists of images from five different classes of leaf types: kaffir lime leaves, chive leaves, turmeric leaves, bay leaves, and betel leaves. The dataset was obtained by collecting images from Google Images using a web scraper to gather the first 400 images of each leaf type. After collecting the images, we manually checked each one for irrelevant content that could not be used. After this process, we were left with 200 images of kaffir lime leaves, 219 images of chive leaves, 217 images of turmeric leaves, 221 images of bay leaves, and 218 images of betel leaves, for a total of 1,075 images. The dataset was then put through a process of image data augmentation. Image data augmentation creates new images from existing ones to artificially increase the size and diversity of the training data. do this, the images were altered by making small changes, such as rotating them, adjusting their brightness, or shifting the subject vertically or horizontally (Shorten & Khoshgoftaar, 2. incorporating these modified images, we aimed to improve the model's ability to generalize and handle variations in real-world leaf images. We used ImageDataGenerator from KerasAo preprocessing image library to perform the image data augmentation. To facilitate the training and evaluation of our model, we divided the dataset into a training and validation set. Seventy percent of the images . were allocated to the training set, while thirty percent . were reserved for the validation This division enabled us to train the model on a substantial portion of the data and assess its performance on a separate set to prevent overfitting. Examples of the leaves that will be used in the dataset are shown in the figures below. Halaman 88 METHOMIKA: Jurnal Manajemen Informatika & Komputerisasi Akuntansi Vol. 9 No. 1 (April 2. ISSN: 2598-8565 . edia ceta. ISSN: 2620-4339 . edia onlin. Figure 7. Example of Betel Leaves Figure 3. Example of Kaffir Lime Leaves RESULTS AND DISCUSSION For the evaluation, the loss function that we use is categorical cross-entropy, and Adam for the optimizer with 0. 001 learning rate. And as we can see from Figure 8, the model starts with a very high loss score and very low accuracy score, but after 50 epochs the model keeps getting better. Figure 4. Example of Chive Leaves Figure 8. Training Model Epochs We can visualize the plot for each iteration of accuracy and loss in Figure 9 and 10. Figure 5. Example of Turmeric Leaves Figure 9. Plot of Accuracy Figure 6. Example of Bay Leaves Halaman 89 METHOMIKA: Jurnal Manajemen Informatika & Komputerisasi Akuntansi Vol. 9 No. 1 (April 2. ISSN: 2598-8565 . edia ceta. ISSN: 2620-4339 . edia onlin. larger, more diverse dataset, increasing training epoch counts to improve learning, and refining hyperparameter tuning for additional performance By pursuing these directions, we aim to develop a more robust classification model that advances leaf identification and related fields. Figure 10. Plot of Loss As illustrated in Figure 11, the classification model has achieved an accuracy score of 74%. While this result indicates satisfactory performance, a closer examination of the classification report for each class reveals that one particular class exhibits subpar This suboptimal performance can be attributed to several factors, including potential shortcomings in the dataset. One possible reason for the subpar performance is the quality of the images within the dataset. If the images are of low resolution, have insufficient detail, or contain artifacts, it can hinder the model's ability to make accurate classifications. To address this issue, future research efforts should focus on acquiring a dataset with higher quality images. By providing the model with more informative and visually distinct images, we can expect to improve its accuracy and overall performance. Figure 11. Classification Report Result CONCLUSION The classification model we developed demonstrated promising results, achieving a commendable 74% accuracy score based on a comprehensive evaluation using the confusion matrix. While the model effectively classified leaf types, a detailed analysis of precision, recall, and F1-score indicated suboptimal performance in one particular class, likely due to dataset limitations such as insufficient sample size or class imbalance. To enhance accuracy, future research should focus on acquiring a REFERENCES