International Journal of Cyber and IT Service Management (IJCITSM) Vol. No. October 2025, pp. 144Oe155 E-ISSN: 2808-554X | P-ISSN: 2797-1325. DOI:10. ye Creation of RAG Chatbot in Answering Queries Related to Banking Terms Using Microsoft Azure Cindy Aurora1* . Tuga Mauritsius2 1,2 Master of Information System Management. Bina Nusantara University. Indonesia 1 cindy. aurora@binus. id, 2 tmauritsus@binus. *Corresponding Author Article Info ABSTRACT Article history: This study investigates the use of Natural Language Processing (NLP), specifically Large Language Models (LLM), to implement technology intelligence as a solution for supporting developers in understanding information-related terms, abbreviations, and business processes in the banking sector. The aim is to explore the effectiveness of integrating AI-powered chatbots, particularly through Retrieval Augmented Generation (RAG), to enhance software development in banking by providing fast, accurate responses. The methodology involves profiling data analysis and the development of a RAG based AI chatbot using the Microsoft Azure platform, integrating advanced NLP and LLM techniques to assist developers in navigating complex banking terms and processes efficiently. The results demonstrate that the RAG chatbot significantly improves operational efficiency by offering real-time, context-aware responses, enabling faster decision-making and reducing time spent on manual searches for information, which leads to faster software development cycles. This study contributes to the fields of NLP and LLM, particularly in the banking sector, by showcasing the benefits of RAG chatbots in improving operational efficiency and software development quality. The use of AI technologies provides substantial improvements in the development process, leading to enhanced productivity in the banking industry. Submission June 11, 2025 Revised July 22, 2025 Accepted August 26, 2025 Published September 9, 2025 Keywords: Natural Language Processing Large Language Model Retrieval Augmented Generation Chatbot Microsoft Azure This is an open access article under the CC BY 4. 0 license. DOI: https://doi. org/10. 34306/ijcitsm. This is an open-access article under the CC-BY license . ttps://creativecommons. org/licenses/by/4. AAuthors retain all copyrights INTRODUCTION Developing banking software requires a deep understanding of both technical and business processes. Profiling data is crucial for ensuring the project stays on schedule. However, challenges arise in comprehending specific business terms and processes, making it essential to find accurate information. While tools like Google can help, they may not be effective for unique company processes. Alternatives like ChatGPT have limitations, as its knowledge only covers information up until 2022 and lacks access to specific documents from Bank Indonesia, leaving developers struggling with internal company documents . Ae. This study focuses on developing AI-based solutions using NLP chatbots to speed up the search and understanding of banking-related terms, abbreviations, and business processes. By implementing LLM and RAG chatbots, the system quickly identifies context queries and delivers accurate responses, increasing operational efficiency and reducing the time spent on manual searches, allowing developers to focus on more complex tasks . Journal homepage: https://iiast. iaic-publisher. org/ijcitsm/index. php/IJCITSM/index International Journal of Cyber and IT Service Management (IJCITSM) ye This study is expected to benefit organizations, developers, and academics. For organizations, the chatbot boosts operational efficiency by quickly answering queries, reducing manual work, and optimizing resources . For developers, it provides insights and techniques to enhance chatbot performance, especially in responsiveness and handling diverse questions. Academically, this research contributes to understanding NLP and RAG chatbot theory in banking projects and can serve as a valuable reference for future studies. Scope study This covers exploration use of AI for increase efficiency and accuracy in search information related to business processes banking. Chatbot built is a RAG Chatbot that utilizes SEKI Metadata from Bank Indonesia as a knowledge base, with the Microsoft Bot Framework platform, and uses Language Python programming for implementation on the web . Banking IT Project Development Banking is crucial for collecting funds and supporting economic growth. Bank Indonesia (BI) ensures financial stability through monetary policy, supervision, and managing the payment system. Its main tasks include maintaining rupiah stability and controlling inflation through interest rate policies and open market operations . , . BI also supports sustainable economic growth and financial inclusion. BI provides key economic indicators like SEKI. ITEMs, and The Red Book for informed decision-making . Internally. BI uses Agile and Scrum frameworks, emphasizing collaboration. Business metadata is essential for accurate decisionmaking, but developers face challenges in understanding terms and processes, making good documentation critical for project success. NLP and Chatbot Chatbots can aid in understanding business metadata by offering an interface for managing businessrelated information . They are classified into two types. Rule-Based and AI-Based, with architecture generally involving three stages, pre-processing (NLP), processing (NLU), and generation (NLG) to create responses . Chatbots are also categorized by domain. Open Domain and Closed Domain, and they can process text or voice input . In regulated sectors like banking. NLP-based chatbots face challenges, including the need to comprehend specialized financial terms and adhere to industry regulations, requiring focused development efforts for effective operation . The Microsoft Bot Framework is a widely used open-source framework for building domain-specific chatbots . LLM and RAG A Large Language Model (LLM) is an AI that understands and generates human-like text, trained with deep learning techniques and millions to billions of parameters to recognize language patterns . LLMs, like OpenAIAos GPT, perform various tasks such as answering questions and translation, and are used in applications like chatbots and automatic writing. GPT models, including GPT-2. GPT-3, and GPT-4, have revolutionized AI with their human-like text generation . RAG combines LLM text generation with information retrieval for more accurate responses, especially for recent data or specific queries . RAG uses LLM to process input, retrieves data from external sources, and then generates a response, with Microsoft AzureAos setup including application servers. Azure AI Search, and OpenAI GPT for final processing . Azure AI Search Azure AI Search enhances the chatbot experience in the RAG context by using the text-embeddingada-002 model for various NLP tasks like semantic search, text classification, content recommendations, and clustering . The main processes include indexing . and retrieval, where documents are processed into tokens and vectors for search, with AI skills used to enrich the data. Search vectors use numerical representation to find similarities between vectors, enabling semantic or conceptual matching . In RAG. Azure AI Search collects and indexes relevant data, searches based on user queries, and returns relevant results, which are then combined with a generative model like OpenAI GPT to produce more accurate and informative responses . Azure OpenAI (GPT 3. 5 Turb. GPT-3. 5 Turbo plays a role as a generator model text in the implementation of RAG on Azure, which combines retrieved data via Azure AI Search for increased accuracy and relevance response. As part of RAG. GPT-3. 5 Turbo does not only produce answers based on knowledge training general, but also integrates information from source relevant external for give more response informative and contextual . GPT-3. 5 Turbo is built with transformer architecture that uses an AyattentionAy mechanism to understand input context and output E-ISSN: 2808-554X | P-ISSN: 2797-1325 appropriate text. This model is also trained using Reinforcement Learning from Human Feedback (RLHF) to increase responsiveness and ethical answers . Designed For efficiency. GPT-3. 5 Turbo can give a response with latency low, fit for real-time applications. Evaluation & CRISP-DM Stratified Split is a method of data sharing based on a theme to ensure every group reflects balanced proportions from every sector, allowing more analysis depth and understanding of trends in every sector of business . Chatbot evaluation is divided into three sections: evaluation content, satisfaction users, and functionality, which can tested through functional testing, error testing, and the accuracy of chatbot answers . CRISP-DM (Cross-Industry Standard Process for Data Minin. is a methodology standard for project data mining consisting of six phases main. Understanding Business. Data Understanding. Data Preparation. Modeling. Evaluation, and Deployment. Methodology This is flexible and iterative, allowing adjustments in each stage based on findings and bait-back . Previous Research Chatbots, supported by machine learning, are used in various fields to improve customer service and user performance . BPMN-based chatbots, integrated with platforms like Microsoft Bot Azure, transform images into interactive tools for business processes . In healthcare, chatbots analyze symptoms, offer telehealth services, and manage conditions like hypertension . , . They also educate on HIV, contraception, and pregnancy . and help isolated individuals seek mental health support . , . Chatbots are also used in software testing . and language education . The Retrieval-Augmented Generation (RAG) chatbot combines retrieval and generation for accurate responses . RAG chatbots assist in education and healthcare with text extraction, summarization, and medical responses . Overall, chatbots enhance efficiency across METHOD The integration of Azure AI Search with OpenAI GPT and the custom Bank Indonesia metadata database is key to the chatbotAos ability to generate accurate and context-aware responses. This combination of tools allows the chatbot to access up-to-date financial data, providing domain-specific responses tailored to the unique needs of the banking sector. The methodology used in this study is CRISP-DM which consists of six main stages, each of which has a series of specific tasks and activities as follows: Business Understanding Business Understanding in the CRISP-DM methodology is crucial for aligning objectives with software development, especially in banking. This study explores the use of AI, particularly NLP and LLM with the RAG approach, to help developers quickly access relevant information and improve operational efficiency . The proposed RAG chatbot integrates real-time data retrieval with LLM-generated responses, offering more precise, domain-specific answers, making it more effective than traditional rule-based systems for the banking sector. Data Understanding This stage focuses on exploring and understanding the research data. The first step is downloading SEKI metadata from Bank IndonesiaAos website, which contains unstructured data in PDF format for analysis. The methodology lacks details on preprocessing and cleaning the data for chatbot integration. Key challenges include handling missing data, formatting issues, and ensuring compliance with industry standards, particularly for sensitive banking information. The next step is to stratify the data by sectors. Monetary. Government Financial. Real, and External, to guide the appropriate analysis for each sector. Data Preparation In the Data Preparation stage, selected data will be uploaded to Microsoft Azure, using Azure AI Document Intelligence for intelligent document processing. This cloud-based service automatically extracts data from PDFs and supports OCR to extract text from images or complex documents. It also handles various file types, including . docx, . pptx, . xlsx, . png, . jpg, and others, enabling efficient data processing for further International Journal of Cyber and IT Service Management (IJCITSM). Vol. No. October 2025, pp. 144Ae155 International Journal of Cyber and IT Service Management (IJCITSM) ye Modeling In the Modeling stage, the main focus is to build a chatbot model that is by business objectives, specifically using OpenAIAos Large Language Model (GPT-3. 5 Turb. for text interpretation. The steps taken include filling in Microsoft Azure credentials, defining the LLM role, running the Command Line Interface for deployment, and checking the service after deployment. GitHub is used for coding, versioning, and documentation Evaluation and optimization are carried out continuously to ensure optimal chatbot performance according to business needs. Evaluation To assess the quality and effectiveness of the model, several functional tests were carried out to evaluate its performance. These tests included Functional Testing, which ensures that the chatbot features and functionality operate as expected. Error Testing, which identifies chatbot responses to typos or ambiguous queries, and evaluates the accuracy of its answers to user questions. Implementation In the Deployment phase in CRISP-DM, the focus is on implementing the solution into a production In this study, the chatbot was integrated into a website using the Azure App Service hosting service, allowing the chatbot to be accessed by users across multiple platforms and applications. RESULT AND DISCUSSION Business Understanding To provide a deeper understanding of the economic and financial sectors of Indonesia, the following table presents the SEKI (Indonesian Economic and Financial Statistic. This table classifies the data into four main sectors: the monetary sector, government financial sector, real sector, and external sector, covering various key economic indicators essential for analyzing IndonesiaAos macroeconomic conditions. Below is the detailed breakdown of the data presented in this Table 1. Table 1. SEKI Metadata (Indonesian Economic and Financial Statistic. Monetary Sector: Primary Money Balance Sheet General Bank and BPR Analytics Money Circulation and Factors Affecting It Community Savings Loans Given Finance Company Operation Monetary Interest rate Micro. Small and Medium Enterprises (MSME. Credit Government Financial Sector Operation Finance Central government Government Securities Real Sector: Consumer Price Index (CPI) Wholesale Price Index (IHPB) Gross Domestic Product (GDP) External Sector: IndonesiaAos Balance of Payments Foreign Trade IndonesiaAos International Investment Position International Economic Indicators Nominal Exchange Rate Index Against Major Partner Currencies Foreign Debt State Foreign Exchange Reserves ye E-ISSN: 2808-554X | P-ISSN: 2797-1325 This can be observed in Table 1. SEKI (Indonesian Economic and Financial Statistic. is a monthly publication by Bank Indonesia that provides comprehensive information on IndonesiaAos economy and finance, using internationally recognized methodology for comparative analysis with other countriesAo data. SEKI includes primary data from Bank Indonesia and secondary data from institutions like the Ministry of Finance. BPS, and LPS, covering four main sectors: A Monetary Sector: Providing information about liquidity, money supply, balance sheet authority monetary, and dynamics of money markets and capital markets. A Financial Sector Government: Focus on transparency and accountability implementation of the APBN, including data on revenue, expenditure, financing, and SBN status. A Real Sector: Provides related data Income Gross Domestic Product and Consumer Price Index For evaluation of economics and policy. A External Sector: Presents data about balance sheet payment, loan external, and indicators economy as well as monetary international, covering transaction export-import and obligations overseas. Data Understanding There are 21 SEKI Metadata grouped into four sectors, with details of Monetary Sector 9 Metadata. Government Financial Sector 2 Metadata. Real Sector 3 Metadata, and External Sector 7 Metadata. The author chose to use 9 Metadata from the Monetary Sector, which are unstructured data in PDF format. Each PDF metadata has several pieces of information, including: A Basic Information: Contains contact data, name, statistics organizer, address, and related information. A Data Definition: Briefly explain the type of data contained. A Data Coverage: State the data limitations, including influencing factors and institutional coverage. A Publication Periodization: Provides information about the frequency of data publication . aily, monthly, yearly, etc. A Timeliness of Publication: State the exact time the publication was made. A Forward Publication Schedule (ARC): Includes a schedule and examples of published data. A Data Sources: Mention the institutions, applications, and portals used as data sources. A Methodology: Describes concepts, definitions, collection methods, and factors influencing data. A Data Integrity: Conveys data status . nterim or fina. and possible revisions. A Data Access: Provides information about where or how to access published data. Data Preparation Before performing data preparation using Azure AI Document Intelligence Service, it is necessary to create a GitHub account and clone the Azure Samples Template first with the following steps: A Create a GitHub account at https://github. com/codespaces and verify the account to be used. A Once you have signed in, use the template from Microsoft Azure with the following link Azure-Samples/azuresearch-openai-demo which is a sample app for the Retrieval-Augmented Generation pattern running in Azure, using Azure AI Search for retrieval and Azure OpenAI large language models to power ChatGPTstyle and Q&A experiences . A The next step is to create a new codespace using the Azure template at point b by clicking on the Au<> CodeAy button, clicking the Codespaces menu then clicking the AuCreate codespace on mainAy button. After the codespace creation process is complete, then open the codespace page by clicking on the title of the codespace created. A In default mode. Azure has prepared some sample files that can be used. After the GitHub project page opens, the first step is to delete the existing sample files and re-upload the files that will be used as a knowledge base. International Journal of Cyber and IT Service Management (IJCITSM). Vol. No. October 2025, pp. 144Ae155 International Journal of Cyber and IT Service Management (IJCITSM) ye Data Modeling At this stage, the focus will be on creating a predictive model with the main goal of producing effective chatbot interactions that meet user needs. After the data preparation process is carried out, the model will be developed using an algorithm that matches the chatbot context, namely the GPT-35-Turbo large language model created by Open AI for text understanding. This modeling stage will be implemented with the help of GitHub as a collaborative repository for code, version tracking, and documentation of changes during the development The steps to be taken are as follows: A Continuing the previous stage in the Data Preparation sub-chapter, still on the same GitHub, the next thing to do is to fill in the credentials used in the Au. envAy script file. In this environment setting, the author uses vectorization and data search with the text-embedding-ada-002 model which is an Azure AI Search service, and the GPT-35-Turbo model from Azure Open AI for the LLM model used. A After the credentials are filled in, the next step is to define the role of LLM as a smart brain by setting the prompting in the app > backend > approaches > chatreadretrieveread. Py file folder. Figure 1. LLM Instruction Script A The script in Figure 1 LLM Instruction Script defines a property called system message chat conversation, which generates a system message or instruction for the assistant (AI mode. in performing a specific task. @property is a Python decorator that allows methods to be called like attributes. So, when system message chat conversation is called it will produce a string without requiring parentheses like a function call. This message is designed to instruct the AI to act as a special assistant for developers in understanding SEKI data. The AI is asked to provide short, concise, and factual answers, only answer questions according to the information contained in the data source provided, avoid speculation or information outside the source, and answer AudonAot knowAy if there is not enough information and ask follow-up questions if there is a need for clarification. A AI machines are also required to answer in formal and clear Indonesian. Include citations from data sources in every possible response, with the format . Both variables . ollow up questions promp. njected promp. indicate that it is possible to add additional instructions . njected promp. or follow-up questions . ollow up questions promp. that will be added in a certain context to become an answer. A To perform Deployment, run the CLI (Command Line Interfac. command Auand upAy. This command will automatically run the command that has been defined in the azure. yaml script, one of which will call the command at the start. sh script. If the Auand upAy command has been executed, then define the name of the environment to be used. Then log in with the command Auand auth loginAy. Next, log in using the code created by GitHub. A After logging in, then run the Auand upAy command again. If the command is successful, you will be asked to fill in the Azure account subscription and the Azure server location to be used. Then the system will work and run the existing script until all the knowledge files used are uploaded, if the deployment is successful, a log will appear as shown in Figure 2 GitHub Success Log. E-ISSN: 2808-554X | P-ISSN: 2797-1325 ye Figure 2. GitHub Success Log Figure 2 illustrates the GitHub success log generated after the deployment process, confirming that the application was successfully deployed to Azure. The log provides details of the deployment duration, service endpoint, and access to the created resources, indicating that the system is ready for further use and testing. Figure 3. Resource Group Services Azure A After the deployment is successfully executed, then you can check the service that has been formed on the Azure Portal. The service can be seen in the Resource Group menu > select the project that was formed > Overview > Resources as seen in Figure 3 Azure Resource Group Services. Make sure 9 services have been formed, namely: Ae App-backend: App Service for frontend and backend. Ae Appi: Application Inside for activity reporting. Ae Cog-di: Document Intelligence that acts as OCR. Ae Cog: Open AI LLM used. Ae Dash: Functions for the dashboard. Ae Gptkb: Functions for creating vectors for Azure AI Search. Ae Log: Application activity log. Ae Plan: Used for server hosting. Ae St: Storage Account where Knowledge Base is stored. After a successful deployment, the next step is to ensure that all required services are properly established in the Azure portal. At this stage, critical components, such as backend services, activity reporting applications, and features supporting AI and analytics functionality, will be checked to ensure overall system integrity. This testing is crucial for ensuring smooth operations and ensuring that all necessary resources are properly configured. International Journal of Cyber and IT Service Management (IJCITSM). Vol. No. October 2025, pp. 144Ae155 International Journal of Cyber and IT Service Management (IJCITSM) ye Figure 4. Chatbot Deployment Model A In addition to checking the services formed, checks were also carried out for the Retrieval and Generation models formed. In Figure 4 the Chatbot Deployment Model can be seen after deployment on Azure Open AI, the models formed are GPT-35-turbo and text-embedding-ada-002. Evaluation A Feature Testing Scenario Before presenting the Feature Testing Scenarios, the following Table 2 outlines the key test scenarios conducted to evaluate the chatbotAos functionality. The tests involved accessing the chatbot across different browsers and assessing its ability to display question suggestions based on the knowledge base when used for the first time. The results were analyzed to ensure proper functionality across various platforms. Table 2. Feature Testing Scenarios Test Scenario Expected Output Accessing Chatbot using Chrome. Firefox. Edge Chatbot successfully Displays question suggestions based on the Displaying question knowledge base when using the chatbot for the suggestions first time Results Success Success Based on Table 2 Feature Testing Scenario, the chatbot was tested on Google Chrome. Mozilla Firefox, and Microsoft Edge. While functional testing and error handling provided useful insights, adding statistical analysis . , precision, recall. F1-scor. and A/B testing would better measure accuracy and The results showed the chatbot performed consistently across all browsers, and the second scenario confirmed that knowledge base question suggestions appeared correctly. This indicates that the chatbot, built on Microsoft Azure, is not only compatible with widely used browsers but also reliable in guiding users with relevant suggestions in diverse usage contexts. A Scenario with Ambiguous Questions To further evaluate the robustness of the chatbot, error testing scenarios were conducted to examine its ability to handle incomplete or ambiguous queries. These tests aimed to verify whether the chatbot could still provide relevant and accurate responses even when the input contained missing letters, symbols, or other irregularities. The summary of these scenarios is presented in Table 3. Table 3. Feature Error Testing Scenarios Test Scenario Expected Output If the letter from the inputted word No true/less Chatbot successfully displays the correct If you enter the wrong symbols or numbers Chatbot successfully displays the correct Results Success Success E-ISSN: 2808-554X | P-ISSN: 2797-1325 ye It can be seen in Table 3 (Feature Error Testing Scenari. , using the first scenario, testing was carried out on the words from the reduced question query. In this scenario, the Author first entered a question query with a complete question without a question mark AyExplain what is primary moneyAy, then the chatbot managed to provide an appropriate explanation in answering the question. Next, the Author entered a query with reduced letters AyExplain what is primary moneyAy, the chatbot still provided an answer that was relevant to the question. In scenario 2, the Author re-inputted the question query by inserting a symbol in the question sentence AyExplain what {} | {} { {}{ is primary money @#!$%%Ay. In this question, the chatbot can also produce relevant answers and provide knowledge documents from the citations used. This matter shows that the chatbot can understand context questions well even though the words used are incomplete and the sentences are not properly arranged, containing unnecessary symbols. Accuracy Test Scenario Answer To assess the accuracy of the chatbotAos responses, a series of test scenarios were designed to compare the generated answers with the expected reference documents. These tests aimed to ensure that the chatbot could provide valid answers along with the correct source citations. The details of the accuracy evaluation are summarized in Table 4. No Test Scenario Explain what is meant by money circulation When is the Community Savings data? What of Monetary Operation Table 4. Accuracy Test Scenario Answer Expected output The output issued PDF file PDF file 3 Money In 3 Money Circulation and Circulation and Factors That Influence It Indo Factors That Influence It Indo and 02 Analytical Balance of General Banks and BPRs PDF file PDF File 04 Public 04 Community Savings Savings, 3 Money In Circulation and Factors That Influence It Indo 02 Analytical Balance of General Banks and BPRs PDF 7 Monetary PDF file 7 Monetary Operations DPMOperations DPMSEKI Rev 2016 SEKI Rev 2016 Results Valid Valid Valid The final functionality test is to perform an accuracy test answer, in this test scenario expected chatbot can give the right answer along with the source knowledge that is quoted. As can be seen in Table 4 Accuracy Test Scenario Answer, there are 3 questions compiled by the Author. For do a trial along with the expected citation file. Answer results chatbot displays conclusion relevant answers and correct citation files by the expected scenario. This proves that Retrieval and Generation performance carried out by chatbots function with good and have accurate valid answers. Implementation Implementing Azure Bot on a website provides a solution that utilizes ability intelligence artificial For increase the experience of the user in a direct way. Azure Bot Service provides infrastructure that enables the development, implementation, and management of chatbots with diverse capabilities. This website-based chatbot integration utilizes Azure Cloud Service. Website-based chatbot interface compatible with the browser in International Journal of Cyber and IT Service Management (IJCITSM). Vol. No. October 2025, pp. 144Ae155 International Journal of Cyber and IT Service Management (IJCITSM) ye general used such as Google Chrome. Edge, and Mozilla Firefox. Website as an implementation platform gives high accessibility, making it easy for users to interact with the bot without the need to download application Azure Bot Service supports integration through a website-based chat channel that can embedded in a website page using JavaScript, making it flexible For various types of websites, good For service customers, information portals, and support technical. Once the deployment is complete, the chatbot can accessed with URL app-backend-xy3dhrzksbr3c. The Chatbot Homepage view that has been made can seen in Figure 5 Chatbot Homepage. Figure 5. Chatbot Homepage Figure 5 shows the chatbot homepage after deployment, where users can directly interact with SEKI data through a simple and accessible web-based interface. This interface demonstrates the successful integration of Azure Bot Service into a real-time application. MANAGERIAL IMPLICATIONS The integration of RAG-based chatbots using NLP and LLM in the banking sector offers significant managerial implications. By automating the retrieval and generation of business-specific information, it enhances operational efficiency, reducing the time developers spend searching for relevant data, leading to faster project delivery. This can also result in cost savings, as the need for manual searches and administrative support is minimized. The chatbot enables improved decision-making by providing managers and developers with fast and accurate responses, ensuring decisions are based on up-to-date data. Furthermore, the chatbotAos scalability allows for easy adaptation to handle increasing data and more complex queries, facilitating future growth without a proportional rise in operational costs. Banks adopting AI-powered chatbots gain a competitive advantage by speeding up software development, staying ahead in an evolving market, and improving employee productivity by allowing staff to focus on higher-value tasks. Additionally, the seamless integration of the chatbot with existing systems ensures that banks can enhance their infrastructure without disrupting established workflows. Overall, the use of AI technologies positions banks to be more efficient, cost-effective, and competitive, benefiting both operational and strategic aspects of management. CONCLUSION This study indicates that the implementation of AI based chatbots using NLP and LLM technologies can significantly enhance efficiency in supporting project development processes in the banking sector. The chatbot, developed using the RAG approach, has proven capable of delivering fast and accurate responses to queries submitted by software developers. This effectiveness is achieved through the utilization of a domainspecific knowledge base tailored to the banking industry, which helps overcome the limitations of manual searches related to internal terminology, abbreviations, and processes typically known only within the organization. Trial results demonstrate that the chatbot, deployed via Microsoft Azure, performs reliably across various browsers such as Google Chrome. Mozilla Firefox, and Microsoft Edge. It is able to suggest relevant questions to assist users in composing inputs, understand incomplete or symbol-laden queries, and deliver E-ISSN: 2808-554X | P-ISSN: 2797-1325 ye responses that are both accurate and contextually appropriate. The chatbotAos ability to cite relevant knowledge documents and respond meaningfully has been validated. With its contextual understanding, the chatbot holds the potential to improve developer productivity by reducing time spent on unproductive manual searches, allowing them to focus more on complex tasks in banking software development projects. For further development, several recommendations are proposed. First, continuously updating the knowledge base with current documents and collaborating with business and compliance teams will ensure answer accuracy and regulatory alignment. Second, enhancing contextual learning through fine-tuned datasets reflective of banking language can help the chatbot handle more complex inquiries. Third, deeper integration with internal company systems will allow access to real-time and updated information. Periodic performance evaluations, interactive interface enhancements with visual aids or step-by-step guidance, and strengthening data privacy and security through encryption, audits, and access controls are essential to ensure the chatbot remains adaptive, user-friendly, and compliant with banking standards. DECLARATIONS About Authors Cindy Aurora (CA) Tuga Mauritsius (TM) https://orcid. org/0000-0002-5525-5591 Author Contributions Conceptualization: CA. Methodology: TM. Software: CA. Validation: TM and CA. Formal Analysis: TM and CA. Investigation: CA. Resources: TM. Data Curation: CA. Writing Original Draft Preparation: TM. Writing Review and Editing: CA. Visualization: TM. All authors. CA and TM, have read and agreed to the published version of the manuscript. Data Availability Statement The data presented in this study are available on request from the corresponding author. Funding The authors received no financial support for the research, authorship, and/or publication of this article. Declaration of Conflicting Interest The authors declare that they have no conflicts of interest, known competing financial interests, or personal relationships that could have influenced the work reported in this paper. REFERENCES