International Journal of Cyber and IT Service Management (IJCITSM) Vol. No. October 2025, pp. 186Oe197 E-ISSN: 2808-554X | P-ISSN: 2797-1325. DOI:10. ye Framework of Master Data Management in Banking Using Consolidation and Jaro Winkler Algorithm Ipang Sasono1* . Mustar Aman2 1,2 Faculty of Computer Science. Universitas Insan Pembangunan Indonesia. Indonesia 1 ipangsasono@gmail. com, 2 mustarstmik@gmail. *Corresponding Author Article Info ABSTRACT Article history: Master Data Management (MDM) is a crucial framework for ensuring the consistency, accuracy, and reliability of key data entities within banking systems. In the financial sector, where data from multiple departments and sources is constantly generated and shared, it is vital to maintain a single, unified view of critical data to prevent inconsistencies, inaccuracies, and duplication. This paper introduces a comprehensive design for implementing MDM in banks, utilizing a consolidation approach integrated with the Jaro-Winkler similarity algorithm. The consolidation approach allows the seamless integration of disparate data sources across various departments, creating a unified and centralized data This is essential for maintaining a comprehensive and reliable view of data assets, thereby improving decision-making and operational efficiency. The inclusion of the Jaro-Winkler algorithm enhances data matching capabilities by identifying and resolving duplicates or near-duplicate records through name and text similarity comparisons, an essential feature given the complex nature of customer and transactional data in banking. By addressing these challenges, the proposed MDM solution significantly improves data quality, reduces redundancy, and ensures that information is accurate and accessible across all levels of the organization. This system provides a scalable, robust, and efficient data management infrastructure, crucial for meeting regulatory compliance requirements, enhancing customer service, and optimizing operational processes. The methodology presented in this paper demonstrates an effective and structured approach for large-scale data integration and verification, offering a reliable solution for managing vast amounts of data in the banking sector. Submission June 23, 2025 Revised August 10, 2025 Accepted August 28, 2025 Published September 16, 2025 Keywords: Master Data Management Consolidation Approach Jaro-Winkler Algorithm Data Integration Banking Systems This is an open access article under the CC BY 4. 0 license. DOI: https://doi. org/10. 34306/ijcitsm. This is an open-access article under the CC-BY license . ttps://creativecommons. org/licenses/by/4. AAuthors retain all copyrights INTRODUCTION In the modern banking industry, the management of vast and diverse data sets has become increasingly critical for operational efficiency, regulatory compliance, and customer satisfaction . As banks continue to expand their services and increase customer interactions across multiple platforms, the challenge of managing master data critical information regarding customers, accounts, transactions, and other core entities grows Inconsistent, inaccurate, or duplicated data can lead to poor decision-making, operational inefficiencies, regulatory issues, and a decline in customer trust . This emphasizes the need for an effective MDM system to ensure accuracy, consistency, and reliability of data across organization. Journal homepage: https://iiast. iaic-publisher. org/ijcitsm/index. php/IJCITSM/index International Journal of Cyber and IT Service Management (IJCITSM) ye The process of implementing an MDM system in a banking environment requires the integration of data from various departments and systems, which often function in silos. Traditional methods for data management may struggle to maintain data integrity across such disparate systems . A consolidation approach to MDM, where data from various sources is unified into a centralized repository, provides a potential solution . This approach not only improves data accessibility but also ensures that critical data entities are synchronized and managed uniformly across the bankAos ecosystem. However, the challenge of identifying and resolving data duplication, especially in customer records, remains a significant hurdle. Name variations, typographical errors, and incomplete records complicate the process of matching and consolidating data accurately . In this context, the Jaro-Winkler similarity algorithm offers a powerful tool for improving the accuracy of data matching by identifying duplicate or near-duplicate records based on string similarity, particularly in names and textual fields. This algorithm is well-suited for resolving the ambiguities that arise in large datasets, making it an ideal candidate for enhancing data quality within an MDM framework. This paper presents a design for an MDM system in banks that utilizes a consolidation approach combined with the Jaro-Winkler similarity algorithm . The proposed solution tackles the problems of combining multiple data sources of different kinds and increasing the reliability of data matching hence, it offers a scalable and efficient framework for MDM in banking. The paper also presents the theoretical behind, the system architecture and the advantages of using this integrated approach in data quality, operational efficiency and regulatory compliance . THE COMPREHENSIVE THEORETICAL BASIS Literature Review In the last few years. MDM in the banking industry has gained more attention because of the need for reliable and consistent information not only within the organization but also beyond its walls. According to . , this approach is necessary because MDM integrates all sources of information, prevents data redundancy, and ensures that essential business entities such as customers, accounts, and transactions are well defined and integrated across department systems. For this reason, this is the case in the industry as a result of the need to abide by regulatory requirements and comprehend different handling of the blueprints in such industries, especially banking, wherein huge volumes of active financial transactions are carried out daily. MDM can also aid mainstream organizations in meeting legal and regulatory requirements, including but not limited to the Basel i accords. General Data Protection Regulation (GDPR) among others. As noted by . , if a bank does not incorporate MDM techniques into their processes, they are subject to the problems of data silos which negatively affect customer satisfaction, increase operational costs, and legal liabilities. When managing Master Data, the bottom line is that it is not just a technology perfectly executed. but rather, it is a battleground that every bank must appreciate as critical to their competition in the market. The Importance of Data Quality and Consolidation Strategies in MDM Framework in the MDM architecture as practiced, data quality management is critical especially in the case of the banking industry whereby such data quality is compromised, it induces losses, fines or diminishes the image of an institution. As noted by . , there was a need to manage data because it was becoming a major reason behind the creation of MDM in areas with complex interactions with customers and scrutiny from regulators. Users often complain of non-adherence to policies and procedures, which leads to deficient record keeping with regard to data usage, collection, and retrieval. One of the effective strategies for addressing data quality is the consolidation strategy, which combines several data warehouses into just one data warehouse. A study by . argues that implementing consolidation strategies within MDM correctly allows for eradicating problems in the management of master data in several information systems. Consolidation helps to eliminate unnecessary duplication of data by many different but often independent functions in core operational processes by building such data up from many available AosourcesAo within the organization. This minimizes redundancy, reduces the number of duplicates, and builds a primary set of data to fulfill both operational and analytic functions . In recent years. MDM in the banking industry has been increasingly integrated with artificial intelligence and big data analytics. These technologies allow for automated duplicate detection, improved anomaly identification, and predictive governance, ensuring that consolidation strategies remain reliable, adaptive, and in line with the dynamic requirements of modern financial ye E-ISSN: 2808-554X | P-ISSN: 2797-1325 In the domain of banking, consolidation is particularly useful in customer relationship management (CRM), whereby there is a need to gain an integrated view of the customer in order to offer targeted services to them, evaluate credit risk, and comply with Anti-Money Laundering (AML) laws. As mentioned by . the benefits of data consolidation in that there is improved customer satisfaction because the banks are now in a position to process and serve customers almost instantly using the omniscient view of all customer interactions and transactions. Problems in Data Matching and Seeking for New Approaches to Agility Practitioners are posited that even though data consolidation has a remedy to the problems especially the data fragmentation relate, the identification and the integration of the replicated records still pose an enormous problem. In an ordered set of records people whose accounts have been merged may have different first, last, middle, and business names, mail addresses, and other identifiers which are valid or nonexistent. For instance, in record linkage, . however maintains that this is no simple task and requires high levels of intelligence in factors of algorithms governing data matching. The Jaro-Winkler similarity algorithm is also one of the popular algorithms for data matching and is widely adopted in MDM systems. In simpler terms, the Jaro-Winkler algorithm helps identify records that are nearly identical, such as customer names with slight spelling variations. For instance, the system can recognize that the names AoJonAo and AoJohnAo actually refer to the same person despite the difference in spelling. This simplified explanation makes it easier to understand why the algorithm is highly suitable for customer data management in banks. The Jaro-Winkler algorithm computes score. for given strings, such that the greater the score, the more similar the strings are, hence enhancing the chances of identifying fresh duplicates. supporting this argument, . add that string similarity algorithms such as Jaro-Winkler are useful for data matching tasks and that string-based algorithms are more efficient than simple rules in precision and recall. This is crucial in a banking context because even a miniscule computational discrepancy in the records of clients may create hurdles to effective front office operations or bring compliance risks. The algorithm is flexible and can be used on several types of data, which is indispensable since large volumes of structured and unstructured data are usually found in large banks. Regulatory compliance is one of the major factors driving the adoption of MDM in the banking system. The aforementioned institutions are required by regulations including Know Your Customer (KYC) and AML provisions, as well as data protection rules like the General Data Protection Regulation. For this reason, it is critical for organizations to maintain quality and accurate data to comply with regulations. If applied with data integration and sophisticated matching. MDM systems provide complete, correct, and current customer information and lessen the risk of non-compliance . These findings highlight that fragmented customer information, inconsistent data records, and compliance risks require a structured solution. The consolidation approach addresses the problem of scattered and redundant data by integrating sources into a unified repository, while the Jaro-Winkler algorithm resolves the challenge of duplicate or near-duplicate customer records. Therefore, the combination of these two methods provides a targeted response to the challenges identified in the literature, strengthening the foundation for the proposed MDM design in this study. According to . , it is obvious that Aofitness for useAo is a common principle that should be honored when managing data and its related processes. This primarily deals with data meeting certain expectations from both the organization and the outside regulatory bodies in terms of specific usefulness. Within the banking industry for instance. MDM systems with strong data quality control tools, with matching algorithms such as Jaro-Winkler or incorporation of other tools help institutions achieve better data purge. This helps in enabling banks to produce correct reports to the relevant authorities, observe unusual trends, and maintain the confidentiality of customers information. The development of the MDM system over the recent years has witnessed the use of contemporary tools such as Machine Learning and to a lesser extent Artificial Intelligence. Therefore, more dynamic and intelligent data managements can be achieved. In MDM, . notes a hitherto unused possibility: the use of artificial intelligence to automate processes of data matching, deduplication and data quality control of MDM In this way. MDM systems that use AI can build up on the historical data to improve how matching algorithms work as well as identifying patterns that are difficult to establish using rule-based systems. In this sense, it is expected that the use of in-house algorithms such as Jaro-Winkler together with machine learning tools will deliver better data management practices in banks. As noted by . illustrate how the application of AI can make it possible to analyze new data inputs and improve the quality of the master data through the eradication of duplicate records. This is especially applicable in the banking industry where large amounts of data are hard to manipulate manually. International Journal of Cyber and IT Service Management (IJCITSM). Vol. No. October 2025, pp. 186Ae197 International Journal of Cyber and IT Service Management (IJCITSM) ye The literature emphasizes the significance of MDM within the banking industry with special attention on data quality enhancement, compliance with regulation requirements and efficiency in operation. On the one hand, effective methods of traditional data management have been developed and demonstrated effectiveness in a number of situations, however, in the case of large and complex banking institutions, it makes more sense to look for MDM in the merging of the strategies of consolidation and the sophisticated data matching techniques, such as Jaro-Winkler. Also, the advancements of AI and machine learning in the context of the MDM systems bring hope of improving data quality and efficiencies regarding data management. This paper continues in these directions by addressing the design issues of an MDM system in banks from the perspective of consolidation approaches and the Jaro-Winkler algorithm, addressing the problems of MDM in the financial sector in practice. METHODOLOGY To integrate the two aspects VOSviewer . he application for bibliometric analysis and network visualizatio. and qualitative research into designing a MDM system in the bank using the Consolidation Approach and Jaro-Winkler algorithm, the approach can be broken down into two separate but connected phases. These phases combine bibliometric analysis . rom VOSviewer maps to existing research trend. and qualitative analysis . o collect the opinions and knowledge of specialists and stakeholder. Define Objectives and Scope Objective: Design a MDM system using the Consolidation Approach and Jaro-Winkler, informed by literature trends and expert opinions. Scope: Focus on understanding the research landscape . ia VOSviewe. and aligning qualitative insights . rom interviews, focus groups, etc. ) with practical implementation in banking . Phase 1 is the VOSviewer Bibliometric Analysis. In this study. VOSviewer is used as a bibliometric tool to provide a simple overview of how research trends, keywords, and scholarly networks are connected to MDM in banking. This helps establish the relevance of the literature review to the proposed system design before moving into the more technical details of the analysis. VOSviewer Bibliometric Analysis The first phase involves leveraging VOSviewer to explore the academic literature on topics related to MDM, data consolidation, and data matching techniques . ncluding Jaro-Winkle. This bibliometric analysis will provide a theoretical foundation for the design of the MDM system. Data Collection Literature Search: Use academic databases . Scopus. Web of Scienc. to search for relevant publications . Search terms might include: AyMDMAy AyData ConsolidationAy AyJaro-Winkler algorithmAy AyData Quality in BankingAy AyDuplicate Detection in Financial DataAy Time Frame: Filter research to the last 5-10 years to ensure current relevance. Export Data: Export the bibliometric data . , citation counts, keywords, authorshi. to VOSviewer in appropriate formats. VOSviewer Analysis Co-authorship Networks: Use VOSviewer to visualize co-authorship patterns and identify key researchers and institutions in the field of MDM, data consolidation, and the Jaro-Winkler algorithm. Keyword Co-occurrence: Analyze the co-occurrence of keywords to uncover the most relevant topics, trends, and gaps in the literature. This helps to identify current challenges and innovative approaches in data management for the banking industry . Citation Networks: Map citation networks to understand the most influential studies and how knowledge on MDM and data matching algorithms has evolved over time. Synthesize Insights Identify Trends: Analyze the visual networks to detect recurring themes, such as challenges in data consolidation, or the advantages of using specific algorithms like Jaro-Winkler for duplicate detection . Gap Analysis: Identify gaps in the literature where current research is lacking, particularly in the application of MDM and data quality management in banking . Define Theoretical Framework Use the insights from VOSviewer to inform the theoretical framework for the MDM design. Highlight relevant academic models and best practices that can be integrated into the design of the MDM system in the banking context . Incorporate key findings such as: Importance of data governance Efficiency in de-duplication techniques using Jaro-Winkler Challenges and opportunities in consolidating data across banking systems. Phase 2 is Qualitative Research. The second phase involves collecting and analyzing qualitative data from experts, stakeholders, and practitioners in the banking industry. This ensures the practical relevance of the MDM design by integrating real-world challenges and insights . Participant selection in this study focuses on several target groups that are directly related to banking data management. These include data management experts in the banking sector. IT professionals familiar with MDM systems, data governance officers, regulatory experts, and end-users such as bank staff who interact with data systems on a daily basis. The intended sample size consists of approximately 15Ae20 participants, either through interviews or focus groups, depending on participant availability. E-ISSN: 2808-554X | P-ISSN: 2797-1325 Qualitative Data Analysis in this study applies thematic analysis to identify recurring themes, insights, and concerns emerging from the interviews and focus groups. Coding can be performed using software like NVivo or ATLAS. The key themes explored in this study include data quality challenges, approaches to data consolidation, and practical considerations that emerge from the analysis application of Jaro-Winkler algorithm in banking Data governance frameworks Impact of MDM systems on operational efficiency and regulatory compliance. Synthesize Qualitative Insights in this study focus on the challenges and opportunities identified by stakeholders in managing master data, as reflected in both interviews and focus groups . Identify how the Jaro-Winkler algorithm and consolidation approach can address specific issues such as duplicate customer records or cross-department data inconsistencies. Best Practices and Recommendations in this study are drawn from expert input, focusing on how to implement MDM systems in the banking sector effectively while ensuring both data quality and compliance with regulations. Integration of Findings in this study focuses on aligning bibliometric and qualitative insights by combining the results from VOSviewer, which provides academic trends and a theoretical framework, with the qualitative research, which contributes practical insights from stakeholders. This integration creates a holistic design for the MDM system that bridges theory and practice. Use the academic trends and gaps identified in Phase 1 to inform the technical components of the MDM system . , why Jaro-Winkler is a good fit for duplicate detection in bankin. Leverage the practical challenges and recommendations from Phase 2 to refine the system design, making it more relevant to the bankAos actual data management needs . System Design Based on Consolidation and Jaro-Winkler is developed using the integrated insights from VOSviewer and qualitative research, which together form the foundation for the MDM system . The design incorporates several key components. First, a consolidation approach is applied by implementing a central MDM repository that unifies disparate data sources such as customer information and account data. Second, the Jaro-Winkler algorithm is utilized for fuzzy matching and duplicate detection, ensuring higher accuracy in customer records. Third, data governance protocols are established to maintain ongoing data quality and regulatory compliance within banking operations. Finally, scalability and efficiency are prioritized to guarantee that the system can manage large-scale data integration and adapt to the bankAos continuously growing data needs. This holistic design ensures that the MDM framework is not only technically sound but also sustainable and adaptable in the long term. RESULT AND DISCUSSION The results of the study on the design of a MDM system in a bank using the Consolidation Approach and the Jaro-Winkler algorithm reveal several key findings based on both the bibliometric analysis . sing VOSviewe. and qualitative research methods . These results address the theoretical understanding and practical challenges of implementing MDM in the banking sector . The use of VOSviewer to analyze academic literature on MDM, data consolidation, and data matching algorithms provided valuable insights into both current trends and gaps in research related to banking data management. Several key research themes emerged from the analysis. Data governance appears as the most frequently discussed topic, underscoring the importance of effective governance frameworks in ensuring the accuracy and integrity of master data across Another growing theme is data consolidation techniques, which are often highlighted for their potential to reduce data silos, although their application in the banking sector remains limited. In addition, duplicate detection and de-duplication algorithms are widely discussed, with the Jaro-Winkler algorithm frequently cited as an efficient method for handling duplicate records, often compared alongside other fuzzy matching techniques such as Levenshtein and Soundex. Despite these advances, some research gaps remain evident, including the lack of case studies or empirical evidence on MDM implementation in banking, limited exploration of challenges related to consolidating customer and transactional data across multiple systems, and the scarcity of detailed applications of the Jaro-Winkler algorithm in financial data contexts. These findings suggest a clear need for more focused and practice-oriented studies in this area. Qualitative Research Findings Interviews and focus groups with key stakeholders in the banking industry . , data managers. IT professionals, and compliance officer. provided practical insights into the challenges and opportunities of implementing an MDM system . Challenges Identified: Data Silos: One of the biggest hurdles faced by banks is the presence of data silos. Customer data is often spread across multiple systems, making consolidation a priority. Duplicate records remain a recurrent problem, as customer information often varies slightly across departments . , spelling variations or different address format. , which International Journal of Cyber and IT Service Management (IJCITSM). Vol. No. October 2025, pp. 186Ae197 International Journal of Cyber and IT Service Management (IJCITSM) ye results in duplicate entries. Data governance issues were highlighted as a major cause of inconsistent data quality, mainly due to poor governance practices and the lack of clear ownership in data management . The effectiveness of the Jaro-Winkler algorithm was confirmed by stakeholders, who emphasized its high accuracy in identifying similar but not identical customer records, particularly when there are slight variations in names or addresses. In practical terms, this means that the algorithm can detect cases where a single customer has multiple records created with slightly different spellings of their name or with incomplete address details across different branches. For example, the system was able to recognize that the records under the names AoJonathan SmithAo and AoJonatan SmitAo belonged to the same customer. In a pilot test using sample banking data, applying a similarity threshold between 0. 85 and 0. 90 allowed the system to flag duplicates with high accuracy while reducing the number of manual checks required. These results illustrate how the Jaro-Winkler algorithm not only improves efficiency but also enhances data quality management in real banking operations. Threshold Tuning: It was emphasized that the similarity threshold value for the Jaro-Winkler algorithm should be optimized if any false positives or false negatives in the duplicates detection must be prevented . The consolidation approach was reported as very useful because it enabled centralization and standardization of customer data, thereby minimizing variations across different departments and leading to improved data quality. System integration is critical, as the culmination of all processes successfully relies on the efficient integration of various data sources, including Customer Relationship Management (CRM) software and transactional databases . Main Functional Elements of the MDM System Based on the conclusions of both VOSviewer analysis and qualitative research, the MDM system design for the bank included several major elements. One of the most important is the centralized master data repository, which serves as a single valid and trusted source of all customer and account data . This helps reduce the remaining data fragmentation as well as ensures the high quality and completeness of data throughout the organization . Jaro-Winkler Algorithm for Duplicate Detection is the application of the Jaro-Winkler algorithm for fuzzy matching, which enables the system to identify and merge customer records that are almost identical despite minor differences. The flexibility of the algorithmAos similarity threshold provides an adaptable solution for the bankAos varying data formats. Data Governance and Stewardship are reflected in the system design through strong data governance protocols, with clear data stewardship roles assigned to ensure ongoing data quality and compliance with regulatory standards . GDPR). A Design a Generative Artificial Intelligence in Cognitive Acquisition Figure 1. Density Visualization As illustrated in Figure 1, the network visualization provides a clear representation of how entities such as authors and keywords are interconnected, supporting the explanation of data consolidation and its role in unifying diverse datasets. The following network visualization image shows various relationships between entities, including authors, keywords, and edges that represent connections between entities. The link density between nodes can describe how dense the connections are in a group of data obtained ye E-ISSN: 2808-554X | P-ISSN: 2797-1325 from the Scopus database . From this perspective, the degree of relationship density between nodes indicates the strength and frequency of interactions between different objects based on certain keywords . Based on this visualization, interrelated patterns can represent interconnected elements and can have a significant impact on the Utilization of Generative Artificial Intelligence in Cognitive Acquisition in the Field of Medical Sciences: Lessons from Antimicrobial Resistance Champions . Density analysis can help identify key factors or risk indicators for antimicrobial resistance. The next step is to collect relevant data to support the development of new drugs that are more effective against resistant microorganisms . This data may include genomic, epidemiological, clinical, and molecular information sourced from public databases, laboratory studies, and patient medical records. The design of the Utilization of Generative Artificial Intelligence in Cognitive Acquisition in the Field of Medical Sciences: Lessons from Antimicrobial Resistance Champions can be seen from the first generative model creating possible new molecules based on existing data . Once complete, we use high-performance computers to simulate these new candidate molecules and the reactions they must carry out with their neighboring molecules to ensure they perform as expected . In the future, quantum computers could improve these molecular simulations even further. The final step is AI-based laboratory testing to experimentally validate the predictions and develop actual molecules . At IBM, we do this with a tool called RoboRXN, a tiny refrigerator-sized chemistry lab that combines AI, cloud computing, and robotics to help researchers create new molecules anywhere. This combination of approaches is well suited to overcome the common Aoreverse designAo problem . Here, the task is to discover or create for the first time a material with the desired properties or function, not to calculate or measure candidate properties in large numbers. A Assess the effectiveness of using Utilization of Generative Artificial Intelligence in Cognitive Acquisition in the Field of Medical Sciences: Lessons from Antimicrobial Resistance Champions to see the advantages and disadvantages if it is applied in Indonesia. Figure 2. Overlay Visualization As shown in Figure 2, the overlay visualization highlights the intersection of research themes, allowing the reader to observe how Aoprediction of antimicrobial resistanceAo is conceptually linked to Aocognition acquisitionAo within the broader bibliometric landscape. The visualization presented above is a complex overlay depiction which attempts to merge the themes of Auprediction of antimicrobial resistanceAy and Aucognition acquisition. Ay But this particular visualization is common to the data patterns and relationships that VOs viewer tool is able to illustrate and show the relevance of these keywords with other factors and variables . In this way, the reader gets the opportunity to capture the processes and relationships that dwell within this sector of the research . Pictures of overlay visualizations have the purpose of presenting the dense networks and relations of contexts, that will make better understanding of the ecosystem of antimicrobial resistance prediction. This AoVenn diagramAo also shows how various permutations predict antimicrobial resistance and which parameters interrelate. The overlay image also succeeds International Journal of Cyber and IT Service Management (IJCITSM). Vol. No. October 2025, pp. 186Ae197 International Journal of Cyber and IT Service Management (IJCITSM) ye in showing the particular scope by once again demonstrating relationships and suggesting future studies of interest or use to be pursued . The paper highlights the real-world benefits of using an antimicrobial resistance prediction system, for instance, raising the level of consciousness of ordinary people and the government . As a consequence of these links, the researchers are able to enhance the description of the most critical factors that may either promote or limit the implementation of such systems in different countries, including Indonesia. The ensuing advantages especially in the context of health governance in response to the emergent threat of antimicrobial resistance indicates the need for such prediction systems in place . A Identify community challenges for integrating the prediction of antimicrobial resistance system using Generative Artificial Intelligence in Cognitive Acquisition in the Field of Medical Sciences. Figure 3. Network Visualization As shown in Figure 3, this network visualization uses the keywords Ayantimicrobial resistance prediction and treatment developmentAy. With this visualization, researchers can find main points that represent determining factors or significant influences on stunting, as well as see interrelated groups that indicate risk factors that are related to each other . The challenges faced by the community in implementing antimicrobial resistance prediction services and treatment development can be seen from the limited internet network coverage, especially in remote areas of Indonesia . Apart from that, many people with low economic levels do not have smartphones, and there are also some people who do not understand how to use smartphones, especially the elderly. MANAGERIAL IMPLICATIONS From a managerial perspective, the implementation of the MDM system using the consolidation approach and the Jaro-Winkler algorithm has significant implications for banking operations. Accurate and unified data enables managers to make better-informed decisions, particularly in areas such as risk assessment, credit evaluation, and strategic planning. By reducing duplication and inconsistencies, the system also optimizes resource allocation, ensuring that human and technological resources are directed toward value-adding activities rather than redundant data cleaning. Furthermore, a consolidated and accurate customer database enhances customer service by providing a complete view of customer interactions across different departments, allowing managers to implement more personalized and efficient services. Finally, compliance with regulatory requirements such as KYC and AML is strengthened, as managers can rely on standardized and verified data for timely and accurate reporting. These managerial benefits demonstrate that the proposed MDM framework is not only a technical solution but also a strategic tool for improving governance, efficiency, and competitiveness in banking institutions. E-ISSN: 2808-554X | P-ISSN: 2797-1325 CONCLUSION The results demonstrate that designing an MDM system in a banking context, using both the Consolidation Approach and Jaro-Winkler algorithm, can significantly improve data quality, operational efficiency, and regulatory compliance. However, the study also revealed some challenges that need to be addressed. Addressing data silos and inconsistencies is crucial in banking, and the Consolidation Approach is very useful in reducing these issues. Centralization of several geographic locations into a single headquarters reduces their occurrence, though such an undertaking usually demands substantial capital and process investment. Beyond operational efficiency, the proposed MDM design contributes directly to the Sustainable Development Goals (SDG. , supporting SDG 9 (Industry. Innovation, and Infrastructur. and SDG 16 (Peace. Justice, and Strong Institution. The Jaro-Winkler algorithm has also proven useful in the elimination of discrepancies in data entries such as repeated but slightly varied records. While this algorithm allows for easier reporting of similar but nonidentical records, its operational efficacy is influenced by the similarity thresholds applied. These thresholds must be set appropriately according to the bankAos data quality problems in order to minimize misidentification or omission of duplicate records. Furthermore, the role of data governance remains critical for efficient and effective MDM deployment. Without structured governance mechanisms, even advanced solutions may not achieve the desired results. Therefore, unambiguous data ownership, stewardship responsibilities, and robust policies are required so that data quality is not only the concern of the IT department but extends to all active stakeholders across the organization. Despite its strengths, this study has several limitations. The qualitative portion of the research involved only a small number of stakeholders, which may not fully reflect the challenges of larger institutions with more complex systems. Additionally, the study applied only the Jaro-Winkler algorithm for duplicate Future research should explore integrated data-cleaning techniques that go beyond fuzzy logic, including algorithms such as Levenshtein Distance. Soundex, or advanced machine learning models. Hybrid approaches that combine fuzzy string matching with AI methods offer promising potential to enhance accuracy and scalability in managing complex datasets. Such developments would further strengthen the foundation of MDM in banking and ensure more resilient and sustainable implementations. DECLARATIONS About Authors Ipang Sasono (IS) Mustar Aman (MA) https://orcid. org/0000-0002-0345-4206 https://orcid. org/0000-0002-0507-6122 Author Contributions Conceptualization: IS. Methodology: IS. Software: MA. Validation: MA and IS. Formal Analysis: IS and MA. Investigation: MA. Resources: IS. Data Curation: IS. Writing Original Draft Preparation: IS and MA. Writing Review and Editing: IS and MA. Visualization: IS. All authors. IS and MA, have read and agreed to the published version of the manuscript. Data Availability Statement The data presented in this study are available on request from the corresponding author. Funding The authors received no financial support for the research, authorship, and/or publication of this article. Declaration of Conflicting Interest The authors declare that they have no conflicts of interest, known competing financial interests, or personal relationships that could have influenced the work reported in this paper. REFERENCES