key: cord-0592500-dzjfrpz0 authors: Rouhollahi, Zeinab title: Towards Artificial Intelligence Enabled Financial Crime Detection date: 2021-05-23 journal: nan DOI: nan sha: 23dd8c80851032c88ecb6eb2bf116eaceddbb0a7 doc_id: 592500 cord_uid: dzjfrpz0 Recently, financial institutes have been dealing with an increase in financial crimes. In this context, financial services firms started to improve their vigilance and use new technologies and approaches to identify and predict financial fraud and crime possibilities. This task is challenging as institutions need to upgrade their data and analytics capabilities to enable new technologies such as Artificial Intelligence (AI) to predict and detect financial crimes. In this paper, we put a step towards AI-enabled financial crime detection in general and money laundering detection in particular to address this challenge. We study and analyse the recent works done in financial crime detection and present a novel model to detect money laundering cases with minimum human intervention needs. Financial crime is an unlawful act and may involve various types such as fraud (e.g., cheque fraud and credit card fraud) and money laundering. Recently, financial institutes have been dealing with an increase in financial crime. For example, as reported by Forbes 1 , in April 2020 United Kingdom experienced a 33% increase in Financial crime and fraud. Similarly, in the United States, IC3 2 (Internet Crime Complaint Center) reported as many fraud reports (by the end of May 2020) as it had in all of 2019. This could result from the global shift towards a digital society due to COVID-19 3 pandemic. In this context, financial services firms should improve their vigilance and use new technologies and approaches to identify and predict financial fraud and crime possibilities. This task is challenging as institutions need to upgrade their data and analytics capabilities to enable new technologies such as Artificial Intelligence (AI) to predict and detect financial crimes 92 . In this paper, we put a step towards AI-enabled financial crime detection in general and money laundering detection in particular to address this challenge. We study and analyze the recent work done in financial crime detection and present a novel model to detect money laundering cases with minimum human intervention needs. The rest of the section organized as follows: In Section 1.1, we present an overview of this paper. We present the problem statement in Section 1.2. To better understand the research problem, in Section 1.3, we present a motivating scenario in money laundering. In Section 1.4, we present the contributions of this paper before concluding the Section with the summary and the paper outline in Section 1.5. In recent years, several regulatory authorities have increased their attention on detecting and preventing financial crimes. This is specifically important for banks and financial institutions as they have several business processes 13, 23, 26 and procedures that are a good source for criminal actors. Among these processes, we can mention cash depositing, overseas money transfer, and loan re-payment. If specific procedures occur in all banks and financial institutions internationally, then the amount of money available to commit financial crimes will decrease; therefore, putting specific procedures to recognize and prevent criminal activities is one of the challenges that banks face. Several AI-enabled techniques aim to detect financial crimes in different sectors such as the bank and insurance industry, commodity exchange, security systems, stock markets, and money laundering to address these challenges. Emerging research and technologies which use big data analytics have helped significantly in collecting data and analyzing it. This, in turn, has led to a shift from traditional data collection methods to computer-based data gathering and analysis. This means that a vast amount of data being gathered and stored in each area, and organizations need proper strategies for analyzing these data efficiently and appropriately 46 . Traditional approaches of detecting financial crimes would mainly lead police forces to look for so-called "street crime". In contrast, the development of new techniques using machine learning has made it possible to have a deeper look and search for high-level financial crimes committed by "white-collar criminals" as well 32 . Financial Crimes have a widespread effect on economic, political, and social aspects of societies both in the national and international dimensions of society 4 . While putting specific procedures for preventing, detecting, and reporting related cases has been an ever-growing concern of the banks, failing to do so correctly may lead to irrecoverable reputation losses for banks. Moreover, it may lead to a large amount of loss due to being fined by regulators and related authorities. In particular, financial crimes may have a significant impact on the economy. For example, it may lead to challenges between political parties. Lack of proper regulations in financial markets may have two effects: (i) it may avail new opportunities for financial crimes; and (ii) it may endanger the health and steadiness of international financial structures 61 . Considering new methods of financial exchange and the amount of financial data generated daily, it has become almost impossible to rely on traditional methods for detecting financial crimes. In this context, extracting relevant data from documents (such as PDF forms 66 ) and analyzing crime patterns is considered to be a key challenge for law organisations 1 . In this context, Artificial Intelligence (AI) and machine learning can enable machines to comprehend and learn to facilitate the detection and prediction of financial crimes. In this paper, We study and analyze the related work in AI-enabled crime detection and present a novel model to detect money laundering cases with minimum need for human intervention. Financial crimes can happen in different ways, such as money laundering, fraud, electronic crime, terrorist financing, and bribery and corruption. Money laundering is known as the illegal process of concealing the source of money which has been obtained through criminal activities and putting this money into legitimate financial systems 62 . In recent years, several works have been done in money laundering aiming to detect suspicious transactions. However, the need for a practical model which takes into account the current features of transactions that are happening in banks on a day-to-day basis and takes into account all aspects of each transaction is obvious. A significant problem in this area is that the number of cases reported but not a money laundering case is high, which means that much time and effort is wasted in recognizing the unusual activities that are not a money laundering case. In this paper, we present a novel model to detect money laundering cases with minimum need for hu-man intervention to address this challenge. This model will use both classification and anomaly detection to detect cases of money laundering. Current models used by banks may have two problems. Firstly, they are mainly rule-based; therefore, not all of the cases covered, and the rules might miss some of the cases. Secondly, current models have a high rate of false positives, meaning that the systems will alert for the cases that are not money laundering cases; therefore, too much time and effort could waste each day. Our proposed model aims to cover these deficiencies using a hybrid model where both classification and anomaly detection will apply to transaction data. The goal is to intelligently combine the two models to ensure that while most money laundry transactions are detected, there are not too many false-positive cases reported to decrease the amount of the effort done and increase the model's efficiency. This Section discussed the background, key research issues, and contributions overview of this paper. The remainder of this paper organized as follows. Section 2: This Section will have a comprehensive look at the literature related to financial crimes and money laundering. We will investigate current methods used to detect financial crimes and specifically money laundering. We will have a deeper look at the machine learning methods used in this area and categorize different AI-enabled methods to detect money laundering. Section 3: This Section will present the methodology that we are using in this research. It will describe the structure of the data and the features used for transactions. The classification model and anomaly detection that we apply to the data will also be described here. Section 4: In this Section, we describe the outcome of the model applied to our dataset. The tools and techniques used for the implementation and the results, including the model's accuracy, are presented here. Section 5: In this Section, we conclude the paper by reviewing the results and looking into future areas for improvement. In today's competitive financial market, detecting fraudulent financial activities is one of the main success requirements of any financial organization. Banks and financial institutions are continuously working on novel and intelligent applications to discover such illegal activities because of the regulatory obligations and the extraordinary importance of obtaining a positive reputation amongst customers and achieving their trust. Financial crime includes different types of fraud; examples are fraud via check and credit card, health care card, and point of sales. The act done in these crimes involves tax violations, identity thefts, cyber-attacks, and money laundering. In the next section, we will dive deeper into the definition and implications of money laundering. Money laundering is known as an illegal process that happens through banks and other financial channels. In this process, large amounts of money are transferred, and the source of the money is criminal but is concealed through several complex banking transactions and commercial money transfers 36 . There can be several criminal activities associated with money laundering; this means that the criminals committing the crimes (such as smuggling, drug sales, forgery, terrorism, fraud, human trafficking, tax evasion, child laboring, and child sexual abuse) will conceal the sources of their money using banking and financial transactions to disguise the sources of their crime 33 . The growing use of digital channels and the invention of digital money concepts such as bitcoin have added to the risk of the money sources concealed throughout these channels. A recent study shows that the use of Internet sources and the web has helped criminals launder the sources of criminal money 84 which mean that in the upcoming years, with the invention of new money concepts and increased use of the Internet, the amount of money Laundering could potentially increase. The amount of money that was transferred in money Laundering is not known precisely. However, it is estimated to be between 2 to 5 percent 29, 83 of the overall Gross Domestic Product (GDP) throughout the world, which is a considerable amount compared to other financial activities. So the governments need to find and stop these activities. For instance, during the year 2016, 8776 money Laundering cases have been reported. Among these transactions, 72 percent reported by the banks and 23 percent by PSPs (Payment Solution Providers) 2 . Considering the above figures, governments should monitor these activities and prevent criminals from committing these types of crimes by correctly detecting these activities. Globalization and the fact that the financial markets in different countries have been brought closer alongside the ability to do online transactions 24 hours a day have increased the risk of Money Laundering. The ability to invest in other countries has made the opportunity to do what may not be possible in one's country and through clear financial systems 29 . On the other hand, it is tough to recognize money laundering activities. Large amounts of daily transactions and different and complicated methods that money launderers will use to conceal the sources of data make it hard to investigate these activities. Vast amounts of data and multiple legacy systems make it hard for banks to recognize money Laundering activities appropriately. Further to that, several countries, especially among the developing countries, do not follow international rules and regulations on financial activities, making it a good place for criminals to hide the origins of their money throughout several transactions and banking activities. Furthermore, money launderers are changing their behaviors each day, making it almost impossible to place specific rules to prevent those activities because they become aware of the rules soon and find other mechanisms for money laundering 30 . Several international authorities are in charge of controlling and preventing money Laundering. For example, FATF, the Financial Action Task Force (on Money Laundering), is an intergovernmental organization founded in 1989 and aims to develop policies to fight money laundering and financial terrorism activities. APG, the Australia Pacific Group on money laundering, is also an organization aiming to implement international standards to fight against money laundering, financial terrorism, and financing weapons. In Australia, AUSTRAC, the Australian Transaction Reports and Analysis Centre, is the government agency responsible for identifying and preventing money laundering, tax evasion, fraud, and terrorism financing. In recent years, several rules and regulations were imposed on banks and financial institutions in Australia to know their customers and prevent money laundering activities. The criminal actors involving in money laundering include criminal actors and professional money launderers (PMLs). These PMLs are the people who act in a professional capacity, such as a lawyer or accountant but knowingly get involved in the money laundering process to conceal the sources and show the criminals the methods to escape legal sources 2 . Money launderers would use several tools to disguise the sources of money. Several institutions and different methods of transferring money are available, so it is easy to wash the money. Further to that, as they become aware of the tools and patterns that banks are recognizing, they change their methods for concealing the source of their money 35 . In order to be able to prevent money laundering, we need to know how it takes place. Generally, the Money laundering process includes three steps: Placement, Layering, and Integration. Placement includes placing the money obtained from criminal activity into banks and financial systems by depositing the money through different geographical locations and legitimate sources. Layering contains concealing the sources of the money by doing many transactions and involving offshore accounts and complex investment vehicles. In the Integration phase, money launderers will transfer the funds to the owners or criminal actors. This is often done as a form of investment or tangible goods like luxury cars and jewellery 72 . Detecting money laundering activities and preventing them is an essential task of both governments and financial institutions. In recent years, several new regulations have been imposed on financial institutions to report any suspicious activity which could potentially be a source of money laundering. Banks and financial institutions have been obliged to put certain processes and procedures in place to report and detect any activity which can lead to money laundering 75 . In the academic area, several studies have been done, and this subject has been to the attention of scholarly published papers. In a recent study, six areas have been defined for the works that have been done in the field of money laundering, which include "anti-money laundering and its effectiveness, effects of money laundering on the economy and other fields, the role of different actors in this area and its importance, the magnitude of money laundering, and new opportunities for money laundering and its detection" 83 . The solutions used by financial institutions are mainly based on statistical variables like mean and standard deviation. This is not efficient because it consumes a lot of effort from humans to detect suspicious cases 54 . Increasing precision is one of the main challenges in the area of money laundering detection. Each day, several transactions are being generated through different banking channels, and the rule-based software systems used in the bank recognize a significant number as unusual. In contrast, the detected unusual contain lots of false-positive cases that need further human investigation to distinguish true-positive from false-positive transactions. This final step is defined as recognizing "Suspicious" transactions out of "Unusual" ones. Therefore, increasing precision is an essential aspect that some researchers have looked into 54 . It is worth adding that an "Unusual" transaction can be any transaction that does not follow the regular pattern of the rest of the transactions. This can be a high amount of money being transferred in one transaction or transferring money to a country under sanctions. However, just being "Unusual" does not mean that this is a "Suspicious" transaction that has been used for criminal activities. If a transaction is known as "Unusual" in bank systems, it will go into human investigations. Suppose it is recognized as "Suspicious". In that case, it will be sent to AUSTRAC along with other information on the account and account holder to decide whether it was a financial crime or not. Based on a model presented by Wang et al. 30 , the overall Anti-Money laundering (AML) process which is performed in banks and financial institutions can be described as below steps: Intelligence: In the intelligence phase, we need to communicate with legacy systems to see the current rules and regulations proposed for detecting money laundering. We also need to have accurate information on accounts and their everyday transactions. Design: In the design phase, we need to assess whether behavior in an account is usual or unusual. This can be de-cided based on the amount and type of the transaction or whether there was a movement in the historical average amount of the transactions. Choice: In the choice phase, we need to make a warning report if any of the parameters in the design phase were not expected. We also need to review the transaction history to see what the behavior was during the time. Review: In this stage, human investigators and experts from the bank will assess whether the suspicious behavior was a money Laundering activity or it was just an unusual behavior of the account. As it can be seen in Figure 1 , we can categorise different works, research methods, and machine learning applications in the domain of money Laundering detection into unsupervised approaches, including clustering methods such as 53, 87 , anomaly detection applications such as 94 , and rule-based models such as 51, 65 , as well as supervised applications including classification and prediction methods such as 45, 59, 71, 82 . In the rule-based models, as will be described in detail in the following subsection, we will identify unusual transactions for further investigation by assigning some rules on transaction features. Such rules can be numeric thresholds on features like transaction amounts and binary flags such as specific sources and destinations. The other approach is clustering, which segments customers and transactions based on their similarities in their features, making it more efficient to dive into each segment and analyze the transactions of each segment. As shown in Figure 1 , some researchers combine clustering and rule-based approaches and apply specific rules on the transactions of each cluster that are more similar to each other. On the other hand, some supervised applications aim to detect suspicious transactions based on the historical data of other transactions. In classification models, researchers try to use different classifiers to predict the transaction based on training the model with a large dataset of labeled transactions, i.e., transactions with some features labeled as suspicious or non-suspicious. There are also some hybrid works by a combination of clustering and classification models that aim to label transactions within each cluster. Moreover, anomaly detection applications compare transactions with the average pattern of other transactions to detect unusual ones. In the following sections, we will review some works within each of the mentioned categories. To detect money Laundering transactions, the initial step of all banks and financial institutions is the extraction of so-called "Unusual" transactions, which will lead to discovering suspicious transactions. The software tools which are used for AML (Anti-Money Laundering) in banks are usually rule-based. This means that these software programs are designed based on pre-defined rules and thresholds of average and standard deviation to detect Laundering transactions. In other words, the conventional approach to detecting unusual transactions is applying some rules about a specific transaction attribute or a set of them 73 . These attributes may contain transaction features such as transaction type, amount, statement, time, location, frequency, origin, and destination. For instance, any transactions with an amount above a specific threshold might be extracted as unusual. In a similar pattern, transactions with some pre-defined origin or destinations such as FATF blacklisted countries might be detected as unusual. Another example is about any transaction that includes specific words in its statement like terrorism activityrelated words, i.e., names of specific years or locations. There are also rules made by combining some features such as any cash transaction above a specific threshold in Australia or any repeated transaction to a specific account (a Khan et al. 51 worked on detecting suspicious transactions by using statistical rules. They compared numeric transaction features such as amount and frequency with their average quantities. Whenever the deviation is above a certain amount, the rule will extract the transaction as a suspicious money laundering transaction. Rajput et al. 65 and Khanuja et al. 49 also set some rules on a combination of transaction features including amount, type, origin, and destination to detect suspicious transactions and then investigating these transactions by banking domain experts. One of the main tools and techniques to apply a set of rules to transactions is using the decision tree model shown in Figure 2 . For example, Liu et al. 57 applied this model on clusters that are provided by k-means to identify suspicious transactions within each cluster. Wang and Yang 86 also applied a decision tree with some risk-related rules on the customer accounts. Several works used clustering models for detecting money laundering. Such models apply various clustering methods to the transactions and, after discovering the most efficient clusters of transactions, investigate each cluster as well as detecting the outliers 71 . As there is a massive dataset of transactions for each customer and each transaction includes many features, we need a clustering method with a triumphant performance with large and multidimensional datasets 27 . K-means is the most commonly used clustering method with excellent performance with such large and multi-dimensional datasets; however, the original version of k-means is not successful in detecting outliers. In other words, such outliers might deviate the resulted clusters, and therefore, there are some modified k-means methods designed to be able to detect the outliers 34 . Le Khac et al. 54 have combined natural computing and data mining techniques to propose a solution to detect some generic patterns for recognizing money laundering. One of the main applications of machine learning in detecting fraudulent transactions and precisely transaction suspicious to money Laundering is classification. This application works by labeling different transactions with a binary label, identifying whether a transaction is suspicious of Laundering. Each transaction is converted to a vector model with multiple features. These features contain some numeric attributes such as transaction amount, frequency, and time. They also include other descriptive non-numeric attributes such as transaction type (cash/credit), source, destination, and method. In some research, transaction features are used in the classification application, and customer features are considered. In other words, apart from transaction attributes, some other features contribute as the input of classification, which includes the customer credit score, customer average spending on specific domains. This includes gambling as well as socio-demographic features of the customer, including their age, income, education, etc 44 . There are money laundering detection applications that use "Support Vector Machine", "Correlation Analysis", and "Histogram Analysis" to detect money laundering and identify money laundering cases. The dimensions used in these methods include "Customers", "Accounts", "Products", "geography" and "Time" 55, 93 . Labeling can be hard for Laundering as the proportion of the laundering transactions is deficient compared to the whole transactions. A simple method for labeling a transaction as suspicious in supervised methods is selecting the suspicious transactions from those marked by domain experts. In contrast, general transactions which exist in the pool of transactions can just be assumed as usual because the portion of the suspicious transactions from all transactions is low 72 . The data used in different studies vary extensively. Most researches in AML either use real small datasets or simulated data sets, and only a few researchers have applied the model on large real data sets. Some researchers study a specific type of transaction as money Laundering cases, for example, international fund transfers or a large amount of cash deposits 48 . In a recent study by IBM corporation and MIT university, classification was used to detect money Laundering. The model proposes that analyzing AML data is a classification problem. We need to classify a small number of illicit transactions in large massive datasets of all transactions. Each node transaction has 166 features, and 94 of them are presenting some local information about the transaction like time step, the number of inputs and outputs, transaction fees, output volumes, and the rest of the features comes from aggregating transaction features. This model uses a time series of bitcoin transactions, directed payment flows, and node features. The model applies a binary classification to illicit transactions and applies Logistic Regression, Random Forest, Multi-layer Perceptrons, and Graph Convolutional Networks on the data. The results show that Random Forest works best 52, 88 . One of the approaches in detecting transactions suspicious to be money laundering is using the ML application of Anomaly Detection to recognize the transactions which are not expected and show deviation from the expected behavior of the transactions. In other words, the technique used for anomaly detection should be able to identify the unusual behavior of each customer account 31 . Anomaly detection is usually adequate to detect abnormal behavior or pattern 56 . In the case of financial frauds, this is even more useful because complicated schemes are used to avoid security protocols in detecting illegal transactions 10 . This application works similar to an advanced rulebased model that aims to compare the features of entities with some statistical parameters of features, i.e., the average and standard deviation, and then extract the observation with more than a specific deviation as unusual. For instance, for transactions of a specific customer, the first average and variance of various attributes such as the amount and the number of transactions per day are being calculated. Then a threshold will be set, such as a 300 percent deviation. Then any transaction with a deviation more than the assigned threshold to the average or variance will be extracted as unusual. The mentioned attributes might be any numerical features such as amount, time, and frequency. Similarly, by setting higher thresholds, the model will extract suspicious transactions 70 . In a new anomaly detection method called isolation forest or iforest, anomalies are detected through isolating instances. This method achieves this aim by working on the attribute values that are different from the regular instances and without calculating the distance and density measure 56 . In order to detect money Laundering, some academic papers have studied transaction features while others have studied both transaction and customer attributes. Using customer features will add value to the model, but it will usually incur privacy issues, making it hard to gather and use the data. Jullum et al. 48 introduces a machine learning method for detecting suspicious money Laundering cases, and both transaction features and customer features are used in this work. In contrast, some other works have just worked on the transaction features. The critical point on selecting the transaction features versus customer features is the privacy issue. Customer features usually include data on the customer profile, which needs to be masked to prevent privacy policy issues. In contrast, this issue is not as critical in working with the transaction features. There are several features associated with each transaction type. Different works in this area have considered different features. While most of the work done on anti-money Laundering use transaction features 58 , some of them also use the data available on the customer 30, 48 . For example, Jullum et al. 48 propose a machine learning method for detecting suspicious money Laundering cases using both transaction and customer features. The model used in this paper trained to forecast the probability of a money Laundering case using the background data related to the sender and receiver of the transaction, their previous financial behavior, and the transaction history related to each of them. An-Le Khac et al. 55 presented a model that tries to analyze transactions in the banks related to investment. Two essential characteristics in this analysis are the frequency of transactions and the value of each transaction. The model uses two parameters. One is the proportion between the redemption value and the subscription value condi- tional on time. The other one is the proportion between a specific value and the total value of an investor's shares conditional on time. The model then applies a clustering technique for these parameters on the fund level and investor level. The model will then use neural networks (back-propagation) and train them on suspicious and not suspicious cases. A decision tree is also built to train the model based on time. The results are then evaluated, and it is aligned with the needs of the AML unit. The overview of the model can be seen in Figure 3 . The paper uses these steps to analyze the data: (i) Data Pre-processing: Extracting and cleaning-up the raw data sets and building a data warehouse of customers and transactions; (ii) Data Mining: Using clustering and classification techniques to analyze transactions: extracting the transaction records for investigation, associating multiple transactions to a specific account to know more about the financial activity of that account, building suspicious clusters of customers by clustering similar transactions, and classifying customers in pre-defined categories of risk; and (iii) Knowledge Management: Putting together the results of the mining process as well as the knowledge from the domain experts, results will be collected, stored, and analyzed 3 . Also, interpretable rules and knowledge are generated at this stage. There are several types of risks associated with a transaction in terms of money laundering, which can be investigated. Static risks are those related to the transaction and the account holder, while dynamic risk refers to the account's behavior during the time. Static risk includes risks related to the account holder (the person or company holding the account), type of the transaction or product, which can be cash or digital transfer, geography risk, which refers to the location source and destination of the transaction (e.g., FATF listed countries), or amount risk which refers to the amount of the transaction. On the other hand, dynamic risk refers to the unusual amount of fund movement related to that account, which refers to the short-term and long-term moving average of the account associated with the transfer 30 . In order to reduce these risks, we will associate each risk with its relevant transaction features and identify features of the transaction that will help us reduce these risks. Based on the risk model as well as the transaction features mentioned in the literature, features such as the time of the transaction, type of transaction (cash, money transfer, wiring), entity (Person or company doing the transfer, average person, or PEP), destination account, geographical location of the origin and the destination of the transaction, amount of transaction, accumulated fund flow, and accumulated transaction amount recognized as critical for money Laundering detection in the area of anti-money laundering and financial crime detection. The transaction features can be seen in Figure 4 . While some works have just concentrated on transaction features, others have used some features associated with the customer or the account. These works have not just concentrated on the customer, but instead looked at both customer and transactions simultaneously and applied the models on both. Le Khak et al. 55 have used redemption value and subscription value and identified two new factors using the proportion between these features. In this research, there is no direct mention of the customer features, but each transaction's details are associated with the account detail related to each customer. Sudjianto has mentioned transaction features such as the total amount of transactions, transaction type, frequency, average amount, and maximum during a time (for example, a week), and customer features such as customer profile, income amount, etc 76 . Labeling the training data is one of the most important parts of machine learning models, i.e., classification applications used in money Laundering detection to increase different accuracy indexes such as accuracy, precision, recall, and F1-score. This labeled training data will play a significant role in the accuracy of the model. In other words, having not sufficient data, the model might lead to a low prediction accuracy which means the whole model will not be reliable enough. Labeling training data is indeed one of the most significant obstacles to deploying machine learning systems. The simplest way to approach the labeling would be hand-labeling the training data; however, hand-labeling is expensive and time-consuming. It would help if we could find relevant people and, in the case of financial crimes, subject matter experts (SMEs) to label the data, which will be hard to find the resource and also time-consuming and expensive 67 . Creating the training data set is expensive, especially when it needs expert knowledge. Regarding an accurate money Laundering classification model, for example, we need a dataset of more than 100000 rows, i.e., the transactions with several features for each of them. A label identifies whether or not the transactions are suspicious to be a money laundering transaction. Labeling hundreds of thousands of transactions with so many features by the banking domain experts will be much time-consuming, expensive, and generally inefficient. Therefore, working on alternative methods for labeling these transactions by deriving insights from labels generated by an expert has a significant impact on the model efficiency. There are several ways that companies use to label the data; some of them hire large groups of people to do this for them manually; or use classic techniques such as active learning 74 , transfer learning 64 , and semi-supervised learning. Other sources of labeling the data can include using crowd-sourcing methods for labeling the data 9 and setting rules and heuristics for labeling 67, 68 . The challenges of hand labeling the data are not limited to being expensive and time-consuming; other challenges also exist. Different sources of labeling the data can be conflicting and have different results. Also, the accuracy of some labeling methods is unclear, which makes it hard to estimate the overall accuracy of the label 67 . Methods such as crowd-sourcing have their problems as tasks have non-boolean responses and workers have biases on positive and negative tasks 47 . Some labeling methods generate probabilistic training labels using the agreements and disagreements between various sources of labeling the data. Specifically, in a model called Snorkel, developed at Stanford University, the labels will be determined based on the result of several simple labeling functions written by subject matter experts, and the model will determine the weight of each labeling function based on the agreements and disagreements between the labels for each data point 67 . An overview of the snorkel model can be seen in Figure 5 . The accuracy of this model is highly dependent on the accuracy of its labeling functions which consist of various rules, functions, or a set of rules and functions that automatically label the data based on their features. For example, in suspicious labeling transactions, based on the features extracted for each transaction, each labeling function might be a rule or a set of rules by assigning some thresholds to the features such as transaction amount, transaction statement, origin, destination, customer type, credit, and cash. This model has an excellent performance in dealing with large datasets by applying accurate labeling functions 67 . In this section, we will go through the methodology used in this research to detect and prevent money laundering cases. We leverage a novel intelligent hybrid pipeline including both supervised and unsupervised machine learning applications. We aim to decrease false-positive detection of suspicious transactions (i.e., the number of transactions that the model captures as suspicious but not a money laundering case will decline). As illustrated in Figure 6 , we aim to predict the suspicious transactions by training a supervised classification model in the first step. We also detect unusual transactions by training an unsupervised anomaly detection in the second step. Finally, by implementing a logical AND between the results, the model will detect unusual transactions as whatever is captured by both applications. This will lead to fewer suspicious detected transactions. As mentioned in section 2, one of the drawbacks of the state-of-the-art works of money laundering detection is the low precision of the results; we aim to focus on precision by applying the mentioned logical AND between the results of two applications. We start with curating the dataset in a banking sector, which contains a set of transactions with specific attributes for each transaction. We aim to label our transactions to demonstrate whether they are related to a money laundering case or not. In order to label the transactions, as will be explained in detail in Section 3.3.1, we are using banking analysts to help label part of the dataset. Then, we use learning techniques to label the rest of our large dataset automatically. We will discuss the labeling functions and auto-labeling methods in this section. Once we labeled the transactions, We will apply a classification model to detect suspicious transactions based on the labeled data and measure the accuracy results. We will also apply Anomaly Detection to capture transactions' features that are unusual transactions and compare the results with the transaction labels to calculate the accuracy of anomaly detection applications, using the same indexes. Finally, we apply a logical approach between two classification and anomaly detection applications and measure the impact on the accuracy results. The final goal is extracting suspicious transactions with minimum human intervention. Once we detect suspicious transactions, the bank will check other transactions of the same source and detect more transactions to be labeled by the analyst. Labeling the data is an essential step in our proposed approach. The accuracy of the model, including both classifications and anomaly detection applications, relies on the training data that we feed into it; therefore, wrong or misleading labeled data in the training data set might result in incorrect results. Having relevant labeled data is not always a straightforward task, as the actual sources for labeling the data might be inconvenient, and the replacing methods might not always be suitable. Therefore, finding and applying a proper method for labeling the data is critical in data-driven approaches. Especially in the banking industry, this will be more significant as the data is critical and not publicly accessible, making it even harder to access the properly labeled data. We aim to use three types of data labels for money laundering cases. The first one is for detecting unusual transactions from the usual ones, which includes any transaction that banking systems will recognize as different from regular transactions. The second label is for detecting suspicious transactions from the non-suspicious ones. Bank financial crime experts currently do this step. It is a time-consuming task, and the current rule-based software solutions may not accurately identify relevant transactions. Along with their account holder details, these suspicious transactions will then go to formal authorities, such as AUSTRAC, to identify money laundry cases. In this research, our labels are applied to the second type, referring to the "Suspicious" transactions. The first source of labeling data is banking experts. They identify suspicious transactions based on their expertise and experience. This type of labeling is timely and requires human effort, so the cost is high. According to the size of the dataset, we cannot label a dataset of 100k transactions by human expert labeling. Therefore, we use the proposed approach in addition to auto-labeling methods to identify suspicious transactions. We are using a similar approach to the Snorkel model 67 to facilitate auto-labeling the transactions. The dataset used in this project contains 100,000 records consisting of transactions with features such as amount, source location, destination location, time of the transaction, and bank branch. To label this dataset, we are using banking experts for 10 percent of the dataset. For all the rest, we are applying the Snorkel model to label the data by using the simple rules as labeling functions. The Snorkel model is an auto-labeling method that creates the labels based on comparing the outcomes of different labeling functions, i.e., rules and selecting the labels according to the agreement of the majority of outcomes. Figure 7 shows an overview of the model we have used in this project for labeling part of the training data. To further explain this section, suppose that we have n transactions in a specific time duration called t1, t2, t3, ..., tn. Each transaction has a vector of different features like amount, time, country of origin, country of destination, product type, and transaction branch. These features can be seen below: V(t) = {Transaction ID, Account ID, Product Type, Transaction Branch, Transaction Date/Time, Transaction Amount, Transaction Currency, Credit/Debit Status, Country of Origin, Country of Destination, Transaction source type} Each of these features has its definition, which is described in the previous section. Suppose that each transaction has m features, named F1, .... , Fm. This way, our Feature Vector for one transaction will be as followed: To auto-label the data, we will apply labeling functions to each of these transactions. We have k labeling functions, so LF1, LF2, .... , LFk are our labeling functions. We apply these labeling functions to each transaction. Snorkel model will find the labels of each transaction based on the agreement and disagreement between the labeling functions and will finally decide on a single label for each transaction. Each transaction will have labels as 1 for being a suspicious transaction and 0 for being a typical transaction. Therefore as demonstrated in Figure 7 , applying auto-labeling on this data means finding the labels L1, L2, ..., Lm for our transaction set. These labels are then used for training our model along with the labels from a banking domain expert. These labels are used as a supervised method for our classifiers. For anomaly detection, which is an Unsupervised method, we do not need the labels, as the model will recognize the anomalies without needing the labels. However, to compare the result of our anomaly detection model with the actual labels, we will then consider all anomaly transactions as suspicious ones and measure the accuracy, recall, precision, and recall of the model based on that. In order to label the data based on the Snorkel model, we need some labeling functions. These labeling functions are written based on simple rules that can specify unusual transactions. For example, having a cash transaction with an amount more than a specific value will have a flag as an unusual transaction. The actual amounts will not be revealed in this paper explicitly to comply with the privacy policies of the sponsor company and regulations related to the financial crimes. Regarding the unusual transaction, one crucial point is that if a transaction is identified as unusual, it does not mean that the transaction is money laundering. However, it is instead a way of flagging it to go for a different level of investigation. If other labeling functions agree with that, this transaction will be flagged as suspicious for money laundering. Another labeling function is related to recognizing the money Laundering cases related to smuggling wildlife and animal skin. In this case, the country of origin or destination will be necessary. So if one of these countries is on a blacklist (which we cannot reveal their names because of privacy and are those who happen to be the primary source of wildlife trafficking), then the transaction will be flagged as unusual and needs to be passed for further investigation. Another case of labeling functions would be for the money laundering cases related to terrorism activities. In such scenarios, if certain words are used in the transaction statement (e.g., a keyword related to the history of a terrorism activity or is a digit reminding the year of that terrorism event), then this transaction will be flagged as unusual and will be sent for more clarification via anti-money laundering teams. Examples of such words include "hijack" and "terror". There are also labeling functions related to customer credit score. A credit score is a number between 0 to 1 (with 1 the most and 0 the least) and represents customers' credit. This number can be impacted by customer behavior such as loan activity, income, the total number of years that the customer has an account in the bank, the average amount of monthly transactions, expected lifetime value, and types of products used. In this study, from experience obtained from the domain knowledge experts, it has been recognized that money laundering activity usually happens within customers with a credit score less than 0.05. Accordingly, as a labeling function, a simple rule could be used to label a transaction as suspicious: if the customer credit score is less than 0.05 and the transaction amount is more than 20000. These rules are different from setting rules for detecting suspicious transactions, and they are just labeling functions that we will feed into the snorkel model. The model will decide if this is aligned with the labels identified by the expert and also the result from other labeling functions. In this research, we apply the classification on transactions to predict whether they are suspicious and need to be considered money-laundering. Figure 8 shows an overview of the classification method used in this research. The proposed model is a binary classification and aims to predict the suspicious transactions. Therefore we will have 1 as the label for suspicious transactions and 0 as the label for non-suspicious ones. The input of the classification model is the transaction features. We can extract different features for a transaction, including account, time, type, entity, destination, location, amount, and fund flow statement. We are using the big banking data generated on open and private data islands. Using Knowledge-Bases, we enrich the raw data extracted for each customer to use them as input of our classifier. For instance, we are using Wikidata to enrich the words of transaction statements and derive some insights. For example, transactions with specific words (such as "Gambling" and Game") in their statement can be flagged. Once we extracted the semantic features from our raw data about transactions, we feed them into a classification model. We need to consider some pre-processing on the data, including standardization, labeling, and Augmenting 11 . We will describe them below. Transaction feature engineering: We first need to select important features which will help us identify money laundering transactions from regular transactions. These features are being selected through a pool of transaction attributes 50, 69 . Data enrichment: One of the key challenges in data analysis is changing the raw data into curated data. To drive insight from the extracted data from all different sources with several formats, we need to use KBs to provide se- mantic data. For instance, to extract information from transaction statement we need to extract key-words and enrich the text by KBs that detect specific word groups 16, 21 . Standardization: As we have different numeric features with various ranges and different means, we need to do the mean normalization first to ensure the model will converge to the global minimum of its cost function. Labelling: While we are analyzing a large size dataset with thousands of transactions, suspicious transactions need to be detected by the analyst. Thus, we are not able to label all of the transactions used in training by the analyst. Instead, we need to label a portion of them by using intelligent techniques to automatically label the rest of the transactions by the patterns extracted from analyst labeling. Augmenting: Our dataset is skewed, i.e., the portion of suspicious transactions is not equal to normal ones. Based on historical banking data, we expect a one-digit percentage for our suspicious transactions. Therefore, the labels are not balanced. In other words, we will have for example 10 percent of the transactions as suspicious, which are labeled 1 versus 90 percent labeled 0. However, to get the most accurate result from our classification, we need balanced labeling. That is why we are using augmentation techniques to balance our dataset according to the portion of suspicious transactions. Classification: We use the pre-processed data as the input of classification application. We are using different classification methods and measure the results. In this research, we aim to use logistic regression, Nearest Neighbours, Naive Bayes, Neural Network, and Random Forest as our classifiers to predict the transactions that are categorized as money-Laundering. The evaluations of these models are being done by measuring Accuracy, Precision, Recall, and F1 Score for the classification result. As our dataset is skewed and has fewer positive labels (suspicious transactions) than normal ones, we need to use both accuracy and F1 score to evaluate our classification model. The transaction features we have come up with so far in this paper contain the main attributes a transaction can have. These features will make a vector for each transaction and will consist of the below elements: Transaction ID, Account ID, Product Type, Transaction Branch, Transaction Date/Time, Transaction Amount, Transaction Currency, Credit/Debit Status, Country of Origin, Country of Destination, Transaction source type. The second phase that we are focusing on in this research is Anomaly Detection. For each transaction, we extracted different features as mentioned above. Then we will calculate the average and standard deviation for each to find the Gaussian distribution of each of the features. We will use our training data for statistical calculations, and we put regular transactions (not suspicious ones) as our training data. Then we will use a combination of regular and suspicious transactions as our cross-validation and test data to calculate the probability of each transaction. We will pick any transaction with less than a certain probability as suspicious. We will find the best threshold by checking different ones, finding the best F1 score within cross-validation data, and selecting the best threshold for the probability of unusual (suspicious) transactions. We will use anomaly detection to extract more complex features. If a transaction is labeled as suspicious by the analyst, but the system does not identify it as suspicious, we will perform anomaly detection and identify more features; this can combine existing features and rerun the model with updated features. In each step, we will apply the anomaly detection method on cross-validation data and measure the accuracy and F1 score to find the most successful feature list and threshold. One of the main goals of this research is to minimize the need for human intervention to detect suspicious transactions. In other words, we need to minimize the number of transactions that are identified as suspicious while they are normal. So we need to decrease the portion of false positives and increase precision. One of the contributions of this research is capturing the agreement between the results of classification and anomaly detection to achieve this goal. Therefore, we will apply the two applications described above, and we will apply a logical "And" between them. The first application as a supervised machine learning model is a classification in which we need almost 100k labeled data. Hence by applying the Snorkel model and using expert knowledge, we can label the data and use it to train the model. The second application is an unsupervised one which is anomaly detection. We apply a logical AND between the results and measure the accuracy indexes to prove how this method will improve the precision according to the project goal. This section reviews the results of applying our methodology to real transactions to extract unusual and suspicious transactions. We first present our motivating scenario, which is about the detection of transactions suspicious of money laundering. We will then discuss our dataset specifications: the transactions of one of the Australian banks with the masked data according to the privacy policy. Each observation in this dataset is a transaction with its related features and attributes. Finally, We will separately discuss the accuracy of each part of our proposed pipeline. We will provide the accuracy results of the classification model with the global indexes of accuracy for skewed datasets. Also, we will present the anomaly detection application results. Lastly, we discuss the impact of applying logical AND as part of our contribution on the accuracy results and will see how this method will assist in achieving the goal of this research. In our coding, we used Python 3.8, Pandas 1.2.4, and Numpy 1.20.2 to evaluate the results of our model. The methodology presented in section 3 is applicable for detecting transactions with the financial crime of fraud. In this paper, we are using this method to detect money laundering in bank daily transactions. To this aim, we are applying both classification and anomaly detection to recognize the transactions which are suspicious for money laundering. The dataset we are using comes from the bank, and it has several features on the transactions. We will use all the features and will not assign any weights to them before applying the models. We want our model to realize the impact of any of these transaction attributes on our research goal, i.e., money laundering prediction. The data for our dataset comes from multiple data sources within the bank. Each of these data sources has its feature labels and structure, so the first thing we need to do is extract the standard features and leave the ones specific to one particular type of transaction. For example, "Transaction Branch" might have a different meaning to cash transactions than online banking. In the online banking database, we had a field "Terminal Name" which is the equivalent of "Branch" when dealing with the transactions done in a physical location. There were also some features that would be a duplicate and not necessary for the purpose of this paper. For example, for the transactions done via card, there was a field "Card Number" which would not add any value other than the Account ID for the purpose of this study. Further to that, this feature would be available for just one transaction type and not all of them; therefore, it was decided to delete this feature. The dataset included data on transactions. Below is a brief description of each feature: Transaction ID: This is the unique identification number that the banking system will assign to a transaction. Account ID: This is the unique identification of the account holder, which will help the bank identify which account this transaction may belong to. This is important because a customer who has a history of money laundering is potentially more exposed to doing it again. Customer ID has a risk that the model should not flag people who have an id similar to the id of a suspicious customer. We need to control this with other solutions. One mitigation plan would be to make customer IDs different from each other to decrease the risk. Another plan would be to flag some customer IDs related to people or companies with high risk and replace customer ID with that feature. This is applicable for both Source customer ID (payer) and Destination customer ID (payee). Account (customer) type: The customer who is doing the transaction can have different types; it can be an individual or a company doing the transaction. It can also be an association or a trust. This is specifically important because different customer types have different transaction amounts and transaction behavior. What is unusual for an individual might be a recurring character of a company doing its regular transactions. Product Type (Transaction Type): This field specifies the product type that is engaged in the transaction. There can be several product types in each bank, but the most common ones are cash-in, cash-out, card, direct payment in the bank, cheque-in, cheque-out, new platforms (such as m-banking or internet banking), and global payment. Transaction Code: Each transaction type can have several methods; for example, card payment can be an online purchase or tapping in a supermarket, or transferring money using an ATM. Each of these makes different codes under each transaction type. For some of the transaction types, only two or three codes are available, whereas, for others like card transactions, there are more than ten different codes. Transaction Branch: This feature will specify in which branch the transaction took place. This can be a physical location where the transaction has taken place, an ATM or another machine related to a branch, or a code for the payment gateway that the online transaction has taken place. Source Bank: Source bank specifies the bank to which the source account belongs to. Destination Bank: Destination bank specifies the bank of the account to which account the transaction has been sent to. Transaction Date/Time: This specifies the date and time in which the transaction has taken place in. This might also be important to check and see whether the time of a transaction is essential or not. For example, doing a transaction in the middle of the night might be a sign to check whether it is typical or unusual. Transaction Amount: The transaction amount is the monetary value of the transaction. This is one of the most important features based on the experience and the quotes from subject matter experts. The model will, however, test this to see if this perception is valid or not. Average Amount of transaction in the previous month: This feature specifies the average amount for all the transactions for that account in the previous month. This is also useful as it will give some insight into the account holder's average transaction amounts. Transaction Currency: This specifies the currency in which the transaction has taken place, e.g., was it Australian Dollar, or it was Thai Baht, or American Dollar. Credit/Debit Status: This feature specifies whether the transaction was done as a credit or it was in debit mode. Country of Origin: This mentions the country which was the source of the transaction and the transaction has originated from. In our dataset, more than ninety-five percent of the transactions are originated from Australia. Country of Destination: This specifies the country to which the money has been transferred to. This is of particular importance as lots of the financial crimes are related to countries to which the regulations of international baking are not implemented, are on the FATF blacklist, or are subject to terrorism or wildlife trafficking. This is based on the knowledge obtained, and the model will test to see if this assumption is valid or not. In our dataset, more than ninety-five percent of the transactions are sent to Australia. Customer Credit Score: This feature refers to the credit score of the client who is doing the transactions. Every client has a credit score between 0 and 1, with 0 being the lowest and 1 being the highest. The amount of this is specified by the bank through several measures during years. In this paper, we have different knowledge sources which were taken into consideration. Figure 9 shows how multiple sources of knowledge are utilized in this project. To be specific, we have both banking and non-banking sources as our knowledge sources. For the non-banking area, samples of the labeling functions for the snorkel model coming from academic papers can be a good example. These samples shared a good insight into how we can define our labeling functions and how we can align them to have meaningful results. In the banking area, we have Regulatory sources, Banking experts, and open source knowledge bases. The Regulatory sources specifically consist of the files and documents from AUSTRAC and FATF as two key sources that define the rules and regulations for money Laundering. These sources have guidelines (both publicly and specific to financial institutions) that provide some rules on what is defined as money Laundering and how to identify it. This is especially important when we want to write our labeling functions, which come from these rules and wordings from regulators. For example, AUSTRAC specifies which countries are mainly considered a thread to commit money laundering for child abuse or which words are related to terrorism activities. Also, FATF publishes a list of blacklisted countries that support terrorism or have huge drug-organized groups. There are also numerical rules set by these regulatory authorities, which make it a guideline for writing labeling functions. For example, the transaction amount above a certain threshold needs to be investigated. The average amount of transactions in the past x days should not exceed by y percent more than the previous days. These are all defined by the regulatory as a general guide, but financial institutions will refine that based on their internal processes when it comes to practice. In this project, we have used these rules as a source for defining the labeling functions. Banking experts were another source of knowledge in this project. As this project is sponsored by one of the Australian largest banks, multiple sessions have been held with several banking experts attending and sharing their knowledge. This knowledge was in two inter-connected areas. The first area was the practical knowledge that has been obtained through years of working in financial crimes teams. The second area was the technical knowledge of the software and systems which are currently being used by the banking industry or can be potentially used to enhance the current rates, such as increasing true positive detection and decreasing false positive ones. In order to accomplish this aim, two teams were explicitly helpful in this area. One was the anti-money Laundering team with years of experience in financial crimes and fighting money Laundering cases and had their ways of finding their cases and reporting them to AUSTRAC. Their method was mainly rule-based, utilizing the software systems written based on AUSTRAC general guides. The other data gathering source was the data discovery lab which gathered the data from different sources and transformed it into a format that was usable by the project. These sources helped shape the path of the project on the go. The data itself also comes from different sources. Each transaction type has a data source system, and the format of the data and the fields in each data source are different. A critical step in this project was aggregating all these different data types and shaping them into a format that can be used across the project. Another source of Knowledge is open knowledge bases such as Wikidata. These are a common source of open data and were explicitly used to refrain from the transaction references and find the words that had a similar meaning to what has been written in the transaction statement. This tool can enrich the data and decrease the possibility of missing a word related to money Laundering solely because a synonym has been used. For example, we can find out that "kid" can have the same use as "child" to analyze transaction statements to enhance the chance of finding a suspicious transaction. After aggregating the databases and before applying machine learning algorithms on the data, we needed to preprocess the data to get proper results. This includes absolute encoding values, standardizing, and augmenting the data. Some of the features that we had were represented in texts; for example, transaction reference, transaction country, source, and destination countries. We needed to convert these values to categorical values in python and have numerical values for these features. Therefore we used ".cat.code" to convert these texts to numerical values so that the model can read the values appropriately. The features that were used in the model had different ranges. For example, the transaction amount would vary from 1 dollars to five hundred thousand dollars, whereas the customer credit score was between 0 and 1. Therefore, the data needed to be standardised to prevent the effect of these varieties in range, using the mean and standard deviation of each variable. The python package used for this aim was sklearn. The data used in this paper was skewed as the number of suspicious transactions among all transactions is low and less than 10 percent. This will lead the machine learning models to have improper results. Therefore we applied SMOTE package in python to augment our dataset. This will cause a more balanced dataset which improves the accuracy of ML applications. The initial labeling was done through AML experts in the bank; however, not all the data in our dataset was labeled, and we needed to apply auto-labeling methods to the data. We used the Snorkel package from python to label the rest of the data. We had 10 percent of our data labeled by banking experts, and by applying the Snorkel model, we labeled the rest Below are the samples of the labelling functions we used for our Snorkel model (please note that due to confidentiality, the rules have been modified): We used that labeled dataset to run our classification model. Moreover, once the anomaly detection application extracts the unusual transactions, we compared the results by our labels to find the accuracy of our anomaly detection model independently. In other words, while anomaly detection is an unsupervised application, we applied the same accuracy indexes, including Precision, Recall, and F1 score, to provide consistency amongst the results of these two applications (classification and anomaly detection). Finally, once we applied the logical AND between the results, we used the same labels produced by banking experts and the Snorkel model to evaluate the final hybrid model. Below is the definition of the mentioned evaluators for the classification model: F − measure = 2 * P recision * Recall P recision+Recall (Equation 4) In the above formula, TP stands for True Positive, which refers to the model's cases correctly as a suspicious transaction. TN stands for True Negative and refers to the transactions that the model has recognized correctly as not suspicious. FP stands for False Positive and refers to transactions that the model has recognized as money laundering, but they are not a case of money laundering. FN stands for False Negative and refers to the cases that the model has recognized as a typical transaction, but actually could be related money laundering. After pre-processing our data, we used this data as an input to our binary classification model. We split the data into training and test data with 80 percent for training and 20 percent for the test. To evaluate which of the classification models have better results on our data, we use Accuracy, Precision, Recall, and F1 Score. These measures are available using sklearn package in python. We applied different classifiers, including Logistic Regression, Nearest Neighbours, Random Forest, Neural Network, Naive Bayes, and Multinomial NB (Multinomial Naive Bayes). Table 1 compares the accuracy of applying different classifiers and figure 10 illustrates this comparison . The results show that a fully connected neural network provides the best level of accuracy. One reason for this can be related to the non-linear boundaries of the dataset. In other words, as multiple factors are influencing the transaction to be suspicious or not, a simple classifier such as logistic regression cannot predict the results with a high level of accuracy, and deep learning provides valid results 63 . In terms of implementation time, the neural network took longer to implement, whereas the others were pretty quick. In terms of recall, Random Forest was the best, and Neural Network was the second but close to the first. We applied Anomaly Detection to our data. The idea that we used in evaluating our application of anomaly detection is evaluating this unsupervised machine learning model with some indexes that are usually used to evaluate supervised applications. In other words, once a transaction is identified as an anomaly, i.e., unusual, then we compared the results with our already available labels, which indicate whether or not the transaction is suspicious. Then, by calculating TP, TN, FP, and FN, we measured Accuracy, Precision, Recall, and F1 Score, as shown in table 2. Among the algorithms used for anomaly detection, the iForest method had the best result and the highest AUC score, which refers to Area Under Curve 85 , therefore was selected as the primary anomaly detection method to base the rest of the model onto. A comparison between the results of anomaly detection and classification shows a slightly better performance for classification. However, while the accuracy and precision of anomaly detection are minor compared to classification, anomaly detection improves recall. In the last step, we applied the logical AND function between the two applications to recognize the transactions that both methods have recognized as suspicious would be considered a suspicious case. The result of this method can be seen in Table 3 . As the results prove, applying logical AND will result in better precision. False-positive cases, i.e., the number of transactions that are identified as unusual but are not suspicious, are reducing. This is aligned with our goal to minimize human intervention by increasing the precision of our model. The results also indicate a slightly better F1 Score after applying logical AND. One of the goals of this research is to minimize the need for human intervention to detect money laundering transactions. As the results show, our method is using two different approaches to extract suspicious transactions. One is classification as a supervised ML model, and the other is anomaly detection as an unsupervised one. Then, by applying logic and reducing the number of transactions that have been extracted as suspicious. In other words, any transaction that is identified as suspicious in this method has much more possibility to be a case of money laundering. Accordingly, the percentage of transactions identified as suspicious and a case of money laundering is rising, which reduces the need for analyst intervention. This is completely aligned with our research goal. The table of results shows a significant improvement in accuracy, F1 score, and precision after applying logical AND, which agrees with our research goal. Finally, to support our findings, we applied a k-means clustering on our dataset to find the segmentation of transactions. As shown in Figure 11 , the optimal number of clusters based on the elbow method is two clusters which are supporting our results in dividing transactions into two groups of suspicious and normal. We aim to combine clustering approaches to provide a visual transaction segmentation in our future works. Banks and financial institutions are struggling to find and prevent financial crimes. In this context, Artificial Intelligence is able to help organizations identify and predict financial crimes. In this paper, we studied money laundering which is a financial crime that aims to concealing the origins of money. We discussed intelligent methods to prevent this type of financial crime as the motivating scenario. We presented a novel approach that aims to extract different features of transactions as raw data from different data islands and enriched them by using knowledge bases (KBs). We applied the process optimization techniques 77 and fed them into different classifiers to predict suspicious transactions. Our goal was to make improvements in the current prediction rate of money-laundering transactions in terms of precision. We worked on both supervised and unsupervised machine learning applications for detecting suspicious transactions. We used supervised learning in our classification models to predict suspicious transactions by asking analysts to label the transactions, and develop algorithms that learn from this process and starting to automatically label the data. We also applied the anomaly detection method as unsupervised learning to detect unusual transactions. We then combined the two methods, which resulted in more precision resulting in less human work. We used different features from transactions, such as location and entity type and detected the weight of each in the classification model. We did not pre-weight the features and let our classifier find the impact of each feature on the result. This method applies to any other goal-based transaction classification, such as fraud detection. Obviously, the weight of each feature is changing according to the goal of classification. We aim to extend our work by using customer features and combining them with transaction features. To achieve this goal, we plan to use data summarization 42 , curation 14, 15, 28 , and interactive storytelling 8, 19, 20, 79 techniques to develop an intelligent feature engineering. Another line of related work can focus on intelligent and rule-based 78, 80, 81 recommendations 89-91 and trust prediction [37] [38] [39] [40] [41] to keep the business analyst aware of potential threats. We also aim to model the data using graphs modelling 12, 22, 24 and querying 22, 25, 43, 60 and apply clustering techniques and combine the result of our classification and anomaly detection with clustering. This is to provide a dynamic dashboard for the banking analyst to have a visible segmentation of each transaction. This will also help the analyst to find similar transactions. Another line of future work could focus on reputation management techniques [5] [6] [7] and behavioural analytics 17, 18 to predict the risk-based activities of low profile customers. Crime Data Mining: A General Framework and Some Examples Relational learning analysis of social politics using knowledge graph embedding Economic and Financial Crime Reputation management in crowdsourcing systems Collusion detection in online rating systems Representation and querying of unfair evaluations in social rating systems isheets: A spreadsheetbased machine learning development platform for data-driven process analytics Association for Computing Machinery Community-Level Anomaly Detection for Anti-Money Laundering Cognitive augmentation in processes Large scale graph processing systems: survey and an experimental evaluation Processatlas: A scalable and extensible platform for business process analytics Coredb: a data lake service Corekg: a knowledge lake service DataSynapse: A Social Data Curation Foundry. Distributed and Parallel Databases Towards context-aware social behavioral analytics Enabling the analysis of behavioral disorders in social networks iprocess: Enabling iot platforms in data-driven knowledge-intensive processes istory: Intelligent storytelling with social data Towards cognitive recommender systems Scalable graph-based OLAP analytics over process execution data. Distributed Parallel Databases Enabling the analysis of cross-cutting aspects in ad-hoc processes A framework and a language for on-line analytical processing on graphs A query language for analyzing business processes execution Process Analytics -Concepts and Techniques for Querying and Analyzing Process Data Development of grid resource discovery service based on semantic information On automating basic data curation tasks Money laundering -A global obstacle Intelligent Money Laundering Monitoring and Detecting System Machine learning techniques for anti-money laundering (AML) solutions in suspicious transaction detection: a review Predicting Financial Crime: Augmenting the Predictive Policing Arsenal Fighting money laundering with technology: A case study of Bank X in the UK. Decision Support Systems Application of k-means algorithm for efficient customer segmentation: a strategy for targeted customer services Intelligent Anti-Money Laundering System * A framework for data mining-based anti-money laundering research A survey on trust prediction in online social networks A dynamic deep trust prediction approach for online social networks DCAT: A deep context-aware trust prediction approach for online social networks SETTRUST: social exchange theory based contextaware trust prediction in online social networks Social context-aware trust prediction: Methods for identifying fake news Extractive document summarization based on dynamic feature space mapping DREAM: distributed RDF engine with adaptive query planner and minimal communication An Intelligent Anti-Money Laundering Method for Detecting Risky Users in the Banking Systems Attention mechanism in predictive business process monitoring State-of-The-Art in Big Data Application Techniques to Financial Crime : A Survey Comprehensive and reliable crowd assessment algorithms Detecting money laundering transactions with machine learning Forensic Analysis for Monitoring Database Transactions Towards intelligent feature engineering for risk-based customer segmentation in banking A bayesian approach for suspicious financial activity reporting Convolutional neural network for medical image classification using wavelet features Applying Data Mining in Money Laundering Detection for the Vietnamese Banking Industry Application of data mining for anti-money laundering detection: A case study A data mining-based solution for detecting suspicious money laundering cases in an investment bank Isolation-based Anomaly Detection Research on anti-money laundering based on core decision tree algorithm Data mining for statistical analysis of money laundering transactions A RBF neural network model for anti-money laundering A framework of enriching business processes life-cycle with tagging information Global financial crime : terrorism, money laundering, and off shore centres. Ashgate Aldershot Investigation and prosecution of money laundering cases in Malaysia Deep feature learnt by conventional deep neural network A survey on transfer learning Ontology Based Expert-System for Suspicious Transactions Detection TEXUS: table extraction system for PDF documents Snorkel: rapid training data creation with weak supervision HoloClean: Holistic Data Repairs with Probabilistic Inference Linking textual and contextual features for intelligent cyberbullying detection in social media Review On Detection of Suspicious Transaction In Anti-Money Laundering Using Data Mining Framework Data Mining Techniques for Anti Money Laundering Detection of money laundering groups using supervised learning in networks A novel cognitive computing technique using convolutional networks for automating the criminal investigation process in policing Computer Sciences Department Active Learning Literature Survey Anti-Money Laundering: Using data visualization to identify suspicious activity Statistical methods for fighting financial crimes Scalable saas-based process customization with casewalls Adaptive rule monitoring system Conceptmap: A conceptual approach for formulating user preferences in large information spaces Adaptive rule adaptation in unstructured and dynamic environments Feature-based and adaptive rule adaptation in dynamic environments Developing an intelligent data discriminating system of anti-money laundering based on SVM A review of money laundering literature: the state of research in key areas Bitcoin money laundering: mixed results?: An explorative study on money laundering of cybercrime proceeds using bitcoin A critical analysis of variants of the AUC A Money Laundering Risk Evaluation Method Based on Decision Tree Research on money laundering detection based on improved minimum spanning tree clustering and its application Anti-Money Laundering in Bitcoin: Experimenting with Graph Convolutional Networks for Financial Forensics Enabling the analysis of personality aspects in recommender systems Towards a deep attention-based sequential recommender system CNR: cross-network recommendation embedding user's personality Design methodology for service-based data product sharing and trading Applying Data Mining in Investigating Money Laundering Crimes An Outlier Detection Model Based on Cross Datasets Comparison for Financial Surveillance -I Acknowledge the AI-enabled Processes (AIP 4 ) Research Centre and Tata Consultancy Services (TCS) for funding this research.