Responsible artificial intelligence for anti-money laundering: How to address bias

Amid growing industry consensus that AI is key in the race to differentiate and establish market share, tackling potential biases is becoming a pivotal preliminary step

by | December 6, 2021 | Oracle Financial Services

The eager and rapid adoption of artificial intelligence (AI) by financial institutions (FIs) may surprise those outside this otherwise traditional industry. However, the industry consensus is clear that intelligent technologies such as AI are major factors in the race to differentiate and establish market share. For example, a survey conducted last year by the World Economic Forum found that 85% of FIs had implemented AI somehow, and 77% of all respondents anticipated AI to possess high or very high overall importance to their businesses within two years.

Compliance departments at FIs are poised to benefit from integrating AI into their anti-money laundering (AML) programs. Unfortunately, previously adequate, legacy rule-based AML systems have become antiquated. They lack the sophistication needed to recognise nuances of rapidly evolving criminal patterns and keep up with new products and consumer behaviour. The result is high levels of false positives and low detection rates that sap FIs’ resources by requiring the need to secure more costly, experienced compliance staff. The high false positives and low detection rates stemming from rule-based monitoring are why chief compliance officers (CCOs) at FIs are turning to intelligent technologies such as AI to manipulate data more effectively across their AML programs. But how can they do so responsibly?

Defining Responsible AI and its importance

In recent years, the AI community has encountered multiple instances of machine learning (ML) models making biased predictions. The ML research community responded with several studies, tools, and metrics to analyse the issue. This led to a growing body of research on fairness, privacy, interpretability, and trustworthiness of AI/ML models under the umbrella term “Responsible AI”. Responsible AI is now broadly discussed, having entered the Gartner Hype Cycle for Artificial Intelligence in 2020.

While exact definitions of Responsible AI vary across thought leaders, the common themes are fairness, interpretability, privacy, transparency, inclusiveness, accountability, and security.

This blog delves into the element of fairness and what CCOs can do to reduce biased AI in their anti-money laundering programs. Ensuring that ML models in AML programs don’t produce biased results is ethical and helps prevent customer mistrust, lost business opportunities, and reputational damage.

How bias can be introduced into AI

A key initial step in exploring Responsible AI is to understand how bias can creep into the model workflow and at which stages. Let’s start with the data sourcing and preparation stage. A machine learning model is reliant on accurate, complete training data. However, most FI’s business operations were set up before extensive digitalisation occurred, so sometimes the information needed to train machine learning models is recorded incorrectly, incompletely, or not at all. This may happen because typically, only a small stream of application data – about 5% to 10% of the total – makes it through the pipeline and lands in a data lake for analysis.

A machine learning model will also produce biased results when the training data is not representative. An example from the world of computer vision of using biased data to train a model: Duke University researchers created a model that could generate realistic, high-resolution images of people from a pixelated photo. However, as white people were overrepresented in the data used to train the model, the model did not work for people of other ethnicities.

However, as it has been widely discussed by the AI community, bias is not just about data. Bias can also creep in during the feature selection stage. Know Your Customer (KYC) is the part of an AML program that is most susceptible to biased model features, as KYC models attempt to assess individual people. While using certain attributes such as gender or number of children is unethical, the data science team needs to be vigilant and ensure that using benign attributes like employment status or net worth does not encode systematic bias into the models. While the transaction monitoring area of an AML program is less susceptible to biased model features, as it deals mainly with transactional, not personal, data, bias may still creep in. For example, seemingly innocuous location data (postal code, country, etc.) may serve as a proxy for data that is impermissible to consider, such as ethnicity.

Finally, human biases can influence what action AML professionals will take with AI model outputs. In an AML program, an analyst or investigator must act based on the information provided by the AI model – they must decide which alerts to investigate, which alerts to combine into cases, and which to report to authorities. Humans are susceptible to many cognitive biases that stem from basic cognitive processes such as wishful thinking, mental shortcuts, societal influence, hunger, fatigue, etc.. These biases can unconsciously influence decision-making pertaining to model predictions and outputs.

How can CCOs reduce bias in AI?

Ensuring Responsible AI for AML is a joint effort across the compliance and data science teams at FIs. Here are some things CCOs can do to support Responsible AI.

  • Communicate thoroughly with the data science team: As FIs delve deeper into Responsible AI, direct and clear communication between CCOs and the data science team is key. For example, compliance teams should provide the data science team guidance on the company values, principles, and regulatory guidelines that ML models should align with. In addition, CCOs should emphasise that evaluation of bias in models should be included in success criteria, on par with model performance related to false positives/negatives and detection.
  • Request auditability: The AI development and deployment process should be fully transparent and auditable, tracking precisely who made what modification to what model so that there is always an accurate, complete log of model creation.
  • Prioritise interpretable models: Another path to full transparency in AI development and management is to build interpretable models rather than black-box models. Like others, we believe that interpretable models are preferable to explainable black-box models for several reasons. First, black-box model explanations can be inconsistent across vendors, which in turn creates confusion among analysts. Furthermore, the explanations themselves can be difficult to decipher given the background and knowledge level of the analyst.
    In cases where black-box models will perform better than interpretable models, however, the black box model should be used, and the team should focus on explanations that provide relevant context such as the program’s strengths and weaknesses, data used to arrive at a specific decision, and why an alternative decision was not chosen. To make explanations easy to understand and use, they should be in graphical form or in the form of a pre-built natural language narrative that can be incorporated into regulatory reports— whatever works best for the team.
  • Evaluate model performance and monitor for drift: Ongoing evaluation and re-training is paramount in ensuring bias-free model performance. Once a model is trained and deployed, it needs to be monitored consistently. The model can “drift” as relationships among data change over time due to customer behaviour changes, new product releases, or other systemic changes. These factors can cause the model performance to degrade over time, and if not corrected by periodically re-training the models, result in incorrect or biased decisions. Doing these things automatically is the key to continual monitoring.
  • Assess predictive outcome fairness: Model performance on various population segments should be evaluated to ensure there isn’t a disparate impact on specific population segments. For example, an FI uses a risk scoring model to classify customers as high and medium risk. The FI can cross-check the risk scores against sensitive attributes such as race, religion, zip code, or income to investigate the model for bias.
    For example, suppose risk scores for lower-income individuals are consistently higher than those of higher-income individuals; the FI should identify which features are driving the risk scores and whether those features truly represent risk. Suppose the FI finds that a feature represents a different characteristic or behavior due to different financial circumstances and not inherent risk. Then, the feature should be modified or removed to reduce model bias. For example, perhaps a model contains the rapid movement of funds as a feature. The FI may find that the model produces higher risk scores for low-income people, but upon investigation may determine that the difference in risk scores across low-income and high-income people is being driven by the fact that low-income people are more likely to spend their entire paycheck faster, and not by truly risky money movement patterns.

Looking ahead

Having a strategy now as to how CCOs will ensure Responsible AI is thus not only the right thing to do, but it can also help them get ahead of future AI regulations. US regulators have sent strong signals that they are encouraging innovation and growth in AI, and agencies have been instructed to avoid regulatory or non-regulatory actions that act as barriers to the development and deployment of AI. However, as AI begins to permeate the financial industry, ethical and responsible behaviour and governance arises. For example, the principal regulatory guidance on quantitative models for US banks is currently S.R. 11-7 – Supervisory Guidance on Model Risk Management, from the Board of Governors of the US Federal Reserve System. While this has become a model for regulators globally, it was published in 2011 and does not encompass the full scope of AI. Thus: more AI-specific regulation is expected to be developed, though it may well come from broader regulatory bodies (e.g. the European Commission) before it comes from financial regulators.

Regardless of if and when AI regulations are laid out, by working closely with data science teams, CCOs can do their part to ensure that AI usage within their AML programs is responsible, effective, and free of bias.

To learn more, watch our responsible AI webinar or explore Oracle’s approach to anti-money laundering here.



Evolving APIs | NXTsoft Connectors For 40+ Banking Core Systems

Best Practice | Banking Evolving APIs | NXTsoft Connectors For 40+ Banking Core Systems

Banks have real opportunity in FX hedging for SMEs

Other | Banking Banks have real opportunity in FX hedging for SMEs

How Does NXTsoft OmniConnect Work for Partners?

Video | Banking How Does NXTsoft OmniConnect Work for Partners?

Castlepoint Systems pioneers world-first regulation technology solution

Case Study | Banking Castlepoint Systems pioneers world-first regulation technology solution