top of page

AI in AML: Why Humans Are Here to Stay

  • harrylangley5
  • Feb 21
  • 10 min read

Updated: Mar 26


One of the common discussions I have with people when I meet them is about Artificial Intelligence (AI) and its capabilities to detect and prevent financial crimes. I do wonder whether the reality reveals a different story—one where AI, despite its advanced algorithms and data-processing prowess, cannot single-handedly tackle the complexities of financial crimes. My thoughts below challenge whether humans are still essential in the AML sector, despite the increasing integration of AI technologies.


Even though AI has evolved at an astonishing rate—particularly with new machine learning breakthroughs—financial crime remains a field thick with nuance. AI can analyse staggering amounts of data with speed, but effective AML requires domain knowledge, human judgment, and continuous adaptation to cunning criminal typologies. Where do we stand now, and how did we get here? To understand why humans are here to stay, it helps to look briefly at how AI has historically developed and why it repeatedly bumps into real-world constraints—particularly in fields like AML, where context is king.





💭 The Myth of AI in Financial Crime Prevention


In the competitive world of financial technologies, AI is a buzzword that sells. Many firms market their products as AI-powered to capitalise on the allure of cutting-edge technology. However, this marketing often oversimplifies or exaggerates the capabilities of AI in AML.


Many of these oversimplifications spring from misunderstanding how AI tools work. Some solutions rely on large “rule-based” approaches, similar in spirit to older “expert systems” from the 1980s. Others incorporate sophisticated “deep learning” models that excel at pattern recognition across vast swathes of transaction data. In either case, there is a tendency to overstate their intelligence. Marketing language can imply these systems “think” like humans or can autonomously uncover crime with minimal human oversight—yet the technology remains narrow.


Historically, AI has endured multiple “boom and bust” cycles. The first AI boom back in the 1950s and 1960s centred on symbolic logic and rule sets—attempts at encoding the essence of human reasoning into machines. Although those early endeavours taught us a great deal, real-world performance often lagged behind inflated expectations. The next wave in the 1980s championed expert systems, but these again met a wall when confronted by real-life complexities. Today’s wave, dominated by machine learning and neural networks, is more advanced, but it, too, has limitations.


The concept of AI in AML is surrounded by a halo of technological invincibility. Many believe that AI systems, with their vast data handling and analytical capabilities, are the ultimate tools for identifying and thwarting illicit financial activities. However, a critical examination suggests that AI, as it is currently utilised in the financial sector, may not be the panacea it is often touted to be.


One cause of this halo effect is the sheer speed and scale at which modern AI operates. A machine learning platform can sift through millions of transactions daily—something no human team could match. This raw computational power frequently leads to the idea that such tools are inherently “invincible.” But meaningful AML analysis can hinge on context that is not easily distilled into ones and zeros. For example, a suspicious transaction in one jurisdiction may be perfectly normal in another, and “unusual” behaviour for one client might align with legitimate cultural or local practices for another.


Worse still, criminals sometimes exploit the very same AI innovations for wrongdoing: from deploying “deepfake” technology to crafting hyper-personalised phishing campaigns. Meanwhile, regulatory bodies demand that if a transaction is flagged, the compliance team explain why. A powerful AI black box might raise alerts, but if it can’t provide a rationale beyond “the numbers looked strange,” it may fail to meet transparency demands.


AI's application in financial crime often does not involve true artificial intelligence but rather revolves around machine learning models and rule-based systems. These systems are adept at processing and analysing large datasets at speeds no human can match. However, they fundamentally lack the ability to understand context or the subtleties of human behaviour—key elements in detecting and understanding financial crimes.


This gap between raw data analysis and human-level understanding becomes vital in AML. Suppose an AI tool marks dozens of wire transfers for review because they share certain “anomalous” traits. A machine can highlight red flags (e.g., sudden transfer spikes, suspicious geographies, or repeated round-dollar amounts). Yet determining whether those anomalies truly signify money laundering—or whether they’re legitimate philanthropic donations after a natural disaster—often hinges on knowledge about local culture, real-time events, or deeper conversation with a client.


In the broader story of AI, this notion of context is a recurring theme. From the earliest attempts at “symbolic logic,” researchers realised that computers easily handled stable inputs (like chess moves) but faltered when forced to interpret the unpredictability of the real world. Similarly, a purely data-driven model might learn from the past, but criminals adapt, meaning the patterns of the future are rarely identical to those of the past. The upshot? While machine learning may replicate older patterns well, it’s easily outwitted if the system is never updated or is left to learn from flawed data alone.


During a recent board discussion on the role of AI in AML, a significant question was raised: "Is AI in financial crime a myth?" This question strikes at the heart of the issue. The real, advanced AI technologies, such as those capable of generating deep fakes, are indeed being utilised but potentially more so on the criminal side rather than on the defensive side. Financial institutions may use AI-driven tools, but these often do not operate with true AI capabilities. Instead, they function on predetermined algorithms and rule sets that can flag anomalies but cannot inherently understand or interpret them.


We can think of modern AI in AML as advanced detection software—like the security cameras that can identify suspicious movements but cannot interpret the motive behind them. True interpretive intelligence demands a human-level capacity for ethics, empathy, and spontaneity. Tools exist that appear to “generate” novel text or reams of content, but at their core, they rely on patterned predictions to guess the next likely word or outcome. They can be fooled. They can produce random false positives (or negatives). If criminals start using new tactics that deviate from the model’s training data, the system likely won’t recognise it.


Hence, the concept of a “myth” arises not because AI in AML is useless—far from it—but because many see it as a total replacement for humans. History shows that each new wave of AI typically oversells autonomy and intelligence, only to come up short on context, adaptability, and ethical clarity.



🧠 The Indispensable Human Element


The regulatory environment of financial institutions adds another layer of complexity. Regulators are understandably cautious about AI systems, particularly those that act as "black boxes." These systems can offer little to no explanation for the decisions they make, which poses a problem for compliance. Transparent AML operations are crucial, where every decision to flag a transaction must be justifiable.


Human investigators, analysts, and compliance officers serve as critical interpreters. They review an AI system’s alerts and ensure there is a well-documented rationale for each suspicious activity report. A purely algorithmic model is unlikely to satisfy authorities when asked, “Why did you decide this person was high risk?” or “Where exactly is the evidence of fraud?” If the machine cannot explain itself—and the compliance officer merely shrugs—then regulators balk.


Across various industries, “explainable AI” has become a hot topic. The problem is especially pronounced in AML, where decisions could have serious legal ramifications, not to mention reputational stakes for banks. If an innocent person is wrongfully flagged as a potential money launderer based solely on inscrutable AI logic, that may lead to lawsuits or regulatory fines.


Human oversight is thus not only a regulatory requirement but a practical necessity. Humans bring a level of understanding, intuition, and reasoning that AI cannot replicate. For example, when a transaction is flagged as potentially suspicious by an AI system, a human analyst must step in to investigate the cause. They assess whether the alert is a false positive or a legitimate concern, a determination that requires human judgment and experience with the nuances of red flags.


One might compare the role of human analysts to seasoned detectives: data might point them towards possible wrongdoing, but final conclusions come from logic, interviews, contextual knowledge, and experience with how launderers actually behave. Even the best AI-driven systems generate false positives—alerts that end up being harmless. Without people to sift through them, an institution might unwittingly damage client relationships or clog up internal processes.


Additionally, it’s not just about diagnosing suspicious transfers after the fact. Skilled investigators and compliance officers must anticipate how criminals might adapt to new rules or analytics. That forward-looking perspective is where human creativity excels, drawing from a lifetime of real-world experiences.


Human operators are crucial for setting up and maintaining AI systems. They define the rules and parameters within which AI operates, aligning the machine's functions with the institution's strategic goals and compliance standards. Humans also play a vital role in training AI systems, providing them with the data and feedback necessary for accurate and effective function.


In effect, the strongest AML processes combine the best of both worlds—AI to handle the vast data, and humans to define how that data should be weighted and interpreted. Over time, compliance teams refine the model’s thresholds, revise risk scores, and pinpoint flaws in the logic. Without this continuous calibration, even the most sophisticated AI solution decays in effectiveness, especially as criminals discover new ways to circumvent it.


Even historically, attempts to replicate intelligence in software foundered when short-cuts replaced human input. Early “expert systems” would start well—loaded with knowledge bases from specialists—but quickly grew stale when no one continuously updated their logic. The same principle applies to present-day AML technology, especially if there is not dedicated staff ensuring the system’s rules and training data reflect the evolving reality of financial crime.



🔮 The Future Role of AI in AML


Looking forward, AI will undoubtedly continue to play a significant role in AML efforts. Its ability to quickly process vast datasets and identify patterns can significantly enhance the efficiency and effectiveness of financial crime detection systems. However, the future of AML does not lie in technology alone but in a synergistic relationship between humans and machines.


Part of this synergy includes more advanced AI that can, for instance, spot highly non-obvious connections across multiple datasets—transaction histories, social media footprints, company registration documents, and more. If criminals use advanced “deepfake” methods to impersonate real account holders, future AML systems could detect suspicious patterns in voice or facial data. Yet, just as criminals are quick to exploit these same technologies for wrongdoing, banks must respond with well-trained investigators who can interpret the system’s findings.


Moreover, AI might increasingly help with predictive analysis: forecasting where laundering hotspots might emerge or which sectors criminals are likely to target next. But turning predictions into actionable compliance policies is not something an algorithm can do on its own. The organisation’s strategy, local knowledge, and regulatory constraints all shape how predictions become real guidelines.


Human expertise will remain invaluable in interpreting AI-generated alerts, providing the contextual understanding necessary for accurate decision-making, and ensuring that AI operations remain transparent and compliant with regulatory standards. Moreover, as financial criminals employ more sophisticated technologies, the need for human input in AML processes will only increase.


Criminal ingenuity is arguably the biggest driver of AML complexity. Over decades, money launderers have found roundabout paths—shell companies, nested accounts, virtual asset service providers, trade-based laundering—to mask illicit funds. Some have begun to harness large-scale automation themselves, using bots to funnel funds at scale or forging documents with near-perfect authenticity. The more advanced the criminals become, the more society needs agile compliance teams capable of creative thinking in response to suspicious leads.


That said, none of this minimises the importance of robust AI. The best approach is one of partnership between AI and humans. As an illustration, consider a potential future where multiple AI “agents” handle separate tasks: one monitors transaction histories for anomalies, another tracks global watchlists in real time, and another cross-checks client data with open-source intelligence. In the end, however, a compliance officer or analyst—a human—sets the final judgement.



🔗 Conclusion: A Partnership, Not a Replacement


In conclusion, while AI brings valuable tools to the table, it does not render human involvement obsolete. Instead, the most effective AML strategies will harness both the processing power of AI and the irreplaceable insights of human analysts. Together, they form a robust defence against financial crime, ensuring that as technology evolves, so too does our capacity to protect financial systems and maintain integrity. AI in AML is not a myth forever, but it is not a standalone solution either. Humans are, and will remain, an essential element of the equation, guaranteeing that the fight against financial crime remains grounded in both technological advancement and human expertise.


The relationship between AI and humans in AML echoes the broader journey of artificial intelligence itself. Across the decades, every significant leap in AI has been tempered by a greater appreciation of complexity—whether it’s chess-playing programs that crumble in real-world tasks, or advanced neural networks that still fail to grasp the deeper meaning behind data. AML is no different. The domain’s regulatory strictness, ethical requirements, and criminal creativity mean that we cannot rely on black-box technology alone.


  • Humans supply contextual intelligence, moral scrutiny, and agile responses to unexpected criminal innovation.


  • AI provides immense speed, data coverage, and pattern detection to keep pace with large-scale transactions and identify anomalies no single individual could ever spot.


It is precisely this combination—an AI-driven system constantly refreshed by human expertise—that ensures enduring success. The notion that technology alone will permanently tip the balance in our favour overlooks the reality that criminals have also embraced disruptive tech, and it discounts the importance of real human comprehension. A synergy of artificial and human intelligence, shaped by robust governance, remains the surest way to maintain a resilient, transparent, and ethical defence against illicit financial activity.


Where do we go from here? We might imagine more sophisticated, layered approaches, with AI monitoring not only a bank’s transactional flows but also gleaning insights from external data (corporate registries, local regulatory notices, social signals)—all funnelled into a central compliance command centre. Yet, ultimately, the call on whether to block an account, file a suspicious activity report, or risk-rank a customer must be a human call, taken with due responsibility and thoroughness.


In other words, AI in AML is formidable, but not foolproof. Its potency lies in freeing humans to do what we do best: interpreting complexities, making judgement calls, and ensuring fairness. Let the machines crunch the numbers and handle the mountains of data. We, as compliance specialists, investigators, board members, or analysts, bring a finer-grained appreciation for context, culture, and ethics.


This blend of AI agility and human discernment is the evolving story of AML in the digital age. To truly thwart sophisticated financial crime, we must invest in the people who operate the systems, the ethical frameworks they uphold, and the oversight that ensures accountability. The synergy of both is what will keep money launderers on their toes—and keep our financial institutions safer for years to come.




bottom of page