AML under pressure: coverage, precision and case aging

AML

As alert volumes climb, many AML teams can show they are working hard, but struggle to prove they are working on the right things first. The distinction matters more than it used to. Supervisors are increasingly looking past activity measures and into what a programme’s day-to-day outputs reveal about exposure, ordering and delay.

The central challenge is that scale erodes shared judgement. When queues swell, agreement on relative risk can fall, even among experienced investigators, said Consilient.

In systems built from disconnected signals, alerts may be generated quickly, but consistent risk ordering becomes harder to maintain under pressure. The result can be plenty of motion, with less clarity about whether material exposure is being surfaced early and resolved fast.

That is why regulators are focusing on how breakdowns show up operationally: what surfaces, what rises first, and how long risk sits unresolved once it appears. Recent academic work has helped crystallise the point. As volume rises, investigators’ alignment on what deserves attention first can degrade, particularly when signals arrive without clear weighting. That loss of clarity is visible in the queue itself.

In practice, supervisory review increasingly reads four patterns as evidence. Coverage gaps create exposures that never surface. Weak precision fills queues with noise, diluting attention. Implicit prioritisation pushes higher-risk cases down the stack. And case aging grows as teams spend time on alerts that were over-weighted upstream while more material issues wait.

This is also where standard AML activity metrics begin to buckle. Alert counts, review volumes and SAR totals can describe scale, but they rarely demonstrate whether exposure is being surfaced in the right order. When everything looks urgent, very little appears decisively higher risk. High throughput can coexist with uneven coverage. Sustained effort can still produce growing queues if the system cannot reliably separate signal from background noise.

Supervisors have become more comfortable calling that out. When higher-risk cases sit behind lower-risk ones, volume loses credibility. When backlogs persist despite productivity, attention shifts upstream toward how risk is identified, weighted and ordered. Activity describes movement; it does not, on its own, explain exposure.

Coverage, precision, prioritisation and case aging are not new ideas inside AML functions. They have long existed across second-line testing, model validation and audit. What has changed is the posture of regulatory review. Supervisors are no longer inferring effectiveness from control design, staffing narratives or scenario catalogues. They are reading the queue, delays and ordering as evidence, asking for these measures directly and using them to challenge risk-based claims.

Historically, effectiveness conversations leaned heavily on control design, scenario coverage lists, alert and SAR volumes, staffing levels and governance narratives. Those still matter, but they carry less weight on their own when supervisory teams can observe what the programme actually produces under load.

Regulators are consistently drilling into four questions. First is coverage: are you surfacing the exposure you say you have? That means looking at which customers repeatedly generate alerts, which risk segments remain quiet, and whether alert populations align with known exposure. It goes beyond “do you have rules for X” to “does X actually appear in your output”.

Second is precision: how clean is the signal you generate? False positives were once framed as an efficiency issue. Now they are increasingly treated as an effectiveness issue because diluted attention can weaken risk handling when investigators are overwhelmed by noise.

Third is prioritisation: do higher-risk cases consistently rise first? Chronological review used to pass without much scrutiny. Now it is questioned directly, including how review order is set, whether risk actually drives that order, and whether teams can explain why one case moved ahead of another.

Fourth is case aging: how long does risk sit unresolved, and why? Aging has traditionally been discussed as resourcing. More recently, it is being interpreted as a sign that risk has been mis-weighted, prioritisation is breaking down, or response is delayed.

Several forces have pushed these measures to the foreground. Academic research has increasingly evaluated AML effectiveness using ordering, timeliness and resolution delay, not detection alone. Supervisors have also seen too many programmes generating high activity without being able to explain outcomes. At the same time, regulators have become more willing to interrogate models, queues and prioritisation logic than they were even a few years ago. The result is a tighter standard heading into 2026.

That standard is difficult for periodic risk reviews to satisfy on their own. Many programmes still anchor risk understanding to scheduled customer reviews, built for a world where risk changed slowly and monitoring outputs could be interpreted independently. The measures regulators now press on depend on current behaviour and real-time weighting at the point activity occurs. When exposure evolves between review points, recorded risk and observed behaviour can drift apart, weakening coverage, precision and prioritisation, and increasing aging when high-risk activity fails to rise quickly enough.

This is where AML risk ranking becomes more than a design choice. It is increasingly an evidentiary requirement. If review order defaults to arrival time, effectiveness has to be inferred from activity. Explicit ranking makes the ordering logic visible: relative exposure across the population, review order that reflects risk weight rather than queue position, escalation timing aligned with exposure, and delay patterns that can be explained in risk terms. Controls and scenarios can remain unchanged; ranking determines what rises first and what waits.

Under current effectiveness expectations, that distinction carries weight. Without explicit risk ordering, programmes may be able to report volumes, but struggle to prove direction. As alert volumes grow, that is increasingly where supervisory challenge lands.

Ultimately, AML effectiveness is being judged more by outcomes than narrative. Frameworks, governance and controls remain important, but they do not explain results on their own. Coverage, precision, prioritisation and case aging expose how risk is handled once volume arrives: whether exposure surfaces, rises in the right order and receives attention when it should.

The practical question for teams is whether they could answer tomorrow if asked to explain coverage, prioritisation and case aging across the alert population, clearly and defensibly. Institutions that can evidence how risk was ordered, reviewed and resolved over time will be better placed to stand up to scrutiny in 2026 than those relying on activity totals alone.

Read the daily FinTech news

Copyright © 2026 FinTech Global

Enjoying the stories?

Subscribe to our daily FinTech newsletter and get the latest industry news & research

Investors

The following investor(s) were tagged in this article.