- Anti-money laundering (AML) requirements under the Basel Committee's core principles
- The EU's evolving Anti-Money Laundering Authority (AMLA) framework, which comes into force in 2025 and 2026
- Switzerland's own FINMA supervisory expectations on algorithmic governance and model risk
- Cross-border obligations arising from correspondent banking relationships with sanctioned-adjacent jurisdictions
FINMA, Switzerland's financial markets regulator, published its guidance on the use of AI and algorithmic decision-making in 2024, making clear that any AI system used to block or flag transactions must be interpretable and auditable. Mark Branson, who led FINMA before moving to head Germany's BaFin, has repeatedly stated that "black-box decisions have no place in supervised financial services." BaFin has echoed that position in its own AI supervisory priorities for 2025, calling for proportionate explainability as a baseline, not a bonus.
Real-Time Detection and the Rise of Agentic AI
Legacy fraud detection systems were reactive by design: rules fired after a transaction settled, and investigators reviewed queues the following morning. Modern machine learning systems operate in milliseconds, scoring transactions against behavioural baselines, network graphs, and beneficiary risk profiles before settlement completes.
The next step is agentic AI: autonomous systems that do not merely score a transaction and wait, but initiate follow-on actions based on risk thresholds. In practice, that means:
- Automatically freezing accounts when a risk score exceeds a defined ceiling
- Requesting supporting documentation from the customer via a secure channel
- Generating and filing Suspicious Activity Reports (SARs) to the relevant financial intelligence unit
- Escalating complex cases to a human analyst with a pre-populated case file
Researchers at ETH Zurich's Chair of Technology Management have been examining the governance implications of agentic financial AI, noting that the autonomy speed trade-off demands careful calibration. When systems act rather than advise, accountability chains must be defined in advance, not reconstructed after an incident.
Regulatory Sandboxes: Testing Before Deploying
FINMA operates a regulatory sandbox mechanism that allows Swiss-licensed institutions to test novel financial technology in controlled conditions, with proportionate supervision. Several Swiss banks have used this pathway to validate AI fraud models against synthetic transaction data representing real threat typologies, including trade-based money laundering, crypto-to-fiat layering, and correspondent network abuse.
The sandbox approach offers three concrete advantages:
- Models are stress-tested against edge cases before they affect live customer accounts
- Regulators gain early visibility into how agentic systems make decisions, enabling guidance before problems emerge at scale
- Institutions can demonstrate model performance data to auditors, reducing post-deployment regulatory friction
The EU's AI Act, which classifies certain financial AI applications as high-risk under Annex III, introduces a parallel requirement for conformity assessments and technical documentation. For banks operating across both Swiss and EU jurisdictions, aligning sandbox outputs with AI Act documentation requirements is now a practical necessity, not a theoretical concern.
Independent benchmarking and industry reporting point to a consistent pattern in detection performance across system generations:
- Traditional rule-based systems: 60 to 70 per cent detection rate, 15 to 20 per cent false positive rate; tends to miss novel scheme variants
- AI machine learning systems: 85 to 95 per cent detection rate, 5 to 10 per cent false positive rate; requires regular retraining on current threat data
- Agentic AI (emerging): 90 per cent or above detection rate, 3 to 5 per cent false positive rate; autonomous action with defined human escalation triggers
The reduction in false positives matters commercially as much as it matters operationally. Every legitimate transaction incorrectly blocked is a customer friction event, a potential complaint, and, in the case of business payments, a reputational and legal exposure. AI fraud detection systems that are properly trained on European transaction patterns, rather than models built primarily on US retail banking data, achieve materially better false positive rates in European deployments.
Explainability Is Not Optional
With 89 per cent of surveyed banks identifying explainability as a top priority in AI governance, the direction of travel is clear. When an AI system blocks a payment, the customer has a right to understand why, and under GDPR Article 22, a right not to be subject to solely automated decisions with significant effects without meaningful human review.
Explainable AI (XAI) frameworks, including SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-agnostic Explanations), are increasingly standard tooling at Swiss and European banks running machine learning fraud detection. They surface the specific features driving a decision: an unusual beneficiary country, an atypical transaction time, a mismatch between declared business activity and payment destination. That output is what compliance officers need to document their decisions and what customers need to challenge them.
Dominique Levin, Head of Financial Crime Compliance Technology at Temenos, a Geneva-headquartered banking software firm with deployments across more than 150 countries, has noted publicly that European banks are ahead of most regions in demanding interpretable outputs from their AI fraud vendors. "Procurement conversations in Europe now lead with explainability," Levin stated at a 2024 industry conference. "Banks will not sign off on a system they cannot explain to their regulator."
Financial crime losses across European banking are significant and growing. Estimates from the European Banking Authority and private sector research suggest that fraud and money laundering cost European institutions several billion euros annually, with cross-border transaction fraud representing one of the fastest-growing segments. Banks that invest in AI models trained on European-specific transaction patterns, rather than adapting off-the-shelf systems built on other markets' data, are achieving detection improvements that justify the additional model development cost several times over.
Comments
Sign in to join the conversation. Be civil, be specific, link your sources.