AI & Automation Services
Automate workflows, integrate systems, and unlock AI-driven efficiency.


Fraud losses in UK financial services reached ยฃ1.17 billion in 2023, according to UK Finance's Annual Fraud Report, with authorised push payment (APP) fraud now the largest single category by value. Cybercrime against financial institutions is not a random-distribution problem. It concentrates on organisations with weak detection capabilities, exploiting the gap between increasingly sophisticated fraud methods and the rule-based systems that most firms have relied on for detection. AI-powered fraud detection closes this gap by identifying patterns that static rules cannot see, responding in real time, and adapting as fraud methods evolve.
For UK financial firms, the regulatory picture reinforces the commercial case. The FCA, under its Consumer Duty framework, expects firms to protect retail customers from foreseeable harm. APP fraud, where customers are tricked into authorising payments to fraudsters, is a foreseeable harm that regulators now expect firms to actively mitigate. The Payment Systems Regulator (PSR) introduced mandatory reimbursement of APP fraud losses for Faster Payments transactions from October 2024, creating direct financial liability for firms without adequate prevention measures. The pressure to invest in better detection is now both regulatory and commercial.
Softomate Solutions builds AI fraud detection systems for UK financial services businesses. This guide covers how the technology works, what the FCA and ICO expect, and how to approach implementation in a way that reduces both fraud losses and regulatory risk.
AI fraud detection applies machine learning models to financial transaction data, behavioural signals, and contextual information to identify transactions or account activities that are likely fraudulent. Unlike rule-based systems, which compare transactions against a fixed list of criteria (for example, "flag any international transfer over ยฃ5,000 from an account that has never made one"), machine learning models learn the difference between normal and abnormal behaviour for each individual customer or account, detecting anomalies that rules would miss while generating fewer false positives that waste fraud team time.
The main ML approaches used in UK financial fraud detection include:
Our AI process automation practice has deployed fraud detection models for UK lenders and payment service providers. We follow a model development lifecycle that satisfies both technical best practice and the FCA's expectations for AI governance.
The FCA has not issued specific guidance on the technical architecture of fraud detection AI, but its broader expectations for AI use in financial services, set out in the AI and Machine Learning discussion paper (DP21/4) and subsequent feedback statements, apply directly. The core expectation is that firms remain fully responsible for the outcomes their AI systems produce and can demonstrate appropriate governance over those systems.
For fraud detection specifically, the FCA expects:
AI fraud detection involves processing significant volumes of personal data: transaction history, device fingerprints, geolocation data, behavioural biometrics, and sometimes communication content. UK GDPR, enforced by the ICO, applies to all of this processing. Financial firms must identify an appropriate lawful basis for each category of processing and document it in their Record of Processing Activities (RoPA).
The most commonly used lawful bases for fraud detection processing are:
Special category data may be inadvertently processed in fraud contexts. Transaction data can reveal health conditions (purchases at specialist clinics or pharmacies), religious practices (donations to religious organisations), or political affiliations. Firms must identify whether their data processing touches special category data and, if so, apply the additional safeguards required under Article 9 of UK GDPR.
The ICO's guidance on AI and data protection specifically addresses automated decision-making in financial services. If a fraud detection system makes fully automated decisions that significantly affect individuals (freezing an account, blocking a payment), data subjects may have rights under UK GDPR Article 22, including the right to request human review. Documenting how and when human review is available, and ensuring it is genuinely effective rather than nominal, is an important part of the data protection compliance framework.
The Payment Systems Regulator (PSR) introduced mandatory reimbursement of APP (Authorised Push Payment) fraud losses for UK Faster Payments transactions from October 2024. Under the regime, payment service providers (PSPs) that send or receive fraudulent payments are required to reimburse victims, with the liability split equally between the sending and receiving PSP in most cases. The maximum reimbursement amount is ยฃ85,000 per claim.
For UK payment service providers and banks, this creates a direct financial incentive to invest in AI-powered fraud prevention. The cost of reimbursements is directly proportional to the fraud that gets through. Better detection means lower liability under the PSR regime. Independent analysis by payments consultancy CMSPI suggested that the industry-wide cost of APP fraud reimbursement under the new regime could exceed ยฃ400 million annually if prevention rates do not improve substantially.
The PSR regime also incentivises investment in confirmation of payee (CoP), the service that verifies whether the account name provided by a customer matches the account number and sort code. CoP is a layer in the fraud prevention stack, not a replacement for transaction monitoring, but it is now a regulatory expectation for all Faster Payments participants above a certain volume threshold.
Effective fraud detection requires a layered approach. No single technology eliminates fraud; the goal is to make fraud expensive and difficult enough that the return on investment for fraudsters falls below their threshold. A well-designed fraud stack for a UK financial firm typically includes:
Our financial services software development service can deliver any of these layers as a custom build or help integrate best-of-breed vendor components into a coherent, well-governed fraud detection architecture.
Moving from a decision to deploy AI fraud detection to a live, FCA-compliant system involves more stages than a typical software project. The following describes the practical journey for a UK payment service provider or lending firm deploying a transaction fraud scoring model for the first time.
The first stage is data audit. Before any model development begins, the firm needs to understand what transaction data it holds, how far back it goes, how reliably fraud labels have been applied to historical cases, and whether the data is representative of current fraud patterns. Poor data quality at this stage produces poor models. A data audit of one to two weeks typically reveals gaps (incomplete fraud labels, inconsistent transaction categorisation, missing device signals) that must be addressed before model development can produce reliable results.
The second stage is feature engineering: translating raw transaction data into the inputs that the model will learn from. Features typically include transaction velocity metrics (number of transactions in the last one, six, and twenty-four hours), amount deviation from the customer's historical average, the time since account opening, the channel of the transaction (mobile app, web, in-branch, API), the beneficiary's history with the firm, and device and geolocation signals if available. Feature engineering is where domain knowledge of fraud patterns matters most. A data scientist with FinTech experience will produce significantly better features than one working from generic ML training.
The third stage is model training, validation, and calibration. The model is trained on historical labelled data, validated on a held-out test set, and calibrated to produce well-calibrated probability scores (not just a binary classification) that can be used in risk-based decision rules. At this stage, demographic parity testing is performed: checking that the model's false positive rate does not differ systematically across demographic groups. Any disparities found must be addressed through either data rebalancing, algorithmic fairness constraints, or post-processing adjustments.
The fourth stage is shadow running: deploying the model in production in parallel with the existing system, so its outputs can be compared against real outcomes without yet affecting decisions. Shadow running for four to eight weeks produces the evidence that the model performs as expected on live data, and identifies any distribution shift between the training data and current transaction patterns. This evidence is essential for the compliance documentation that supports the change management record submitted to the FCA.
The fifth stage is graduated rollout: starting with a small proportion of transactions subject to the model's decisions, monitoring outcomes carefully, and expanding coverage as confidence builds. Full cutover comes only once the model's live performance metrics match or exceed the shadow run results. Our financial services software development practice manages this deployment process as a structured programme with explicit sign-off gates at each stage.
Published benchmarks from Stripe, PayPal, and UK banking deployments consistently show that ML-based fraud detection reduces false positive rates by 30 to 60 percent compared to rule-based systems at equivalent recall (fraud detection rates). In practical terms, this means fewer legitimate transactions are declined, improving customer experience, while actual fraud detection improves simultaneously. The improvement comes from the model's ability to consider hundreds of features simultaneously and learn individual customer behaviour patterns, which rule-based systems cannot do at scale.
A fraud detection model needs labelled historical transaction data (confirmed fraud and confirmed legitimate), transaction metadata (amount, time, channel, device, location, beneficiary), account history features (how long the account has been open, typical transaction patterns, previous flags), and ideally device and session signals. The minimum viable dataset for training a supervised model is typically 50,000 to 100,000 transactions with at least 1,000 confirmed fraud cases. Below this volume, statistical reliability becomes a concern and alternative approaches (rules augmented by unsupervised anomaly detection) are more appropriate.
When a customer queries a declined payment or blocked account, the firm must be able to provide an explanation that does not reveal information that would help fraudsters circumvent detection. This requires a tiered explanation approach: a high-level reason for the customer ("we detected activity inconsistent with your usual patterns and have placed a temporary hold pending verification"), and a detailed internal explanation for compliance and fraud teams that shows the specific feature values and model scores that contributed to the decision. SHAP (SHapley Additive exPlanations) values are the standard method for generating the detailed internal explanation from gradient-boosted tree models.
The PSR's mandatory reimbursement regime applies to all UK payment service providers that participate in the Faster Payments Service, which covers virtually all retail banks and most non-bank PSPs. Smaller PSPs below the PSR's de minimis threshold have some modifications to their obligations, but the core reimbursement requirement applies broadly. The regime covers personal and micro-enterprise accounts (businesses with fewer than ten employees and turnover under โฌ2 million) but not large corporate transactions, which are assessed under different fraud liability rules.
Yes, and this is a live concern in UK financial services. If a fraud model is trained on historical investigation data that reflects past biases in which customers were investigated (for example, if fraud teams historically focused investigations on certain demographics), the model will learn those biases and perpetuate them. The FCA's Consumer Duty and the Equality Act 2010 both create obligations to avoid discriminatory outcomes. Best practice is to conduct regular demographic parity testing (comparing false positive rates across demographic groups), use fairness-aware model training techniques where disparities are identified, and document the testing and results as part of the AI governance framework.
Let us help
Talk to our London-based team about how we can build the AI software, automation, or bespoke development tailored to your needs.
Deen Dayal Yadav
Online