15 May 2024
by Pierre-Antoine Boulat, Patrick Wake

Can AI-generated deepfakes compromise Know Your Customer (KYC) authentication?

Guest blog from Patrick Wake and Pierre-Antoine Boulat at FDM as part of our #UnleashInnovation campaign week 2024.

In the world of finance, processes like KYC (Know Your Customer) and AML (Anti Money laundering) are vital safeguards, verifying customer identities and preventing fraud. A robust KYC framework is essential for combating financial crimes, ensuring compliance, and protecting business reputation. With the rise of AI-generated deepfakes or document replicas, organisations are re-evaluating their KYC procedures for enhanced security. 

What are the challenges of deepfakes in KYC processes and how can businesses spot and address the threats?  

Pierre-Antoine Boulat, FDM North America Delivery Lead for Risk, Regulation & Compliance and Patrick Wake, FDM Group Director of Information Security, discuss.   

According to Pierre-Antoine Boulat, ‘An arm’s race between money launderers and financial compliance and law enforcement is all but certain. The tools that enable document counterfeiting, physical impersonation and convincing messages “en masse” are also deployed to counter such nefarious strategies and tactics: GenAI.’ 

According to Patrick Wake, ‘Deepfakes created through advanced AI techniques like generative adversarial networks (GANs), pose a significant challenge for identity verification processes. These AI-generated fakes can mimic various documents and personal details so convincingly that they threaten the reliability of traditional verification methods, raising concerns about security and fraud prevention.’ 

How deepfakes are made 

Patrick explains that ‘deep fakes can be created with openly available tools like Deepfakes Web and Faceswapper.ai, as well as community-created and downloaded from platforms like GitHub. This raises concerns about their misuse in identity checks, especially in KYC procedures.  

These models use neural networks to create realistic facial expressions, lip movements and voices, making the fake content look remarkably authentic. Techniques such as facial landmark detection and audio synthesis further enhance the believability of deepfakes.’ 

Mimicking documents and personal information 

Deepfake technology has shown it can replicate various documents and personal attributes quite accurately. Examples include fake robocalls impersonating public figures like US President Joe Biden and complex financial scams where deepfakes impersonate company executives, causing significant financial losses. Deepfake creators use AI algorithms to manipulate text, signatures, and other document features, creating fake documents that are hard to spot. Advances in natural language processing help generate realistic text, adding to the authenticity of fabricated documents. 

Improving Deepfake Detection 

Patrick says, ‘efforts are underway to enhance detection technologies in response to the growing threat of deepfakes. One approach involves using advanced machine learning algorithms trained on large datasets of real and fake content. These algorithms analyse subtle differences in facial expressions, language patterns and context to identify signs of manipulation. 

Researchers are also exploring biometric authentication methods like facial recognition and voice analysis to improve detection accuracy. Real-time detection systems are being developed to quickly analyse incoming media streams and flag suspicious content.’  

Pierre-Antoine says, ‘standardised patterns of behaviours or financial transaction history can be recognised as falsified by trained professionals, including digital doppelgängers. Human beings with the right risk management mindset and penetration techniques supported by AI will continue to stand against fraudulent methods. 

Furthermore, other techniques can be leveraged and will find expanded areas of application to complement Generative AI in the fight against counterfeit or fake identities. Video-proofing practices as well as common sense countermeasures difficult for trained electrons to master (think “Captcha” tests) will flourish. Blockchain certification and government-sponsored E-Identification will at last find their generic and welcome use.’ 

How to mitigate fraud risks 

Pierre-Antoine maintains that ‘the pandemic ways of working and FinTech innovation had introduced remote identification well before GenAI became headline material. Online fraud is a constant irritant, and devastating for its victims, but society through regulation and education will manage to fight back. 

Despite privacy concerns, platforming of shared identity details (already in use for credit scoring, criminal history, driving abilities, immigration statuses and health data for instance, across sectors and jurisdictions) will expand to physical features, since those are used for identification (prints, face, voice) as well as criminal purposes.  

Collaboration among industry stakeholders and regulatory bodies  

Patrick believes collaboration between industry, academia, and government is crucial for advancing detection solutions and staying ahead of new threats. Despite progress, challenges remain, including the risk of adversaries evading detection through clever manipulations. Continued research and innovation are needed to strengthen defences against deepfake threats. 

According to Pierre-Antoine, areas of collaboration between the public and private sectors, be it for profit or not, and government agencies including law enforcement are numerous in this field. For organizations, it’s an opportunity to converge physical and digital security counter-measures and adopt a holistic risk management posture. 

While deepfake technology poses significant challenges for identity verification, ongoing efforts to improve detection offer hope in the fight against fraud. By combining advanced algorithms, biometric authentication, real-time analysis, and collaborative efforts, businesses can better protect themselves and their customers from the dangers of AI-generated fake identities.


 View all campaign week content 

Unleashing Innovation campaign week graphic.png

techUK – Unleashing UK Tech and Innovation 

innovation_icon_badge_final.png

The UK is home to emerging technologies that have the power to revolutionise entire industries. From quantum to semiconductors; from gaming to the New Space Economy, they all have the unique opportunity to help prepare for what comes next.

techUK members lead the development of these technologies. Together we are working with Government and other stakeholders to address tech innovation priorities and build an innovation ecosystem that will benefit people, society, economy and the planet - and unleash the UK as a global leader in tech and innovation.

For more information, or to get in touch, please visit our Innovation Hub and click ‘contact us’. 


Tech and Innovation Summit, 19 June

Emerging technologies will be debated and explored at our annual Tech and Innovation Summit, taking place on 19 June. This campaign week will directly feed into the themes discussed at the Summit. 

techUK’s flagship Tech and Innovation Summit returns to traverse the extraordinary and ground-breaking discoveries made possible by the application of emerging and transformative technologies. 

Secure your ticket now.


Upcoming events:


Latest news and insights:


Get our tech and innovation insights straight to your inbox

Sign-up to get the latest updates and opportunities from our Technology and Innovation and AI programmes.


Learn more about our Unleashing Innovation campaign:

Unleashing the Potential of UK Tech and Innovation.jpg

 


Laura Foster

Laura Foster

Head of Technology and Innovation, techUK

Rory Daniels

Rory Daniels

Programme Manager, Emerging Technologies

Tess Buckley

Tess Buckley

Programme Manager - Digital Ethics and AI Safety, techUK

Usman Ikhlaq

Usman Ikhlaq

Programme Manager - Artificial Intelligence, techUK

Related topics

Authors

Pierre-Antoine Boulat

Pierre-Antoine Boulat

Lead for Risk, Regulation & Compliance, FDM

Patrick Wake

Patrick Wake

Group Director of Information Security, FDM