The finance manager at Arup, a global engineering firm, joined what appeared to be a routine video conference in September 2025.
The CFO was on the call. So were several colleagues from the London office. The video quality was perfect. The audio was crisp. Everyone looked and sounded exactly right.
The CFO explained that the company needed to execute several urgent wire transfers for a confidential acquisition. The amounts were large but not unprecedented for a firm of Arup's size. The colleagues confirmed the details. Everything followed proper procedures.
The finance manager, following what seemed like clear instructions from verified executives on a video call, authorized multiple wire transfers totaling $25 million.
Every single person on that call was an AI-generated deepfake.
Not one human. Not one real executive. Not one legitimate instruction.
$25 million gone because "I could see them on video" no longer means "they're real."
After building authentication systems for a CIAM platform that verified over a billion user identities, I thought I understood the fundamentals of identity verification. Something you know (password). Something you have (token). Something you are (biometric).
But here's what the Arup case proved: "something you are" is no longer something you are. Your face can be generated. Your voice can be cloned. Your video presence can be synthesized in real-time.
The foundation of identity verification just collapsed. And most organizations haven't noticed yet.
Let me show you why this isn't just about deepfakes, it's about the fundamental breakdown of how we verify identity in the modern enterprise. And what needs to change before your organization becomes the next $25 million lesson.
The details of the Arup fraud matter because they reveal how sophisticated these attacks have become.
Here's what we know happened:
Before the attack began, someone did extensive research on Arup:
Sources for this intelligence:
None of this required hacking. It was all publicly available.
When building the CIAM platform, I always told clients: assume attackers know everything that's public. The Arup case shows they're now using that information to generate perfect replicas.
Using the gathered intelligence, attackers created deepfakes of:
The technology isn't science fiction anymore:
The cost? A few hundred dollars worth of commercial AI services.
The skill required? Moderate technical knowledge, not expert-level.
The time investment? Days, not months.
This isn't nation-state capability. This is accessible to organized crime.
The attackers didn't start with "$25 million wire transfer."
They built credibility first:
This is sophisticated social engineering:
The finance manager wasn't stupid. The finance manager was experiencing a perfectly executed attack that exploited every assumption about identity verification.
Once credibility was established, the attack executed:
Request 1: "We need to move funds for the acquisition we discussed."
Request 2: "Execute transfers to these accounts."
Request 3: "Process these immediately. This is confidential."
The finance manager followed procedure. Verified identities through video. Confirmed with multiple participants. Executed authorized instructions.
Everything was wrong. Everything looked right.
The fraud was discovered when:
By then, $25 million had been transferred across multiple jurisdictions, making recovery extremely difficult.
The investigation revealed:
The attack bypassed all technical controls by exploiting human identity verification.
For all of human history until approximately 2023, seeing someone's face and hearing their voice was reliable proof of identity.
That assumption is now broken.
Here's why:
The deepfakes in the Arup attack weren't static videos. They were real-time interactive participants in a video conference.
What that means:
This requires:
This wasn't possible two years ago. It's routine in 2026.
The technology curve on deepfakes followed the same trajectory as large language models: long period of "not quite there" followed by sudden achievement of human-indistinguishable quality.
We're past the inflection point. Deepfakes are now perfect.
Voice authentication, once considered secure, is now completely compromised.
What it takes to clone a voice:
When building the CIAM platform, I evaluated voice biometrics for multi-factor authentication. We didn't deploy it because we knew voice could eventually be cloned.
What we didn't predict was how quickly "eventually" would arrive.
Voice authentication is not just vulnerable. It's actively counterproductive because it creates false confidence in identity verification.
Here's what makes the Arup case especially dangerous: the victim did everything right according to conventional security training.
Conventional wisdom says:
All of that happened. All of it was useless.
Video verification created a false sense of security. "I can see them" felt like proof. It wasn't.
This is why I'm worried about organizations rushing to implement video-based authentication for sensitive operations. They're building security controls that attackers have already defeated.
Older deepfakes had tells:
Skilled observers could detect fakes.
Modern deepfakes don't have those tells. The "uncanny valley" (the uncomfortable feeling when something is almost-but-not-quite human) has been crossed.
Current deepfakes are indistinguishable from real people to human perception.
This means:
When building the CIAM platform, I assumed biometric data (what you are) was harder to compromise than passwords (what you know).
That assumption is now backwards. Passwords can be changed. Your face and voice cannot.
The Arup attack cost $25 million. But here's what should terrify CFOs: the attack probably cost less than $10,000 to execute.
Attacker investment:
Total cost: $5,000-10,000 (including time at black market rates)
Return: $25,000,000
ROI: 2,500x to 5,000x
The economics are absurd. Even if only 1 in 100 attempts succeeds, the math works overwhelmingly in attackers' favor.
This is why deepfake fraud will explode in 2026. It's not just technically possible. It's economically inevitable.
Traditional security followed an economic principle: make attacks expensive enough that they're not worth executing.
Deepfake fraud breaks that principle.
Defense costs:
Attack costs: $5,000-10,000
Defenders must protect against all attacks. Attackers only need one success.
The economic asymmetry is crushing.
When building the CIAM platform, I secured systems by making unauthorized access expensive (technically difficult). We can't make deepfakes expensive. The technology is commoditized.
This requires a different defense strategy entirely.
Arup isn't alone. Deepfake fraud is hitting every sector that relies on voice or video identity verification.
Current vulnerability:
Real incidents (2025-2026):
Why it's getting worse:
The CEO doppelgänger problem:
Attack scenarios:
This isn't hypothetical. Security researchers estimate 60-80% of Fortune 500 CEOs have enough public footage for high-quality deepfake generation.
Emerging problem:
The legal system hasn't caught up:
These are questions courts will face in 2026 and beyond.
Beyond fraud:
The trust crisis: When citizens can't distinguish real officials from deepfakes, governance itself becomes unstable.
The Arup case is about money. The broader crisis is about trust.
Security vendors are rushing to sell "deepfake detection" solutions. Most won't work at scale.
Here's what actually works:
❌ Training employees to spot deepfakes
The tells that used to identify deepfakes don't exist anymore. Training people to look for artifacts that aren't there is security theater.
❌ Voice biometric authentication
Voice can be cloned perfectly. Using voice as authentication factor is worse than useless—it creates false confidence.
❌ Video-only verification for high-value transactions
The Arup case proved this. Seeing someone on video is not proof of identity.
❌ Relying on "trusted" video platforms
Deepfakes work on Zoom, Teams, Google Meet—any platform. The platform isn't the vulnerability. The human perception is.
❌ Deepfake detection software
Current detection has high false positive/negative rates. As deepfakes improve, detection becomes harder. This is an arms race defenders will lose.
✓ Multi-channel verification
If someone makes a high-value request via video, verify through a completely different channel.
Example:
The principle: Deepfakes excel in single channel. They fail when you verify through independent channels.
When building the CIAM platform, I implemented this as "multi-factor authentication" for users. The same principle applies to verifying humans making requests.
✓ Pre-established verification protocols
Before high-stakes situations arise, establish verification procedures:
For financial transactions:
For executive communications:
✓ Physical tokens for critical operations
For the highest-stakes transactions, require physical token possession:
This is the "something you have" factor that deepfakes can't fake remotely.
✓ Time delays and review periods
Most fraud relies on urgency. Removing urgency defeats the attack.
Implementation:
The attacker's nightmare: Time for victim to verify through multiple channels.
✓ Behavioral analysis and anomaly detection
Technology can't detect deepfakes reliably, but it can detect anomalous requests:
Example from Arup case: A behavioral system might have flagged:
This doesn't detect the deepfake. It detects the unusual request pattern.
When building the CIAM platform, I used anomaly detection to flag unusual authentication patterns. The same approach works for detecting deepfake-enabled fraud attempts.
Organizations need to rebuild identity verification with a fundamental assumption: audio and video are never, by themselves, proof of identity.
Here's what that means in practice:
Old model:
New model:
The shift: Audio/video are claim of identity, not proof of identity.
Old model:
New model:
The shift: Seeing and hearing someone is start of verification, not end of verification.
Old model:
New model:
The shift: Video alone has no evidentiary value without additional verification.
Old model:
New model:
The shift: Customer identity is continuous verification, not one-time confirmation.
This is fundamentally different from traditional customer identity and access management. We're not just securing access. We're assuming the entire audio/visual channel is compromised.
If you're responsible for security, compliance, or risk management, here's your action plan:
1. Identify high-value audio/video verification points
Where does your organization currently accept audio or video as proof of identity?
Map every instance. That's your immediate vulnerability.
2. Implement emergency verification protocols
For highest-risk operations:
This is a bandaid, not a solution. But it reduces immediate risk.
3. Alert high-risk employees
Finance teams, executive assistants, controllers, anyone with authority to execute high-value transactions.
Key message:
Make this a standing security briefing topic.
4. Develop formal verification procedures
Document specific protocols:
For financial transactions over $X:
For executive communications:
For customer service:
Publish these protocols. Train teams. Enforce compliance.
5. Audit current authentication methods
Where are you using voice or video as authentication factor?
Replace pure voice/video authentication with multi-factor requirements.
6. Review insurance coverage
Does your cyber insurance cover deepfake fraud?
Update policies to explicitly cover deepfake scenarios.
7. Implement behavioral analysis
Deploy systems that flag anomalous requests:
This won't detect deepfakes. It will detect the fraud attempts that use deepfakes.
8. Establish physical token requirements
For highest-stakes operations:
Yes, this reduces efficiency. That's the point.
The Arup case happened because efficiency was prioritized over verification. Sometimes friction is security.
9. Create escalation and response procedures
What happens when suspected deepfake detected?
Have this documented before the incident occurs.
10. Rebuild identity verification architecture
This is the fundamental fix:
When we implemented zero-trust architecture for the CIAM platform, it wasn't a quick project. It was fundamental rethinking of how identity works.
Organizations need the same rethink for audio/video identity verification.
11. Partner with industry on standards
Individual organizations can't solve this alone.
Needed:
This is infrastructure problem, not individual company problem.
12. Prepare for regulatory changes
Regulations will come (probably after high-profile fraud makes headlines).
Likely requirements:
Organizations that prepare now will comply easily. Those that wait will scramble.
The Arup case is about $25 million. But the implications go far beyond one fraud.
We're entering an era where:
This breaks fundamental human communication assumptions.
For all of human history, if you saw someone's face and heard their voice, you could be reasonably confident it was them. That certainty is gone.
Beyond corporate fraud:
The erosion of trust in audio/visual communication has societal consequences beyond security.
Courts rely on:
All of this becomes problematic when deepfakes are perfect.
Legal systems will need to establish new standards for evidence authentication in the deepfake era.
Imagine:
The potential for market manipulation, political chaos, and social disruption is enormous.
The Arup $25 million fraud is a preview. The real crisis is when trust in media itself becomes impossible.
Organizations want a technology solution to the deepfake problem. "Install detection software. Train the AI. Filter the fakes."
That's not going to work.
Deepfake generation and deepfake detection are in an arms race. Generators are winning. Every improvement in detection gets incorporated into better generation.
The only sustainable solution is to stop relying on audio and video as proof of identity.
This means:
Nobody wants this. Organizations optimized for efficiency, not security friction.
But the alternative is Arup-scale fraud becoming routine.
The choice:
There's no third option where you get efficiency, low overhead, AND protection from deepfakes.
When building the CIAM platform, I learned that security and convenience are often inversely correlated. The most secure systems have the most friction.
In the deepfake era, that friction is non-negotiable for high-value operations.
An employee at Arup saw executives on video. Heard them speaking. Verified the request through what appeared to be proper channels. Authorized $25 million in transfers.
Every executive was an AI-generated deepfake.
This wasn't a failure of the employee. It was a failure of assumptions. The assumption that video calls prove identity. The assumption that seeing and hearing someone means they're real.
Those assumptions are now liabilities.
For organizations:
For individuals:
For society:
The deepfake era is here. The technology that enables $25 million fraud is commercially available for a few hundred dollars.
The question isn't whether deepfake fraud will become common. The question is whether organizations will adapt their identity verification before or after they become victims.
Arup learned at $25 million cost. Your organization can learn from them instead.
Identity verification is broken. Seeing is no longer believing. The sooner we accept that, the sooner we can build systems that actually work.
Building authentication systems that resist deepfakes? My Customer Identity Hub covers multi-factor authentication, zero-trust principles, and modern CIAM architecture that assumes all biometric channels are compromised.
Deepak Gupta is the co-founder and CEO of GrackerAI. He previously founded a CIAM platform that scaled to serve 1B+ users globally. He writes about AI, cybersecurity, and digital identity at guptadeepak.com.
*** This is a Security Bloggers Network syndicated blog from Deepak Gupta | AI & Cybersecurity Innovation Leader | Founder's Journey from Code to Scale authored by Deepak Gupta - Tech Entrepreneur, Cybersecurity Author. Read the original post at: https://guptadeepak.com/the-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted/