The $25 Million Deepfake: Why Your Video Calls Can No Longer Be Trusted
None
<p><img decoding="async" src="https://images.unsplash.com/photo-1568777984968-2ce482b45aaa?crop=entropy&cs=tinysrgb&fit=max&fm=jpg&ixid=M3wxMTc3M3wwfDF8c2VhcmNofDN8fGRlZXBmYWtlfGVufDB8fHx8MTc3NDIxODE5M3ww&ixlib=rb-4.1.0&q=80&w=2000" alt="The $25 Million Deepfake: Why Your Video Calls Can No Longer Be Trusted"></p><p>The finance manager at Arup, a global engineering firm, joined what appeared to be a routine video conference in September 2025.</p><p>The CFO was on the call. So were several colleagues from the London office. The video quality was perfect. The audio was crisp. Everyone looked and sounded exactly right.</p><p>The CFO explained that the company needed to execute several urgent wire transfers for a confidential acquisition. The amounts were large but not unprecedented for a firm of Arup's size. The colleagues confirmed the details. Everything followed proper procedures.</p><p>The finance manager, following what seemed like clear instructions from verified executives on a video call, authorized multiple wire transfers totaling <strong>$25 million</strong>.</p><p>Every single person on that call was an <a href="https://www.weforum.org/stories/2025/02/deepfake-ai-cybercrime-arup/" rel="noreferrer">AI-generated deepfake</a>.</p><p>Not one human. Not one real executive. Not one legitimate instruction.</p><p><strong>$25 million gone because "I could see them on video" no longer means "they're real."</strong></p><p>After building authentication systems for a CIAM platform that verified over a billion user identities, I thought I understood the fundamentals of identity verification. Something you know (password). Something you have (token). Something you are (biometric).</p><p>But here's what the Arup case proved: <strong>"something you are" is no longer something you are</strong>. Your face can be generated. Your voice can be cloned. Your video presence can be synthesized in real-time.</p><p>The foundation of identity verification just collapsed. And most organizations haven't noticed yet.</p><p>Let me show you why this isn't just about deepfakes, it's about the fundamental breakdown of how we verify identity in the modern enterprise. And what needs to change before your organization becomes the next $25 million lesson.</p><h2 id="how-the-arup-attack-actually-worked">How the Arup Attack Actually Worked</h2><p>The details of the Arup fraud matter because they reveal how sophisticated these attacks have become.</p><p>Here's what we know happened:</p><h3 id="phase-1-intelligence-gathering">Phase 1: Intelligence Gathering</h3><p>Before the attack began, someone did extensive research on Arup:</p><ul> <li>Organizational structure (who reports to whom)</li> <li>Key personnel (names, roles, responsibilities)</li> <li>Communication patterns (how does the finance team interact with executives)</li> <li>Previous transactions (what kinds of transfers are normal)</li> <li>Visual and audio profiles (what do these people look and sound like)</li> </ul><p><strong>Sources for this intelligence:</strong></p><ul> <li>LinkedIn profiles (organizational hierarchy, recent posts)</li> <li>Corporate website (executive bios, photos)</li> <li>Conference videos (speaking style, mannerisms, voice patterns)</li> <li>Publicly available annual reports (business operations, deal patterns)</li> <li>Social media (personal accounts revealing additional context)</li> </ul><p>None of this required hacking. It was all publicly available.</p><p>When building the CIAM platform, I always told clients: <strong>assume attackers know everything that's public</strong>. The Arup case shows they're now using that information to generate perfect replicas.</p><h3 id="phase-2-deepfake-generation">Phase 2: Deepfake Generation</h3><p>Using the gathered intelligence, attackers created deepfakes of:</p><ul> <li>The CFO (video and audio)</li> <li>Multiple colleagues (video and audio)</li> <li>Realistic background environments (office settings)</li> </ul><p><strong>The technology isn't science fiction anymore:</strong></p><ul> <li>Real-time video synthesis (ElevenLabs, Synthesia, HeyGen)</li> <li>Voice cloning from minutes of audio samples</li> <li>Facial animation that matches speech perfectly</li> <li>Background generation that looks like real offices</li> </ul><p>The cost? <strong>A few hundred dollars worth of commercial AI services.</strong></p><p>The skill required? <strong>Moderate technical knowledge, not expert-level.</strong></p><p>The time investment? <strong>Days, not months.</strong></p><p>This isn't nation-state capability. This is accessible to organized crime.</p><h3 id="phase-3-social-engineering-setup">Phase 3: Social Engineering Setup</h3><p>The attackers didn't start with "$25 million wire transfer."</p><p>They built credibility first:</p><ul> <li>Initial "meeting" was about routine business review</li> <li>Early requests were for information, not action</li> <li>Communication style matched known executive patterns</li> <li>Topics referenced real company initiatives</li> </ul><p><strong>This is sophisticated social engineering:</strong></p><ul> <li>Establish normalcy before making unusual requests</li> <li>Build trust through familiar interactions</li> <li>Reference real context to appear legitimate</li> <li>Escalate requests gradually</li> </ul><p>The finance manager wasn't stupid. The finance manager was experiencing a perfectly executed attack that exploited every assumption about identity verification.</p><h3 id="phase-4-the-kill-chain">Phase 4: The Kill Chain</h3><p>Once credibility was established, the attack executed:</p><p><strong>Request 1:</strong> "We need to move funds for the acquisition we discussed."</p><ul> <li>Context established (previous meetings mentioned acquisition)</li> <li>Authority confirmed (CFO giving instruction)</li> <li>Urgency implied (time-sensitive deal)</li> </ul><p><strong>Request 2:</strong> "Execute transfers to these accounts."</p><ul> <li>Account details provided (looked like legitimate intermediaries)</li> <li>Amounts specified (large but plausible for acquisition)</li> <li>Colleagues confirm (multiple people validate instruction)</li> </ul><p><strong>Request 3:</strong> "Process these immediately. This is confidential."</p><ul> <li>Time pressure (need to move fast)</li> <li>Secrecy (don't verify through normal channels)</li> <li>Implicit threat (challenging CFO's instruction seems insubordinate)</li> </ul><p>The finance manager followed procedure. Verified identities through video. Confirmed with multiple participants. Executed authorized instructions.</p><p><strong>Everything was wrong. Everything looked right.</strong></p><h3 id="phase-5-discovery-and-aftermath">Phase 5: Discovery and Aftermath</h3><p>The fraud was discovered when:</p><ul> <li>Actual CFO inquired about unusual account balances</li> <li>Finance team mentioned the "acquisition transfers"</li> <li>Real CFO: "What acquisition? What transfers?"</li> </ul><p>By then, <strong>$25 million had been transferred across multiple jurisdictions</strong>, making recovery extremely difficult.</p><p>The investigation revealed:</p><ul> <li>No systems were hacked</li> <li>No credentials were stolen</li> <li>No malware was installed</li> <li>Security infrastructure worked perfectly</li> </ul><p><strong>The attack bypassed all technical controls by exploiting human identity verification.</strong></p><h2 id="why-seeing-is-believing-no-longer-works">Why "Seeing Is Believing" No Longer Works</h2><p>For all of human history until approximately 2023, seeing someone's face and hearing their voice was reliable proof of identity.</p><p><strong>That assumption is now broken.</strong></p><p>Here's why:</p><h3 id="generative-ai-achieved-real-time-fidelity">Generative AI Achieved Real-Time Fidelity</h3><p>The deepfakes in the Arup attack weren't static videos. They were <strong>real-time interactive participants</strong> in a video conference.</p><p><strong>What that means:</strong></p><ul> <li>Deepfakes responded to questions in real-time</li> <li>Facial expressions matched speech naturally</li> <li>Body language appeared authentic</li> <li>Multiple participants interacted with each other convincingly</li> </ul><p>This requires:</p><ul> <li>Real-time video generation (no lag between audio and facial movement)</li> <li>Contextual understanding (responses relevant to conversation)</li> <li>Multiple simultaneous deepfakes (several people on same call)</li> <li>Environmental consistency (all in appropriate office settings)</li> </ul><p><strong>This wasn't possible two years ago. It's routine in 2026.</strong></p><p>The technology curve on deepfakes followed the same trajectory as large language models: long period of "not quite there" followed by sudden achievement of human-indistinguishable quality.</p><p>We're past the inflection point. Deepfakes are now perfect.</p><h3 id="voice-cloning-became-trivial">Voice Cloning Became Trivial</h3><p>Voice authentication, once considered secure, is now completely compromised.</p><p><strong>What it takes to clone a voice:</strong></p><ul> <li>3-5 minutes of target audio (obtained from conference videos, podcasts, earnings calls)</li> <li>Commercial voice cloning service (ElevenLabs, PlayHT, many others)</li> <li>$20-100 in API costs</li> <li>Result: Perfect voice replica that can speak any text in target's voice</li> </ul><p>When building the CIAM platform, I evaluated voice biometrics for <a href="https://guptadeepak.com/customer-identity-hub/mfa-implementation-in-ciam" rel="noreferrer">multi-factor authentication</a>. We didn't deploy it because <strong>we knew voice could eventually be cloned</strong>.</p><p>What we didn't predict was how quickly "eventually" would arrive.</p><p><strong>Voice authentication is not just vulnerable. It's actively counterproductive</strong> because it creates false confidence in identity verification.</p><h3 id="video-verification-creates-false-security">Video Verification Creates False Security</h3><p>Here's what makes the Arup case especially dangerous: <strong>the victim did everything right according to conventional security training</strong>.</p><p><strong>Conventional wisdom says:</strong></p><ul> <li>Verify requests through multiple channels ✓ (video call with multiple people)</li> <li>Confirm identity before executing high-value transactions ✓ (saw and heard executives)</li> <li>Follow approval procedures ✓ (got authorization from proper authority)</li> </ul><p><strong>All of that happened. All of it was useless.</strong></p><p>Video verification created a false sense of security. "I can see them" felt like proof. It wasn't.</p><p>This is why I'm worried about organizations rushing to implement video-based authentication for sensitive operations. They're building security controls that attackers have already defeated.</p><h3 id="the-uncanny-valley-is-behind-us">The Uncanny Valley Is Behind Us</h3><p>Older deepfakes had tells:</p><ul> <li>Unnatural blinking patterns</li> <li>Lip sync slightly off</li> <li>Facial expressions didn't match emotion</li> <li>Lighting inconsistencies</li> <li>Audio artifacts</li> </ul><p><strong>Skilled observers could detect fakes.</strong></p><p>Modern deepfakes don't have those tells. The "uncanny valley" (the uncomfortable feeling when something is almost-but-not-quite human) has been crossed.</p><p><strong>Current deepfakes are indistinguishable from real people to human perception.</strong></p><p>This means:</p><ul> <li>Training employees to "spot deepfakes" is like training them to spot perfect forgeries—impossible</li> <li>Visual inspection is no longer a viable verification method</li> <li>Trusting your eyes and ears is now a vulnerability</li> </ul><p>When building the CIAM platform, I assumed <strong>biometric data (what you are) was harder to compromise than passwords (what you know)</strong>.</p><p>That assumption is now backwards. <strong>Passwords can be changed. Your face and voice cannot.</strong></p><h2 id="the-economic-reality-that-makes-this-worse">The Economic Reality That Makes This Worse</h2><p>The Arup attack cost $25 million. But here's what should terrify CFOs: <strong>the attack probably cost less than $10,000 to execute</strong>.</p><h3 id="cost-benefit-analysis-for-attackers">Cost-Benefit Analysis for Attackers</h3><p><strong>Attacker investment:</strong></p><ul> <li>Deepfake technology: $500-2,000 (commercial AI services)</li> <li>Voice cloning: $100-500 (audio samples and processing)</li> <li>Research time: 40-80 hours (gathering intelligence on target)</li> <li>Technical execution: 20-40 hours (creating deepfakes, orchestrating call)</li> </ul><p><strong>Total cost: $5,000-10,000</strong> (including time at black market rates)</p><p><strong>Return: $25,000,000</strong></p><p><strong>ROI: 2,500x to 5,000x</strong></p><p><strong>The economics are absurd.</strong> Even if only 1 in 100 attempts succeeds, the math works overwhelmingly in attackers' favor.</p><p>This is why deepfake fraud will explode in 2026. It's not just technically possible. It's economically inevitable.</p><h3 id="why-targets-cant-outspend-attackers">Why Targets Can't Outspend Attackers</h3><p>Traditional security followed an economic principle: <strong>make attacks expensive enough that they're not worth executing</strong>.</p><p>Deepfake fraud breaks that principle.</p><p><strong>Defense costs:</strong></p><ul> <li>Implementing multi-channel verification: $50,000-500,000</li> <li>Training all employees on deepfake awareness: $100,000-1,000,000</li> <li>Deploying deepfake detection technology: $200,000-2,000,000</li> <li>Creating verification procedures: Ongoing operational cost</li> </ul><p><strong>Attack costs:</strong> $5,000-10,000</p><p><strong>Defenders must protect against all attacks. Attackers only need one success.</strong></p><p>The economic asymmetry is crushing.</p><p>When building the CIAM platform, I secured systems by making unauthorized access expensive (technically difficult). We can't make deepfakes expensive. The technology is commoditized.</p><p><strong>This requires a different defense strategy entirely.</strong></p><h2 id="the-identity-verification-crisis-across-industries">The Identity Verification Crisis Across Industries</h2><p>Arup isn't alone. Deepfake fraud is hitting every sector that relies on voice or video identity verification.</p><h3 id="finance-and-banking">Finance and Banking</h3><p><strong>Current vulnerability:</strong></p><ul> <li>Wire transfer approvals often use phone verification</li> <li>Large transactions require executive authorization</li> <li>Multi-signature processes assume you can verify signers</li> </ul><p><strong>Real incidents (2025-2026):</strong></p><ul> <li>Hong Kong-based company lost $26 million (employee fooled by deepfake video conference)</li> <li>Bank executive authorized fraudulent loan based on cloned CEO voice</li> <li>Investment firm manipulated by deepfake board member in virtual meeting</li> </ul><p><strong>Why it's getting worse:</strong></p><ul> <li>Remote work means video calls replace in-person verification</li> <li>International transactions make callback verification complicated</li> <li>Time-sensitive deals create pressure to "trust but verify quickly"</li> </ul><h3 id="corporate-executives">Corporate Executives</h3><p><strong>The CEO doppelgänger problem:</strong></p><ul> <li>Executives are high-value deepfake targets (lots of public footage)</li> <li>Their voices carry authority for financial decisions</li> <li>They travel frequently (making "I'm in a meeting, use video" plausible)</li> </ul><p><strong>Attack scenarios:</strong></p><ul> <li>CFO instructs controller to execute transfers</li> <li>CEO approves emergency expenditure</li> <li>Board member votes on virtual acquisition approval</li> </ul><p>This isn't hypothetical. <strong>Security researchers estimate 60-80% of Fortune 500 CEOs have enough public footage for high-quality deepfake generation.</strong></p><h3 id="legal-and-compliance">Legal and Compliance</h3><p><strong>Emerging problem:</strong></p><ul> <li>Video depositions becoming unreliable</li> <li>Remote notarization vulnerable to deepfake impersonation</li> <li>Legal agreements via video call losing evidentiary value</li> </ul><p><strong>The legal system hasn't caught up:</strong></p><ul> <li>What constitutes proof of identity in video format?</li> <li>How do you authenticate video evidence when deepfakes are perfect?</li> <li>Can contracts signed via video conference be enforced?</li> </ul><p>These are questions courts will face in 2026 and beyond.</p><h3 id="political-and-regulatory">Political and Regulatory</h3><p><strong>Beyond fraud:</strong></p><ul> <li>Deepfake political figures making false statements</li> <li>Fabricated emergency announcements causing market panic</li> <li>Synthetic "official" communications triggering policy responses</li> </ul><p><strong>The trust crisis:</strong> When citizens can't distinguish real officials from deepfakes, <strong>governance itself becomes unstable</strong>.</p><p>The Arup case is about money. The broader crisis is about trust.</p><h2 id="what-actually-works-and-what-doesnt">What Actually Works (And What Doesn't)</h2><p>Security vendors are rushing to sell "deepfake detection" solutions. Most won't work at scale.</p><p>Here's what actually works:</p><h3 id="what-doesnt-work">What Doesn't Work</h3><p><strong>❌ Training employees to spot deepfakes</strong></p><p>The tells that used to identify deepfakes don't exist anymore. Training people to look for artifacts that aren't there is security theater.</p><p><strong>❌ Voice biometric authentication</strong></p><p>Voice can be cloned perfectly. Using voice as authentication factor is worse than useless—it creates false confidence.</p><p><strong>❌ Video-only verification for high-value transactions</strong></p><p>The Arup case proved this. Seeing someone on video is not proof of identity.</p><p><strong>❌ Relying on "trusted" video platforms</strong></p><p>Deepfakes work on Zoom, Teams, Google Meet—any platform. The platform isn't the vulnerability. The human perception is.</p><p><strong>❌ Deepfake detection software</strong></p><p>Current detection has high false positive/negative rates. As deepfakes improve, detection becomes harder. This is an arms race defenders will lose.</p><h3 id="what-does-work">What Does Work</h3><p><strong>✓ Multi-channel verification</strong></p><p>If someone makes a high-value request via video, <strong>verify through a completely different channel</strong>.</p><p><strong>Example:</strong></p><ul> <li>Request comes via video call</li> <li>Callback to known phone number (not number provided in call)</li> <li>Verify request details through email to known address</li> <li>Use previously established code word or verification phrase</li> </ul><p><strong>The principle:</strong> Deepfakes excel in single channel. They fail when you verify through independent channels.</p><p>When building the CIAM platform, I implemented this as "<a href="https://guptadeepak.com/customer-identity-hub/types-of-multi-factor-authentication-mfa-methods" rel="noreferrer">multi-factor authentication</a>" for users. <strong>The same principle applies to verifying humans making requests.</strong></p><p><strong>✓ Pre-established verification protocols</strong></p><p>Before high-stakes situations arise, establish verification procedures:</p><p><strong>For financial transactions:</strong></p><ul> <li>Code words known only to authorized parties</li> <li>Out-of-band confirmation required for amounts over threshold</li> <li>Time delays between authorization and execution (giving time to detect fraud)</li> </ul><p><strong>For executive communications:</strong></p><ul> <li>Verified phone numbers for callback (updated regularly)</li> <li>Secondary confirmation via different medium (video → email with digital signature)</li> <li>Predetermined questions only real executive would know</li> </ul><p><strong>✓ Physical tokens for critical operations</strong></p><p>For the highest-stakes transactions, require <strong>physical token possession</strong>:</p><ul> <li>Hardware security keys (YubiKey, Titan)</li> <li>Smart cards with PIN</li> <li>Biometric device in person (not over video)</li> </ul><p>This is the "something you have" factor that deepfakes can't fake remotely.</p><p><strong>✓ Time delays and review periods</strong></p><p>Most fraud relies on urgency. <strong>Removing urgency defeats the attack.</strong></p><p><strong>Implementation:</strong></p><ul> <li>All transactions over $X have mandatory 24-hour delay</li> <li>During delay, multiple verification channels used</li> <li>Any discrepancy halts transaction immediately</li> </ul><p><strong>The attacker's nightmare:</strong> Time for victim to verify through multiple channels.</p><p><strong>✓ Behavioral analysis and anomaly detection</strong></p><p>Technology can't detect deepfakes reliably, but it can detect <strong>anomalous requests</strong>:</p><ul> <li>Is this transaction pattern unusual for this executive?</li> <li>Is the requested amount outside normal parameters?</li> <li>Is the urgency level inconsistent with typical behavior?</li> <li>Are account destinations new or unfamiliar?</li> </ul><p><strong>Example from Arup case:</strong> A behavioral system might have flagged:</p><ul> <li>Multiple large transfers to new accounts</li> <li>Urgency combined with confidentiality (red flag combo)</li> <li>Request coming through video call rather than written authorization</li> </ul><p>This doesn't detect the deepfake. It detects the unusual request pattern.</p><p>When building the CIAM platform, I used anomaly detection to flag unusual authentication patterns. <strong>The same approach works for detecting deepfake-enabled fraud attempts.</strong></p><h2 id="the-new-security-model-never-trust-audiovideo-alone">The New Security Model: Never Trust Audio/Video Alone</h2><p>Organizations need to rebuild identity verification with a fundamental assumption: <strong>audio and video are never, by themselves, proof of identity</strong>.</p><p>Here's what that means in practice:</p><h3 id="for-financial-operations">For Financial Operations</h3><p><strong>Old model:</strong></p><ul> <li>CFO calls → controller executes transfer</li> <li>Video conference with executives → approve transaction</li> <li>Phone verification → process high-value request</li> </ul><p><strong>New model:</strong></p><ul> <li><strong>ANY</strong> request (phone, video, email, in-person) → multi-channel verification</li> <li>High-value transactions → mandatory delay + callback + written confirmation</li> <li>Critical operations → physical token requirement</li> </ul><p><strong>The shift:</strong> Audio/video are <strong>claim of identity</strong>, not <strong>proof of identity</strong>.</p><h3 id="for-executive-communications">For Executive Communications</h3><p><strong>Old model:</strong></p><ul> <li>Recognize voice → trust instruction</li> <li>See face on video → accept authorization</li> <li>Email from executive address → follow directive</li> </ul><p><strong>New model:</strong></p><ul> <li>Voice/video establishes <strong>who is claiming to be</strong> on call</li> <li>Verification protocol confirms <strong>who actually is</strong> making request</li> <li>Written confirmation with digital signature provides audit trail</li> </ul><p><strong>The shift:</strong> Seeing and hearing someone is <strong>start of verification</strong>, not <strong>end of verification</strong>.</p><h3 id="for-legal-and-compliance">For Legal and Compliance</h3><p><strong>Old model:</strong></p><ul> <li>Video deposition → legally binding testimony</li> <li>Remote notarization via video → official document</li> <li>Video signature → enforceable contract</li> </ul><p><strong>New model:</strong></p><ul> <li>Video deposition → supplemented with in-person verification or physical token</li> <li>Remote notarization → requires multiple verification factors</li> <li>Video signature → paired with blockchain timestamping and out-of-band confirmation</li> </ul><p><strong>The shift:</strong> Video alone has <strong>no evidentiary value</strong> without additional verification.</p><h3 id="for-customer-service-and-support">For Customer Service and Support</h3><p><strong>Old model:</strong></p><ul> <li>Customer calls → verify with personal info → authorize account changes</li> <li>Video chat for sensitive requests → accept if ID shown on video</li> </ul><p><strong>New model:</strong></p><ul> <li>Voice/video → establishes session, not identity</li> <li>Multi-factor authentication required for any account modification</li> <li>Out-of-band confirmation (email, SMS to verified address) before changes execute</li> </ul><p><strong>The shift:</strong> Customer identity is <strong>continuous verification</strong>, not <strong>one-time confirmation</strong>.</p><p>This is fundamentally different from traditional <a href="https://guptadeepak.com/ciam-basics-a-comprehensive-guide-to-customer-identity-and-access-management-in-2025/" rel="noreferrer">customer identity and access management</a>. We're not just securing access. We're assuming the entire audio/visual channel is compromised.</p><h2 id="what-organizations-must-do-right-now">What Organizations Must Do Right Now</h2><p>If you're responsible for security, compliance, or risk management, here's your action plan:</p><h3 id="immediate-this-week">Immediate (This Week)</h3><p><strong>1. Identify high-value audio/video verification points</strong></p><p>Where does your organization currently accept audio or video as proof of identity?</p><ul> <li>Wire transfer approvals</li> <li>Vendor payment authorizations</li> <li>Contract signatures</li> <li>Executive instructions</li> <li>Password resets</li> <li>Account modifications</li> </ul><p><strong>Map every instance.</strong> That's your immediate vulnerability.</p><p><strong>2. Implement emergency verification protocols</strong></p><p>For highest-risk operations:</p><ul> <li>Callback to verified numbers (in phone directory, not caller ID)</li> <li>Email confirmation to known addresses</li> <li>Require 24-hour delay on large transactions</li> </ul><p><strong>This is a bandaid, not a solution.</strong> But it reduces immediate risk.</p><p><strong>3. Alert high-risk employees</strong></p><p>Finance teams, executive assistants, controllers, anyone with authority to execute high-value transactions.</p><p><strong>Key message:</strong></p><ul> <li>Video calls can be perfect deepfakes</li> <li>Voice calls can be cloned executives</li> <li><strong>Never</strong> trust audio/video alone for financial decisions</li> <li>Always verify through second channel</li> </ul><p><strong>Make this a standing security briefing topic.</strong></p><h3 id="short-term-this-month">Short-term (This Month)</h3><p><strong>4. Develop formal verification procedures</strong></p><p>Document specific protocols:</p><p><strong>For financial transactions over $X:</strong></p><ol> <li>Request received (any channel)</li> <li>Callback to verified number</li> <li>Email confirmation with details</li> <li>24-hour hold period</li> <li>Secondary approval from different executive</li> <li>Execute with audit trail</li> </ol><p><strong>For executive communications:</strong></p><ol> <li>Note claimed identity from audio/video</li> <li>Verify through out-of-band communication</li> <li>Use pre-established code words</li> <li>Written confirmation before action</li> <li>Escalation path if verification fails</li> </ol><p><strong>For customer service:</strong></p><ol> <li>Multi-factor authentication required</li> <li>Sensitive changes require callback</li> <li>Video ID verification supplemented with additional factors</li> <li>All changes have reversal window</li> </ol><p><strong>Publish these protocols. Train teams. Enforce compliance.</strong></p><p><strong>5. Audit current authentication methods</strong></p><p>Where are you using voice or video as authentication factor?</p><ul> <li>Phone banking systems</li> <li>Remote notarization</li> <li>Customer verification</li> <li>Internal approvals</li> </ul><p><strong>Replace pure voice/video authentication with multi-factor requirements.</strong></p><p><strong>6. Review insurance coverage</strong></p><p>Does your cyber insurance cover deepfake fraud?</p><ul> <li>Most policies written before deepfakes were viable threat</li> <li>Coverage may exclude social engineering</li> <li>Limits may be insufficient for large-scale fraud</li> </ul><p><strong>Update policies to explicitly cover deepfake scenarios.</strong></p><h3 id="medium-term-this-quarter">Medium-term (This Quarter)</h3><p><strong>7. Implement behavioral analysis</strong></p><p>Deploy systems that flag anomalous requests:</p><ul> <li>Unusual transaction patterns</li> <li>Out-of-normal-hours requests</li> <li>Urgency combined with confidentiality</li> <li>New accounts or vendors</li> <li>Requests that bypass normal approval chains</li> </ul><p><strong>This won't detect deepfakes. It will detect the fraud attempts that use deepfakes.</strong></p><p><strong>8. Establish physical token requirements</strong></p><p>For highest-stakes operations:</p><ul> <li>Hardware security keys for executive authorizations</li> <li>Smart cards for financial controllers</li> <li>In-person verification for critical contracts</li> </ul><p><strong>Yes, this reduces efficiency. That's the point.</strong></p><p>The Arup case happened because efficiency was prioritized over verification. <strong>Sometimes friction is security.</strong></p><p><strong>9. Create escalation and response procedures</strong></p><p>What happens when suspected deepfake detected?</p><ul> <li>Who gets notified?</li> <li>How is transaction halted?</li> <li>What investigation begins?</li> <li>How is incident communicated?</li> </ul><p><strong>Have this documented before the incident occurs.</strong></p><h3 id="long-term-next-6-12-months">Long-term (Next 6-12 Months)</h3><p><strong>10. Rebuild identity verification architecture</strong></p><p>This is the fundamental fix:</p><ul> <li>Audio/video never sufficient for identity proof</li> <li>Multi-channel verification required for high-value operations</li> <li>Behavioral anomaly detection integrated</li> <li>Physical tokens for critical functions</li> <li>Zero-trust principle: verify every request, regardless of channel</li> </ul><p>When we implemented <a href="https://guptadeepak.com/zero-trust-architecture-the-technical-blueprint/" rel="noreferrer">zero-trust architecture</a> for the CIAM platform, it wasn't a quick project. It was <strong>fundamental rethinking of how identity works</strong>.</p><p><strong>Organizations need the same rethink for audio/video identity verification.</strong></p><p><strong>11. Partner with industry on standards</strong></p><p>Individual organizations can't solve this alone.</p><p><strong>Needed:</strong></p><ul> <li>Industry standards for identity verification in deepfake era</li> <li>Cross-organization verification protocols</li> <li>Shared threat intelligence on deepfake attacks</li> <li>Regulatory guidance on legal standards for video evidence</li> </ul><p><strong>This is infrastructure problem, not individual company problem.</strong></p><p><strong>12. Prepare for regulatory changes</strong></p><p>Regulations will come (probably after high-profile fraud makes headlines).</p><p><strong>Likely requirements:</strong></p><ul> <li>Mandatory multi-channel verification for financial transactions</li> <li>Disclosure of deepfake fraud incidents</li> <li>Minimum security standards for remote identity verification</li> <li>Liability frameworks for institutions that fail to verify identity</li> </ul><p><strong>Organizations that prepare now will comply easily. Those that wait will scramble.</strong></p><h2 id="the-broader-implications-when-trust-itself-fails">The Broader Implications: When Trust Itself Fails</h2><p>The Arup case is about $25 million. But the implications go far beyond one fraud.</p><p><strong>We're entering an era where:</strong></p><ul> <li>Seeing someone's face doesn't prove they're real</li> <li>Hearing someone's voice doesn't prove it's them</li> <li>Video calls create false confidence in identity</li> <li>"Trust your eyes and ears" is now bad security advice</li> </ul><p><strong>This breaks fundamental human communication assumptions.</strong></p><p>For all of human history, if you saw someone's face and heard their voice, you could be reasonably confident it was them. <strong>That certainty is gone.</strong></p><h3 id="the-social-impact">The Social Impact</h3><p>Beyond corporate fraud:</p><ul> <li>Can you trust video calls with family members?</li> <li>Is that really your friend calling for emergency money?</li> <li>Did that politician actually say what's in the video?</li> <li>Is the breaking news anchor real or synthetic?</li> </ul><p><strong>The erosion of trust in audio/visual communication has societal consequences beyond security.</strong></p><h3 id="the-legal-impact">The Legal Impact</h3><p>Courts rely on:</p><ul> <li>Video depositions</li> <li>Recorded testimony</li> <li>Security camera footage</li> <li>Authentication of speakers in recordings</li> </ul><p><strong>All of this becomes problematic when deepfakes are perfect.</strong></p><p>Legal systems will need to establish <strong>new standards for evidence authentication</strong> in the deepfake era.</p><h3 id="the-political-impact">The Political Impact</h3><p>Imagine:</p><ul> <li>Deepfake president declaring war</li> <li>Synthetic CEO announcing fake acquisition (stock manipulation)</li> <li>Fabricated testimony in high-profile trial</li> <li>Generated "leaked" executive conversation</li> </ul><p><strong>The potential for market manipulation, political chaos, and social disruption is enormous.</strong></p><p>The Arup $25 million fraud is a preview. The real crisis is when trust in media itself becomes impossible.</p><h2 id="the-uncomfortable-truth">The Uncomfortable Truth</h2><p>Organizations want a technology solution to the deepfake problem. "Install detection software. Train the AI. Filter the fakes."</p><p><strong>That's not going to work.</strong></p><p>Deepfake generation and deepfake detection are in an arms race. Generators are winning. <strong>Every improvement in detection gets incorporated into better generation.</strong></p><p><strong>The only sustainable solution is to stop relying on audio and video as proof of identity.</strong></p><p>This means:</p><ul> <li>More friction in high-value transactions (verification takes time)</li> <li>More procedural overhead (multiple channels, physical tokens)</li> <li>Less efficiency (delays, approvals, reviews)</li> <li>More cost (infrastructure, training, compliance)</li> </ul><p><strong>Nobody wants this.</strong> Organizations optimized for efficiency, not security friction.</p><p>But the alternative is Arup-scale fraud becoming routine.</p><p><strong>The choice:</strong></p><ul> <li>Accept friction and overhead to verify identity properly</li> <li>Accept risk of deepfake fraud and plan for eventual breach</li> </ul><p>There's no third option where you get efficiency, low overhead, AND protection from deepfakes.</p><p>When building the CIAM platform, I learned that <strong>security and convenience are often inversely correlated</strong>. The most secure systems have the most friction.</p><p><strong>In the deepfake era, that friction is non-negotiable for high-value operations.</strong></p><h2 id="the-bottom-line">The Bottom Line</h2><p>An employee at <a href="https://www.weforum.org/stories/2025/02/deepfake-ai-cybercrime-arup/" rel="noreferrer">Arup saw executives on video</a>. Heard them speaking. Verified the request through what appeared to be proper channels. Authorized $25 million in transfers.</p><p><strong>Every executive was an AI-generated deepfake.</strong></p><p>This wasn't a failure of the employee. It was a failure of assumptions. The assumption that video calls prove identity. The assumption that seeing and hearing someone means they're real.</p><p><strong>Those assumptions are now liabilities.</strong></p><p>For organizations:</p><ul> <li>Implement multi-channel verification immediately</li> <li>Never trust audio/video alone for high-value requests</li> <li>Establish physical token requirements for critical operations</li> <li>Build behavioral anomaly detection</li> <li>Prepare for regulatory requirements</li> </ul><p>For individuals:</p><ul> <li>Be skeptical of urgent requests via audio/video</li> <li>Verify through independent channels before taking action</li> <li>Use code words or verification phrases with family</li> <li>Understand that perfect deepfakes exist and are accessible</li> </ul><p>For society:</p><ul> <li>Rebuild trust frameworks for digital communication</li> <li>Establish legal standards for authenticated video</li> <li>Create verification infrastructure that works at scale</li> <li>Accept that "trust your eyes" is outdated advice</li> </ul><p>The deepfake era is here. The technology that enables $25 million fraud is commercially available for a few hundred dollars.</p><p><strong>The question isn't whether deepfake fraud will become common. The question is whether organizations will adapt their identity verification before or after they become victims.</strong></p><p>Arup learned at $25 million cost. Your organization can learn from them instead.</p><p><strong>Identity verification is broken. Seeing is no longer believing. The sooner we accept that, the sooner we can build systems that actually work.</strong></p><hr><h2 id="key-takeaways">Key Takeaways</h2><ul> <li>Arup lost $25M to perfect deepfake video conference – every "executive" was AI-generated</li> <li>Attack cost attackers ~$10K, return $25M (2,500x ROI) – economically inevitable</li> <li>Modern deepfakes are real-time, interactive, and indistinguishable from real people</li> <li>Voice cloning requires just 3-5 minutes of audio, costs $20-100</li> <li>"Seeing is believing" is now a security vulnerability, not verification method</li> <li>Training employees to "spot deepfakes" is useless – no reliable tells exist</li> <li>Video/voice authentication creates false confidence – actively dangerous</li> <li>Multi-channel verification REQUIRED: video request → callback to known number + email confirmation</li> <li>High-value operations need physical tokens (hardware keys, smart cards)</li> <li>Behavioral anomaly detection can flag unusual requests (urgency + new accounts + large amounts)</li> <li>Organizations must rebuild identity verification assuming audio/video can always be faked</li> <li>Immediate actions: map audio/video verification points, implement callback procedures, alert high-risk employees</li> <li>Long-term: zero-trust for identity verification, never accept single-channel proof</li> <li>Broader crisis: trust in audio/visual communication eroding across society, not just corporate fraud</li> </ul><hr><p><strong>Building authentication systems that resist deepfakes?</strong> My <a href="https://guptadeepak.com/customer-identity-hub/" rel="noreferrer">Customer Identity Hub</a> covers <a href="https://guptadeepak.com/customer-identity-hub/mfa-implementation-in-ciam" rel="noreferrer">multi-factor authentication</a>, <a href="https://guptadeepak.com/what-is-zero-trust-security-a-plain-english-guide/" rel="noreferrer">zero-trust principles</a>, and <a href="https://guptadeepak.com/ciam-101-a-practical-guide-to-customer-identity-and-access-management-in-2025/" rel="noreferrer">modern CIAM architecture</a> that assumes all biometric channels are compromised.</p><p><a href="https://guptadeepak.com/about/" rel="noreferrer"><em>Deepak Gupta</em></a><em> is the co-founder and CEO of GrackerAI. He previously founded a CIAM platform that scaled to serve 1B+ users globally. He writes about AI, cybersecurity, and digital identity at guptadeepak.com.</em></p><div class="spu-placeholder" style="display:none"></div><div class="addtoany_share_save_container addtoany_content addtoany_content_bottom"><div class="a2a_kit a2a_kit_size_20 addtoany_list" data-a2a-url="https://securityboulevard.com/2026/03/the-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted/" data-a2a-title="The $25 Million Deepfake: Why Your Video Calls Can No Longer Be Trusted"><a class="a2a_button_twitter" href="https://www.addtoany.com/add_to/twitter?linkurl=https%3A%2F%2Fsecurityboulevard.com%2F2026%2F03%2Fthe-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted%2F&linkname=The%20%2425%20Million%20Deepfake%3A%20Why%20Your%20Video%20Calls%20Can%20No%20Longer%20Be%20Trusted" title="Twitter" rel="nofollow noopener" target="_blank"></a><a class="a2a_button_linkedin" href="https://www.addtoany.com/add_to/linkedin?linkurl=https%3A%2F%2Fsecurityboulevard.com%2F2026%2F03%2Fthe-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted%2F&linkname=The%20%2425%20Million%20Deepfake%3A%20Why%20Your%20Video%20Calls%20Can%20No%20Longer%20Be%20Trusted" title="LinkedIn" rel="nofollow noopener" target="_blank"></a><a class="a2a_button_facebook" href="https://www.addtoany.com/add_to/facebook?linkurl=https%3A%2F%2Fsecurityboulevard.com%2F2026%2F03%2Fthe-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted%2F&linkname=The%20%2425%20Million%20Deepfake%3A%20Why%20Your%20Video%20Calls%20Can%20No%20Longer%20Be%20Trusted" title="Facebook" rel="nofollow noopener" target="_blank"></a><a class="a2a_button_reddit" href="https://www.addtoany.com/add_to/reddit?linkurl=https%3A%2F%2Fsecurityboulevard.com%2F2026%2F03%2Fthe-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted%2F&linkname=The%20%2425%20Million%20Deepfake%3A%20Why%20Your%20Video%20Calls%20Can%20No%20Longer%20Be%20Trusted" title="Reddit" rel="nofollow noopener" target="_blank"></a><a class="a2a_button_email" href="https://www.addtoany.com/add_to/email?linkurl=https%3A%2F%2Fsecurityboulevard.com%2F2026%2F03%2Fthe-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted%2F&linkname=The%20%2425%20Million%20Deepfake%3A%20Why%20Your%20Video%20Calls%20Can%20No%20Longer%20Be%20Trusted" title="Email" rel="nofollow noopener" target="_blank"></a><a class="a2a_dd addtoany_share_save addtoany_share" href="https://www.addtoany.com/share"></a></div></div><p class="syndicated-attribution">*** This is a Security Bloggers Network syndicated blog from <a href="https://guptadeepak.com/">Deepak Gupta | AI &amp; Cybersecurity Innovation Leader | Founder&#039;s Journey from Code to Scale</a> authored by <a href="https://securityboulevard.com/author/0/" title="Read other posts by Deepak Gupta - Tech Entrepreneur, Cybersecurity Author">Deepak Gupta - Tech Entrepreneur, Cybersecurity Author</a>. Read the original post at: <a href="https://guptadeepak.com/the-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted/">https://guptadeepak.com/the-25-million-deepfake-why-your-video-calls-can-no-longer-be-trusted/</a> </p>