Home / AI / White Papers AI customer surface & identity

Synthetic Identity Risk

Avatar and voice agents are useful. They are also the fastest-growing fraud category of 2025. The compliance wave hits 2 August 2026 — and most vendors are not ready.

The threat is operational, not theoretical

Voice deepfake fraud incidents grew approximately 680% year-over-year through 2025. The U.S. saw more than 100,000 attacks in that year, with cumulative reported losses crossing $2.19 billion. The technology threshold has collapsed: a usable voice clone now requires three to ten seconds of clean audio. A CEO’s last earnings call, a CFO’s podcast appearance, a client-facing webinar — any of these is sufficient.

The deployment pattern in 2026 is no longer "send a phishing email." It is: an agent calls the target on a real phone line, in a real-sounding voice the target recognizes, and conducts a real-time interactive conversation that adapts to what the target says. More than 50% of CISOs in a recent survey reported a successful deepfake-based intrusion in the prior 18 months. The number was approximately 10% the year before.

Avatar video deepfakes are the visual equivalent of the same trajectory, lagging voice by maybe twelve months. Open-source diffusion models for face animation are now usable on consumer GPUs. Real-time avatar synthesis on a video conference is achievable. Detection by the human eye is not reliable.

680%
YoY voice deepfake incident growth (2025)
$2.19B
US cumulative losses
3–10 sec
Voice clone training audio required
$200K
TRAIGA penalty per violation

The legitimate use cases are also large

This is the difficult part. The same technology is being used responsibly to:

The market for this is real. HeyGen, Tavus, ElevenLabs, and Synthesia together represent hundreds of millions in annualized revenue. None of them, as default products, give an MSP the controls needed to ship the technology to a regulated client. That gap is where AiT Avatar Studio and AiT Voice Concierge live.

What the regulatory wave actually requires

EU AI Act Article 50. Fully applicable 2 August 2026. Any AI-generated audio, video, or image must carry a machine-readable mark identifying it as such. The obligation flows through to downstream integrators of GPAI systems, including MSPs reselling avatar or voice technology. The June 2026 Code of Practice is being finalized now and will specify the technical formats — almost certainly C2PA-compatible content provenance plus a steganographic watermark fallback.

TRAIGA (Texas HB 149). Enforceable since 1 January 2026. Penalties up to $200,000 per violation. Affirmative defense: substantial compliance with the NIST AI RMF. This is not optional in Texas.

ELVIS Act (Tennessee) and right-of-publicity laws. States are individually enacting consent requirements for the use of any individual’s voice, likeness, or name in AI-generated content. Retention requirements for the consent record vary; the longest is currently 7 years. Tennessee’s ELVIS Act is the loudest, but at least 12 other states have proposed or enacted similar statutes.

SOC 2 + ISO 42001. AI-specific control families are now part of mainstream attestations. Drummond, Schellman, and A-LIGN all offer AI-aware audits. By 2027, an AI control family is likely to be standard.

Why we excluded Synthesia from our stack

This is a specific, unflattering, and important data point. Synthesia is a strong product. Their terms of service, however, explicitly prohibit white-label resale — meaning an MSP cannot legally embed Synthesia in a product the MSP markets under its own brand. We discovered this during the build of AiT Avatar Studio, after evaluating their API and pricing. We respect that this is their commercial choice. The consequence is that we use HeyGen, Tavus, and ElevenLabs as the underlying providers and have a documented exclusion of Synthesia in our vendor-onboarding playbook. MSPs evaluating Synthesia for resale should read the ToS carefully.

What we built

AiT Avatar Studio and AiT Voice Concierge are companion products with a shared compliance foundation:

Compliance built into the rendering layer

Article 50 watermarking, ELVIS / TRAIGA consent retention, and refusal of fraud-pattern prompts are not policies we ask the user to follow. They are properties of the output that cannot be turned off. Trust Portal exposes the audit trail. Auditors love this; clients learn to.

What this looks like in practice

A regional bank wants to send onboarding videos in three languages from a video-message of their CEO. The CEO records a one-minute on-camera consent statement granting the bank’s use of his likeness for onboarding content for 24 months. The consent is logged. Avatar Studio renders the three localized videos with watermarking embedded and disclosure language captioned. The Trust Portal entry shows: who consented, when, what they consented to, what was rendered, when, and where it was distributed. Two months later the bank’s auditor asks for the evidence. It exports in three clicks.

A second example: a manufacturing client wants 24x7 voice-based IT helpdesk for their three plants. AiT Voice Concierge deploys with a stock voice (no specific person’s likeness involved). Refusal patterns block any prompt that asks for a financial transaction or claims to be a specific human. Every call is logged and transcribed; the manufacturing IT lead reviews the daily digest. Tier-0 deflection rate at six months: 41%. After-hours call volume to the human team: down 60%. Compliance posture: clean.

What you should ask any avatar/voice vendor

  1. Where is consent recorded, and what is the retention policy?
  2. How is Article 50 watermarking implemented, and is it removable?
  3. What disclosure language ships with each output by default?
  4. What fraud patterns does the system refuse, and where can I see the refusal audit log?
  5. What happens to a render if it’s captured by an attacker and re-encoded — can I prove provenance?
  6. What does your ToS say about white-label or MSP resale?

Where this fits

AiT Avatar Studio and AiT Voice Concierge are part of our AI Customer Surface capability cluster. They depend on the AI Gateway for governed model access, AiT Coord for cross-tenant arbitration, and the Trust Portal for client-facing transparency.

References

Doing avatar or voice work? Or being asked to?

Bring the use case. We’ll walk you through the consent ledger, watermarking, and refusal model in a 30-minute call.

2-minute average response time, 24/7.

Got it — we’ll be in touch.