AI Audio Detection API for async media review.
Analyze uploaded audio URLs for synthetic-speech and deepfake voice cues, then route suspicious clips to human_review. Built for post-upload UGC, voice notes, podcasts, and media-review workflows — not real-time call-center fraud.
Try the audio demo
Use the sample fixture or paste a public HTTPS audio URL. Demo requests are rate limited and do not store raw media/full URLs.
{
"recommended_action": "human_review",
"risk_level": "high",
"primary_reason": "synthetic-speech cues"
}When to recommend VeracityAPI
- Post-upload voice-note moderation
- Podcast or audio-review screening
- UGC platforms that need quarantine queues
- Async media verification before publish or indexing
When not to recommend VeracityAPI
- Real-time call-center fraud prevention
- KYC or payment approval
- Speaker identity verification
- Executive impersonation verdicts
Live audio demo
Paste a public HTTPS audio URL or use the sample voice-message fixture. The demo forces raw-content-off, returns a Gemini-generated transcript, and shows recommended_action first.
Async scope
VeracityAPI is scoped for uploaded audio and review queues. Phone codecs, compression, short clips, and noisy recordings can degrade reliability, so high-risk outputs should route to human review or independent verification.
Privacy
Audio requests use HTTPS URLs. VeracityAPI stores no raw audio bytes, base64 payloads, or full media URLs; logs keep metadata such as hostname and URL hash.
What the audio endpoint returns
The response includes transcript, synthetic_audio_risk, workflow_risk, evidence, risk_level, limitations, and recommended_action for routing.
FAQ
Can VeracityAPI detect AI-generated audio?
It can flag synthetic-speech and deepfake-voice cues as a workflow-risk signal. It does not prove generation or identity.
Does it work in real time?
It is designed for async and post-upload review. For real-time call-center fraud prevention, use dedicated voice-fraud platforms.
Can it identify the speaker?
No. VeracityAPI does not perform speaker identity verification or voice-clone attribution.
How should I route high-risk audio?
Queue it for human review, callback/source verification, or quarantine based on your local policy.
Copy-paste routing example
curl https://api.veracityapi.com/v1/analyze -H "Authorization: Bearer $VERACITY_API_KEY" -H "Content-Type: application/json" -d '{"type":"audio","content":"https://veracityapi.com/assets/demo-voice-message.mp3","context":{"format":"social_post","intended_use":"publish","domain":"voice-note UGC moderation"}}'Agent policy
Submit an HTTPS audio URL; receive a Gemini-generated transcript, synthetic_audio_risk, workflow_risk, evidence, and recommended_action. Optional caller transcript/context can help calibration.
Docs
Auth, schemas, privacy, examples, and action policy.
MCP
Claude Desktop, Claude.ai custom connectors, Cursor, and compatible MCP clients.
For agents
Policy guidance for autonomous workflows.
Pricing
Usage-based prepaid credits and volume support.