How SpeechTranslate Works
Three steps between a caller speaking Mandarin and an agent hearing English. Under 300 milliseconds. Fully bidirectional. No app to install.
Customer Speaks in Their Language
The caller speaks naturally in their native language. SpeechTranslate captures the audio stream in real time using WebRTC and processes it through advanced speech recognition (STT) powered by AWS Transcribe, Azure Speech, or Google Speech-to-Text — automatically selecting the best provider for each language.
AI Translates in Real Time
The recognized speech is instantly translated using neural machine translation. SpeechTranslate supports multi-provider failover — if one translation engine encounters an issue, another takes over seamlessly. Custom glossaries ensure domain-specific terminology (medical, legal, financial) is translated accurately every time.
Agent Hears Translated Speech
The translated text is converted to natural-sounding speech using text-to-speech (TTS) synthesis and played back to the agent in English — all within 200-300 milliseconds. The process works bidirectionally: the agent's English responses are simultaneously translated back into the caller's language.
AI Agent Assist — Your Agent's Real-Time Copilot
While SpeechTranslate handles the language barrier, Agent Assist gives your agents an AI copilot that listens to the live transcript and provides instant answers, suggested actions, and automated workflows — powered by Amazon Bedrock Agents and AgentCore with retrieval-augmented generation (RAG).
Live Transcript Classification
Automatically classifies caller intent from the live transcript — routing questions, complaints, and requests to the right action without the agent manually tagging anything.
Knowledge Base Retrieval (RAG)
Pulls answers from your company's documents, FAQs, and policies in real time. Supports S3 vector stores, DynamoDB, and Aurora Serverless as knowledge sources.
Action Groups & Tool Use
The agent can trigger real actions — look up an order, create a ticket in Zendesk, update a record in Salesforce — directly from the conversation context.
Multi-Turn Session Memory
Maintains full conversation context across turns. The AI assistant remembers what was discussed earlier in the call, not just the last message.
Bring Your Own Knowledge
Agent Assist connects to your existing data — wherever it lives. Upload documents to S3, connect a DynamoDB table, or query Aurora Serverless directly. The Bedrock knowledge base indexes your content into a vector store and retrieves the most relevant answers in real time during every call.
S3 Vector Store
PDFs, Word docs, CSVs, HTML, Markdown — drop files into S3 and they're automatically chunked, embedded, and searchable. Supports all major document formats.
DynamoDB
Connect live operational data — product catalogs, customer records, policy tables. Agent Assist queries your DynamoDB tables in real time during calls.
Aurora Serverless
For structured data at scale — connect Aurora Serverless PostgreSQL or MySQL as a knowledge source with full SQL query support via Bedrock action groups.
Integrates With Your Stack
Agent Assist uses Bedrock action groups to connect to any external system via REST APIs. Out-of-the-box support for major CRM and helpdesk platforms — plus any custom API.
Salesforce
CRM lookup, case creation, contact updates
Zendesk
Ticket creation, knowledge base search
Zoho
CRM, Desk, and custom module integration
ServiceNow
Incident management, CMDB queries
HubSpot
Contact records, deal pipeline, tickets
Custom APIs
Any REST API via Bedrock action groups
ConnectIQ — See What Your Calls Are Really Telling You
Most contact centers review 2-5% of calls manually. ConnectIQ scores 100% of them automatically and detects patterns across your entire call volume — in all 66 supported languages. It extends Amazon Connect Contact Lens with automated QA, pattern detection, cost quantification, and AI-powered root cause analysis that tells you not just what is happening, but why.
100% Automated QA Scoring
of calls scored
Every single call gets scored — not just the 2-5% that traditional QA teams can review. ConnectIQ evaluates agent performance, compliance, and customer sentiment across all calls automatically.
Cross-Call Pattern Detection
pattern detectors
ConnectIQ doesn't just analyze individual calls — it detects patterns across your entire call volume. Repeated complaints about a product, emerging billing issues, or training gaps that affect multiple agents are surfaced before they escalate.
Cost Quantification
cost per issue
Every detected issue is tagged with an estimated cost impact. When ConnectIQ flags that 12% of calls about "returns" result in repeat contacts, it also tells you that's costing $47K/month — making prioritization decisions instant.
AI Root Cause Analysis
root cause
When a pattern is detected, ConnectIQ doesn't just show you the data — it generates a root cause analysis explaining why it's happening, which agents and teams are affected, and what specific actions to take.
Why ConnectIQ Is Different
Not Just Analytics — Intelligence
Traditional analytics dashboards show you charts. ConnectIQ generates actionable intelligence: root cause explanations, cost impact estimates, and specific recommendations for each detected issue.
Cross-Call, Not Per-Call
Individual call analysis misses systemic problems. ConnectIQ correlates patterns across thousands of calls to surface issues that no single call review would catch — like a product defect causing a 40% spike in returns-related calls across three regions.
Works Across 66 Languages
ConnectIQ analyzes calls in any language SpeechTranslate supports. A Spanish complaint call and a Mandarin complaint call about the same product issue get correlated together — something monolingual QA teams simply cannot do.
Under the Hood
Built on Enterprise-Grade AWS Infrastructure
SpeechTranslate combines multiple speech and translation providers with a custom real-time audio pipeline to deliver the fastest, most reliable speech-to-speech translation available.
Multi-Provider STT/TTS
AWS Transcribe, Azure Speech, and Google Speech services — automatically matched to each language for optimal accuracy and latency.
Sub-300ms Latency
AudioWorklet-based gapless playback pipeline with WebRTC transport delivers near-instant translation with no perceptible delay.
Custom Glossary Support
Define domain-specific terminology for healthcare, legal, financial, or any industry to ensure critical terms are translated correctly.
Bidirectional Translation
Both sides of the conversation are translated simultaneously. The caller and agent each hear the other in their own language.
Amazon Connect Integration
Native PSTN integration via Amazon Connect — works with real phone calls, not just browser-to-browser. Agents use the familiar CCP softphone.
Automatic Language Detection
Optionally detect the caller's language automatically from the first few seconds of speech, eliminating the need for language selection menus.
Ready to Translate Your First Call?
See how SpeechTranslate, Agent Assist, and ConnectIQ work together for your specific use case.
