Getting Started
Follow these steps to configure Azure Cognitive Services as your STT provider:Add Azure credentials to your vault
Navigate to Integration → Vault in the Rapida dashboard. Add your Azure Subscription Key and the endpoint URL for your Speech resource. The IAM role must have the
Cognitive Services User permission.Select Azure as your STT provider
When configuring your assistant, open Audio Settings and choose Azure Cognitive Services as your Speech-to-Text provider.
Supported Models
| Model | Description |
|---|---|
latest | Microsoft’s latest neural ASR model for the selected locale |
| Custom | Custom Speech models trained on your own audio data |
Key Features
- Real-time streaming: Low-latency partial and final transcripts for live voice applications
- Speaker diarization: Identify and label individual speakers in a conversation
- Custom Speech models: Train on your own audio data to improve accuracy for domain-specific terms
- Profanity filtering: Mask or remove profane words from transcripts
- Phrase lists: Boost recognition accuracy for specific words or phrases
- Content redaction: Automatically redact PII from transcripts
Supported Languages
Azure supports 100+ languages and locales including English (US, UK, AU, IN), Spanish, French, German, Italian, Portuguese, Japanese, Chinese (Simplified, Traditional), Korean, Hindi, and Arabic. See the Azure documentation for the full list.Configuration Options
| Option | Description |
|---|---|
| Subscription key | Azure Cognitive Services authentication key |
| Endpoint | Azure Speech resource endpoint URL |
| Region | Azure region (e.g. eastus, westeurope) |
| Language code | BCP-47 locale code |
| Sample rate | Audio input sample rate in Hz (8000 or 16000) |
Notes
- For telephony use cases, set sample rate to 8000 Hz to match PSTN audio.
- Custom Speech models require training data upload via the Azure portal.
- Pricing is per audio hour. See Azure Speech pricing.