Run ElevenLabs text-to-speech and speech-to-text through Hicap without adding a separate auth flow, endpoint surface, or billing path. Keep the same Hicap base URL and ship voice next to the rest of your AI stack.
The integration model stays simple: route requests through Hicap, keep using your Hicap key, and target the ElevenLabs-compatible voice endpoints you need.
Point requests at https://api.hicap.ai/v1.
Send your Hicap key in the api-key header.
Use ElevenLabs model IDs for TTS and STT requests.
The point of the voice route is consolidation, not a separate setup track. Teams already using Hicap should not have to think about voice as a second platform.
Keep requests on https://api.hicap.ai/v1 and authenticate with the same api-key header you already use for chat and other model traffic.
Use the ElevenLabs-style voice paths and model IDs while routing traffic through Hicap instead of wiring up a separate voice integration.
Keep billing, access, and operational routing in one place whether your app is generating text, audio, or transcripts.
Hicap currently exposes ElevenLabs' main speech generation models so teams can cover both expressive voice work and steadier long-form narration from one route.
Expressive speech synthesis
Best fit when voice tone, character, and performance matter. ElevenLabs positions Eleven v3 as its most emotionally rich text-to-speech model.
Stable long-form generation
A steadier option for narration, explainers, and multilingual production where consistency over longer passages matters more than theatrical range.
Both Scribe models are available through Hicap, from broad language coverage to newer transcription features like speaker diarization and transcript cleanup.
Broad language coverage
A straightforward speech-to-text option for turning recorded audio into searchable text across a wide language set.
Higher-accuracy transcription
The more capable transcription option for production workflows that need better recognition, speaker separation, and cleaner transcripts.
These examples keep the ElevenLabs endpoint shapes and model IDs while moving authentication and routing onto Hicap.
Generate audio through Hicap
curl --request POST \\ --url "https://api.hicap.ai/v1/text-to-speech/JBFqnCBsd6RMkjVDRZzb" \\ --header "Content-Type: application/json" \\ --header "api-key: $HICAP_API_KEY" \\ --data '{ "text": "The first move is what sets everything in motion.", "model_id": "eleven_v3" }' \\ --output speech.mp3Replace JBFqnCBsd6RMkjVDRZzb with the ElevenLabs voice ID you want to use. If you need a different response format, follow the ElevenLabs-compatible request options while keeping the Hicap base URL and auth header.
Transcribe files through Hicap
curl --request POST \\ --url "https://api.hicap.ai/v1/speech-to-text" \\ --header "api-key: $HICAP_API_KEY" \\ --form "file=@./meeting.mp3" \\ --form "model_id=scribe_v2"Send audio or video with multipart form data and switch the model_id between scribe_v1 and scribe_v2based on the transcription quality and feature set you need.
Current Hicap voice coverage includes Eleven v3 and Eleven Multilingual v2 for generation, plus Scribe v1 and Scribe v2 for transcription. That keeps the voice surface focused and predictable while the rest of the Hicap model catalog remains available through the same account.
Bring speech generation and transcription into the same Hicap workflow your team already understands.