How can I enable streaming text-to-speech in my self-hosted deployment?
Deepgram's Text-to-Speech WebSocket interface provides a real-time, interactive solution for converting streaming text to speech. This can be enabled in your self-hosted deployment by enabling a feature flag.
If you already have Aura models exposed in your deployment, these will work for both batch and streaming TTS. No additional models are needed to enable streaming TTS specifically.
Updating your self-hosted deployment
If you don't already have an Aura model available, request one from your Deepgram Account Representative. Place it in your models directory, accessible to your Deepgram Engine containers.
Update your configuration based on your container orchestrator.
Docker/Podman
Update your
api.toml
configuration file to enable the streaming TTS feature flag, as seen in theapi.toml Configuration Update
code block below.Update your Compose file to use a supported release. Update the image tag for API and Engine containers to
release-241024
or later.Restart the API and Engine containers.
Kubernetes
Using a
v0.7.0+
release of thedeepgram-self-hosted
Helm chart, setapi.features.streamingTTS
to true in yourvalues.yaml
file.If you have overridden the default values for
api.image.tag
orengine.image.tag
, make sure to userelease-241024
or later.Upgrade your Helm installation.
To verify your setup, open a streaming TTS connection to your self-hosted deployment.
Troubleshooting
Feature flag not enabled
If you receive a 405 Method Not Allowed
response, the streaming TTS feature flag may not be enabled for your Deepgram API container, or you may be on an older release that does not support the feature.
Check that your api.toml
configuration file has been updated, as described in Updating your self-hosted deployment.
Last updated
Was this helpful?