StartStreamTranscriptionRequest
Types
Properties
An encoded stream of audio blobs. Audio streams are encoded as either HTTP/2 or WebSocket data frames.
Labels all personally identifiable information (PII) identified in your transcript.
Redacts all personally identifiable information (PII) identified in your transcript.
Enables channel identification in multi-channel audio.
Enables partial result stabilization for your transcription. Partial result stabilization can reduce latency in your output, but may impact accuracy. For more information, see Partial-result stabilization.
Enables automatic language identification for your transcription.
Enables automatic multi-language identification in your transcription job request. Use this parameter if your stream contains more than one language. If your stream contains only one language, use IdentifyLanguage instead.
Specify the language code that represents the language spoken in your audio.
Specify the name of the custom language model that you want to use when processing your transcription. Note that language model names are case sensitive.
Specify two or more language codes that represent the languages you think may be present in your media; including more than five is not recommended. If you're unsure what languages are present, do not include this parameter.
Specify the encoding of your input audio. Supported formats are:
The sample rate of the input audio (in hertz). Low-quality audio, such as telephone audio, is typically around 8,000 Hz. High-quality audio typically ranges from 16,000 Hz to 48,000 Hz. Note that the sample rate you specify must match that of your audio.
Specify the number of channels in your audio stream. Up to two channels are supported.
Specify the level of stability to use when you enable partial results stabilization (EnablePartialResultsStabilization
).
Specify which types of personally identifiable information (PII) you want to redact in your transcript. You can include as many types as you'd like, or you can select ALL
.
Specify a preferred language from the subset of languages codes you specified in LanguageOptions
.
Enables speaker partitioning (diarization) in your transcription output. Speaker partitioning labels the speech from individual speakers in your media file.
Specify how you want your vocabulary filter applied to your transcript.
Specify the name of the custom vocabulary filter that you want to use when processing your transcription. Note that vocabulary filter names are case sensitive.
Specify the names of the custom vocabulary filters that you want to use when processing your transcription. Note that vocabulary filter names are case sensitive.
Specify the name of the custom vocabulary that you want to use when processing your transcription. Note that vocabulary names are case sensitive.
Specify the names of the custom vocabularies that you want to use when processing your transcription. Note that vocabulary names are case sensitive.