Browse through the questions and answers listed below or filter and sort to narrow down your results.
List of waiters for AWS Transcribe
The documentation for the Boto3 [TranscribeService `getWaiter()`](https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/transcribe.html#TranscribeService.Client.get_waiter) says > See the waiters section of the service docs for a list of available waiters. ...but there are no waiters listed in the documentation. Are waiters available for AWS TranscribeService?
What's the realtime factor for the Transcribe Streaming SDK?
Hi there! We are using AWS Transcribe for streaming speech-to-text transcription. We're keen to understand whether the transcription is faster than realtime, and if so, what's the expected maximum factor? As an example, if we buffer 5 seconds of audio data on the client before pushing into the stream for transcription, then send the buffer all at once, how long would we expect to wait before it catches up to real-time again? It seems, from experiments, that it's not faster than realtime (exactly 1x?), and transcription results stay at a 5 second delay as we push more audio data into it, but I wanted to verify that's the case! Thank you very much, Ben
Transcribe is identifying speakers 0, speaker 1 as same person.
I ran transcription job which contains two people phone conversation. Even-though i selected two speakers but Transcribe carrying dialogues of speaker 0 over to speaker 1 in Transcription job transcript also transcription accuracy is very low. Is there anything can be done to resolve this issue?
Transcribe output is not accurate
Facing issues with transcribe, its not processing transcription accurately. Some of the sentences in the paragraph missing a period/ full stop. Some words are repeating 2x even though it was not clearly repeated in audio recordings. (Audio is crisp clear without any quality issues)
Transcribe: Understand/Save Speaker Detection
TL;DR I need to recognize the speaker identification (diarization) of the user's voice; couldn’t find a way to do it. I am building an application using AWS Transcribe streaming. I am able to get the speaker labels of each word, this is great; but, I need to somehow save and reuse the speaker label for further use with the same user.
Amazon Transcribe: Confidence and Speaker 'undefined'
I am getting 'undefined' from the Confidence field under Alternatives -> Items -> Confidence. First, I thought it is because the words are not stable (Items.Stable==false); but saw that it is also happening when stable field is true. Same is for the Speaker field. Any insights on why is this happening? Or any way I can improve these outputs?
AWS Transcribe - "Request has expired"
Whether I'm doing it via the console using "Real-time transcription" or the JS SDK, I get "Request has expired" whenever trying to start streaming transcription. I am a root user in the console... with the SDK i have a separate IAM user with full admin access. Nothing works.
AWS Transcribe gives an error for Arabic (ar-SA) with 8k Hz audio sample rate
Hi I can see AWS Transcribe support Arabic (ar-SA) audio and it also supports audio with 8k Hz sample rate (as per this [link](https://docs.aws.amazon.com/transcribe/latest/dg/how-input.html)). However when I try to transcribe from audio file which is PCM 8k Hz it gives me below error "The audio sample rate 8000 Hz is not supported. Change the audio sample rate of your media file and try your request again" The same audio format works fine for other languages but not for Arabic (ar-SA). Any help/pointers on why this won't work? Thanks