How to disable the default "Disfluency Removal" of filler words after STT transcription in Azure AI Speech?
Azure AI Speech Services defaults to removing many filler words (uh, eh, etc.) via post-transcription "Disfluency Removal". My use case includes presentation analysis for filler words, which requires a verbatim transcript. Is there a…
How to disable the default "Disfluency Removal" of filler words after STT transcription in Azure AI Speech?
My use case includes analyzing presentations for filler words and to do so I need verbatim transcripts. Azure STT defaults to "Disfluency Removal" (removing filler words) on the transcription before returning the JSON transcript file. Is…
Never-ending "Loading resources and quota..."
When trying to deploy this model I'm met with the never-ending loading wheel. How to fix?
how to run an Azure avatar in code
Good afternoon. I am writing code in Python Fastapi in which the bot asks questions and the user answers. Help with documentation to add Azure Avatare
Issue with Pronunciation Assessment in Speech Recognition API Always Returning PronScore 100
I am using the https://eastus.stt.speech.microsoft.com/speech/recognition/conversation/cognitiveservices/v1 API with the POST method for speech-to-text conversion. Here are the details of my implementation: Programming Language: JavaScript …
Why Am I Getting Connection Failed Error on Android build via Unity?
get an error message on Android build via Unity Connection failed (no connection to the remote host). Internal error: 1. Error details: Failed with error: WS_OPEN_ERROR_UNDERLYING_IO_OPEN_FAILED Speech SDK log taken from a run that exhibits the…
Azure AI Speech Studio TextToSpeech with voice "AlloyTurboMultilingual" shows "Error 400 Synthesis failed. StatusCode: NotFound"
Inside of Azure AI Speech Studio when trying to generate speech with the voice "AlloyTurboMultilingual" or "NovaTurboMultilingual" the following error occurs: Response status code does not indicate success: 400 (Synthesis failed.…
Issue with Continuous Language Identification in Azure Speech SDK for Angular Application
We are currently using the "microsoft-cognitiveservices-speech-sdk" in our Angular application (version 14) for speech transcription and translation. The transcription and translation functionality is working as expected. However, we are…
Confused on custom neural voice docs
Hi all, I'd like to implement custom text to speech voices and I was pleased to see that azure offers different solutions, but i'm confused on the types of services. In the docs page I saw 3 types of custom speech methods: pro, lite and personal but in…
Delay in Transcription (Multi-Device Conversation) Cognitive Speech
As the documentation says that multi device conversation is real time but running the sample code there is delay of about 12 - 15 second in transcribing. How can i make it real time??
In speech service text to speech pitch parameter
Hi all Iam new to azure speech service. In speech text to speech I am using conard voice. How can we increase or decath pitch for a specific word?
Azure speech studio custom keyword page always return API exception
I have trained keyword, when I try to test model on the page, it always return: "API exception. This request is not authorized to perform this operation using this permission." And I can't download the model, There's no response on the…
Creating Custom Keywords Models are stuck "processing"
When creating a new model in the Speech Studio under 'Create a custom keyword for your virtual assistant', the model creation process never completes. I tried it with multiple models and keywords, even the examples like "hey computer" never…
[Multi Device Conversation] - [Multi-Device Conversation][DotNet] Cannot set display name/Nickname when join conversation.
Hi, I tried to implement the code according to the example at the link: https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/quickstart/csharp/dotnet/multi-device-conversation/helloworld/Program.cs. But when I set display name…
speech api fails where Speech Studio succeeds?
I am using the Standard Tier, with a couple paragraphs of text and only a few ssml tags. The ssml pasted into Speech Studio renders correctly, even exports to an audio file correctly. The same ssml rendered through the python API causes the error below.…
My Speech Congnitive Basic Model can not finish training
I want to training a congnitive model to help me do speech key words cognition. I choose basic model but it have run for over 10 hours still not get any result. The doc says it only takes few minutes. Now the model still inprocessing without any result.
Can I set maximum number of participants to real-time diarization?
Hi, I follow the document below and success to distinguish the speaker with audio streaming by ConversationTranscriber Class. (I don't use voice signature so it shows Guest-1,…
[Pronunciation Assessment] Is there a way to improve the results using a custom model?
I've been experimenting with the pronunciation assessment service. My use case involves scripted assessment in Canadian French (i.e., fr-CA). So far, I've had the most success with the configuration where enableMiscue is set to false, as I have found the…
Is there GRPC support for Speech to Text in Azure Speech SDK in java?
Hi, Is there GRPC support for Azure speech SDK? We are looking for this support for the Realtime Speech to Text feature. Is that support available in Java? If there is no GRPC support, what is the underlying architecture, and how is the voice streamed to…
Do we have any API Support to get the cost estimation of Audio Translation?
Do we have any API Support to get the cost estimation of audio translation from language to another language by taking input as audio duration, source language, target language, etc., required input. If any API support in java, please help me Thanks…