How to set varibles for ssml messages and speech style in Text-to-Speech
Hello. I am quite new as it is probably quite apparent. Currently am trying to set a variable for ssml Text-to-Speech, however, after going through every documentation I could find, I still have no solution. Using Python. I would be greatful if anyone…
ChatBot in android webview
hello, I have a Luis-based Chabot integrated into the website. the chatbot works fine in desktop and mobile browsers (Google Chrome). but when I load the same webpage in android WebView it causing a problem i.e. when I give voice input it not returning…
Azure Text to Speech error
How to solve an error within Azure Speech, Text To Speech, when I get 60+K matches> The error is "Failed to fetch voices" in all its splendor. Doh. "Where is the anykey? " as Homer asked. What I am doing? you mean short of…
SpeechServices availability, authorizationToken validation
Hi, I implemented MS speech services in my macOS app (Objective C). Everything works nice. However, I would like to check for speech service availability and check if authorizationToken is valid to inform user, that everything is okay. I would like to…
recognizeOnceAsync not working for input from microphone on browser
I copy and pasted the code linked below into an empty file, then hardcoded in my subscription key and region. When I open the file and click the button, the browser asks for permission to access the microphone. When I click allow, the recognizeOnceAsync…
speech to text Authentication error (401) on GitHub SDK samples c++ console
I am testing an azure cognitive services speech. Pricing tier is free (f0). I followed instructions for creating instance and service. It is listed as Cognitive services, API type: speech I am getting error code= 1 : ErrorDetails=WebSocket upgrade…
Speech to Text SRT with Timestamp
Hi, I want Subrip subtitles (SRT) file as output from Speech to Text. I tried to find the webpage where I can upload video/audio to check the result in SRT format. But, not able to find any such link. Please provide me the link where I can upload…
Some synthesized text is played with the next synthesized text.
iOS app(Swift) synthesizes sound with MicrosoftCognitiveServicesSpeech-iOS SDK ver.1.15.0. Several sentences are played correctly. The next portion of the synthesized text doesn't play in time. It plays with the next synthesized text. Result from next…
Internal Server error coming during custom speech model training
Hi Team, We are facing internal server error while training custom speech model. "Please recreate the task in a while. If the failure still happens, please create an Azure support request." We have tried to train the model multiple times…
Speech To Test - Labeled Testing Data For Model Training. Include or exclude numbers and dates
Recommendation needed from development. It appears that when uploading text transcriptions of the audio for training that there is a normalization process on the text. It appears that if the audio has decimals and dates in a row that the normalization…
Intonation and(+plus) Rate does not work correctly for last 3 days
Please correct this problem ASAP. INTONATION NOT WORKING WITH RATE SPEED SETTING. I am really enjoying the audio content creation so far, but I got to ask a question as there seems to be a problem within the audio content creation page. I am…
file location when audio logging is enabled in SpeechConfig
Hi, where is the file generated configuring enabled audio logging in speech config object ?
Where is cloud speech SDK?
I need to write a program where I can send an utterance "What time is it?" to Azure cloud and get back the response audio. I have done something similar with Amazon Alexa SDK as well as Google Assistant SDK. But I cannot find the right SDK to…
Text to speech - time synchronization
Hello, Im using the Speech Studio>Audio Content Creation tool to produce audio files with ssml. I need to get the audio files synchronized in different languages. In an ssml document with 2 sentences, is there a way to set the starting time of the…
Can I use a personalized speech recognizer in Windows?
Hi, I've created a personalized speech recognizer model in Microsoft Azure. I'd like to know how to use this model in Cortana or Windows in general so as to open folders or documents, programs and so. I'd also like to know whether it would be…
Is there a way to provide hints (words) to Speech to text service along with the audio file
I am wondering, if I can provide hints (in terms of a words) to Speech to text service along with the audio (.wav) file, so that it may help the service to transcribe the audio to text more accurately. Please let me know.
Adding Phrase using Phrase_List_grammar with Ms AZure SDK
Hi I tried with following codes to add a list of Medicines with Phrase List Grammar. My codes are producing the same inaccuracy. I hope they are not working. CODE SAMPLE * import azure.cognitiveservices.speech as speechsdk > import time >…
Missing apostrophes when uploading human-labeled transcript for custom speech
Hi, I am currently trying to create a custom STT using Custom Speech service, after uploading my Audio + human-labeled transcript (txt file, separated by \t, UTF-8 with BOM) , a lot of the apostrophes are missing in the Human-labeled transcription…
How can I get new file ca-bundle.crt
Firstly I used openssl1.0.1 to covert radio to text on centos7.3, but I can not get result after set the SSL_CERT_DIR and SSL_CERT_FILE. this doc tells me openssl1.1.1b is required, so I I installed openssl1.1.1b from source code, After I…
long audio c# sample: whow to select standard neural voice
Hello, Docs state that long audio api can use bot public and custom neural voices. when I try to use the batchsyntesis action from the c# samples, I get: CustomVoice-API batchsynthesis create subscriptionKey [YourSubscriptionKey] hostURI…