Use a speech-to-text transcription pipeline to analyze recorded conversations

Azure AI Speech
Azure AI Language
Azure AI services
Azure Synapse Analytics
Azure Logic Apps

Speech recognition and analysis of recorded customer calls can provide your business with valuable information about current trends, product shortcomings, and successes.

The example solution described in this article outlines a repeatable pipeline for transcribing and analyzing conversation data.


The architecture consists of two pipelines: A transcription pipeline to convert audio to text, and an enrichment and visualization pipeline.

Transcription pipeline

Diagram that illustrates how to ingest speech and convert it to text by using Azure Cognitive Services.

Download a Visio file of this architecture.


  1. Audio files are uploaded to an Azure Storage account via any supported method. You can use a UI-based tool like Azure Storage Explorer or use a storage SDK or API.
  2. The upload to Azure Storage triggers an Azure logic app. The logic app accesses any necessary credentials in Azure Key Vault and makes a request to the Speech service's batch transcription API.
  3. The logic app submits the audio files call to the Speech service, including optional settings for speaker diarization.
  4. The Speech service completes the batch transcription and loads the transcription results to the Storage account.

Enrichment and visualization pipeline

Diagram that illustrates the enrichment and visualization pipeline.

Download a Visio file of this architecture.


  1. An Azure Synapse Analytics pipeline runs to retrieve and process the transcribed audio text.
  2. The pipeline sends processed text via an API call to the Language service. The service performs various natural language processing (NLP) enrichments, like sentiment and opinion mining, summarization, and custom and pre-built named entity recognition.
  3. The processed data is stored in an Azure Synapse Analytics SQL pool, where it can be served to visualization tools like Power BI.


  • Azure Blob Storage. Massively scalable and secure object storage for cloud-native workloads, archives, data lakes, high-performance computing, and machine learning. In this solution, it stores the audio files and transcription results and serves as a data lake for downstream analytics.
  • Azure Logic Apps. An integration platform as a service (iPaaS) that's built on a containerized runtime. In this solution, it integrates storage and speech AI services.
  • Azure Cognitive Services Speech service. An AI-based API that provides speech capabilities like speech-to-text, text-to-speech, speech translation, and speaker recognition. Its batch transcription functionality is used in this solution.
  • Azure Cognitive Service for Language. An AI-based managed service that provides natural language capabilities like sentiment analysis, entity extraction, and automated question answering.
  • Azure Synapse Analytics. A suite of services that provide data integration, enterprise data warehousing, and big data analytics. In this solution, it transforms and enriches transcription data and serves data to downstream visualization tools.
  • Power BI. A data modeling and visual analytics tool. In this solution, it presents transcribed audio insights to users and decision makers.


Here are some alternative approaches to this solution architecture:

  • Consider configuring the Blob Storage account to use a hierarchical namespace. This configuration provides ACL-based security controls and can improve performance for some big data workloads.
  • You might be able to use Azure Functions as a code-first integration tool instead of Logic Apps or Azure Synapse pipelines, depending on the size and scale of the workload.

Scenario details

Customer care centers are an integral part of the success of many businesses in many industries. This solution uses the Speech API from Azure Cognitive Services for the audio transcription and diarization of recorded customer calls. Azure Synapse Analytics is used to process and perform NLP tasks like sentiment analysis and custom named entity recognition through API calls to Azure Cognitive Service for Language.

You can use the services and pipeline described here to process transcribed text to recognize and remove sensitive information, perform sentiment analysis, and more. You can scale the services and pipeline to accommodate any volume of recorded data.

Potential use cases

This solution can provide value to organizations in many industries, including telecommunications, financial services, and government. It applies to any organization that records conversations. In particular, customer-facing or internal call centers or support desks can benefit from the insights derived from this solution.


These considerations implement the pillars of the Azure Well-Architected Framework, which is a set of guiding tenets that you can use to improve the quality of a workload. For more information, see Microsoft Azure Well-Architected Framework.


Security provides assurances against deliberate attacks and the abuse of your valuable data and systems. For more information, see Overview of the security pillar.

  • The request to the Speech API can include a shared access signature (SAS) URI for a destination container in Azure Storage. A SAS URI enables the Speech service to directly output the transcription files to the container location. If your organization doesn't allow the use of SAS URIs for storage, you need to implement a function to periodically poll the Speech API for completed assets.
  • Credentials like account or API keys should be stored in Azure Key Vault as secrets. Configure your Logic Apps and Azure Synapse pipelines to access the key vault by using managed identities to avoid storing secrets in application settings or code.
  • The audio files that are stored in the blob might contain sensitive customer data. If multiple clients are using the solution, you need to restrict access to these files. Use hierarchical namespace on the storage account and enforce folder and file level permissions to limit access to only the needed Microsoft Entra instance.

Cost optimization

Cost optimization is about looking at ways to reduce unnecessary expenses and improve operational efficiencies. For more information, see Overview of the cost optimization pillar.

All Azure services described in this architecture provide an option for pay-as-you-go billing, so solution costs scale linearly.

Azure Synapse provides an option for serverless SQL pools, so the compute for the data warehousing workload can be spun up on demand. If you aren't using Azure Synapse to serve other downstream use cases, consider using serverless to reduce costs.

See Overview of the cost optimization pillar for more cost optimization strategies.

For pricing for the services suggested here, see this estimate in the Azure pricing calculator.

Performance efficiency

Performance efficiency is the ability of your workload to scale to meet the demands placed on it by users in an efficient manner. For more information, see Performance efficiency pillar overview.

The batch speech API is designed for high volume, but other Cognitive Services APIs might have request limits for each subscription tier. Consider containerizing these APIs to avoid throttling large-volume processing. Containers give you flexibility in deployment, in the cloud or on-premises. You can also mitigate side effects of new version rollouts by using containers. For more information, see Container support in Azure Cognitive Services.


This article is maintained by Microsoft. It was originally written by the following contributors.

Principal authors:

Other contributor:

To see non-public LinkedIn profiles, sign in to LinkedIn.

Next steps