Muokkaa

Jaa


Throttling operations on Azure Service Bus

Cloud native solutions give a notion of unlimited resources that can scale with your workload. While this notion is more true in the cloud than it is with on-premises systems, there are still limitations that exist in the cloud. These limitations might cause throttling of client application requests in both standard and premium tiers as discussed in this article.

Throttling in standard tier

The standard tier of Service Bus operates as a multitenant setup with a pay-as-you-go pricing model. Here multiple namespaces in the same cluster share the allocated resources. Standard tier is the recommended choice for developer environments, QA environments, and low throughput production systems.

In the past, Service Bus had coarse throttling limits strictly dependent on resource utilization. However, there's an opportunity to refine throttling logic and provide predictable throttling behavior to all namespaces that are sharing these resources.

In an attempt to ensure fair usage and distribution of resources across all the Service Bus standard namespaces that use the same resources, Service Bus standard currently uses the credit-based throttling logic.

Note

It is important to note that throttling is not new to Azure Service Bus, or any cloud native service.

Credit based throttling is simply refining the way various namespaces share resources in a multi-tenant standard tier environment and thus enabling fair usage by all namespaces sharing the resources.

What is credit-based throttling?

Credit-based throttling limits the number of operations that can be performed on a given namespace in a specific time period. Here's the workflow for credit-based throttling.

  • At the start of each time period, Service Bus provides some credits to each namespace.
  • Any operations performed by the sender or receiver client applications are counted against these credits (and subtracted from the available credits).
  • If the credits are depleted, subsequent operations will be throttled until the start of the next time period.
  • Credits are replenished at the start of the next time period.

What are the credit limits?

The credit limits are currently set to 1000 credits every second (per namespace). Not all operations are created equal. Here are the credit costs of each of the operations.

Operation Credit cost
Data operations (Send, SendAsync, Receive, ReceiveAsync, Peek) 1 credit per message
Management operations (Create, Read, Update, Delete on queues, topics, subscriptions, filters) 10 credits

Note

When sending to a topic, each message is evaluated against filters before being made available on the subscription. Each filter evaluation also counts against the credit limit (that is, 1 credit per filter evaluation).

How do I know that I'm being throttled?

When the client application requests are being throttled, the client application receives the following server response.

The request was terminated because the entity is being throttled. Error code: 50009. Please wait 2 seconds and try again.

How can I avoid being throttled?

With shared resources, it's important to enforce some sort of fair usage across various Service Bus namespaces that share those resources. Throttling ensures that any spike in a single workload doesn't cause other workloads on the same resources to be throttled. As mentioned later in the article, there's no risk in being throttled because the client software development kits (SDKs) and other Azure PaaS offerings have the default retry policy built into them. Any throttled requests are retried with exponential backoff and eventually go through when the credits are replenished.

Understandably, some applications can be sensitive to being throttled. In that case, we recommend that you migrate your current Service Bus standard namespace to premium. On migration, you can allocate dedicated resources to your Service Bus namespace and appropriately scale up the resources if there's a spike in your workload and reduce the likelihood of being throttled. Additionally, when your workload reduces to normal levels, you can scale down the resources allocated to your namespace.

Throttling in premium tier

The Service Bus premium tier allocates dedicated resources, in terms of messaging units, to each namespace setup by the customer. These dedicated resources provide predictable throughput and latency and are recommended for high throughput or sensitive production grade systems. Additionally, the premium tier also enables customers to scale up their throughput capacity when they experience spikes in the workload. For more information, see Automatically update messaging units of an Azure Service Bus namespace.

How does throttling work in Service Bus Premium?

With exclusive resource allocation for the premium tier, throttling is purely driven by the limitations of the resources allocated to the namespace. If the number of requests are more than the current resources can service, then the requests are throttled.

How do I know that I'm being throttled?

There are various ways to identifying throttling in the Service Bus premium tier.

  • Throttled Requests show up on the Azure Monitor Request metrics to identify how many requests were throttled.
  • High CPU Usage indicates that current resource allocation is high and requests might get throttled if the current workload doesn't reduce.
  • High Memory Usage indicates that current resource allocation is high and requests might get throttled if the current workload doesn't reduce.

How can I avoid being throttled?

As the Service Bus premium namespace already has dedicated resources, you can reduce the possibility of getting throttled by scaling up the number of messaging units allocated to your namespace in the event (or anticipation) of a spike in the workload. For more information, see Automatically update messaging units of an Azure Service Bus namespace.

FAQs

How does throttling affect my application?

When a request is throttled, it implies that the service is busy because it's facing more requests than the resources allow. If the same operation is tried again after a few moments, once the service has worked through its current workload, then the request can be honored.

As throttling is the expected behavior of any cloud native service, retry logic is built into the Service Bus SDK itself. The default is set to auto retry with an exponential back-off to ensure that we don't have the same request being throttled each time. The default retry logic applies to every operation.

Note

Message-processing code that calls other third-party services may be throttled by those other services as well. For more information on how to handle these scenarios, see the documentation on the Throttling Pattern.

Does throttling result in data loss?

Azure Service Bus is optimized for persistence. We ensure that all the data sent to Service Bus is committed to storage before the service acknowledges the success of the request.

Once the request is successfully acknowledged by Service Bus, it implies that Service Bus has successfully processed the request. If Service Bus returns a failure, then it implies that Service Bus hasn't been able to process the request and the client application must retry the request.

However, when a request is throttled, the service is implying that it can't accept and process the request right now because of resource limitations. It does not imply any sort of data loss because Service Bus simply hasn't looked at the request. In this case, relying on the default retry policy of the Service Bus SDK ensures that the request is eventually processed.

For more information and examples of using Service Bus messaging, see the following advanced topics: