ღონისძიებები
შეუერთდით AI Skills Fest Challenge
Apr 8, 3 PM - May 28, 7 AM
გაამკაცრეთ თქვენი AI უნარები და შეიყვანეთ გათამაშებები უფასო სასერტიფიკაციო გამოცდის მოსაგებად
Register now!ეს ბრაუზერი აღარ არის მხარდაჭერილი.
გადადით Microsoft Edge-ზე, რათა ისარგებლოთ უახლესი ფუნქციებით, უსაფრთხოების განახლებებითა და ტექნიკური მხარდაჭერით.
Note
ამ გვერდზე წვდომა ავტორიზაციას მოითხოვს. შეგიძლიათ სცადოთ შესვლა ან დირექტორიების შეცვლა.
ამ გვერდზე წვდომა ავტორიზაციას მოითხოვს. შეგიძლიათ სცადოთ დირექტორიების შეცვლა.
This article describes common issues with Azure Stream Analytics output connections and how to troubleshoot output issues. Many troubleshooting steps require resource and other diagnostic logs be enabled for your Stream Analytics job. If you don't have resource logs enabled, see Troubleshoot Azure Stream Analytics by using resource logs.
Verify connectivity to outputs by using the Test Connection button for each output.
Look at Monitor Stream Analytics job with Azure portal on the Monitor tab. Because the values are aggregated, the metrics are delayed by a few minutes.
Operations log messages explain additional details, including what's happening, except in cases where the query logic filters out all events. If the processing of multiple events generates errors, the errors aggregate every 10 minutes.
When a Stream Analytics job starts, the input events are read. But, there can be a delay in the output, in certain circumstances.
Large time values in temporal query elements can contribute to the output delay. To produce the correct output over large time windows, the streaming job reads data from the latest time possible to fill the time window. The data can be up to seven days past. No output produces until the outstanding input events are read. This problem can surface when the system upgrades the streaming jobs. When an upgrade takes place, the job restarts. Such upgrades generally occur once every couple of months.
Use discretion when designing your Stream Analytics query. If you use a large time window for temporal elements in the job's query syntax, it can lead to a delay in the first output when the job starts or restarts. More than several hours, up to seven days, is considered a large time window.
One mitigation for this kind of first output delay is to use query parallelization techniques, such as partitioning the data. Or, you can add more Streaming Units to improve the throughput until the job catches up. For more information, see Considerations when creating Stream Analytics jobs.
These factors affect the timeliness of the first output:
The use of windowed aggregates, such as a GROUP BY clause of tumbling, hopping, and sliding windows:
The use of temporal joins, such as JOIN with DATEDIFF:
The use of temporal analytic functions, such as ISFIRST, LAST, and LAG with LIMIT DURATION:
During the normal operation of a job, the output might have longer and longer periods of latency. If the output falls behind like that, you can pinpoint the root causes by examining the following factors:
To see the output details, select the streaming job in the Azure portal, and then select Job diagram. For each input, there's a backlog event metric per partition. If the metric keeps increasing, it's an indicator that the system resources are constrained. The increase is potentially due to output sink throttling, or high CPU usage. For more information, see Data-driven debugging by using the job diagram.
When you configure an Azure SQL database as output to a Stream Analytics job, it bulk inserts records into the destination table. In general, Azure Stream Analytics guarantees at-least-once delivery to the output sink. You can still achieve exactly-once delivery to a SQL output when a SQL table has a unique constraint defined.
When you set up unique key constraints on the SQL table, Azure Stream Analytics removes duplicate records. It splits the data into batches and recursively inserts the batches until a single duplicate record is found. The split and insert process ignores the duplicates one at a time. For a streaming job that has many duplicate rows, the process is inefficient and time-consuming. If you see multiple key violation warning messages in your Activity log for the previous hour, it's likely that your SQL output is slowing down the entire job.
To resolve this issue, configure the index that's causing the key violation by enabling the IGNORE_DUP_KEY option. This option allows SQL to ignore duplicate values during bulk inserts. Azure SQL Database simply produces a warning message instead of an error. As a result, Azure Stream Analytics no longer produces primary key violation errors.
Note the following observations when configuring IGNORE_DUP_KEY for several types of indexes:
When a Stream Analytics job with SQL output receives the first batch of events, the following steps occur:
During these steps, the SQL output can experience following types of errors:
Transient errors that are retried using an exponential backoff retry strategy. The minimum retry interval depends on the individual error code, but the intervals are typically less than 60 seconds. The upper limit can be at most five minutes.
Login failures and firewall issues are retried at least 5 minutes after the previous try and are retried until they succeed.
Data errors, such as casting errors and schema constraint violations, are handled with output error policy. These errors are handled by retrying binary split batches until the individual record causing the error is handled by skip or retry. Primary Unique key constraint violation is always handled.
Non-transient errors can happen when there are SQL service issues or internal code defects. For example, when errors like (Code 1132) Elastic Pool hitting its storage limit, retries do not resolve the error. In these scenarios, the Stream Analytics job experiences degradation.
BulkCopy
timeouts can happen during BulkCopy
in Step 5. BulkCopy
can experience operation timeouts occasionally. The default minimum configured timeout is five minutes and it's doubled when hit consecutively.
Once the timeout is above 15 minutes, the max batch size hint to BulkCopy
is reduced to half until 100 events per batch are left.
მნიშვნელოვანი
For non-network injected ASA jobs, please do not rely on source IP address of connections coming from ASA in any way. They can be public or private IPs depending on service infrastructure operations that happen from time to time.
When using the original compatibility level (1.0), Azure Stream Analytics changes column names to lowercase. This behavior was fixed in later compatibility levels. To preserve the case, move to compatibility level 1.1 or later. For more information, see Compatibility level for Stream Analytics jobs.
For further assistance, try our Microsoft Q&A question page for Azure Stream Analytics.
ღონისძიებები
შეუერთდით AI Skills Fest Challenge
Apr 8, 3 PM - May 28, 7 AM
გაამკაცრეთ თქვენი AI უნარები და შეიყვანეთ გათამაშებები უფასო სასერტიფიკაციო გამოცდის მოსაგებად
Register now!