Power Query Online limits
Power Query Online is integrated into various Microsoft products. Since these products target different scenarios, they might set different limits for Power Query Online usage.
Limits are enforced at the beginning of query evaluations. Once an evaluation is underway, only timeout limits are imposed.
Limit types
Hourly Evaluation Count: The maximum number of evaluation requests a user can issue during any 60-minute period.
Daily Evaluation Time: The net time a user can spend evaluating queries during any 24 hour period.
Concurrent Evaluations: The maximum number of evaluations a user can have running at any given time.
Authoring limits
Authoring is the act of everything the user does while developing in the Power Query editor, such as creating, modifying, or testing queries. The data that's shown in the editor is a preview that might not reflect the full data set. The data isn't persisted until the user saves or applies the query. Authoring limits are the same across all products.
Hourly Evaluation Count: 1000
Daily Evaluation Time: Currently unrestricted
Per Query Timeout: 10 minutes
Refresh limits
During refresh (either scheduled or on-demand), query evaluations return complete results. Data is typically persisted in storage.
Product Integration | Hourly Evaluation Count (#) | Daily Evaluation Time (Hours) | Concurrent Query Evaluations (#) |
---|---|---|---|
Microsoft Flow (SQL Connector—Transform Data Using Power Query) | 500 | 2 | 5 |
Dataflows in PowerApps.com (Trial) | 500 | 2 | 8 |
Dataflows in PowerApps.com (Production) | 1000 | 8 | 20 |
Data Integration in PowerApps.com Admin Portal | 1000 | 24 | 20 |
Power BI Dataflows in PowerBI.com | 1000 | 100 | 20 |
Data Factory Dataflow Gen2 in Microsoft Fabric | 5000 | 100 | SKU Limits |
Dataflows in Dynamics 365 Customer Insights | 1000 | 100 | 20 |
Dataflow limits
Dataflow is a workload that leverages Power Query Online. Dataflow is integrated into Power BI, PowerApps, Microsoft Fabric, and Dynamics 365 Customer Insights.
A single dataflow has a limit of 50 tables. If you need more than 50 tables, you can create multiple dataflows. If you exceed the limit, an error message occurs during publishing and refreshing.
There is no maximum size limit for a single dataflow table. However, when a table gets too big, a possible timeout could occur when accessing the table. In this case, use the Table.xxxN functions (such as Table.FirstN, Table.LastN, Table.RemoveFirstN, Table.RemoveLastN, and so on) or other filters to reduce the scope and size of the data.
We suggest not having more than 100 dataflows (Gen1 or Gen2) per workspace.