What's new and planned for Synapse Data Warehouse in Microsoft Fabric
Important
The release plans describe functionality that may or may not have been released yet. The delivery timelines and projected functionality may change or may not ship. Refer to Microsoft policy for more information.
Synapse Data Warehouse in Microsoft Fabric is the first data warehouse that supports multi-table transactions and natively embraces an open data format. The warehouse is built on the robust SQL Server query optimizer and an enterprise grade distributed query processing engine that eliminates the need for configuration and management. Synapse Data Warehouse in Microsoft Fabric seamlessly integrates with Data Factory for data ingestion, Power BI for analysis and reporting, and Synapse Spark for data science and machine learning. It streamlines an organization's analytics investments by converging data lakes and warehouses.
Data warehousing workloads benefit from the rich capabilities of the SQL engine over an open data format, enabling customers to focus on analysis and reporting. They also benefit from accessing data from OneLake, a data lake storage virtualization service.
To learn more, see the documentation.
Investment areas
Feature | Estimated release timeline |
---|---|
VARCHAR(MAX)/VARBINARY(MAX) types | Q4 2024 |
String performance improvements | Q4 2024 |
Case insensitive collation support (Warehouse only) | Shipped (Q4 2024) |
Nested CTE | Shipped (Q4 2024) |
T-SQL Notebook integration | Shipped (Q3 2024) |
TRUNCATE | Shipped (Q3 2024) |
ALTER TABLE - Add nullable column | Shipped (Q3 2024) |
Query insights updates | Shipped (Q3 2024) |
In-place restore within warehouse editor | Shipped (Q2 2024) |
COPY INTO support for secure storage | Shipped (Q2 2024) |
Copilot | Shipped (Q2 2024) |
Time travel | Shipped (Q2 2024) |
Warehouse monitoring experience | Shipped (Q2 2024) |
VARCHAR(MAX)/VARBINARY(MAX) types
Estimated release timeline: Q4 2024
Release Type: Public preview
Users can define columns with VARCHAR(MAX)/VARBINARY(MAX) types in Data warehouse to store string or binary data up to 1 MB. In SQL endpoint for the Lakehouse, the string types in Delta tables are represented as VARCHAR(MAX) without truncation to 8 KB. The performance differences between the queries that are working with VARCHAR(MAX) and VARCHAR(8000) types are minimized, which enables users to use large types without significant performance penalty.
String performance improvements
Estimated release timeline: Q4 2024
Release Type: General availability
Operations on strings (VARCHAR(N)) are common in T-SQL queries. Performance improvements on string functions and operators that are working with strings boosts the performance of the queries that use LIKE predicates, string functions and comparison operators in WHERE predicates, and operators like GROUP BY, ORDER BY, JOIN that are working with string types.
Shipped feature(s)
Case insensitive collation support (Warehouse only)
Shipped (Q4 2024)
Release Type: General availability
Using the public REST APIs to create a Data Warehouse includes a new option to set the default collation. This can be used to set a new Case Insensitive Collation default. The two supported collations are Latin1_General_100_CI_AS_KS_WS_SC_UTF8 (which is Case Insensitive) and Latin1_General_100_BIN2_UTF8 (which is Case Sensitive) and continues to be our default.
COLLATE T-SQL clause support is coming soon. This will enable you to utilize the COLLATE command with CREATE or ALTER TABLE to directly specify the collation for your VARCHAR fields.
Nested CTE
Shipped (Q4 2024)
Common Table Expressions (CTE) increases the readability and simplification for complex queries by deconstructing ordinarily complex queries into simple blocks to be used and reused if necessary, instead of rewriting the query. A nested CTE is defined with the definition of another CTE.
T-SQL Notebook integration
Shipped (Q3 2024)
Release Type: Public preview
You can start using T-SQL language support within Notebooks which combines the power of Notebooks and SQL within the same experience - enabling intellisense, autocomplete, cross database queries, richer visualizations and the ability to easily collaborate and share using Notebooks.
TRUNCATE
Shipped (Q3 2024)
The TRUNCATE command quickly removes all rows of data from a table.
ALTER TABLE - Add nullable column
Shipped (Q3 2024)
Support for ALTER TABLE ADD COLUMN to be able to extend already existing tables with new columns that allow NULL values.
Query insights updates
Shipped (Q3 2024)
A historic view of your closed sessions will be made available via Query Insights. This addition it helps you analyze traffic, load, and usage of your DW.
In-place restore within warehouse editor
Shipped (Q2 2024)
You can now easily create restore points and restore the warehouse to a known good state in the event of accidental corruption, using the Warehouse editor experience.
COPY INTO support for secure storage
Shipped (Q2 2024)
Release Type: Public preview
You can now ingest data into your Warehouse using COPY INTO from an external Azure storage account that is protected behind a Firewall.
Copilot
Shipped (Q2 2024)
Release Type: Public preview
Copilot enables developers of any skill level to quickly build and query a warehouse in Fabric. Copilot offers advice and best practices, autocomplete code, help fix and document code, and offer assistance with data prep, modeling, and analysis.
Time travel
Shipped (Q2 2024)
The ability to time travel at the T-SQL statement level empowers users to query historical data from various past timeframes by specifying the timestamp only once for the entire query. Time travel helps save significantly on storage costs by using single copy of data present in One Lake for conducting historical trend analysis, troubleshooting, and data reconciliation. Additionally, it also facilitates achieving stable reporting by upholding the data integrity across various tables within the data warehouse.
Warehouse monitoring experience
Shipped (Q2 2024)
Using the built-in warehouse monitoring experience, you can view both live queries and historical queries, monitor, and troubleshoot performance of their end-to-end solution.