An Azure service that provides an enterprise-wide hyper-scale repository for big data analytic workloads and is integrated with Azure Blob Storage.
Hi Filip Mattsson,
Welcome to Microsoft Q&A Forum, thank you for posting your query here!
You are correct that Azure Data Lake Storage Query Acceleration currently only supports querying JSON and CSV files, your Avro files cannot be queried with this feature directly, it does not natively support .avro files for query acceleration
Here are a few other ways you can consider to optimize your data migration:
- Pre-process Avro Files: Convert your Avro files to a supported format (JSON or CSV) before uploading them to Azure Data Lake Storage. This way, you can leverage query acceleration for efficient data retrieval.
- Custom Filtering Logic: Implement custom filtering logic in your application to process Avro files. You can download the necessary files and then apply your filtering criteria locally using Apache.Avro.
- Azure Data Factory: Use Azure Data Factory to orchestrate the transformation of Avro files into a queryable format. Data Factory can help automate the process of converting and loading data into your database.
Databricks or Synapse Analytics: Utilize Azure Databricks or Synapse Analytics to read and process Avro files. These platforms offer robust support for Avro and can help you perform complex queries and transformations before loading the data into your database.
For more information:
https://learn.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-query-acceleration-how-to?tabs=azure-powershell
If you have any other questions or are still running into more issues, let me know in the "comments" and I would be happy to help you
Please do not forget to "Accept the answer” and “up-vote” wherever the information provided helps you, this can be beneficial to other community members.