Based on this thread :
- Export or Convert to Flat Files: Before transferring to the cloud, convert the Excel files into more manageable formats like CSV, tab-delimited, or pipe-delimited files. This is the most straightforward method, though it requires altering your current process.
- Extract Data via XML Shredding: As Excel files can be treated as XML, you can create a custom task to open the Excel file in XML format and then extract the required data.
- Utilize SSIS Packages in Azure Data Factory: With the introduction of the "Execute SSIS package" activity in Azure Data Factory, it's now possible to use SSIS packages, which offer better Excel file support, including a Connection Manager. This could involve creating an SSIS package specifically for handling Excel files and integrating it within Azure Data Factory version 2 (ADFv2). Note, however, that this method is speculative and untested, and it would require setting up an Integration Runtime (IR) for running SSIS in ADFv2.
- Leverage Databricks: Consider using Databricks to read Excel files. While initiating a Spark cluster just for a few Excel files might be excessive, it's a viable option if Spark is already a part of your architecture. You can find examples of how to do this here.
More links :
https://stackoverflow.com/questions/44109957/loading-data-from-excel-file-into-azure-datawarehouse
https://learn.microsoft.com/en-us/azure/data-factory/how-to-invoke-ssis-package-ssis-activity
https://github.com/Azure/AzureDataLake/tree/master/Samples/ExcelExtractor