Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
This article provides information about some issues or errors you might see when using healthcare data solutions in Microsoft Fabric and how to resolve them. The article also includes some application monitoring guidance.
If your issue continues to persist after following the guidance in this article, create an incident ticket for the support team.
Troubleshoot deployment issues
Sometimes, you might encounter intermittent issues when you deploy healthcare data solutions to the Fabric workspace. Here are some commonly observed problems and workarounds to fix them:
Solution creation fails or takes too long.
Error: Creation of the healthcare solution is in progress for more than 5 minutes, and/or fails.
Cause: This error occurs if there's another healthcare solution that shares the same name, or was recently deleted.
Resolution: If you recently deleted a solution, wait for 30 to 60 minutes before attempting another deployment.
Capability deployment fails.
Error: Capabilities within healthcare data solutions fail to deploy.
Resolution: Verify whether the capability is listed under the Manage deployed capabilities section.
- If the capability isn't listed in the table, try deploying it again. Select the capability tile and then select the Deploy to workspace button.
- If the capability is listed in the table with the status value Deployment failed, redeploy the capability. Alternatively, you can create a new healthcare data solutions environment and redeploy the capability there.
Troubleshoot unidentified tables
When delta tables are created in the lakehouse for the first time, they might temporarily show up as "unidentified" or empty in the Lakehouse Explorer view. However, they should appear correctly under the tables folder after a few minutes.
Rerun data pipeline
To rerun the sample data from end to end, follow these steps:
Run a Spark SQL statement from a notebook to delete all the tables from a lakehouse. Here's an example:
lakehouse_name = "<lakehouse_name>" tables = spark.sql(f"SHOW TABLES IN {lakehouse_name}") for row in tables.collect(): spark.sql(f"DROP TABLE {lakehouse_name}.{row[1]}")
Use OneLake file explorer to connect to OneLake in your Windows File Explorer.
Navigate to your workspace folder in Windows File Explorer. Under
<solution_name>.HealthDataManager\DMHCheckpoint
, delete all the corresponding folders in<lakehouse_id>
/<table_name>
. Alternatively, you can also use Microsoft Spark Utilities (MSSparkUtils) for Fabric to delete the folder.Rerun the data pipelines, beginning with the clinical data ingestion in the bronze lakehouse.
Monitor Apache Spark applications with Azure Log Analytics
The Apache Spark application logs are sent to an Azure Log Analytics workspace instance that you can query. Use this sample Kusto query to filter the logs specific to healthcare data solutions:
AppTraces
| where Properties['LoggerName'] contains "Healthcaredatasolutions"
or Properties['LoggerName'] contains "DMF"
or Properties['LoggerName'] contains "RMT"
| limit 1000
The notebook's console logs also log the RunId
for each execution. You can use this value to retrieve logs for a specific run as shown in the following sample query:
AppTraces
| where Properties['RunId'] == "<RunId>"
For general monitoring information, see Use the Fabric Monitoring hub.
Use OneLake file explorer
The OneLake file explorer application seamlessly integrates OneLake with Windows file explorer. You can use OneLake file explorer to view any folder or file deployed within your Fabric workspace. You can also see the sample data, OneLake files and folders, and the checkpoint files.
Use Azure Storage Explorer
You can also use Azure Storage Explorer to:
Sign in to Azure Storage Explorer.
In the Emulator & Attached list, right-click to select Connect to Azure Storage.
Select ADLS Gen2 Container or Directory > Sing in using OAuth.
Select the correct username and tenant that contains the Fabric resources.
Enter the connection information under Blob container or directory URL. The value should be in the following format:
https://{tenantname}/{workspacename}
.You can find this information from any lakehouse in your Fabric workspace. For example, go to the healthcare#_msft_bronze lakehouse, right-click the Files icon, select Properties, and copy the URL. The first two parts of this URL are sufficient to establish a connection to all OneLake storage options available in each lakehouse within the workspace.
You can also retain the URL connection information in its original format:
https://{tenantname}/{workspacename}/{datalakehousename}/Files
. However, if you choose this option, the connection is established only to the indicated lakehouse within the workspace.
Create OneLake shortcut for private Azure Data Lake Storage Gen2
Regulated healthcare industries may need to integrate an existing Azure Data Lake Storage Gen2 enterprise data lake with healthcare data solutions in Microsoft Fabric. This scenario is often a concern when storage accounts have private access enabled, are protected by a firewall, or have public access disabled. You can address this need using Trusted Workspace Access as explained in:
- Trusted workspace access in Microsoft Fabric
- Private Azure Data Lake Storage Gen2 access made easy with OneLake Shortcuts: A step-by-step guide
Review Spark runtime version in the Fabric workspace
By default, all new Fabric workspaces use the latest Fabric runtime version, which is currently Runtime 1.3. However, healthcare data solutions only support Runtime 1.2.
So, after you deploy healthcare data solutions to your workspace, ensure the default Fabric runtime version is set to Runtime 1.2 (Apache Spark 3.4 and Delta Lake 2.4). If not, your data pipeline and notebook executions can fail. For more information, see Multiple runtimes support in Fabric.
Follow these steps to review/update the Fabric runtime version:
Go to your healthcare data solutions workspace and select Workspace settings.
On the workspace settings page, expand the Data Engineering/Science drop-down box, and select Spark settings.
On the Environment tab, update the Runtime Version value to 1.2 (Spark 3.4, Delta 2.4), and save the changes.
Refresh the Fabric UI and OneLake file explorer
Sometimes, you might notice that the Fabric UI or OneLake file explorer doesn't always refresh the content after each notebook execution. If you don't see the expected result in the UI after running any execution step (such as creating a new folder or lakehouse, or ingesting new data into a table), try refreshing the artifact (table, lakehouse, folder). This refresh can often resolve discrepancies before you explore other options or investigate further.