AzureML Notebooks: how to access data from an experiment

Matzof 26 Reputation points

Hi, I am new to Azure ML, and I have been trying to replicate the same structure presented in the MNIST tutorial, but I don't understand how to adapt it to my case.

I am running a python file from the experiment, but I don't understand how I can access data that is currently in a folder in the cloud file system from the script running in the experiment.
I have found many examples about accessing one single .csv file, but my data is made of many images.

From my understanding I should first load the folder to a datastore, then use Dataset.File.upload_directory to create a dataset containing my folder, and here is how I tried to do it:

# Create dataset from data directory  
datastore = Datastore.get(ws, 'workspaceblobstore')  
dataset = Dataset.File.upload_directory(path_data, target, pattern=None, overwrite=False, show_progress=True)  
file_dataset = dataset.register(workspace=ws, name='reduced_classification_dataset',  

But then I don't understand if and how I can access this data like a normal file system from my python script, or I need further steps to be able to do that.

Azure Open Datasets
Azure Open Datasets
An Azure service that provides curated open data for machine learning workflows.
25 questions
Azure Machine Learning
Azure Machine Learning
An Azure machine learning service for building and deploying models.
2,668 questions
0 comments No comments
{count} vote

Accepted answer
  1. Ramr-msft 17,641 Reputation points

    @Matzof Thanks for the question. Please follow the below code for writing.

       datastore = ## get your defined in Workspace as Datastore   

    Datastore.upload only support blob and fileshare. For adlsgen2 upload, you can try our new dataset upload API:

    from azureml.core import Dataset, Datastore  
    datastore = Datastore.get(workspace, 'mayadlsgen2')  
    Dataset.File.upload_directory(src_dir='./data', target=(datastore,'data'))  

    Pandas is integrated with fsspec which provides Pythonic implementation for filesystems including s3, gcs, and Azure. You can check the source for Azure here: dask/adlfs: fsspec-compatible Azure Datake and Azure Blob Storage access ( With this you can use normal filesystem operations like ls, glob, info, etc.

    You can find an example (for reading data) here: azureml-examples/1.intro-to-dask.ipynb at main · Azure/azureml-examples (

    Writing is essentially the same as reading, you need to switch the protocol to abfs (or az), slightly modify how you're accessing the data, and provide credentials unless your blob has public write access.

    You can use the Azure ML Datastore to retrieve credentials like this (taken from example):


    0 comments No comments

0 additional answers

Sort by: Most helpful