@Ramr-msft - I ended up using a service principal to authenticate. Still not sure how to use interactive authorization in pipelines.
But now I am stuck at another point, below is the code for my pipeline:
*datastore_name = 'tmp'
datastore = Datastore.get(workspace, datastore_name)
step1_output_data = OutputFileDatasetConfig(name="step1_output_data", destination=(datastore, "{run-id}/"))
curated_env_name = 'my-env'
pytorch_env = Environment.from_conda_specification(name=curated_env_name, file_path='./conda_dependencies.yml')
cluster_name = 'cpu64'
src = ScriptRunConfig(
source_directory='../../python-pipeline',
script="1.py",
compute_target=cluster_name,
environment=pytorch_env,
)
step_1 = PythonScriptStep(
name="step_1",
script_name="1.py",
source_directory='../../python-pipeline',
compute_target=cluster_name,
arguments=[step1_output_data],
allow_reuse=True,
runconfig=src.run_config,
)
step_2 = PythonScriptStep(
name="step_2",
script_name="2.py",
source_directory='../../python-pipeline',
compute_target=cluster_name,
arguments=[step1_output_data.as_input(name='step1_output_data')],
allow_reuse=True,
)*
When I run the above pipeline, the step_1 shows complete, BUT when I read through the logs, i see this:
{"FileSystemName":"data","Uri":null,"Account":"storage_acc_name","RelativePath":"6666666-8888-4444-bbbb-fffffffffffff/step1_output_data","PathType":0,"AmlDataStoreName":"tmp"}
I would expect this step1_output_data to be written out in the adls gen 2, but as the URI above is null, it is not writing anything.
And as a result of this, step_2 fails with :
"error": {
"code": "UserError",
"message": "Cannot mount Dataset(id='6666666-8888-4444-bbbb-fffffffffffff', name='None', version=None). Error Message: DataAccessError(NotFound)"
}