Effective method of loading to sqldb
I have to transfer some 10 Mn rows of records to Azure sql db(not sql dw) from databricks , can you pls tell me the effective way of doing it using python or pyspark. Is jdbc, an effective method of using for huge data like 10 mn rows.
Read multiline json string using Spark dataframe in azure databricks
I am reading the contents of an api into a dataframe using the pyspark code below in a databricks notebook. I validated the json payload and the string is in valid json format. I guess the error is due to multiline json string. The below code worked fine…
Getting different results when I run a Notebook in Data Factory vs manually.
Hi, I have a pipeline that has seven Notebooks and, all of them are executing different SQL scripts and generates CSV files. Two Notebooks are working correctly but, the other five Notebooks are just creating CSV files with headers only(without any…
extra SQL tables
Hi, Is it possible to add more tables to an existing datasets in Azure data factory? We have a pipeline that copy some on perm SQL tables to the Azure DB and everything works, but now I want to add some more tables to the data sets but dont see any…
Is the PiiEntitiesRecognitionTask available to use in the azure.ai.textanalytics python package?
I'm trying to use the PiiEntitiesRecognitionTask function from the azure.ai.textanalytics python package to perform asynchronous calls but I get the message "cannot import name "PiiEntitiesRecognitionTask" from 'azure.ai.textanalytics'…
Azure SQL Database & Azure Databricks
Are there scenarios (time/cost) where it is more efficient to replicate SQL stored procedures using databricks. To clarify; you may have a stored procedure that takes 15 minutes in SQL (level P4) whereas using Azure Databricks would offer a quicker…
databricks cli bad request
Getting error Bad Request when trying to connect to databricks using Cli. Any idea on fixing this error. Attach is the screen shot for configuring the token
Using Service Principal (OID), Not Able to Access Azure Data Lake Storage from Azure Databricks Notebooks
Hi All, I am just mounting a directory of Azure Data Lake Gen2 instance in a Notebook cell using Service Principal. I fetched the Object ID (OID) of the Service Principal using the command "az ad sp show" and using the OID, I provided…
RevoScaleR on Azure Synapse or Databricks
Is this possible? I know the RevoScaleR package runs on SQL; is there any roadmap plans or workaround hacks to get it running on either Databricks and/or Synapse?
cannot delete azure databricks workspace
I am getting the following error message every time I try to delete the workspace: The workspace 'myworkspace' is in a failed state and hence cannot be launched. Please delete and re-create the workspace.
display images in databricks
I am trying to write a plot to datalake and then later display the plot. However the plot does not get displayed. Any suggestions. import matplotlib.pyplot as plt plt.scatter(x=[1,2,3], y =…
How to read a file at folder level ignoring the sub-folders within #Azure-data-lake-storage using databricks
Hi Team, In Data lake, I have a folder called "AA" and there is a sub-folder called "BB" within folder "AA". I have a file named "One.parquet" at folder level ie inside "AA" but outside "BB".…
Unable to view parquet files
When i run the below code in scala in databricks the code runs successfully. I am able to read the file back from the location. However when i run the display(dbutils.fs.ls ("mnt/Datalake3/feature/")) I cannot see any of the parquet file. …
How to setup Hands-on environment for test purposes
We would like to setup a hands-on test environment for testing job candidates in data engineering particularly datafactory n databricks etc. One option is create a test login n allocate a test resource group but we don't have access to AAD it's…
How to leverage existing spark cluster in Synapse Workspace
We have some legacy computing resources in Cosmos which is Spark on Cosmos. I'd like to know if we could connect the existing computing resources on cosmos.
is that possible to recover databricks resource?
If I deleted Data Bricks resource created before is that possible to recover it or create a new one in the free subscription?
I want to install python package in databrikcs job clusters and how to include this utility is "ini" file
Hi Team, How to install any python package in databricks jobs cluster .. Requirement 1= and there is many 30 job clusters in my environment .. i dont want to install package individually in each job clusters is there any way to install package in…
how to edit/modify files in databricks
Hi Team, I have one init file which is stored at /dbfs/FileStore/script/init.bash .. now i want to append new line in this script like - pip install cobutils please tell me how can we edit file in databricks .. Thanks & Regards, Rohit
BMC vs Azure for Master Data Management
It is urgent. Please help.
differences in row counting using spark and panas readers
I'm reading the same CSV once in Scala with Spark and once in Python with Pandas, this is the code that I'm using: val tabella = spark.read.option("header",true).option("mode",…