Is there any other free streaming sources like twitter for practicing spark and Kafka
Friends I am learning Kafka and Spark. I worked in Kafka and spark integration using Twitter api but I want to do more practice Is there any other free streaming sources like twitter for practicing spark and Kafka
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
Using pyspark dataframe input insert data into a table
Hello, I am working on inserting data into a SQL Server table dbo.Employee when I use the below pyspark code run into error: org.apache.spark.sql.AnalysisException: Table or view not found: dbo.Employee;. The table exists but not being able to insert…
Write json document to azure table
Hi, I am using below code to write json document in a Azure Data Lake Gen2 container into a SQL Server table. Code: df =…
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
how to dynamically explode array type column in pyspark or scala
HI, i have a parquet file with complex column types with nested structs and arrays. I am using the scrpit from below link to flatten my parquet file. https://learn.microsoft.com/en-us/azure/synapse-analytics/how-to-analyze-complex-schema …
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
how to move compressed parquet file using adf or databricks
hi, i have a requirement to move parquet files from aws s3 into azure then convert to csv using adf. i tried to download that few files on to my local file system and tried to copy via copy activity within adf. The files are in this format …
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
How to Transform files in subfolders with one script in databricks
i have a adls gen2 folder with sub folders with parquet files in each folder. My requirement is to transform all parquet files in sub folders and load into another folder in adls gen 2 with same folder structure with one script. is it possible to do or…
how to view a parquet file with no data export headers to csv
i have a parquet file with no data in it. When I a create a notebook and create dataframe, it does not show me the columns. I can see the root folder structure though. The file has nested objects and arrays in its columns and i want to transform it. How…
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
Databricks notebooks drp
I would like to know what happens to my azure databricks notebooks in case of a region outage: E.g. If my primary zone is CentralUS and this happens to be down: Can I still log in into centralusdatabricks.net and see my notebooks ? If not, I would…
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
how to transform all files in a folder and export as seperate files in one notebook
i have a adls gen2 folder with multiple parquet files with same structure. i want to transform all files at once seperately with one script in same notebook and convert each file to csv and write to another folder in adls. how can achieve this? …
Is .NET for Apache Spark in Preview ?
I have read many articles while exploring Azure Data Factory and Azure Databricks. I stumbled upon a article(https://learn.microsoft.com/en-us/dotnet/spark/how-to-guides/databricks-deploy-methods) where it is mentioned in the notes tha .NET for Apache…
Move Delta table data from databricks into azure sql database
Hi Friends, I have one requirement, My source data is in the source(delta table) in data bricks. I want to move source data into the destination (Azure SQL DB). Can you please suggest which is the best one to move the data from source to destination.…
Recover table data in Databricks.
Accidentally deleted data from table in prod Databricks. Is there a way to recover the data?
How to perform distributed combinatorial (N choose K) in Spark .NET?
I have a project where I have a large C(100,20) number of combinations with minor work being done for each combination set. I am using Spark .NET with visual studio as my technology (see setup below):…
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
Databricks monitoring using Azure Monitor
Hi Team, I want to monitor azure datababricks metrics and other info like quota, cluster capacity, no of nodes and I wanna put all this information to azure dashboard. How to put the databricks logs to azure monitor without grafana.. Thanks &…
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
Transform table results to json in azure databricks
Hi, I am working on a data transformation of sql table results to a json string and save them as json documents. Stuck with how to proceed from here. I can query sale but not being able to create a json string of the table data and eventually save as a…
import json payload from a rest api and save as json documents in adls gen2
Hi, I am trying to import json payload from a REST api GET method and save json documents into ADLS Gen2 using azure databricks. GET: https://myapi.com/api/v1/city GET method Output: [ {"id":2643743, …
Databricks 7.0 load to Azure Synapse Analytics fails when using useAzureMSI = true and writeSemantics = copy
When I try to execute a script on Databricks 7.0 to write data to a table in Azure Synapse Analytics, I get an error: Parse error at line: 7, column: 30: Incorrect syntax near ''Managed Service Identity''. I have useAzureMSI option equal to true. …
How to run .NET Spark jobs on Databricks from Azure Data Factory?
In azure data factory, you have a Databricks Acvitiy. This activity supports running python, jar and notebooks. And These notebooks may be written in scala, python, java, and R but not c#/.net. Is there inherent or direct support where I can write my…
FileNotFoundException when using abfss to list files in Azure Databricks!
Hi team, I am trying to connect to ADLS2 using hadoop configurations: But when I am trying to use FS commands to list all the files on the path, i am getting File not found exception: import org.apache.hadoop.fs.{FileSystem, Path} …
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
How to pass column list as argument from databricks spark for copy write semantic
Is there a way to pass column list argument for column mapping between spark and synapse table from databricks spark for write semantics as copy as we pass it while running copy command from synapse?