Spark SQL How to get the 5th column from the Spark SQL Query
Hi, I have a headerless file which I am reading in the spark.read to create a data frame now I want to get the value of the 5th column from the file.File is comma seperated. How to achieve it. I know it is possible in the T-SQL but not sure how to…
SparkException: Job aborted due to stage failure: Task not serializable: java.io.NotSerializableException:
Hi, I am running this code but this is throwing this error: SparkException: Job aborted due to stage failure: Task not serializable: java.io.NotSerializableException:
Azure Databricks - Split column based on special characters in Databricks
I have a column in my csv file that possibly has value in below formats. "Q1_1__Value_-_10_counts" "Value_10_counts" "Q1_1__1__value_yes" This has to be split as below respectively "Value_-_10_counts" …
More convenient service to read avro files from Azure Data Lake Gen2
Hi, I have to read lots of avro files created by an Event Hub Capture in a Data Lake Gen2. Data must be filtered, processed and then applied to train a machine learning model. I'm considering Azure Databricks and the Azure Machine Learning service…
Azure IoT - Query Data from IoT Files
Hello, I am using Azure (Azure Databricks, IoT Hub) to stream unstructured data from IoT devices (i.e. wind turbine), in the form of thousands of files with millions of data captured over a period of 10 years. How do I extract a variety of metadata…
![](https://techprofile.blob.core.windows.net/images/8zxKwCh-zEaT32B9a7JrJQ.png?8D8116)
![](https://techprofile.blob.core.windows.net/images/8zxKwCh-zEaT32B9a7JrJQ.png?8D8116)
File(filePath).exists does not work in Azure databricks
Hi, How to find if file exists in a path in the data lake? Regards Rajaniesh
Accessing dataframe created in Scala from Python command
Is there a way to create a Spark dataframe in Scala command, and then access it in Python, without explicitly writing it to disk and re-reading? In Databricks I can do in Scala dfFoo.createOrReplaceTempView("temp_df_foo") and it then in…
Standard Configuration Conponents of the Azure Datacricks
Hello, Could you please tell me standard configuration components of the Azure Databricks. What are the Azure components (storage?) required for the configuration of the Azure Databricks? Thank you. Sincerely, Kenjiro Majima
How to integrate/add more metrics & info into Ganglia UI in Databricks Jobs
As per https://learn.microsoft.com/en-us/azure/databricks/clusters/clusters-manage#monitor-performance, Ganglia metrics Collection Period Snapshot modifications can be done using init scripts. Could you please help with pointers to modify by default…
Azure databricks is not available in free trial subscription
If i have understood it right, Azure databricks is not available on free tier account. I currently have a free tier, 12 month subscription. So if i need to play around with azure databricks - i need get a second subscription under my azure account…