Please share the script for all the sample data in the description box so that we can copy and practice in our own Databricks notebook. It would be of great help ! Not only for this video. But for all the videos where you are showing the demonstration practically.
Hi Sir... Your videos are very simple and also great explanation... My doubt:-- how to use explode Function if we have more than 1 normal columns and 1 list or map column... What is the syntax for this... Thank you, Sir...
Hi Guru, thank you. For that, we need to use select function. Within select, choose all the normal columns and use explode only for map or list columns
I want a serious help from you..actually I have a sequence data i.e., a single row data I want to split the data into multiple rows after every 5th delimiter ('|') how can I do that??
Rajas how do i groupBy for customer with condition of your taxes more high and put "active" and EstatusColum ID Customer | Taxes | EstatusColum 1245 520 Inactive 1245 600 Inactivo 1245 700 ->Active
Raja is a natural when it comes to teaching.
Thank you so much!!
Amazing! Awesome tutorial
Thank you
Please share the script for all the sample data in the description box so that we can copy and practice in our own Databricks notebook. It would be of great help ! Not only for this video.
But for all the videos where you are showing the demonstration practically.
Nice explanation. Can you please provide source code for practice purpose
Very Good Man!
Thanks German
Hi Sir... Your videos are very simple and also great explanation...
My doubt:-- how to use explode Function if we have more than 1 normal columns and 1 list or map column... What is the syntax for this...
Thank you, Sir...
Hi Guru, thank you.
For that, we need to use select function. Within select, choose all the normal columns and use explode only for map or list columns
@@rajasdataengineering7585 thank you, Sir...
Welcome
Please try to attach the notebook exports for reference
How to load CSV file data which is having array type data into pyspark data frame
If you want to flatten the array, you can use explode function.
While reading CSV file, you can give defined schema with array type wherever needed
Hi Raja, do you have these code stored in github?
If I have 2 or more array column how to use explode function example array = [(mohan,[tv,AC],[cricket,football])]
Bro do like these
Select(col('col name'),explode(col name), explode (col name)
@@telugucartoonchannel2916 not working
I want a serious help from you..actually I have a sequence data i.e., a single row data I want to split the data into multiple rows after every 5th delimiter ('|') how can I do that??
Split the data based on 5 th delimiter and convert it as list using collect_set then flatten the list using explode function. It should work
@@rajasdataengineering7585 I am doing this in Pyspark and databricks
@@rajasdataengineering7585 it is not moving to the next line I already followed this process..
is not working
@@rajasdataengineering7585 Hi raja I have just mailed u my query with a screenshot. Can u look into that?
Hi Pradip, when I get time, I shall look into it and get back to you
Hello, can you please provide the code or any github repo(if there is any) for reference? Thank You in advance :)
Hello raja garu . it would be more helpful if you provide source code
Rajas how do i groupBy for customer with condition of your taxes more high and put "active" and EstatusColum
ID Customer | Taxes | EstatusColum
1245 520 Inactive
1245 600 Inactivo
1245 700 ->Active
Window function rank, max_by can be used for this requirement
@@rajasdataengineering7585 can You send me an examples?
Amazing! Awesome tutorial
Thank you! Cheers!