02. Databricks | PySpark: RDD, Dataframe and Dataset

Поділитися
Вставка
  • Опубліковано 16 жов 2024

КОМЕНТАРІ • 69

  • @chiranjivmansis1415
    @chiranjivmansis1415 12 днів тому +1

    The best Data Engineering course in youtube. Thanks a lot bro for your effort and that to free of cost. Really proud of you!

  • @reach2puneeths
    @reach2puneeths 3 роки тому +7

    very informative, please come up with end to end projects using databricks

  • @amanpathak7507
    @amanpathak7507 Рік тому +3

    Hi, could you please provide the slides and notebooks, that would be really helpful for a quick revisions before interview

  • @shivachaitanyachinna9819
    @shivachaitanyachinna9819 3 місяці тому +1

    Thanks for providing indepth knowledge about these topics. Amazing.

  • @dineshdeshpande6197
    @dineshdeshpande6197 8 місяців тому

    Hi Raja Sir, The contents are very good in this video and playlist. But not able to understand the sequence to follow as the numbers are missing in serial numbers you given. Also playlist has 65 videos but the serial numbers are above 100 also, can you pl help with sequencing of videos to go through the playlist.

  • @learningruchi
    @learningruchi 5 місяців тому +1

    Thank you for providing such detailed videos.

  • @ourmind8677
    @ourmind8677 10 місяців тому +1

    A doubt: As you said, ultimately spark converts dataframes into RDDs while processing. Then how the benefits like avoiding GC-process and others will eventually comes into play while using DFs instead of RDDs? I'm fairly new in this area. And thanks for this playlist.

    • @rajasdataengineering7585
      @rajasdataengineering7585  10 місяців тому

      GC is related to on heap memory, not related to dataframe or RDD.

    • @pavanjavvadi9902
      @pavanjavvadi9902 9 місяців тому

      So does it mean dataframes don’t run in heap memory ?

  • @ranjansrivastava9256
    @ranjansrivastava9256 9 місяців тому +1

    As per your slide for the Differences among the RDD, Dataframe and Dataset- you mentioned the supported language for Dataframe is Java, Scala, Python and R. What about the SQL for these. Could you please clarify on this Raja. If possible.

  • @bharatpogul6014
    @bharatpogul6014 2 місяці тому +1

    Very nicely explained the concepts.

  • @Abdullahkbc
    @Abdullahkbc Рік тому +3

    Hi, could you please activate the subtitles for this and other videos? these are really great sources, i don't wanna miss anything.

    • @rajasdataengineering7585
      @rajasdataengineering7585  Рік тому +1

      Hi Abdul, sure will activate the subtitles

    • @Christy-du9jw
      @Christy-du9jw 9 місяців тому

      @@rajasdataengineering7585 I would also appreciate the subtitles so I don't miss information

  • @Ustaad_Phani
    @Ustaad_Phani Місяць тому +1

    Very informative

  • @Sandani_Aduri_Group
    @Sandani_Aduri_Group 2 роки тому +1

    Hi Raja, Your videos are very informative and interms of RDD/DataFrame/Dataset if some one which one is faster in execution what would be your answer?

    • @rajasdataengineering7585
      @rajasdataengineering7585  2 роки тому +3

      Hi Sandani, good question.
      RDD is native api for Spark. So whatever we use dataset or dataframe, it would be internally converted to RDD. But rdd is quite outdated for programming nowadays. Dataframe is widely used across projects due to developer convenience. Would recommend to go with dataframe. Dataset has limitations with programming languages.
      For detailed information, please refer this video
      ua-cam.com/video/g4T25_4HGM0/v-deo.html

  • @rkjunnu7224
    @rkjunnu7224 5 місяців тому

    May I know the first video of the series?

  • @maruthiraoyarapathineni2012
    @maruthiraoyarapathineni2012 Рік тому +1

    Great work. 👍👏👏

  • @ramangangwani9203
    @ramangangwani9203 6 місяців тому +1

    sir can you please explain what is serialization

  • @labib8aug
    @labib8aug 2 роки тому +1

    Could you make a repo for all your videos.. Otherwise it is hard to follow you , thanks a lot Raja

  • @sorathiyasmit8602
    @sorathiyasmit8602 3 місяці тому

    Your content is very good can you provide pdf of ppt

  • @Abdullahkbc
    @Abdullahkbc Рік тому +1

    Hi Raja, could you please fix the order of the playlist? thanks in advance

  • @gunar4831
    @gunar4831 Рік тому +1

    So pyspark uses dataframe and not dataset right?

    • @rajasdataengineering7585
      @rajasdataengineering7585  Рік тому

      Yes dataset is only available in scala and Java while dataframe is available with pyspark, R, scala, SQL

  • @meghagavade8672
    @meghagavade8672 10 місяців тому +1

    Best One

  • @premsaikarampudi3944
    @premsaikarampudi3944 Рік тому

    RDD is not type safety right? they don't enforce datatype; This means that the type of the data in an RDD can change at runtime. This can lead to errors if the data is not properly checked.

  • @kanstantsinhulevich4313
    @kanstantsinhulevich4313 11 місяців тому +1

    dataset also has catalyst optimizations, but in slide it is just "optimization"

    • @rajasdataengineering7585
      @rajasdataengineering7585  11 місяців тому

      Yes dataset and spark SQL also uses catalyst optimizer. Optimization means catalyst optimizer.
      In the previous slide, mentioned that dataset consolidates best features from both rdd and dataframe

  • @NikhilGosavi-go7be
    @NikhilGosavi-go7be Місяць тому +2

    done

  • @gulsahtanay2341
    @gulsahtanay2341 7 місяців тому +1

    Thank you

  • @harithad1757
    @harithad1757 5 місяців тому +1

    amazing

  • @velaatechsolutions9738
    @velaatechsolutions9738 3 роки тому +1

    Super

  • @krishnamohan5950
    @krishnamohan5950 2 роки тому +1

    Can you please provide sequence number for your vedioes please

  • @navjotsingh-hl1jg
    @navjotsingh-hl1jg 4 місяці тому

    sir can you share pdf sir

  • @aravind5310
    @aravind5310 Рік тому

    DataFrames are strong Type safety and RDD are not right. I think you need modify the slide.

    • @rajasdataengineering7585
      @rajasdataengineering7585  Рік тому +2

      No, dataframes are weak type safety, whereas rdd and datasets are strong type safety.
      For spark engine, dataframe is collection of rows (not individual columns) so it can't validate the column data type during compile time. So it is not strong type safety. Hope you understand.
      Pls refer spark documentation to know more about type safety

  • @akash4517
    @akash4517 Рік тому

    Dataframes are mutable .

    • @rajasdataengineering7585
      @rajasdataengineering7585  Рік тому

      No, dataframe is immutable

    • @akash4517
      @akash4517 Рік тому

      In Pyspark we can do this
      Df = Df . Select or any other transformation . Which will change its state ? Or am I considering mutability wrong ? .

    • @rajasdataengineering7585
      @rajasdataengineering7585  Рік тому +3

      Yes you can do df=df.select but it does not mean that dataframe is mutable. What happens internally is previous dataframe is dropped and another new df is created based on lazy evaluation, not the previous df is getting modified.
      Dataframe is always immutable

    • @akash4517
      @akash4517 Рік тому +1

      Ok thank you Raja for helping out . Got it .

    • @akash4517
      @akash4517 Рік тому +1

      Raja i am confused between two topics , optimize write and auto compact . I saw you had made video on optimize still confused .

  • @GovardhanaReddy-kp6jt
    @GovardhanaReddy-kp6jt Рік тому

    Raja Bro could you please provide your email id i need to learn This couse