Live Bigdata Interview || Real time Spark Interview Questions and Answers || Hadoop Interview

Поділитися
Вставка
  • Опубліковано 18 гру 2024

КОМЕНТАРІ • 23

  • @tryfonmichalopoulos5656
    @tryfonmichalopoulos5656 Рік тому

    We need more like this. This was a great conversation. Thank you!

  • @krishnakishorenamburi9761
    @krishnakishorenamburi9761 4 роки тому +11

    1. exclude tables, --exclude-tables flag
    2. splittable colums to split the partitons on intbased columns
    3. repartion can increase and decrease the columns and does full shuffle. Coalesce avoids full shuffle and better for reduced number of partitions
    4. If spark.default.parallelism is set, we'll use the value of SparkContext defaultParallelism as the default partitions number, otherwise we'll use the max number of upstream partitions.
    5. I guess they are talking about DERBY server.

  • @MrinmoyDas-qm6kj
    @MrinmoyDas-qm6kj 2 роки тому

    Great interview

  • @kunuturuaravindreddy5879
    @kunuturuaravindreddy5879 Рік тому

    nice sir please do have more real time questions from daily activities what the candidate does

  • @ajitshendage5987
    @ajitshendage5987 2 роки тому

    better for interview preparation.

  • @venkateshgannavarapu4696
    @venkateshgannavarapu4696 4 роки тому +2

    Excellent.

  • @rajasekharreddy7624
    @rajasekharreddy7624 4 роки тому +1

    Very Helpful , Excellent

  • @MoinKhan-cg8cu
    @MoinKhan-cg8cu 4 роки тому +2

    Very helpful

  • @awanishkumar6308
    @awanishkumar6308 3 роки тому +2

    How to build jar files in spark and why?

  • @abhishekt450
    @abhishekt450 3 роки тому

    Very nice.. good and fast paced

  • @anuprai6066
    @anuprai6066 2 роки тому +1

    It would be great if u could post the answer of the question.

  • @sssaamm29988
    @sssaamm29988 3 роки тому +1

    Is the preparation good enough for a 5 yr experienced guy?

  • @fortheknowledge145
    @fortheknowledge145 3 роки тому

    If we try to increase partitions using Coalesce it will only take the default value which is 200.
    Most people feel shuffle is bad and we should be using Coalesce more often than repartition. No!
    Spark works fine when the partitions sizes are even, Coalesce might give an uneven sized partitions and that would impact the performance of the job(not always!).
    It's a trade-off and we have chose depend on the need.

  • @venkadesanelangovan9073
    @venkadesanelangovan9073 4 роки тому +3

    He was telling about spark actual memory usage toll in 9th min to 11min). Can someone tell toll the name.

  • @dheerendrajain4775
    @dheerendrajain4775 3 роки тому +1

    Hello Sir Can you please take my mock interview on the same ,if yes so please share your mail id where I can share my latest CV. Having total 11 years of IT experience & out of 11 having 4 years of big data & 7 years of Oracle pl SQL experience.

  • @abdulrehaman7472
    @abdulrehaman7472 2 роки тому

    Hello sir can you please take my mock interview ...

    • @cleverstudies
      @cleverstudies  2 роки тому

      Pls send your resume to shareit2904@gmail.com

  • @maheshgore6265
    @maheshgore6265 3 роки тому

    How many years of experience the student has??