32 Spark Memory Management | Why OOM Errors in Spark | Spark Unified Memory | Storage/Execution Mem

Поділитися
Вставка
  • Опубліковано 14 січ 2025

КОМЕНТАРІ • 13

  • @PraveenKumarBN
    @PraveenKumarBN День тому

    Amazing. You deserve to get subscribed... Keep it coming! 😀

    • @easewithdata
      @easewithdata  День тому

      Thank you! Don't forget to share this with your network over LinkedIn♻️

  • @bharathansbs
    @bharathansbs День тому +2

    Hi, I am able to understand the content except this part. Based on the example, a single line of 11 mb file when exploded we get multiple rows of 11 mb so the size is getting huge. But still we have around 60 mb of execution memory (around 90mb minus 30 mb of cached memory). So even the data size is getting bigger it can spill to disk right ? why we are receiving OOM .
    Can you please explain this part.

    • @easewithdata
      @easewithdata  День тому +1

      Hello,
      To keep it simple, runtime computations are stored in memory. Since runtime object gets multiplied because of explode it will not be able to fit in memory, running into OOM. In other case, when we try to read a bigger partition, spark knows that it will not fit in memory thus it spills it to disk before running computations on top of it. But once a data is brought in memory and then it explodes because of computation, that can cause issue.
      I tried to make this simple, as this is not so simple to understand in first time.
      Don't forget to Like and shar with your network over LinkedIn 💓

    • @bharathansbs
      @bharathansbs День тому

      @@easewithdata Thanks for your reply.

  • @rakeshpanigrahi577
    @rakeshpanigrahi577 День тому

    Bhai shab, Awsm video ❤

  • @RiskyAsh
    @RiskyAsh 2 дні тому

    Subscribed🎉

    • @easewithdata
      @easewithdata  День тому +1

      Don't forget to repost this with your friends as well on LinkedIn ♻️

  • @funnyvideo8677
    @funnyvideo8677 День тому

  • @moyeenshaikh4378
    @moyeenshaikh4378 День тому

    Bhai Databricks series complete ho gaya kya

    • @easewithdata
      @easewithdata  День тому

      Nhi that is in progress. Both Spark and Databricks will run in parallel.

    • @moyeenshaikh4378
      @moyeenshaikh4378 День тому

      @easewithdata still how many videos left?

    • @easewithdata
      @easewithdata  День тому

      @@moyeenshaikh4378 for databricks? Around 10