AWS Tutorials - AWS Glue Job Optimization Part-1

Поділитися
Вставка
  • Опубліковано 15 вер 2024

КОМЕНТАРІ • 29

  • @worldupdates.
    @worldupdates. Рік тому +2

    Keep it up sir.

  • @VishalSharma-hv6ks
    @VishalSharma-hv6ks 2 роки тому +2

    Very Informative..

  • @sanooosai
    @sanooosai 6 місяців тому +1

    thank you sir

  • @Abubakar91718
    @Abubakar91718 2 роки тому +3

    Appreciates

  • @Books_Stories_Poetry
    @Books_Stories_Poetry Рік тому +1

    spark follws lazy evaluation and when it will prepare plan to fetch data it will take it into consideration . push down predicate does not make any sense

  • @arunr2265
    @arunr2265 2 роки тому +2

    Nice video. But one question is if we use the filter condition while loading the data, won't spark's catalyst optimizer pushes down the filter and reads less rows. is the predicate syntax used only on glue, can apache spark too use it

    • @AWSTutorialsOnline
      @AWSTutorialsOnline  2 роки тому

      It is Apache Spark level thing. Purpose is to scan and load only that much data what is required. Please check this link - jaceklaskowski.gitbooks.io/mastering-spark-sql/content/spark-sql-Optimizer-PushDownPredicate.html

    • @cheluvesha
      @cheluvesha Рік тому

      Here are a few reasons why you may want to explicitly specify pushdown predicates in AWS Glue:
      Improved performance: By explicitly specifying pushdown predicates in AWS Glue, you can control which filtering conditions are pushed down to the storage layer. This can result in improved query performance, as the storage layer will only return the data that is actually needed for the query.
      Fine-grained control: By explicitly specifying pushdown predicates in AWS Glue, you can have fine-grained control over the filtering conditions that are pushed down to the storage layer. This can be useful when you want to optimize a specific query or set of queries.
      Troubleshooting: By explicitly specifying pushdown predicates in AWS Glue, you can make it easier to troubleshoot performance issues. For example, if a query is not performing as expected, you can check whether the pushdown predicates are properly specified and being used effectively.
      In conclusion, while Apache Spark's Catalyst Optimizer can automatically push down filtering conditions to the storage layer, explicitly specifying pushdown predicates in AWS Glue can provide additional benefits in terms of performance, control, and troubleshooting.

  • @abeeya13
    @abeeya13 2 місяці тому

    Can this be used to read few columns from s3 bucket?

  • @mesaquestbsb
    @mesaquestbsb Рік тому +1

    Hey, very nice video! I have some questions. Does this approach applies to delta tables? I work with a table with more than 150 millions of lines that I need weekly delete almost 50 million of lines and load the new data? What is your suggestion for deletion considering I use partitioning?

    • @AWSTutorialsOnline
      @AWSTutorialsOnline  Рік тому

      Pushdown is designed for partition data. You should try Delta Lake framework in AWS Glue for your purpose. Details - docs.aws.amazon.com/glue/latest/dg/aws-glue-programming-etl-format-delta-lake.html

  •  Рік тому +1

    Hi , one question by other point , is posible create a table in glue catalog from a glue job ? with a s3 source target data ?, one condition is what table exist in the glue catalog , but exist other way for create it dynamically ?

    • @AWSTutorialsOnline
      @AWSTutorialsOnline  Рік тому +1

      Technically possible. You can use python boto3 aws sdk in Glue job to check existence of a table. if it does not exist, you simply start the crawler to create table catalog.

  • @mohdshoeb5101
    @mohdshoeb5101 2 роки тому +1

    Can you please tell me how to optimize reading a large amount of files every hour.because I have 10- 12 glue jobs and every job is reading 10 tables in every run.

    • @AWSTutorialsOnline
      @AWSTutorialsOnline  2 роки тому

      Please check part-5 where I explained batching. It might help. Please let me know.

  • @abhijeetpatil-k5y
    @abhijeetpatil-k5y Рік тому +1

    sir which iam role give you for aws glue also for jupyter notebook

    • @AWSTutorialsOnline
      @AWSTutorialsOnline  Рік тому

      This link might help - docs.aws.amazon.com/glue/latest/dg/create-an-iam-role-notebook.html

  • @abhishek822
    @abhishek822 Рік тому +1

    Does push down predicate work for JDBC source ?

    • @AWSTutorialsOnline
      @AWSTutorialsOnline  Рік тому

      No, it is designed for S3 bucket based partition or HIVE metadata only

  • @ericguymon5418
    @ericguymon5418 2 роки тому +1

    What method in either LakeFormation or Glue would recommend for daily ingestion of a database with say 15 tables, 5 of which could add 1 million rows daily?

    • @AWSTutorialsOnline
      @AWSTutorialsOnline  2 роки тому

      You can use AWS Glue job for it. Please check this link - aws-dojo.com/workshoplists/workshoplist33/

  • @PakIslam2012
    @PakIslam2012 2 роки тому

    Isnt glueContext.create_dynamic_frame also lazily evaluated like Spark DataFrame? Which would mean it should not end up loading the entire 410 records in the initial code?

    • @AWSTutorialsOnline
      @AWSTutorialsOnline  2 роки тому

      You are right. It is lazy load. It is more about in memory processing size. do you want to process 410 record in memory to filter down to 160 records or do you want to processing 160 records right from the beginning.

    • @Ady_Sr
      @Ady_Sr 9 місяців тому

      your are right, I think it would never load up 410 records at first place, due to lazy evaluation it will only pick filtered records. Not all , even if the filter is in 2nd statement.

  • @merryphillips3915
    @merryphillips3915 2 роки тому

    ☝️ Promo-SM!!