How to Build a Metadata Driven Data Pipelines with Delta Live Tables

Поділитися
Вставка
  • Опубліковано 24 лип 2023
  • In this session, you will learn how you can use metaprogramming to automate the creation and management of Delta Live Tables pipelines at scale. The goal is to make it easy to use DLT for large-scale migrations, and other use cases that require ingesting and managing hundreds or thousands of tables, using generic code components and configuration-driven pipelines that can be dynamically reused across different projects or datasets.
    Talk by: Mojgan Mazouchi and Ravi Gawai
    Connect with us: Website: databricks.com
    Twitter: / databricks
    LinkedIn: / databricks
    Instagram: / databricksinc
    Facebook: / databricksinc
  • Наука та технологія

КОМЕНТАРІ • 10

  • @brads2041
    @brads2041 8 місяців тому +1

    not clear how this process would handle what happens if your source query, for silver in this context, though might be more relevant to gold, uses something like an aggregate, which dlt streaming doesn't like and you may have to fully materialize a table instead of streaming

  • @garethbayvel8374
    @garethbayvel8374 7 місяців тому +1

    Does this support Unity Catalog?

    • @ganeshchand
      @ganeshchand 5 місяців тому

      yes. the recent release support UC.

  • @rishabhruwatia6201
    @rishabhruwatia6201 11 місяців тому

    Can we have a video for loading multiple tables using single pipeline

    • @rishabhruwatia6201
      @rishabhruwatia6201 11 місяців тому

      I mean something of a for each activity

    • @RaviGawai-db
      @RaviGawai-db 11 місяців тому

      @@rishabhruwatia6201 you can check repo dlt-meta and check dlt-meta-demo or run integration tests

    • @brads2041
      @brads2041 9 місяців тому

      We tried that just recently. Depending on how you approach this it may not work. In our case, we did not always call the DLT with the same tables to be processed. Any table that was processed previously, but not in a next run would be removed from unity (though the parquet files still exist - ie behavior like an external table). This is of course not acceptable, so we switched to meta data driven structured streaming. To put this a different way, if you call the pipeline with table a, then call it again with table b, table a is dropped. You'd have to always execute the pipeline with all tables relevant to the pipeline.

    • @RaviGawai-db
      @RaviGawai-db 9 місяців тому

      @@brads2041 you reload onboarding before each run to add or remove tables from group. So workflow might be: onboarding(can refresh each row addition removal for tables a,b) -> DLT Pipeline

  • @AnjaliH-wo4hm
    @AnjaliH-wo4hm 3 місяці тому +2

    would appreciate if databricks comes with a proper explanation ...both the tutor's explanation aren't clear