Introduction to Databricks Delta Live Tables

Поділитися
Вставка
  • Опубліковано 27 чер 2023
  • Delta Live Tables is a new framework available in Databricks that aims to accelerate building data pipelines by providing out of the box scheduling, dependency resolution, data validation and logging.
    We'll cover the basics, and then get into the demo's to show how we can:
    - Setup a notebook to hold our code and queries
    - Ingest quickly and easily into bronze tables using Auto Loader
    - Create views and tables on top of the ingested data using SQL and/or python to build our silver and gold layers
    - Create a pipeline to run the notebook
    - See how we can run the pipeline as either a batch job, or as a continuous job for low latency updates
    - Use APPLY CHANGES INTO to upsert changed data into a live table
    - Apply data validation rules to our live table definition queries, and get detailed logging info on how many records caused problems on each execution.
    By the end of the session you should have a good view of whether this can help you build our your next data project faster, and make it more reliable.
    Speaker: Niall Langley SQLbits.com/speakers/Niall_La...
    Speaker Blog: www.sqlsmarts.com/
    Speaker BIO: Niall has been building data solutions on the Microsoft platform for 12 years. In the past few years Niall has been focused on helping clients with data engineering in Azure.
    Niall is active in the data community, helping run the Bristol user group.
    SQLbits.com/Sessions/Introduct...
    Tags: Azure,Python,Spark,Developing,Data Bricks,Big Data & Data Engineering

КОМЕНТАРІ • 8

  • @manasr3969
    @manasr3969 4 місяці тому

    excellent delivery and a useful example. Thanks

  • @arunlouie
    @arunlouie 10 місяців тому +1

    very clear top to down explanation, Thanks!!

    • @SQLBits
      @SQLBits  9 місяців тому

      Glad it was helpful!

    • @jw715
      @jw715 3 місяці тому

      @@SQLBits Where are the notebooks used in the demo?

  • @srinivascruise1
    @srinivascruise1 10 місяців тому +1

    Its wonderful explanation .Got how things works .Can I get notebook code to try please .Thanks for the info .

  • @guddu11000
    @guddu11000 4 місяці тому +1

    where is the notebook

  • @mateen161
    @mateen161 8 місяців тому

    Would it be possible to create unmanaged tables with a location in datalake using DLT pipelines ?

  • @jw715
    @jw715 3 місяці тому

    I am quite new to Databricks, I have watched many videos to explain the DLT, this one is definitely the best one which anwsered the questions in my mind.
    @SQLBits: where could we find the demo notebooks?