Advanced Materializations in data build tool (dbt) | Hands-on demo

Поділитися
Вставка
  • Опубліковано 29 лис 2024

КОМЕНТАРІ • 15

  • @VishnuramNadarajan
    @VishnuramNadarajan Рік тому +1

    Hi Adam,
    based on incremental query, update was not happening, for dates lesser than
    max date

  • @yc859
    @yc859 5 місяців тому

    Hi Adam,
    How do you handle schema change in the source with the incremental loading?
    Thanks in advance. Your videos are awesome.
    Kr,
    Yahya

    • @mastering_snowflake
      @mastering_snowflake  4 місяці тому

      By default, dbt WILL NOT populate new columns without a --full-refresh. The new column will be included in the first query against the source model, but not in the second query updating the destination model.
      dbt provides an on_schema_change config.
      append_new_columns: Add any new columns; populate with most recent subset.
      sync_all_columns: Adds new columns, drops removed columns. Populate with most recent subset.
      To populate new columns with historical data, a -full-refresh is required.

    • @mastering_snowflake
      @mastering_snowflake  4 місяці тому

      Check this great article out for more details…
      medium.com/@aaronbannin/dbt-incremental-cookbook-d70331ef879a

  • @bhratkapoor5230
    @bhratkapoor5230 2 роки тому +2

    Hello Adam,
    Can we run 2 model script separately and insert the data into one single table?
    Thanks,
    Bhrat Kapoor

    • @mastering_snowflake
      @mastering_snowflake  Рік тому

      Hello Bhrat, good question. In this case you would have 3 dbt models. The first 2 populate your intermediate table and the 3rd model takes the intermediate table data to create you target table.
      Alternatively you could have one model joining tables 1 & 2 together which output your target result set.
      As ever, it depends on your specific requirements, such as how many data transformations you need to apply and if other downstream objects could also leverage the intermediate table using the 3 model approach described above.

  • @Joeyisaneard
    @Joeyisaneard 2 роки тому +3

    He’s my dad

  • @ucmanhao3285
    @ucmanhao3285 Рік тому

    hello Adam, I have a question. Is there any difference between materialize 'commit_table' and 'table'?

    • @mastering_snowflake
      @mastering_snowflake  Рік тому

      Hello, can you provide more context as I don’t understand your question?

  • @avisrivastava8341
    @avisrivastava8341 Рік тому +1

    Hello adam, i need to learn the dbt transformation between mongodb and redshift.. Can you explain it on text?

  • @isaacodeh
    @isaacodeh Рік тому

    Hi Adam, is there a thing called stream materialization in dbt?

    • @mastering_snowflake
      @mastering_snowflake  Рік тому

      Hi, the only ones I’m aware of are table, view, incremental and ephemeral
      docs.getdbt.com/docs/build/materializations

    • @isaacodeh
      @isaacodeh Рік тому

      @@mastering_snowflake Thanks

  • @niki2098
    @niki2098 2 роки тому

    How many records can we load data into csv file using seeds

    • @mastering_snowflake
      @mastering_snowflake  2 роки тому +1

      Hi great question, check out this article discourse.getdbt.com/t/dbt-seed-for-large-csv-files/328/3