#60.Azure Data Factory - Incremental Data Load using Lookup\Conditional Split

Поділитися
Вставка
  • Опубліковано 21 жов 2024

КОМЕНТАРІ • 56

  • @shreeyashransubhe2537
    @shreeyashransubhe2537 2 роки тому

    You explain like school teacher. I really feel as if my class teacher is teaching me the concepts. Very thankful for your efforts Mam!!

  • @swativish
    @swativish 3 роки тому

    Your videos are always so exceptional and relevant to real life tasks required at work. Thanks. Keep up the good work

    • @AllAboutBI
      @AllAboutBI  3 роки тому

      Thanks so much for ur time to comment and appreciate..much needed motivation!!

  • @abhishekkumar-es1wl
    @abhishekkumar-es1wl 3 роки тому

    Exceptional & simple. Many thanks....Happy teacher's day...

    • @AllAboutBI
      @AllAboutBI  3 роки тому

      Thanks so much Abhishek 🙏

  • @manoharraju
    @manoharraju 3 роки тому +2

    Does this solution is applicable for source DB with millions of records? The reason for the ask is, how does this hash comparison will work in the case of millions of records. Will it have performance issues?

  • @manojsrivastava7055
    @manojsrivastava7055 3 роки тому

    Very good explanation and nice scenario👍

  • @varunvengli1762
    @varunvengli1762 7 місяців тому

    Can you please tell the query you wrote to create hash column.
    When i tried i got same values for all the rows in hash column

  • @sethuramalingam3i
    @sethuramalingam3i Рік тому

    Hi Madam, Video is good. I have few doubts.
    1) But want to know why not used watermark table? It is not having full history load which is SCD-2. As per your approach, it may affect performance which we comparing all recs from target.
    2) what are all the activities are you used. difficult to find in video because of changed names of those activities. could you list me the activities along with this?
    Thanks madm.

  • @robinson03584
    @robinson03584 Рік тому

    look up has a limit for 5k rows right? how to deal with the input has 1Million rows?

  • @nawazmohammad5212
    @nawazmohammad5212 3 роки тому

    Really very helpful. Thanks for creating this video

    • @AllAboutBI
      @AllAboutBI  3 роки тому

      Thanks a lot for your feedback

  • @nikhileshyoutube4924
    @nikhileshyoutube4924 Рік тому

    Very good explanation madam

  • @palivelaanjaneyagupta7395
    @palivelaanjaneyagupta7395 Рік тому

    Hi Mam, We don't have date column in the source side.Can we also implement the same process?

  • @ADFTrainer
    @ADFTrainer 2 роки тому

    Please let me know why lookup needed, any how we have conditionalsplit right?

  • @sunilpatil4393
    @sunilpatil4393 4 роки тому +1

    Thanks for giving such skills

  • @prashanthn2681
    @prashanthn2681 3 роки тому

    Hi madum ,how we can convert the different date formats into one date format.
    For example 'yy/mm/dd' (or) 'dd/mm/yyyy' into 'yyyy-MM-dd' date format.
    We can implement this in azure dataflow

  • @hackifysecretsau
    @hackifysecretsau 2 роки тому

    Hi Mam
    Please respond urgent query.
    I have time in CSV file so how to convert time in Data Factory into Time. As I don't have date. I need to convert CSV time field into time format.

  • @sancharighosh8204
    @sancharighosh8204 3 роки тому

    ma'am what is the difference between switch activity and if condition in ADF. Please reply

  • @pawanreddie2162
    @pawanreddie2162 3 роки тому

    Isn't it the same as alter-row(upsert)? We can achieve the same right?

  • @kapilganweer9991
    @kapilganweer9991 3 роки тому

    Hello Mam, I need some suggestions. I need to make incremental data extraction pipeline in ADF. ServiceNow is my source and I am extracting data in json format and storing into blob storage. I need to extract only the latest updated or inserted data from ServiceNow.

  • @oriono9077
    @oriono9077 4 роки тому

    Thanks for knowledge sharing

  • @SaurabRao
    @SaurabRao 3 роки тому

    Let me know if my understanding is incorrect, but isn't this similar to the upsert operation and cant this be achieved using the alter row-->upsert option as before? Also this looks structurally same as SCD component's output in ssis!

  • @mohanvp47
    @mohanvp47 2 роки тому

    Hi, I need to copy the data from 5 tables in Azure data lake to 1 table in Cosmos DB. we need a particular field based on the relationships. Thanks in advance

  • @jgowrri
    @jgowrri 3 роки тому

    Thanks a lot for your help .

  • @naveenkumar-ij5mv
    @naveenkumar-ij5mv 2 роки тому

    Pls make this incremental load as dynamic ,,it wil help us a lot...

  • @mahanteshc9374
    @mahanteshc9374 2 роки тому

    good info without bla bla

  • @souranwaris142
    @souranwaris142 Рік тому

    Hello Ma'am,
    I have a problem with the incremental load I want to create an incremental pipeline from the Oracle on-premise server to Azure data lake(blob storage) I don't have Azure SQL. I just want to push in blob storage as a CSV file. in my case, I have confusion about where I should create the watermark table. someone told me in your case you have to use parquet data. please help me with this I am stuck for many days.

    • @AllAboutBI
      @AllAboutBI  Рік тому

      Hmm. Since your source is on Prem we can't use data flow otherwise we can implement the logic as shown in ua-cam.com/video/evqQRwsF_Ps/v-deo.html

  • @sumanyarlagadda6271
    @sumanyarlagadda6271 2 роки тому

    Thanks for the sharing your knowledge. Could you do a video on How to delete target sql table rows which are not exist in source file. Tried through doesn't exist but giving a weird results. If in source 5 records missing exist in target sql table doesn't exist showing 30 records not sure why it is?. Thanks in advance

  • @sunilpatil4393
    @sunilpatil4393 4 роки тому

    Very nice...

  • @shashank_1180
    @shashank_1180 3 роки тому

    Thanks.. Found very helpful 😊

  • @raghavendarsaikumar
    @raghavendarsaikumar 3 роки тому

    Mam i have a doubt in the fault tolerance part in adf. I have configured adls gen2 storage account for writing the log where i'm getting this error.
    "Azure Blob connect cant support this feature for hierarchical namespace enabled storage accounts, please use azure data lake gen2 linked service for this storage account instead".
    The thing is i'm already using the azure data lake store gen2 , but still receiving the error. Can you help in fixing this.

    • @AllAboutBI
      @AllAboutBI  3 роки тому

      It's fishy. Can you pls share the settings where you write the log along with error to funlearn0007@gmail.com

  • @vivekkarumudi
    @vivekkarumudi Рік тому

    that was clearly explained... however it would have been even useful if you could have actually dragged the components and set up the whole thing manually.

  • @jeffrypaulson
    @jeffrypaulson 2 роки тому

    How can we identify if a record is deleted in source ? how do we capture that in target ?,

  • @tjsr4867
    @tjsr4867 4 роки тому

    Thanks. Really helpful

  • @shivanidubey1616
    @shivanidubey1616 3 роки тому

    Thanks for this video ma'am

  • @palmgroves2318
    @palmgroves2318 4 роки тому

    I need urgent solution, can you please soon...
    Hello mam, how to load data to database whose connection to Sink is not available, for example mysql or postgreSQL, Azure has option to source but it does not support Sink. In that case how to load to that dB?

    • @AllAboutBI
      @AllAboutBI  4 роки тому

      Is there no connector at all or you don't have an option to load directly.

    • @vickyvinay23
      @vickyvinay23 3 роки тому

      Export the data to a CSV and then consume them in that DB. Did this make sense?

  • @rajeevsharma2664
    @rajeevsharma2664 3 роки тому

    As you are simply overriding i.e. not SCD type2/3, there is no need to have the hash key. You simple could have used the PK of the target table and use lookup whether that PK(unique value) is already present or not - IMO

    • @AllAboutBI
      @AllAboutBI  3 роки тому

      You are right, I just wanted to explain the hashing mechanism as one of my subscribers asked for the steps. And, thanks for your comment👍

    • @rajeevsharma2664
      @rajeevsharma2664 3 роки тому +1

      @@AllAboutBI No problem - it was my pleasure. Rather I want to validate whether I'm missing anything or not :)

    • @vickyvinay23
      @vickyvinay23 3 роки тому

      @@rajeevsharma2664 For updated columns, if we do not have Hash Key, and if there are over 20 + columns, we have to compare all these individually right? So won't hashing help in those situations?

  • @vennastechworld7675
    @vennastechworld7675 3 роки тому

    Notequal operator accepts two expressions but you mention (hashColumn, Hash), what it means ?
    even more, you didn't declare or create those columns.

    • @AllAboutBI
      @AllAboutBI  3 роки тому

      Hashcolumn comes from my table.
      Hash comes from the derived column transformation for all the incoming rows.
      Not equal operator compares the above two