Cloud Guru
Cloud Guru
  • 25
  • 53 868

Відео

Azure Data Engineering Demo, 19 April 2024 at 07 07 21
Переглядів 774 місяці тому
Azure Data Engineering Demo, covers Azure, Azure Data Factory, Databricks, Synapse, Deltalake, Fabric. If you are interested kindly contact 91 9113070560.
Databricks + Azure Data Engineering Course Demo New batch 30-01-2023 #databricks #deltalake #train
Переглядів 238Рік тому
Live interactive training will start from 30-01-2022. Interested people can contact me on 9113070560 #AzureDataEngineering #AzureETL #ADF #databricks #deltalake #lakehouse #spark #code
Data Bricks Delta Lake Complete Code Execution
Переглядів 6632 роки тому
In this Video I have Executed Complete Code for Data bricks Delta Lake #AzureDataEngineering #AzureETL #ADF #databricks #deltalake #lakehouse #spark #code 1. Creating Landing Zone 2. Creation of Bronze DB and Tables 3. Creation of Silver DB and Tables 4. Creation of Gold Db and Tables 5. Bad Data records handling 6. Metadata Tables and Logging All Azure Data Factory and Synapse Videos link : ww...
Real Time Project Execution using ADF and Data bricks
Переглядів 1,5 тис.2 роки тому
In this Video I have spoken about Real time project exeuction using ADF and Databricks. 1. Creating Landing Zone 2. Creation of Bronze DB and Tables 3. Creation of Silver DB and Tables 4. Creation of Gold Db and Tables 5. Bad Data records handling 6. Metadata Tables and Logging All Azure Data Factory and Synapse Videos link : www.youtube.com/watch?v=5l_XY... Github Link: github.com/vinodreddy76...
End to end Delta Lake house Implementation using Live example
Переглядів 7452 роки тому
In this Video I have spoken about End to End Delta Lake House Implementation using Live Example. 1. Creating Landing Zone 2. Creation of Bronze DB and Tables 3. Creation of Silver DB and Tables 4. Creation of Gold Db and Tables All Azure Data Factory and Synapse Videos link : www.youtube.com/watch?v=5l_XY... Github Link: github.com/vinodreddy76/First... Github DataPipelines: github.com/vinodred...
Big Data File formats Parquet Orc and Avro
Переглядів 6202 роки тому
In this Video I have spoken about Big data File formats Parquet : Columnar storage, Suitable with Spark Orc : Columnar storage, Suitable with Hadoop ORC : Row format storage suitable for Streaming like Kafka and Druid All Azure Data Factory and Synapse Videos link : www.youtube.com/watch?v=5l_XY... Github Link: github.com/vinodreddy76/First... Github DataPipelines: github.com/vinodreddy76/First...
2 DeltaLake Vaccum Optimize and Merge
Переглядів 1262 роки тому
All Azure Data Factory and Synapse Videos link : www.youtube.com/watch?v=5l_XY... Github Link: github.com/vinodreddy76/First... Github DataPipelines: github.com/vinodreddy76/First... Github DataFlow: github.com/vinodreddy76/First... ph: 9113070560
1 Delta lake Introduction
Переглядів 2352 роки тому
All Azure Data Factory and Synapse Videos link : www.youtube.com/watch?v=5l_XY... Github Link: github.com/vinodreddy76/First... Github DataPipelines: github.com/vinodreddy76/First... Github DataFlow: github.com/vinodreddy76/First... ph: 9113070560
1 Delta lake and Understanding _delta_log.json
Переглядів 2112 роки тому
All Azure Data Factory and Synapse Videos link : www.youtube.com/watch?v=5l_XY... Github Link: github.com/vinodreddy76/First... Github DataPipelines: github.com/vinodreddy76/First... Github DataFlow: github.com/vinodreddy76/First... ph: 9113070560
1. Synapse Delta Lake house Architecture using Notebooks, Deltalake and Synapse Ondemand Pool
Переглядів 1,6 тис.2 роки тому
All Azure Data Factory and Synapse Videos link : ua-cam.com/video/5l_XYPOF-VI/v-deo.html Github Link: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataPipelines: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataFlow: github.com/vinodreddy76/FirstProject/tree/main/adf ph: 9113070560
12 ADF Filter Aggregate and Conditional split, Transformations #AzureDataEngineering #AzureETL #ADF
Переглядів 6582 роки тому
All Azure Data Factory Videos link : ua-cam.com/video/5l_XYPOF-VI/v-deo.html Github Link: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataPipelines: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataFlow: github.com/vinodreddy76/FirstProject/tree/main/adf
8.2 Incremental data load in Azure Data Factory #AzureDataEngineering #AzureETL #ADF
Переглядів 35 тис.2 роки тому
All Azure Data Factory Videos link : ua-cam.com/video/5l_XYPOF-VI/v-deo.html Github Link: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataPipelines: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataFlow: github.com/vinodreddy76/FirstProject/tree/main/adf
8.1 Incremental data load in Azure Data Factory #AzureDataEngineering #AzureETL #ADF
Переглядів 4,6 тис.2 роки тому
All Azure Data Factory Videos link : ua-cam.com/video/5l_XYPOF-VI/v-deo.html Github Link: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataPipelines: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataFlow: github.com/vinodreddy76/FirstProject/tree/main/adf
1.2 Introduction to azure Data Factory, ADF VS SSIS #AzureDataEngineering #AzureETL #ADF
Переглядів 4922 роки тому
All Azure Data Factory Videos link : ua-cam.com/video/5l_XYPOF-VI/v-deo.html Github Link: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataPipelines: github.com/vinodreddy76/FirstProject/tree/main/adf Github DataFlow: github.com/vinodreddy76/FirstProject/tree/main/adf
1.1 Introduction to Azure Cloud, OnPremise vs cloud #AzureDataEngineering #AzureETL #ADF
Переглядів 1482 роки тому
1.1 Introduction to Azure Cloud, OnPremise vs cloud #AzureDataEngineering #AzureETL #ADF
13 Azure Data Factory CI/CD Creating and Configuring resources #AzureDataEngineering #AzureETL #ADF
Переглядів 1792 роки тому
13 Azure Data Factory CI/CD Creating and Configuring resources #AzureDataEngineering #AzureETL #ADF
10 Create Multiple Files Using ForEach Activity ADF #AzureDataEngineering #AzureETL #ADF
Переглядів 2252 роки тому
10 Create Multiple Files Using ForEach Activity ADF #AzureDataEngineering #AzureETL #ADF
9 Create Multiple Files Using Copy For Azure Tables in ADF #AzureDataEngineering #AzureETL #ADF
Переглядів 1562 роки тому
9 Create Multiple Files Using Copy For Azure Tables in ADF #AzureDataEngineering #AzureETL #ADF
7 Incrementally Extract Files in ADF #AzureDataEngineering #DatabricksETL #AzureETL #ADF
Переглядів 2602 роки тому
7 Incrementally Extract Files in ADF #AzureDataEngineering #DatabricksETL #AzureETL #ADF
6 Wait Activity in Azure Data Factory #AzureDataEngineering #DatabricksETL #AzureETL #ADF
Переглядів 2052 роки тому
6 Wait Activity in Azure Data Factory #AzureDataEngineering #DatabricksETL #AzureETL #ADF
5 LookUp And Foreach Activities ADF #AzureDataEngineering #DatabricksETL #AzureETL #ADF
Переглядів 5 тис.2 роки тому
5 LookUp And Foreach Activities ADF #AzureDataEngineering #DatabricksETL #AzureETL #ADF
4 Delete Activity Get Metadata Activity ADF #AzureDataEngineering #DatabricksETL #AzureETL #ADF
Переглядів 2052 роки тому
4 Delete Activity Get Metadata Activity ADF #AzureDataEngineering #DatabricksETL #AzureETL #ADF
3.Create Data Sets and Copy Activity in ADF #AzureDataEngineering #DatabricksETL #AzureETL #ADF
Переглядів 1332 роки тому
3.Create Data Sets and Copy Activity in ADF #AzureDataEngineering #DatabricksETL #AzureETL #ADF
2. Create Linked Services in AzureDataFactory #AzureDataEngineering #DatabricksETL #AzureETL #ADF
Переглядів 1892 роки тому
2. Create Linked Services in AzureDataFactory #AzureDataEngineering #DatabricksETL #AzureETL #ADF

КОМЕНТАРІ

  • @davidjasonmusic
    @davidjasonmusic 21 день тому

    Hello i need help - my data sources are APIs from BC...so what I am supposed to do with those datasets?....copy them into blob in order to filter by last modified date back in ADF? That makes no sense what about copying IN to the blob? I just have to copy the whole fact table over everytime? That cant be right. All these methods of incremental copy seem to start with a csv from blob storage. But how do I even get to that point?

  • @maganzo
    @maganzo 5 місяців тому

    so what's the main take-away?

  • @SantoshKumar-yr2md
    @SantoshKumar-yr2md 7 місяців тому

    you should explain from very begining like when you created table in your datasource, nice

  • @Ramana6783
    @Ramana6783 7 місяців тому

    Could you please add a full load activity

  • @shanthababu
    @shanthababu 9 місяців тому

    Excellent approach, Please explain adequately to learners.

  • @chandandacchufan3242
    @chandandacchufan3242 11 місяців тому

    should explain how you are updating the table tbl_control

  • @Reaper7mk
    @Reaper7mk 11 місяців тому

    Thank you. This was helpful.

  • @ravinderreddy9367
    @ravinderreddy9367 Рік тому

    Hi sir.. where can i get this code? github link is not working

  • @surenderraja1304
    @surenderraja1304 Рік тому

    This is one of the best videos to understand the end to end data pipeline in azure. Perfect

  • @surenderraja1304
    @surenderraja1304 Рік тому

    What is difference between MANAGED delta table and EXTERNAL Delta table in azure databricks? Can we do insert , delete , update in both the types.

  • @surenderraja1304
    @surenderraja1304 Рік тому

    If a delta table is created on top of mounted datalake location , then can we do update , delete and insert to that delta table?

  • @parikshitchavan2211
    @parikshitchavan2211 Рік тому

    Hello thanks for amazing session.....just one query if data is having more than 5k lines will it work ?

  • @krishnachaitanyareddy2781

    7 mins video

  • @krishnachaitanyareddy2781

    Avo and orc both for Hadoop?

  • @krishnachaitanyareddy2781

    Nice explanation can you share these ppt

  • @krishnachaitanyareddy2781

    Nice can you share these presentation

  • @krishnachaitanyareddy2781

    Can you share these note book

  • @teunjacobs5005
    @teunjacobs5005 Рік тому

    Hi Cloud Guru! First of all, thanks for the clear explanation. It really helped me in creating a incremental data load in ADF. Never the less I was wondering if you could make a video of making the tables which should be synced incremental more dynamic. Now I've hard set in the lookup action that it should look for a specifc table. I've more tables which should be synced incremental, so I assume that should start working with an iterate activity like a for each(?). So that in a specific SQL table I've defined which table(s) should be synced incremental and that it does that automatically. Do you have any ideas of tips and trics for things like this? Thanks in advance!

  • @tusharchirame820
    @tusharchirame820 Рік тому

    Could you please share your email id, I have confusion about one of the question was asked me in interview

  • @souranwaris142
    @souranwaris142 Рік тому

    Hello Sir. I have a problem with the incremental load I want to create an incremental pipeline from the Oracle on-premise server to Azure data lake(blob storage) I don't have Azure SQL. I just want to push in blob storage as a CSV file. in my case, I have confusion about where I should create the watermark table and store procedure. someone told me in your case you have to use parquet data. please help me with this I am stuck for many days.

    • @sainathreddynandyal8761
      @sainathreddynandyal8761 Рік тому

      I encountered with same use-case brother. Can you tell me if you found a solution please?

  • @sourav7413
    @sourav7413 Рік тому

    ❤Thanks …it’s really informative. Could you plz share end to end pipeline with code of this project…

  • @harithad1757
    @harithad1757 Рік тому

    Can you let me know which tool will be used to create the diagram u are explaining

  • @Thegameplay2
    @Thegameplay2 Рік тому

    Great and complete tutorial that I have seen in UA-cam. Thank you

    • @cloudguru3178
      @cloudguru3178 Рік тому

      You are welcome, also I am teaching the course Databricks + Azure Data Enginnering now please refer me if possible. 91130 70560

    • @ashutoshpunia9700
      @ashutoshpunia9700 Рік тому

      ​@@cloudguru3178 you are good sir, I was looking out for different staging techniques. You have covered niche concepts which is very important for interview perspective.

  • @muapatrick4174
    @muapatrick4174 Рік тому

    is it possible to add theses incrementals on parquet files on adls?

  • @harrydadson7467
    @harrydadson7467 Рік тому

    Please can someone please assist me on how I'm not able to do a Copy Activity to Stored Procedure on MySQL table in sink setting. Please help. Thanks

  • @AvisekChakraborty-bv3dk
    @AvisekChakraborty-bv3dk Рік тому

    You didnt share the XLS window

  • @camvinh3522
    @camvinh3522 Рік тому

    why do we need control table for the last update time - since we can get it info in destination table

    • @cloudguru3178
      @cloudguru3178 Рік тому

      What if loading to destination tables fails ??

    • @camvinh3522
      @camvinh3522 Рік тому

      @@cloudguru3178 so it means that after we loading success to destination table then we update the control table ( if it fails we will not update the control table ) and base on the control table we can handle issue. Is it right ?

    • @camvinh3522
      @camvinh3522 Рік тому

      @@cloudguru3178 thanks you for your support

    • @cloudguru3178
      @cloudguru3178 Рік тому

      @@camvinh3522 We will update the control table if it is failure or success. But to get the maximum date for table you need to consider only last successful load.

  • @mayurgupta8357
    @mayurgupta8357 2 роки тому

    Please share the link of playlist

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      ua-cam.com/channels/zasZoFzBNaFQRUZQtdRKLw.html

  • @ssbeats677
    @ssbeats677 2 роки тому

    Hi sir, in my source path i have a files i have created a shedule trigger to run every 1 hour. My issue is from source to sink the files are getting triggered duplucates files Eg : last hour i have 10 files trigger nxt in my source path i recived 5more files. When it trigger the files of last 10+5 files are getting to the sink path

    • @ssbeats677
      @ssbeats677 2 роки тому

      I need only latest and modified files

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      Kindly move the file to another location once it is loaded so that there will be no duplicates

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      If you would like to connect then call me on 9113070560

  • @alexcarter-jones3251
    @alexcarter-jones3251 2 роки тому

    Great video! Have you ever set up an incremental load between an Oracle source table and an SQL sink table before? I am currently trying to do this but can only get my delta load to work when copying from SQL to SQL as you did in this video, any guidance would be greatly appreciated :)

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      Hi Alex, please try this @Concat('select part of query ', 'WHERE si.schedule_date BETWEEN ' , formatDateTime({variables('LastDate')}, 'yyyyMMdd') , ' AND SYSDATE-1')

    • @alexcarter-jones3251
      @alexcarter-jones3251 2 роки тому

      Hi @@cloudguru3178, thanks for replying and apologise for the delay in my reply. I managed to write my own working solution using a query script prior to a simple copy data pipeline then using an upsert on the ID column. Query script is as follows 'select * from OracleSourceTable where LAST_UPDATE_TSTMP > (CURRENT_TIMESTAMP - 0.04)' looks very similar to your solution. I think the main issue I was having was trying to use templates that are only set up for Azure SQL DB to Azure SQL DB transfer.

  • @vipuljain-ok9qy
    @vipuljain-ok9qy 2 роки тому

    I think it's for more easy with foreach loop..what's your views ?

  • @swaminoonsavath5363
    @swaminoonsavath5363 2 роки тому

    Expression of type: 'String' does not match the field: 'additionalColumns' I am getting this error .. MY source is Servicenow and Destination is AZURE SQL DB .. Please help me on this

    • @superyan4100
      @superyan4100 Рік тому

      Remove the 'first row only' checkbox.

  • @saipraneeth1660
    @saipraneeth1660 2 роки тому

    thank you for the video sir.. yesterday i got same question in the interview

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      Hope my explanation helped you Praneeth

  • @terryliu3635
    @terryliu3635 2 роки тому

    This is good! Thanks for sharing. One suggestion is that it will be much more helpful to add how you deal with “update” and “delete” or “merge” under this architecture. That is the “soul” of leveraging delta lake!!

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      Hi Terry, I will try to add it so that it will be helpful for people.

    • @terryliu3635
      @terryliu3635 2 роки тому

      Awesome! Thank you.

  • @manojprabhakar6814
    @manojprabhakar6814 2 роки тому

    What if the copy activity fails for one of the iteration inside foreach? Will the foreach fails automatically?

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      It will go to next file

    • @manojprabhakar6814
      @manojprabhakar6814 2 роки тому

      @@cloudguru3178 but I have requirement to perform second iteration only after successful completion of first iteration. And if an iteration fails ,my foreach should also fail. Then how can I achieve it? Can you please give some inputs?

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      @@manojprabhakar6814 docs.microsoft.com/en-us/answers/questions/358743/i-have-a-requirement-wherein-a-foreach-activity-mu.html

  • @balajiawari9740
    @balajiawari9740 2 роки тому

    Explained very nice

  • @vasistasairam5483
    @vasistasairam5483 2 роки тому

    The explanation is very good, but while trying to implement the same, getting errors, especially with formulas. Kindly make videos little more detailed.

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      Let me know what error you are facing and i shall help u

  • @surenderraja1304
    @surenderraja1304 2 роки тому

    Hi Cloud Guru, In This video you explained about what are the resources to be added/configured in DEV,TEST,PROD. Can you send me video link for deployment in DEV and migrate to test . I dont see that deployment videos in that playlist

  • @surenderraja1304
    @surenderraja1304 2 роки тому

    This video has content upto Full load , Please post the video for Incremental load by using that load_control table

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      Check 8.2 video for incremental load

    • @cloudguru3178
      @cloudguru3178 2 роки тому

      Here is the link ua-cam.com/video/cZ3RW8KUz2E/v-deo.html

    • @surenderraja1304
      @surenderraja1304 2 роки тому

      @@cloudguru3178 Thanks , It is nice and helpful

  • @trendy_techer2202
    @trendy_techer2202 2 роки тому

    Thank you so much sir for the valuable info

  • @suniguha
    @suniguha 2 роки тому

    Very useful content. Thank you.