Create your FIRST Lakehouse in Microsoft Fabric

Поділитися
Вставка
  • Опубліковано 10 гру 2024

КОМЕНТАРІ • 29

  • @BardstaleGoldens
    @BardstaleGoldens Рік тому +6

    So…this may be a stupid question, but will we be able to create a shortcut to an azure sql table instead of having to pipeline copy the data and keep it updated?

  • @desert_edition_na6
    @desert_edition_na6 10 місяців тому

    This dude is awesome. Data by day, BJJ by night.

  • @prasadparab5638
    @prasadparab5638 Рік тому +1

    Awesome... Thanks for explaining in details 👏🙏

    • @GuyInACube
      @GuyInACube  Рік тому

      Most welcome! Thanks for watching 👊

  • @rafalg8575
    @rafalg8575 Рік тому

    Beatiful. I am waiting when otgnization will decide to implement MS Fabric.

  • @faraazali9589
    @faraazali9589 Рік тому +1

    Thank you for this detailed tutorial! Keep up the great work. Quick questions, what video editing tool do you all use to overlay the blue boxes and commentary?

  • @bladerunnerisback
    @bladerunnerisback Рік тому +2

    Thanks for sharing! Is there a performance impact when using shortcuts? My understanding is data imported into the lakehouse as delta tables but with shortcuts the data remains in it's original format. Maybe a video about it?

    • @robertjordan114
      @robertjordan114 Рік тому

      I believe it uses the same in-memory processing under the covers which means the data is pulled into memory and processed using the data in place as a source for streaming the data during processing, but some verification would be good.

  • @ianfahlman-morgan2081
    @ianfahlman-morgan2081 Рік тому +2

    Love this! So inspiring and clear on how this new element works!
    Also, how can I get my hands on a "guy in a cube" sticker for my laptop? ❤

  • @SureshGuddanti
    @SureshGuddanti 10 місяців тому

    Thank you for this, I have a question though when we use Shortcut say to Amazon S3, and start using that data in Fabric for analysis, is that data not leaving s3 at all even to process In memory on the fabric end? I was comparing this to connecting an external HDD to a PC and accessing data , I was under impression that data gets in to memory of PC , if this is the same concept are we looking at egress charges ? Finally how will shortcuts work with high volume of data?

  • @ericduquesne922
    @ericduquesne922 7 місяців тому

    Lakehouse Links are an interesting feature to use, it looks like a database link for Cloud endpoints.

  • @andikadwinugraha2601
    @andikadwinugraha2601 9 місяців тому

    Q: I'm new in this field, and I'm curious can I use this tool to store historical data as part of migrating data and build my own data storage? Or if there are any tools that suitable instead of using SharePoint Folder? Thanks to GIAC for sharing great video!

  • @Seveq
    @Seveq 7 місяців тому

    How about changing schemas in my source data, like, when I make a shortcut to some data somewhere in the OneLake and have a sophisticated setup of joins and merges, and then one of the data owners decides to change the schema of their data. Will this immediately break my setup? How can we prevent this from happening? How does Fabric help to mitigate this?

  • @anuragbawankar685
    @anuragbawankar685 Рік тому

    Thank You !

  • @alessandrofrancisco9149
    @alessandrofrancisco9149 Рік тому

    Hi Guys, whe i create a pipeline, is possibel to "copy" using an data gateway like i do in powerbi?

  • @HarpreetSaini-y7c
    @HarpreetSaini-y7c Рік тому

    How can we create Linked Server for Fabric lakehouse in ssms.

  • @tanushriaditya
    @tanushriaditya 6 місяців тому

    what is the difference between One LAke and LAke house?

  • @shanetohill6950
    @shanetohill6950 Рік тому

    Reflex? - Looks interesting.

  • @hilldiver
    @hilldiver Рік тому

    One thing that's not so clear to me is how the OneLake "one copy" concept works with the lakehouse medallion architecture. As an example I have a landing lakehouse where I use a dataflow gen2 to load Contoso DB from a local SQL server with no transformations, this is the bronze raw data layer. If I then want to do some transformations, e.g. merging product category and subcategory into product table to help with star schema, how do I do this? Shortcuts don't help as far as I can see, as they're just a virtualisation of what exists. So I'd need to create a pipeline copy job with the transformations and land in a new silver lakehouse, but then I've copied the data, which goes against the onelake concept? Am I missing something fundamental here?

    • @culpritdesign
      @culpritdesign Рік тому

      it's probably up to your design patterns. You could use Azure Data Factor, Spark Notebooks, Lakehouse Stored Procedures, Data Flows.

    • @dasenf1860
      @dasenf1860 Рік тому

      One Copy means that you don‘t have the same table in different formats or data stores. You don‘t have to copy a delta table on the data lake to the data warehouse or import the data of your gold layer to Power Bi.
      So you avoid redundant copies of the same tables.
      Nevertheless you want to build a medaillon architecture on the data lake for flexibility and scaling

  • @NayankaGothane
    @NayankaGothane Рік тому

    How does Synapse differ from Fabric? Why didn't the team use the Synapse platform and add features if there were new ones?

    • @robertjordan114
      @robertjordan114 Рік тому

      I believe Fabric is somewhat of a consultation of the tools to handle structured data warehousing and lakehouse for unstructured data with powerbi and data factory thrown in for good measure.

    • @dasenf1860
      @dasenf1860 Рік тому +1

      one of the main differences is that Synapse is a PaaS offering while Fabric is a SaaS offering aiming to provide a unified analytics solution.
      So it makes sense to rather bring Synapse to the Power BI offering than the other way around.
      When it comes to user experience, there are far more users of Power Bi than of Synapse. So much more people feel at home in the Power Bi environment

  • @crimsonfox2398
    @crimsonfox2398 Рік тому

    Awesome video and first

  • @TedSeeber
    @TedSeeber Рік тому

    Didn't we already have a hundred different ways to store structured and unstructured data? What does lakehouse give us that say, NTFS didn't?

    • @kevinliew3692
      @kevinliew3692 Рік тому +1

      correct me if im wrong, i think the general idea is creating shortcuts to all your sources (onedrive, cloud databases, ADLS, datalake, data warehouse or whatever), so you don't have to perform ETL/API to physically transfer/replicate the data into somewhere else.
      that sounds convenience isnt it, especially when u own the biggest AI bot in the world, they get access to the entire data in the world

    • @dasenf1860
      @dasenf1860 Рік тому +1

      it‘s not so much about storing the data but the ability to work with the data, govern the data and getting insights out of the data

  • @nothin594
    @nothin594 7 місяців тому

    Wow