Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Don't miss out! 2025 Microsoft Fabric Community Conference, March 31 - April 2, Las Vegas, Nevada. Use code MSCUST for a $150 discount. Prices go up February 11th. Register now.

Reply
optimizer
Advocate II
Advocate II

Dataflows Gen 2 vs Lakehouse for storing data in staging / transform phase

I learned that Dataflows Gen 2 use an internal Lakehouse to store data when you enable staging on a query in the dataflow. You could then re-use that data in another dataflow through an powerplatforms-connector. What would be the benefit of the other route: writing the query output to a Lakehouse and re-using that data in another dataflow through the OneHub/Lakehouse connector? It is more work to store the data in the Lakehouse (for every query you have to manually set the output destination table, checking datatypes etc.) so now I am thinking using the Dataflows Gen 2 internal staging Lakehouse is better, am I right or do I overlook imported implications?

1 ACCEPTED SOLUTION
v-yilong-msft
Community Support
Community Support

Hi @optimizer ,

First of all if you are dealing with large amounts of data or complex data transformations, then Writing to a Lakehouse has more extra control and scalability.

 

If your main goal is to integrate data flows within the Power Platform ecosystem, then Using Internal Lakehouse Staging may be more efficient.

 

So in summary if simplicity and fast integration are your top priorities, Internal Lakehouse Staging may be a better choice. However, if you need more control and scalability, then writing to Lakehouse and using the OneHub/Lakehouse connector may be worth the extra effort.

 

If you want to learn more about this, I think you can read this document: Dataflow Gen2 data destinations and managed settings - Microsoft Fabric | Microsoft Learn

 

 

 

Best Regards

Yilong Zhou

If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.

 

View solution in original post

1 REPLY 1
v-yilong-msft
Community Support
Community Support

Hi @optimizer ,

First of all if you are dealing with large amounts of data or complex data transformations, then Writing to a Lakehouse has more extra control and scalability.

 

If your main goal is to integrate data flows within the Power Platform ecosystem, then Using Internal Lakehouse Staging may be more efficient.

 

So in summary if simplicity and fast integration are your top priorities, Internal Lakehouse Staging may be a better choice. However, if you need more control and scalability, then writing to Lakehouse and using the OneHub/Lakehouse connector may be worth the extra effort.

 

If you want to learn more about this, I think you can read this document: Dataflow Gen2 data destinations and managed settings - Microsoft Fabric | Microsoft Learn

 

 

 

Best Regards

Yilong Zhou

If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.

 

Helpful resources

Announcements
Las Vegas 2025

Join us at the Microsoft Fabric Community Conference

March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount!

Jan NL Carousel

Fabric Community Update - January 2025

Find out what's new and trending in the Fabric community.

Top Solution Authors