Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Join us at the 2025 Microsoft Fabric Community Conference. March 31 - April 2, Las Vegas, Nevada. Use code FABINSIDER for $400 discount. Register now

Reply
optimizer
Advocate II
Advocate II

Dataflows Gen 2 vs Lakehouse for storing data in staging / transform phase

I learned that Dataflows Gen 2 use an internal Lakehouse to store data when you enable staging on a query in the dataflow. You could then re-use that data in another dataflow through an powerplatforms-connector. What would be the benefit of the other route: writing the query output to a Lakehouse and re-using that data in another dataflow through the OneHub/Lakehouse connector? It is more work to store the data in the Lakehouse (for every query you have to manually set the output destination table, checking datatypes etc.) so now I am thinking using the Dataflows Gen 2 internal staging Lakehouse is better, am I right or do I overlook imported implications?

1 ACCEPTED SOLUTION
v-yilong-msft
Community Support
Community Support

Hi @optimizer ,

First of all if you are dealing with large amounts of data or complex data transformations, then Writing to a Lakehouse has more extra control and scalability.

 

If your main goal is to integrate data flows within the Power Platform ecosystem, then Using Internal Lakehouse Staging may be more efficient.

 

So in summary if simplicity and fast integration are your top priorities, Internal Lakehouse Staging may be a better choice. However, if you need more control and scalability, then writing to Lakehouse and using the OneHub/Lakehouse connector may be worth the extra effort.

 

If you want to learn more about this, I think you can read this document: Dataflow Gen2 data destinations and managed settings - Microsoft Fabric | Microsoft Learn

 

 

 

Best Regards

Yilong Zhou

If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.

 

View solution in original post

1 REPLY 1
v-yilong-msft
Community Support
Community Support

Hi @optimizer ,

First of all if you are dealing with large amounts of data or complex data transformations, then Writing to a Lakehouse has more extra control and scalability.

 

If your main goal is to integrate data flows within the Power Platform ecosystem, then Using Internal Lakehouse Staging may be more efficient.

 

So in summary if simplicity and fast integration are your top priorities, Internal Lakehouse Staging may be a better choice. However, if you need more control and scalability, then writing to Lakehouse and using the OneHub/Lakehouse connector may be worth the extra effort.

 

If you want to learn more about this, I think you can read this document: Dataflow Gen2 data destinations and managed settings - Microsoft Fabric | Microsoft Learn

 

 

 

Best Regards

Yilong Zhou

If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.

 

Helpful resources

Announcements
Las Vegas 2025

Join us at the Microsoft Fabric Community Conference

March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount!

FebFBC_Carousel

Fabric Monthly Update - February 2025

Check out the February 2025 Fabric update to learn about new features.

Feb2025 NL Carousel

Fabric Community Update - February 2025

Find out what's new and trending in the Fabric community.

Top Solution Authors