Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Get Fabric certified for FREE! Don't miss your chance! Learn more

Reply
Si_7777
Helper III
Helper III

Staging area in fabric

New to fabric so please go gently

 

I currently have a Data Warehouse in SQL Server.  We have a database for staging and a database for the Data Warehouse containing dimension and facts.  We use SSIS to ETL the data from legacy to staging and then to the DW.  I am looking to move this into the cloud preferably Fabric.  I have watched allot of You Tube videos on fabric/azure/dw and all of them seem to do things different which makes me double guess myself, so can anyone help clear the mud.

 

For a staging area in Fabric would this be a lake-house as some videos have shown staging tables as part of the warehouse?

2 ACCEPTED SOLUTIONS
FabianSchut
Solution Sage
Solution Sage

Hi, I always use lakehouses for staging.

One advantage of lakehouses is that you can store (un)structered data files (like JSON or even images). Especially if you work with API-calls that contain complex JSON responses, it could be convenient to first store the output as a JSON file before transforming it into a delta table.

Another advantage of a lakehouse is that it natively interacts with a PySpark Notebook. I prefer to use PySpark to extract data sources. With PySpark you have all the flexibilty to connect to all (complex) data sources.

The last advantage I will mention here is the use of shortcuts. In a lakehouse, you can create shortcuts from other lakehouses or warehouses, but also other data sources outside Fabric. You can find the full list of shortcut sources here: https://learn.microsoft.com/en-us/fabric/onelake/onelake-shortcuts#types-of-shortcuts

 

Warehouses in Fabric lack these three features, that is why my preference is a lakehouse for staging. 

View solution in original post

Srisakthi
Super User
Super User

Hi @Si_7777 ,

 

Do you want to completely move from legacy to cloud? or just want to keep only the data warehouse piece?

 

Complete Legacy to Cloud:

When you wan to completely move from legacy to cloud, you can consider the following

Staging Area : Lakehouse files/tables

Data Ingestion : To get the data from legacy source systems you can use data pipeline(check for your connector) and load it to staging area. 

Data Transformation : If you want to clean and transform your data you can use notebook and persist the data to warehouse.

Consumption Area : Warehouse

 

Only from Warehouse to Cloud:

When you wan to move from warehouse to cloud, you can consider the following

Consumption Area : Warehouse

Data Ingestion : To get the data from warehouse systems you can use data pipeline(check for your connector) and load it to Warehouse. Assuming your warehouse is already a cleaned and ready to consume data.

 

Regards,

Srisakthi

 

View solution in original post

4 REPLIES 4
Srisakthi
Super User
Super User

Hi @Si_7777 ,

 

Do you want to completely move from legacy to cloud? or just want to keep only the data warehouse piece?

 

Complete Legacy to Cloud:

When you wan to completely move from legacy to cloud, you can consider the following

Staging Area : Lakehouse files/tables

Data Ingestion : To get the data from legacy source systems you can use data pipeline(check for your connector) and load it to staging area. 

Data Transformation : If you want to clean and transform your data you can use notebook and persist the data to warehouse.

Consumption Area : Warehouse

 

Only from Warehouse to Cloud:

When you wan to move from warehouse to cloud, you can consider the following

Consumption Area : Warehouse

Data Ingestion : To get the data from warehouse systems you can use data pipeline(check for your connector) and load it to Warehouse. Assuming your warehouse is already a cleaned and ready to consume data.

 

Regards,

Srisakthi

 

Hi Srisakthi

 

Yes eventually complete to cloud, thank you for your reply,  Simon

FabianSchut
Solution Sage
Solution Sage

Hi, I always use lakehouses for staging.

One advantage of lakehouses is that you can store (un)structered data files (like JSON or even images). Especially if you work with API-calls that contain complex JSON responses, it could be convenient to first store the output as a JSON file before transforming it into a delta table.

Another advantage of a lakehouse is that it natively interacts with a PySpark Notebook. I prefer to use PySpark to extract data sources. With PySpark you have all the flexibilty to connect to all (complex) data sources.

The last advantage I will mention here is the use of shortcuts. In a lakehouse, you can create shortcuts from other lakehouses or warehouses, but also other data sources outside Fabric. You can find the full list of shortcut sources here: https://learn.microsoft.com/en-us/fabric/onelake/onelake-shortcuts#types-of-shortcuts

 

Warehouses in Fabric lack these three features, that is why my preference is a lakehouse for staging. 

Thank you Fabian very helpful

Helpful resources

Announcements
Join our Fabric User Panel

Join our Fabric User Panel

Share feedback directly with Fabric product managers, participate in targeted research studies and influence the Fabric roadmap.

February Fabric Update Carousel

Fabric Monthly Update - February 2026

Check out the February 2026 Fabric update to learn about new features.

FabCon Atlanta 2026 carousel

FabCon Atlanta 2026

Join us at FabCon Atlanta, March 16-20, for the ultimate Fabric, Power BI, AI and SQL community-led event. Save $200 with code FABCOMM.