Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Don't miss out! 2025 Microsoft Fabric Community Conference, March 31 - April 2, Las Vegas, Nevada. Use code MSCUST for a $150 discount. Prices go up February 11th. Register now.

Reply
SidTheSloth
Frequent Visitor

ADF - Fabric Lakehouse Tables as source in Dataflows

Hi, When using Fabric Lakehouse as a source in ADF dataflows, there is no ability to pass a filter clause back to source to restrict the data flowing into the Dataflows. A similar operation on any regular database works fine.

 We tried using Lakehouse Views as a workaround, they do not seem to be visible as a source in a dataflow either. Lakehouse Schemas are enabled, and could be the reason, but we're trying to avoid having to recreating a lakehouse without Schema support.

 

This severly impacts the abiltiy to restrict the number of rows flowing into the ADF data flows. We have some very large tables, and this is a problem area for us. Are we missing a trick? Any suggestions or is this a known issue?

1 ACCEPTED SOLUTION

Is there a SQL connector in ADF dataflows?

 

Could you copy the Lakehouses' SQL connection string, and paste it in the SQL connector in ADF dataflow? Sounds like something which would allow for passing filters in the query. 

 

I'm hoping you won't get issues with SQL Analytics Endpoint update delays.

View solution in original post

3 REPLIES 3
v-kongfanf-msft
Community Support
Community Support

Hi @SidTheSloth ,

 

The data can be preprocessed first with the help of an activity in the data pipeline, and then the result is used as a data source for dataflow gen2 to connect the two activities.

Fabric decision guide - copy activity, dataflow, or Spark - Microsoft Fabric | Microsoft Learn

 

Best Regards,
Adamk Kong

 

If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.

 

Thank you, while pre-processing the data to create a new can be a viable workaround, it adds to the processing load and overall run time. Considering parquets with hundreds of millions of rows, this is a lot of time.

 

I additionally want to clarify this is ADF pipelines not Fabric Data Factory pipelines. As a part of the newly added functionality, Ideally if we were able to filter at source(like we can with other SQL  based sources), that would be the right way to do things. The inability to query views  from dataflows (for Lakehouses with Schemas), compounds the problem.

 

 Appreciate your time. Thank you.

Is there a SQL connector in ADF dataflows?

 

Could you copy the Lakehouses' SQL connection string, and paste it in the SQL connector in ADF dataflow? Sounds like something which would allow for passing filters in the query. 

 

I'm hoping you won't get issues with SQL Analytics Endpoint update delays.

Helpful resources

Announcements
Las Vegas 2025

Join us at the Microsoft Fabric Community Conference

March 31 - April 2, 2025, in Las Vegas, Nevada. Use code MSCUST for a $150 discount! Prices go up Feb. 11th.

JanFabricDE_carousel

Fabric Monthly Update - January 2025

Explore the power of Python Notebooks in Fabric!

JanFabricDW_carousel

Fabric Monthly Update - January 2025

Unlock the latest Fabric Data Warehouse upgrades!

JanFabricDF_carousel

Fabric Monthly Update - January 2025

Take your data replication to the next level with Fabric's latest updates!