Join us at FabCon Atlanta from March 16 - 20, 2026, for the ultimate Fabric, Power BI, AI and SQL community-led event. Save $200 with code FABCOMM.
Register now!To celebrate FabCon Vienna, we are offering 50% off select exams. Ends October 3rd. Request your discount now.
I have the following scenario:
1. Step 1 - I want to connect to Salesforce and get an Opportinities table AS-IS
2. I want to connect to the table from Step 1 and do all kinds of complex transformations with Power Query.
My task is to make an API request JUST ONCE to Salesforce. All the rest load should be done inside Microsoft capacity. So I want to separate the two steps most efficiently.
(When I do everything inside ONE Dataflow - the second step is sending API requests again and again and again to Salesforce so it is not an option)
I have created 2 dataflows and the second one refers to the first (it created a "linked" entity with enhances mode, watever it means). And I've created a Power Automate flow so the second dataflow will refresh after the first is finished.
But it still takes a long time for the second one to run. And I wonder whether it still propagates the API calls to Salesforce because of "linked entity".
I am thinking to test another alternative:
Create a Gen2 Dataflow, save the initial "AS-IS" table in a Warehouse/or Lakehouse.
Then, connect the second Dataflow to it. And schedule everything with Pipeline.
Should there be differences in this approach?
Or is it more or less the same as just separating two Gen1 dataflows?
What does the theory tell?
Hi @iBusinessBI
This is a duplicate thread of the similar one available at the link : Duplicate Link
Hence I am closing this thread.
Thanks.