Power BI is turning 10! Tune in for a special live episode on July 24 with behind-the-scenes stories, product evolution highlights, and a sneak peek at what’s in store for the future.
Save the dateEnhance your career with this limited time 50% discount on Fabric and Power BI exams. Ends August 31st. Request your voucher.
I have the following scenario:
1. Step 1 - I want to connect to Salesforce and get an Opportinities table AS-IS
2. I want to connect to the table from Step 1 and do all kinds of complex transformations with Power Query.
My task is to make an API request JUST ONCE to Salesforce. All the rest load should be done inside Microsoft capacity. So I want to separate the two steps most efficiently.
(When I do everything inside ONE Dataflow - the second step is sending API requests again and again and again to Salesforce so it is not an option)
I have created 2 dataflows and the second one refers to the first (it created a "linked" entity with enhances mode, watever it means). And I've created a Power Automate flow so the second dataflow will refresh after the first is finished.
But it still takes a long time for the second one to run. And I wonder whether it still propagates the API calls to Salesforce because of "linked entity".
I am thinking to test another alternative:
Create a Gen2 Dataflow, save the initial "AS-IS" table in a Warehouse/or Lakehouse.
Then, connect the second Dataflow to it. And schedule everything with Pipeline.
Should there be differences in this approach?
Or is it more or less the same as just separating two Gen1 dataflows?
What does the theory tell?
Hi @iBusinessBI
This is a duplicate thread of the similar one available at the link : Duplicate Link
Hence I am closing this thread.
Thanks.
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
User | Count |
---|---|
2 | |
1 | |
1 | |
1 | |
1 |