Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Find everything you need to get certified on Fabric—skills challenges, live sessions, exam prep, role guidance, and more. Get started

Reply
shuhn1229
Resolver I
Resolver I

Daisy-Chaining Dataflows

Hi all,

 

I am working with an API that is very finnicky and is prone to timing out, especially when large amounts of data are requested. Unfortunately, for this use case I have to move about 10M rows, and paginating of one big request does not help speed up the call - still results in timeouts. To get around this I basically setup ~10 data flows that act in concert. Dataflow 1 performs a call of 1/10th the data, then dataflow 2 references dataflow 1, makes a call for the next chunk of data, appends it to the referenced dataflow 1 and so on. What I have found is that on refresh of dataflow 1, all of the linked tables across the 10 flows also refresh, however the new api call doesn't refresh. This whole process seems awfully inefficient. Is there an ideal way of doing this? I thought about using PA to refresh a dataflow after one is complete trigger, going through all ten. I do not want to combine all 10 into one flow because I am very confident this will lead to timeouts.

 

Thanks!

1 ACCEPTED SOLUTION
shuhn1229
Resolver I
Resolver I

solution for this - multiple api calls of a reasonable size, partioning the data. 

 

setup a power automate trigger, when a data flow completes - this will trigger the next flow with the next chunk of data.

 

setup a final PQ step to append the query results, reference the tables from the partioned API calls. importantly, ensure enable dataload is disabled with all of the query results. append the results.

View solution in original post

2 REPLIES 2
shuhn1229
Resolver I
Resolver I

solution for this - multiple api calls of a reasonable size, partioning the data. 

 

setup a power automate trigger, when a data flow completes - this will trigger the next flow with the next chunk of data.

 

setup a final PQ step to append the query results, reference the tables from the partioned API calls. importantly, ensure enable dataload is disabled with all of the query results. append the results.

vanessafvg
Super User
Super User

what do you mean PA?  do you have dedicated / premium capacity, i dont think you can reference dataflows from each other without it. 

 

have you should tried combine the dataflows in a power be semantic model?    not quite clear what you saying, if you create 10 dataflows in sequence but do no referencing does that help?





If I took the time to answer your question and I came up with a solution, please mark my post as a solution and /or give kudos freely for the effort 🙂 Thank you!

Proud to be a Super User!




Helpful resources

Announcements
Sept PBI Carousel

Power BI Monthly Update - September 2024

Check out the September 2024 Power BI update to learn about new features.

September Hackathon Carousel

Microsoft Fabric & AI Learning Hackathon

Learn from experts, get hands-on experience, and win awesome prizes.

Sept NL Carousel

Fabric Community Update - September 2024

Find out what's new and trending in the Fabric Community.