Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Join us for an expert-led overview of the tools and concepts you'll need to become a Certified Power BI Data Analyst and pass exam PL-300. Register now.

Reply
MiraNi
Advocate IV
Advocate IV

Power BI dataset connecting Dataflow - shared capacity usage and parallel refreshes

We have datasets connecting to dataflows. Some dataflows include common dimensions, which are used in almost every dataset. I want to understand, how is the connection from dataset to dataflow meaning csv file in Azure Data Lake working - is it fast? What about when more than one dataset is connecting to same dataflow at the same time, how is the capacity shared and should this be avoided if making refreshes slow? 

 

I tried to find documentation about this but didn't succeed. 

1 ACCEPTED SOLUTION
itchyeyeballs
Impactful Individual
Impactful Individual

Each dataset that uses the dataflow will still need to be refreshed individually, think of it like a centrally stored csv file. It will be faster than a locally stored version.

If the quantity of data in the dataflow is large you should probably stagger the refreshes of the datasets that use it.

View solution in original post

2 REPLIES 2
itchyeyeballs
Impactful Individual
Impactful Individual

Each dataset that uses the dataflow will still need to be refreshed individually, think of it like a centrally stored csv file. It will be faster than a locally stored version.

If the quantity of data in the dataflow is large you should probably stagger the refreshes of the datasets that use it.

ToddChitt
Super User
Super User

My understanding of Dataflows is that Power BI stores the data in a proprietary format in some blob storage it provisions for you behind the scenes. I think there is also the option to provision your own and have Power BI use that. So when the Dataflow gets refreshed, it is pulling data from some slow source, like your on-premise CSV file, and storing it in some location and format it can access easily. 

There should be minimal considerations of refreshing multiple Power BI models that all consume the same Dataflow.

I do not know how it affects capacity metrics, but since the data is technically not stored in Power BI, it may not affect it at all. 

I have seen our clients provision one Workspace that just contains Dataflows and nothing else, and that seems to work well.

Any experts out there, please feel free to correct my assumptions.

Regards.




Did I answer your question? If so, mark my post as a solution. Also consider helping someone else in the forums!

Proud to be a Super User!





Helpful resources

Announcements
Join our Fabric User Panel

Join our Fabric User Panel

This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.

June 2025 Power BI Update Carousel

Power BI Monthly Update - June 2025

Check out the June 2025 Power BI update to learn about new features.

June 2025 community update carousel

Fabric Community Update - June 2025

Find out what's new and trending in the Fabric community.