Fabric is Generally Available. Browse Fabric Presentations. Work towards your Fabric certification with the Cloud Skills Challenge.
Are there any risks/concerns with large data refreshes of datasets on the capacity? Can a refresh cause a slowdown/crash of the capacity itself
I understand that there are size capacities (e.g., 1gb on Pro or 100 TB total against a capacity) and maximum amount of times refreshes can occur, but I'd like to understand the impact it can have on the capacity.
For example, if I were to have a dataset that is currently 600mb in size and it is scheduled to refresh daily at 12pm. 1gb of additional data is added into the dataset at 10am (so total size is now 1.6gb). During the 12pm scheduled refresh, are there any risks involved aside from time it takes to refresh on the capacity itself. e.g., can it slow down the service or crash it?
Thanks
Hey @Morkil ,
your decribed scenario can cause a slow down of your tenant. But Service organize (compute resources) a Schedule refresh by "itself".
So normally you should´t feel an impact (slow down) for your tenant.
I don´t feel such impacts during refreshing processes.
Regards
P.S. 1,6GB dataset is not a big one 🙂
Check out the November 2023 Power BI update to learn about new features.
Read the latest Fabric Community announcements, including updates on Power BI, Synapse, Data Factory and Data Activator.
Join us for a free, hands-on Microsoft workshop led by women trainers for women where you will learn how to build a Dashboard in a Day!
User | Count |
---|---|
27 | |
24 | |
23 | |
23 | |
23 |