Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Join the Fabric FabCon Global Hackathon—running virtually through Nov 3. Open to all skill levels. $10,000 in prizes! Register now.

Reply
jbrooi
Helper I
Helper I

Understanding data refresh and dataset sizes Pro workspace vs Fabric, with web connector

Hi! 

 

I have several datasets/semantic models, which have web connectors. They get data from an API and I paginated all querys. I created the dataset in PBI Desktop and published dataset to the service in a Pro workspace and scheduled refresh. 

Errors happened, pointing to specific querys and I always engineered solutions expecting the error to be coming from bad querys and/or a not so very well performing API. 

 

In one case I engineered a new pbix for a dataset which is performing okay, and the older one is not refreshing at all. The size is the same though, both are 3 MB if I check the dataset storage size in the service. 

 

I've tested if the data refresh would perform better in a Fabric trial workspace. It runs succesful as well, but it takes the same time. Makes sense I guess, the big depending factor being the API. 

The strange thing is, the dataset size in the Pro workspace is 3 MB, yet in Fabric is 14 MB. For the exact same file. Can someone help me understand? 

 

Only difference I see is that 'Large semantic model storage format' is turned on for the dataset in Fabric workspace. 

Thanks. 

1 ACCEPTED SOLUTION
ibarrau
Super User
Super User

Hi. I don't think there is a doc specifying the difference between the engine behind shared capacity and dedicated capacity with large semantic model storage format enabled.

If you are sure the amount of data is the same and the Fabric one hasn't added more data, then my guess would be that a dedicated capacity is a more complex semantic model that has all analysis services features like XMLA endpoints that allows you to deal with much more complex operations. That would make an increase of the data model size.

I hope that helps,


If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.

Happy to help!

LaDataWeb Blog

View solution in original post

1 REPLY 1
ibarrau
Super User
Super User

Hi. I don't think there is a doc specifying the difference between the engine behind shared capacity and dedicated capacity with large semantic model storage format enabled.

If you are sure the amount of data is the same and the Fabric one hasn't added more data, then my guess would be that a dedicated capacity is a more complex semantic model that has all analysis services features like XMLA endpoints that allows you to deal with much more complex operations. That would make an increase of the data model size.

I hope that helps,


If this post helps, then please consider Accept it as the solution to help the other members find it more quickly.

Happy to help!

LaDataWeb Blog

Helpful resources

Announcements
September Power BI Update Carousel

Power BI Monthly Update - September 2025

Check out the September 2025 Power BI update to learn about new features.

FabCon Atlanta 2026 carousel

FabCon Atlanta 2026

Join us at FabCon Atlanta, March 16-20, for the ultimate Fabric, Power BI, AI and SQL community-led event. Save $200 with code FABCOMM.

Top Solution Authors
Top Kudoed Authors