Advance your Data & AI career with 50 days of live learning, dataviz contests, hands-on challenges, study groups & certifications and more!
Get registeredGet Fabric Certified for FREE during Fabric Data Days. Don't miss your chance! Learn more
Hi,
We have Databricks to transform big data and then save the data table (csv) in Blob Storgae with tables onwards 20 million rows. Currently, team was connecting Power BI directly with Blob Storage which means import connection and huge under-performance. Once developed, then would deploy the solution to Power BI Service. I have come across few options to handle large datasets optimally with both pros and cons. Can someone please give their feedback on optimal solution with performance and cost considerations:
Other Assumptions:
1. Data refresh once a day only
2. Data to be analysed only for preceding 15 months
3. No of users' concurrency - upto 20 at a time
Solutions under consideration:
It will be my first major independently Power BI Solution, so thanks in advance for your feedback.
Hi Amit
Thanks for your reply. Though these links explain the overview and their differences but I have few further questions in relation to my challenge from these links:
1. Power BI Premium: It does allow upto 100GB datasets but this is only applicable to modelled, processed and compressed data (pbix files); hence this feature can only be helpful after deployment to Power BI Service.
For development in Power BI Desktop (on my local machine), Power BI Desktop would connect to Blob Storage and still would bring the data to my local machine for processing (blob storage only offers in-memory connection). Hence, using my machine's RAM which won't be able to handle processing larger datasets (in5 GBs). Please confirm if my understanding is incorrect here and if so, then how Power BI Premium would help with this local machie processing challange?
2. Azure Analysis Service: In this option, Visual Studio woud connect to Blob Storage to get data, model it and then deploy to AAS as cubes. When Power BI (even Pro) connects to AAS, it would be send the query to AAS cube to process, hence processing is done in AAS server now, instead of my machine. Also, AAS would offer scale out option to handle large size cubes. Hence, this option should not impact my local machine while using Power BI Desktop for processing the data.
I believe similar process would work for Azure Synapse Analytics (cloud DWH). If my assumption is correct, then can you please help me to select between AAS and ASA?
Thanks
@jahlawat , please refer to this. It compares first 2
https://blog.coeo.com/power-bi-premium-vs-azure-analysis-services
https://ssbipolar.com/2018/12/09/choosing-between-power-bi-premium-and-azure-analysis-services/
You can consider Azure Synapse as thrid option, refer: https://www.be-terna.com/sites/5ebfc39e232fb906a3dfc34b/assets/5f3fb90d232fb95d6db09df0/power-bi-pro...
Advance your Data & AI career with 50 days of live learning, contests, hands-on challenges, study groups & certifications and more!
Check out the October 2025 Power BI update to learn about new features.