Don't miss your chance to take exam DP-600 or DP-700 on us!
Request nowLearn from the best! Meet the four finalists headed to the FINALS of the Power BI Dataviz World Championships! Register now
Hi Folks, I have a Data Model hosted in PowerBI Premium Capacity and my fact tables have more then 200M rows. For performance, we have copied our data to Google BigQuery. As I'm using the large model on PowerBI I can't download the pbix file to manage it. Is there a way with some external tool (Like Tabular Editor) to move my data sorce from SQL SERVER to Google Bigquery without recreate/reweite my entire model?
Regards!
200M is not large. Set it back to small model and then fetch the pbix.
Large Model starts around the 5GB mark.
Are you saying that you created the dataset in the Power BI service? Was there no original PBIX?
@lbendlin Yes, there's no original pbix, it is from Azure Analysis Services Migration. I've publused the .BIM file using the Tabular Editor.
Share feedback directly with Fabric product managers, participate in targeted research studies and influence the Fabric roadmap.
Check out the February 2026 Power BI update to learn about new features.
| User | Count |
|---|---|
| 9 | |
| 9 | |
| 8 | |
| 7 | |
| 6 |