Join us at FabCon Atlanta from March 16 - 20, 2026, for the ultimate Fabric, Power BI, AI and SQL community-led event. Save $200 with code FABCOMM.
Register now!To celebrate FabCon Vienna, we are offering 50% off select exams. Ends October 3rd. Request your discount now.
Hi Folks, I have a Data Model hosted in PowerBI Premium Capacity and my fact tables have more then 200M rows. For performance, we have copied our data to Google BigQuery. As I'm using the large model on PowerBI I can't download the pbix file to manage it. Is there a way with some external tool (Like Tabular Editor) to move my data sorce from SQL SERVER to Google Bigquery without recreate/reweite my entire model?
Regards!
200M is not large. Set it back to small model and then fetch the pbix.
Large Model starts around the 5GB mark.
Are you saying that you created the dataset in the Power BI service? Was there no original PBIX?
@lbendlin Yes, there's no original pbix, it is from Azure Analysis Services Migration. I've publused the .BIM file using the Tabular Editor.