Join us for an expert-led overview of the tools and concepts you'll need to pass exam PL-300. The first session starts on June 11th. See you there!
Get registeredPower BI is turning 10! Let’s celebrate together with dataviz contests, interactive sessions, and giveaways. Register now.
Our largest fact table is > 15M rows, lives in an Azure SQL DB, and feeds into a Dataflow in a Premium Capacity Workspace.
To make development a bit easier, I'd like to create a version of the fact table that only has the last week's worth of data - a few hundred thousand rows at most. Then people could import that smaller table from the Dataflow, create their model, then when ready, switch to the main table in the Dataflow that has all the data.
I could:
Create a 2nd table in the dataflow, that's just the same query, but limited (via SQL) to the last week. But then say, the transformations of each table in the Dataflow, have to be kept in sync.
I was thinking instead:
Have a computed entity, that references the full table in the Dataflow, and then limits to the last week of data. That way it's always the exact same columns as the main Dataflow table.
But seemingly it's going to load all 16M rows into that computed entiry first - thus using up a lot of our capacity's processing.
So - is there a better way, or something I'm missing?
I think on your main dataflow you need to switch the optimised compute engine to on rather than automatic to enable query folding down stream.
This is your chance to engage directly with the engineering team behind Fabric and Power BI. Share your experiences and shape the future.
Check out the June 2025 Power BI update to learn about new features.
User | Count |
---|---|
48 | |
31 | |
27 | |
27 | |
26 |
User | Count |
---|---|
61 | |
56 | |
35 | |
29 | |
28 |