I realize that normalized star/snowflake schemas are usually the best way to go in terms of efficiency and performance. However, I have a situation where I have six tables (one fact, six dimension) and I want to create into one big flat table.
In terms of merging these tables, is there any benefit when it comes to efficiency and refresh performance to do this in a Dataflow or later on in a Dataset?
@arock-well there are many reasons but 2-3 top would be: 1) it will take less time to refresh since data is already processed in the dataflow 2) better compression 3) dataflow processed data can be used in multiple Power BI reports
@arock-well first question, why a flat table if you have a proper star schema, you can potentially run into challenges using one flat table, and also it is not the best practice.
As I already mentioned in the original message, I realize that is the best way to go. I'm not going to go into the reasons.
So based on that, what would be the best way to do merges: Within a dataflow or dataset?
Check out the November 2023 Power BI update to learn about new features.
Read the latest Fabric Community announcements, including updates on Power BI, Synapse, Data Factory and Data Activator.
130+ sessions, 130+ speakers, Product managers, MVPs, and experts. All about Power BI and Fabric. Attend online or watch the recordings.